From 58a172076cd3d339c080051927d706b5a4f3163b Mon Sep 17 00:00:00 2001 From: Sunghyun Park Date: Mon, 8 Jan 2024 23:59:39 +0000 Subject: [PATCH] fix --- serve/mlc_serve/utils.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/serve/mlc_serve/utils.py b/serve/mlc_serve/utils.py index ba9c18bba7..e47596e784 100644 --- a/serve/mlc_serve/utils.py +++ b/serve/mlc_serve/utils.py @@ -56,11 +56,11 @@ def create_mlc_engine(args: argparse.Namespace): "max_decode_steps": args.max_decode_steps, } ) - + # type: off if args.use_staging_engine: engine = StagingInferenceEngine( tokenizer_module=HfTokenizerModule(args.model_artifact_path), - model_module_loader=PagedCacheModelModule, + model_module_loader=PagedCacheModelModule, model_module_loader_kwargs={ "model_artifact_path": args.model_artifact_path, "engine_config": engine_config, @@ -68,12 +68,12 @@ def create_mlc_engine(args: argparse.Namespace): ) engine.start() else: - engine = SynchronousInferenceEngine( - PagedCacheModelModule( + engine = SynchronousInferenceEngine( + PagedCacheModelModule( model_artifact_path=args.model_artifact_path, engine_config=engine_config, ) ) - + # type: on return engine