diff --git a/examples/offline_inference/qwen2_5_omni/only_thinker.py b/examples/offline_inference/qwen2_5_omni/only_thinker.py index c2c28d5ae6ae..52b6e977eaa2 100644 --- a/examples/offline_inference/qwen2_5_omni/only_thinker.py +++ b/examples/offline_inference/qwen2_5_omni/only_thinker.py @@ -140,7 +140,7 @@ def main(args): print(generated_text) -if __name__ == "__main__": +def parse_args(): parser = FlexibleArgumentParser( description='Demo on using vLLM for offline inference with ' 'audio language models') @@ -155,5 +155,9 @@ def main(args): default=None, help="Set the seed when initializing `vllm.LLM`.") - args = parser.parse_args() + return parser.parse_args() + + +if __name__ == "__main__": + args = parse_args() main(args) diff --git a/examples/online_serving/disaggregated_serving/disagg_proxy_demo.py b/examples/online_serving/disaggregated_serving/disagg_proxy_demo.py index 1bf4d50e2c92..c6d26778ee49 100644 --- a/examples/online_serving/disaggregated_serving/disagg_proxy_demo.py +++ b/examples/online_serving/disaggregated_serving/disagg_proxy_demo.py @@ -414,7 +414,7 @@ def run_server(self): server.run() -if __name__ == "__main__": +def parse_args(): # Todo: allow more config parser = argparse.ArgumentParser("vLLM disaggregated proxy server.") parser.add_argument("--model", @@ -445,6 +445,10 @@ def run_server(self): default=8000, help="Server port number", ) - args = parser.parse_args() + return parser.parse_args() + + +if __name__ == "__main__": + args = parse_args() proxy_server = ProxyServer(args=args) proxy_server.run_server()