@@ -1586,6 +1586,14 @@ async def init_app_state(
15861586 state .vllm_config = vllm_config
15871587 model_config = vllm_config .model_config
15881588
1589+ if envs .VLLM_USE_V1 :
1590+ supported_tasks = await engine_client \
1591+ .get_supported_tasks () # type: ignore
1592+ else :
1593+ supported_tasks = model_config .supported_tasks
1594+
1595+ logger .info ("Supported_tasks: %s" , supported_tasks )
1596+
15891597 resolved_chat_template = load_chat_template (args .chat_template )
15901598 if resolved_chat_template is not None :
15911599 # Get the tokenizer to check official template
@@ -1647,7 +1655,7 @@ async def init_app_state(
16471655 reasoning_parser = args .reasoning_parser ,
16481656 enable_prompt_tokens_details = args .enable_prompt_tokens_details ,
16491657 enable_force_include_usage = args .enable_force_include_usage ,
1650- ) if "generate" in model_config . supported_tasks else None
1658+ ) if "generate" in supported_tasks else None
16511659 state .openai_serving_chat = OpenAIServingChat (
16521660 engine_client ,
16531661 model_config ,
@@ -1664,7 +1672,7 @@ async def init_app_state(
16641672 reasoning_parser = args .reasoning_parser ,
16651673 enable_prompt_tokens_details = args .enable_prompt_tokens_details ,
16661674 enable_force_include_usage = args .enable_force_include_usage ,
1667- ) if "generate" in model_config . supported_tasks else None
1675+ ) if "generate" in supported_tasks else None
16681676 state .openai_serving_completion = OpenAIServingCompletion (
16691677 engine_client ,
16701678 model_config ,
@@ -1673,40 +1681,38 @@ async def init_app_state(
16731681 return_tokens_as_token_ids = args .return_tokens_as_token_ids ,
16741682 enable_prompt_tokens_details = args .enable_prompt_tokens_details ,
16751683 enable_force_include_usage = args .enable_force_include_usage ,
1676- ) if "generate" in model_config . supported_tasks else None
1684+ ) if "generate" in supported_tasks else None
16771685 state .openai_serving_pooling = OpenAIServingPooling (
16781686 engine_client ,
16791687 model_config ,
16801688 state .openai_serving_models ,
16811689 request_logger = request_logger ,
16821690 chat_template = resolved_chat_template ,
16831691 chat_template_content_format = args .chat_template_content_format ,
1684- ) if "encode" in model_config . supported_tasks else None
1692+ ) if "encode" in supported_tasks else None
16851693 state .openai_serving_embedding = OpenAIServingEmbedding (
16861694 engine_client ,
16871695 model_config ,
16881696 state .openai_serving_models ,
16891697 request_logger = request_logger ,
16901698 chat_template = resolved_chat_template ,
16911699 chat_template_content_format = args .chat_template_content_format ,
1692- ) if "embed" in model_config . supported_tasks else None
1700+ ) if "embed" in supported_tasks else None
16931701 state .openai_serving_classification = ServingClassification (
16941702 engine_client ,
16951703 model_config ,
16961704 state .openai_serving_models ,
16971705 request_logger = request_logger ,
1698- ) if "classify" in model_config . supported_tasks else None
1706+ ) if "classify" in supported_tasks else None
16991707
1700- enable_serving_reranking = ("classify" in model_config .supported_tasks
1701- and getattr (model_config .hf_config ,
1702- "num_labels" , 0 ) == 1 )
1708+ enable_serving_reranking = ("classify" in supported_tasks and getattr (
1709+ model_config .hf_config , "num_labels" , 0 ) == 1 )
17031710 state .openai_serving_scores = ServingScores (
17041711 engine_client ,
17051712 model_config ,
17061713 state .openai_serving_models ,
17071714 request_logger = request_logger ,
1708- ) if ("embed" in model_config .supported_tasks
1709- or enable_serving_reranking ) else None
1715+ ) if ("embed" in supported_tasks or enable_serving_reranking ) else None
17101716
17111717 state .openai_serving_tokenization = OpenAIServingTokenization (
17121718 engine_client ,
@@ -1721,13 +1727,13 @@ async def init_app_state(
17211727 model_config ,
17221728 state .openai_serving_models ,
17231729 request_logger = request_logger ,
1724- ) if "transcription" in model_config . supported_tasks else None
1730+ ) if "transcription" in supported_tasks else None
17251731 state .openai_serving_translation = OpenAIServingTranslation (
17261732 engine_client ,
17271733 model_config ,
17281734 state .openai_serving_models ,
17291735 request_logger = request_logger ,
1730- ) if "transcription" in model_config . supported_tasks else None
1736+ ) if "transcription" in supported_tasks else None
17311737 state .task = model_config .task
17321738
17331739 state .enable_server_load_tracking = args .enable_server_load_tracking
0 commit comments