File tree Expand file tree Collapse file tree 3 files changed +22
-11
lines changed Expand file tree Collapse file tree 3 files changed +22
-11
lines changed Original file line number Diff line number Diff line change 2424    RequestResponseMetadata , ToolCall , UsageInfo )
2525from  vllm .entrypoints .openai .reasoning_parsers  import  (ReasoningParser ,
2626                                                       ReasoningParserManager )
27- from  vllm .entrypoints .openai .serving_engine  import  OpenAIServing 
27+ from  vllm .entrypoints .openai .serving_engine  import  (OpenAIServing ,
28+                                                     clamp_prompt_logprobs )
2829from  vllm .entrypoints .openai .serving_models  import  OpenAIServingModels 
2930from  vllm .entrypoints .openai .tool_parsers  import  ToolParser , ToolParserManager 
3031from  vllm .entrypoints .openai .tool_parsers .mistral_tool_parser  import  (
@@ -844,7 +845,7 @@ async def chat_completion_full_generator(
844845            model = model_name ,
845846            choices = choices ,
846847            usage = usage ,
847-             prompt_logprobs = final_res .prompt_logprobs ,
848+             prompt_logprobs = clamp_prompt_logprobs ( final_res .prompt_logprobs ) ,
848849        )
849850
850851        return  response 
Original file line number Diff line number Diff line change 2323                                              RequestResponseMetadata ,
2424                                              UsageInfo )
2525# yapf: enable 
26- from  vllm .entrypoints .openai .serving_engine  import  OpenAIServing 
26+ from  vllm .entrypoints .openai .serving_engine  import  (OpenAIServing ,
27+                                                     clamp_prompt_logprobs )
2728from  vllm .entrypoints .openai .serving_models  import  OpenAIServingModels 
2829from  vllm .logger  import  init_logger 
2930from  vllm .outputs  import  RequestOutput 
@@ -394,13 +395,7 @@ def request_output_to_completion_response(
394395        for  final_res  in  final_res_batch :
395396            prompt_token_ids  =  final_res .prompt_token_ids 
396397            assert  prompt_token_ids  is  not None 
397-             prompt_logprobs  =  final_res .prompt_logprobs 
398-             if  prompt_logprobs :
399-                 for  logprob_dict  in  prompt_logprobs :
400-                     if  logprob_dict :
401-                         for  logprob_values  in  logprob_dict .values ():
402-                             if  logprob_values .logprob  ==  float ('-inf' ):
403-                                 logprob_values .logprob  =  - 9999.0 
398+             prompt_logprobs  =  clamp_prompt_logprobs (final_res .prompt_logprobs )
404399            prompt_text  =  final_res .prompt 
405400
406401            token_ids : GenericSequence [int ]
Original file line number Diff line number Diff line change 4242from  vllm .pooling_params  import  PoolingParams 
4343from  vllm .prompt_adapter .request  import  PromptAdapterRequest 
4444from  vllm .sampling_params  import  BeamSearchParams , SamplingParams 
45- from  vllm .sequence  import  Logprob 
45+ from  vllm .sequence  import  Logprob ,  PromptLogprobs 
4646from  vllm .tracing  import  (contains_trace_headers , extract_trace_headers ,
4747                          log_tracing_disabled_warning )
4848from  vllm .transformers_utils .tokenizer  import  AnyTokenizer , MistralTokenizer 
@@ -535,3 +535,18 @@ def _get_model_name(self,
535535        if  model_name  is  None :
536536            return  self .models .base_model_paths [0 ].name 
537537        return  model_name 
538+ 
539+ 
540+ def  clamp_prompt_logprobs (
541+     prompt_logprobs : Union [PromptLogprobs ,
542+                            None ]) ->  Union [PromptLogprobs , None ]:
543+     if  prompt_logprobs  is  None :
544+         return  prompt_logprobs 
545+ 
546+     for  logprob_dict  in  prompt_logprobs :
547+         if  logprob_dict  is  None :
548+             continue 
549+         for  logprob_values  in  logprob_dict .values ():
550+             if  logprob_values .logprob  ==  float ('-inf' ):
551+                 logprob_values .logprob  =  - 9999.0 
552+     return  prompt_logprobs 
 
 
   
 
     
   
   
          
    
    
     
    
      
     
     
    You can’t perform that action at this time.
  
 
    
  
    
      
        
     
       
      
     
   
 
    
    
  
 
  
 
     
    
0 commit comments