File tree Expand file tree Collapse file tree 3 files changed +22
-11
lines changed Expand file tree Collapse file tree 3 files changed +22
-11
lines changed Original file line number Diff line number Diff line change 2424 RequestResponseMetadata , ToolCall , UsageInfo )
2525from vllm .entrypoints .openai .reasoning_parsers import (ReasoningParser ,
2626 ReasoningParserManager )
27- from vllm .entrypoints .openai .serving_engine import OpenAIServing
27+ from vllm .entrypoints .openai .serving_engine import (OpenAIServing ,
28+ clamp_prompt_logprobs )
2829from vllm .entrypoints .openai .serving_models import OpenAIServingModels
2930from vllm .entrypoints .openai .tool_parsers import ToolParser , ToolParserManager
3031from vllm .entrypoints .openai .tool_parsers .mistral_tool_parser import (
@@ -844,7 +845,7 @@ async def chat_completion_full_generator(
844845 model = model_name ,
845846 choices = choices ,
846847 usage = usage ,
847- prompt_logprobs = final_res .prompt_logprobs ,
848+ prompt_logprobs = clamp_prompt_logprobs ( final_res .prompt_logprobs ) ,
848849 )
849850
850851 return response
Original file line number Diff line number Diff line change 2323 RequestResponseMetadata ,
2424 UsageInfo )
2525# yapf: enable
26- from vllm .entrypoints .openai .serving_engine import OpenAIServing
26+ from vllm .entrypoints .openai .serving_engine import (OpenAIServing ,
27+ clamp_prompt_logprobs )
2728from vllm .entrypoints .openai .serving_models import OpenAIServingModels
2829from vllm .logger import init_logger
2930from vllm .outputs import RequestOutput
@@ -394,13 +395,7 @@ def request_output_to_completion_response(
394395 for final_res in final_res_batch :
395396 prompt_token_ids = final_res .prompt_token_ids
396397 assert prompt_token_ids is not None
397- prompt_logprobs = final_res .prompt_logprobs
398- if prompt_logprobs :
399- for logprob_dict in prompt_logprobs :
400- if logprob_dict :
401- for logprob_values in logprob_dict .values ():
402- if logprob_values .logprob == float ('-inf' ):
403- logprob_values .logprob = - 9999.0
398+ prompt_logprobs = clamp_prompt_logprobs (final_res .prompt_logprobs )
404399 prompt_text = final_res .prompt
405400
406401 token_ids : GenericSequence [int ]
Original file line number Diff line number Diff line change 4242from vllm .pooling_params import PoolingParams
4343from vllm .prompt_adapter .request import PromptAdapterRequest
4444from vllm .sampling_params import BeamSearchParams , SamplingParams
45- from vllm .sequence import Logprob
45+ from vllm .sequence import Logprob , PromptLogprobs
4646from vllm .tracing import (contains_trace_headers , extract_trace_headers ,
4747 log_tracing_disabled_warning )
4848from vllm .transformers_utils .tokenizer import AnyTokenizer , MistralTokenizer
@@ -535,3 +535,18 @@ def _get_model_name(self,
535535 if model_name is None :
536536 return self .models .base_model_paths [0 ].name
537537 return model_name
538+
539+
540+ def clamp_prompt_logprobs (
541+ prompt_logprobs : Union [PromptLogprobs ,
542+ None ]) -> Union [PromptLogprobs , None ]:
543+ if prompt_logprobs is None :
544+ return prompt_logprobs
545+
546+ for logprob_dict in prompt_logprobs :
547+ if logprob_dict is None :
548+ continue
549+ for logprob_values in logprob_dict .values ():
550+ if logprob_values .logprob == float ('-inf' ):
551+ logprob_values .logprob = - 9999.0
552+ return prompt_logprobs
You can’t perform that action at this time.
0 commit comments