-
-
Notifications
You must be signed in to change notification settings - Fork 11k
Description
Your current environment
deepseek_r1_reasoning_parser.py exception
🐛 Describe the bug
I used vllm 0.7.3 V1 version
model is deepseek distill 14B , use llmcompressor.transformers Quantization to FP8
when muilt request to the vllm server
sometimes server raise exception :
await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/starlette/_exception_handler.py", line 53, in wrapped_app
raise exc
File "/usr/local/lib/python3.10/dist-packages/starlette/_exception_handler.py", line 42, in wrapped_app
await app(scope, receive, sender)
File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 715, in call
await self.middleware_stack(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 735, in app
await route.handle(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 288, in handle
await self.app(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 76, in app
await wrap_app_handling_exceptions(app, request)(scope, receive, send)
File "/usr/local/lib/python3.10/dist-packages/starlette/_exception_handler.py", line 53, in wrapped_app
raise exc
File "/usr/local/lib/python3.10/dist-packages/starlette/_exception_handler.py", line 42, in wrapped_app
await app(scope, receive, sender)
File "/usr/local/lib/python3.10/dist-packages/starlette/routing.py", line 73, in app
response = await f(request)
File "/usr/local/lib/python3.10/dist-packages/fastapi/routing.py", line 301, in app
raw_response = await run_endpoint_function(
File "/usr/local/lib/python3.10/dist-packages/fastapi/routing.py", line 212, in run_endpoint_function
return await dependant.call(**values)
File "/usr/local/lib/python3.10/dist-packages/vllm/entrypoints/utils.py", line 56, in wrapper
return handler_task.result()
File "/usr/local/lib/python3.10/dist-packages/vllm/entrypoints/openai/api_server.py", line 410, in create_chat_completion
generator = await handler.create_chat_completion(request, raw_request)
File "/usr/local/lib/python3.10/dist-packages/vllm/entrypoints/openai/serving_chat.py", line 267, in create_chat_completion
return await self.chat_completion_full_generator(
File "/usr/local/lib/python3.10/dist-packages/vllm/entrypoints/openai/serving_chat.py", line 731, in chat_completion_full_generator
reasoning_parser.extract_reasoning_content(
File "/usr/local/lib/python3.10/dist-packages/vllm/entrypoints/openai/reasoning_parsers/deepseek_r1_reasoning_parser.py", line 137, in extract_reasoning_content
reasoning_content = self.reasoning_regex.findall(model_output)[0]
IndexError: list index out of range
Before submitting a new issue...
- Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
Metadata
Metadata
Assignees
Labels
Type
Projects
Status
