You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
INFO 11-27 20:01:13 logger.py:37] Received request chatcmpl-fde0bff308f34171ac7fdf1d80e1d8cc: prompt: '[gMASK]<|user|>\n描述这张图片<|assistant|>', params: SamplingParams(n=1, presence_penalty=0.0, frequency_penalty=0.0, repetition_penalty=1.0, temperature=0.7, top_p=1.0, top_k=-1, min_p=0.0, seed=None, stop=[], stop_token_ids=[], bad_words=[], include_stop_str_in_output=False, ignore_eos=False, max_tokens=128, min_tokens=0, logprobs=None, prompt_logprobs=None, skip_special_tokens=True, spaces_between_special_tokens=True, truncate_prompt_tokens=None, guided_decoding=None), prompt_token_ids: None, lora_request: None, prompt_adapter_request: None.
INFO 11-27 20:01:13 preprocess.py:215] Your model uses the legacy input pipeline instead of the new multi-modal processor. Please note that the legacy pipeline will be removed in a future release. For more details, see: vllm-project#10114
INFO 11-27 20:01:13 engine.py:285] Aborted request chatcmpl-fde0bff308f34171ac7fdf1d80e1d8cc.
INFO: 10.161.63.218:48780 - "POST /v1/chat/completions HTTP/1.1" 500 Internal Server Error
ERROR: Exception in ASGI application
Traceback (most recent call last):
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_utils.py", line 76, in collapse_excgroups
yield
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 189, in call
response_sent.set()
File "/usr/local/python3.9/lib/python3.9/site-packages/anyio/_backends/_asyncio.py", line 763, in aexit
raise BaseExceptionGroup(
exceptiongroup.ExceptionGroup: unhandled errors in a TaskGroup (1 sub-exception)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/python3.9/lib/python3.9/site-packages/uvicorn/protocols/http/httptools_impl.py", line 409, in run_asgi
result = await app( # type: ignore[func-returns-value]
File "/usr/local/python3.9/lib/python3.9/site-packages/uvicorn/middleware/proxy_headers.py", line 60, in call
return await self.app(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/fastapi/applications.py", line 1054, in call
await super().call(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/applications.py", line 113, in call
await self.middleware_stack(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/errors.py", line 187, in call
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/errors.py", line 165, in call
await self.app(scope, receive, _send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 189, in call
response_sent.set()
File "/usr/local/python3.9/lib/python3.9/contextlib.py", line 137, in exit
self.gen.throw(typ, value, traceback)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_utils.py", line 82, in collapse_excgroups
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 187, in call
response = await self.dispatch_func(request, call_next)
File "/workspace/vllm/vllm/entrypoints/openai/api_server.py", line 489, in add_request_id
response = await call_next(request)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 163, in call_next
raise app_exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 149, in coro
await self.app(scope, receive_or_disconnect, send_no_error)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/cors.py", line 85, in call
await self.app(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/exceptions.py", line 62, in call
await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 42, in wrapped_app
await app(scope, receive, sender)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 715, in call
await self.middleware_stack(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 735, in app
await route.handle(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 288, in handle
await self.app(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 76, in app
await wrap_app_handling_exceptions(app, request)(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 42, in wrapped_app
await app(scope, receive, sender)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 73, in app
response = await f(request)
File "/usr/local/python3.9/lib/python3.9/site-packages/fastapi/routing.py", line 301, in app
raw_response = await run_endpoint_function(
File "/usr/local/python3.9/lib/python3.9/site-packages/fastapi/routing.py", line 212, in run_endpoint_function
return await dependant.call(**values)
File "/workspace/vllm/vllm/entrypoints/openai/api_server.py", line 346, in create_chat_completion
generator = await handler.create_chat_completion(request, raw_request)
File "/workspace/vllm/vllm/entrypoints/openai/serving_chat.py", line 226, in create_chat_completion
return await self.chat_completion_full_generator(
File "/workspace/vllm/vllm/entrypoints/openai/serving_chat.py", line 586, in chat_completion_full_generator
async for res in result_generator:
File "/workspace/vllm/vllm/utils.py", line 401, in iterate_with_cancellation
item = await awaits[0]
File "/workspace/vllm/vllm/engine/multiprocessing/client.py", line 633, in _process_request
raise request_output
KeyError: 'prompt'
Before submitting a new issue...
Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
The text was updated successfully, but these errors were encountered:
Your current environment
tokenizers 0.20.4
tomli 2.1.0
torch 2.5.1
torch-npu 2.5.1rc1
torchvision 0.20.1
tqdm 4.67.1
transformers 4.46.3
typing_extensions 4.12.2
Model Input Dumps
No response
🐛 Describe the bug
curl "http://10.161.63.218:50077/v1/chat/completions"
-H "Content-Type: application/json"
-d '{
"model": "GLM-4V",
"messages": [
{
"role": "user",
"content": [
{ "type": "text", "text": "描述这张图片" },
{ "type": "image_url", "image_url": { "url": "data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAACgAAAAfCAIAAAAa4xgvAAAACXBIWXMAABJ0AAASdAHeZh94AAAAEXRFWHRTb2Z0d2FyZQBTbmlwYXN0ZV0Xzt0AAAGDSURBVEiJ7ZZNboJAFMefxnMIsvQAphtkgQt7AxIJLMzEnoGkLtqEM2gIi6oLjsACmkC78hBWepHpYuwEBuTDjNSk/sJiPvnPe3nvzXSG4wdoGQwA0GtZj9JtTzjLnwlzdTWuXkK5u7oZTXzLwFrsrVykG2dlmO+XyVjZ++HeD721W1O4zGJv5UqCWDgVxJFlv9jWcjJWyMhoqtJD2NaycH1dYe1pnu6imaHKirbIDDq7N2e3ye+l5wAApBuDPmtAj86hmUnakiCS9uhRhatxEna2G2e7AQBv7YYfUaERfLm9dPLWrmW/HpJj+X6km0g3aTeIo/fP6EJh8i9tMa9UhaLgonFeV5jmxrlA5UuXqn59H0dTtY6VXDhZzGR3C/CMakkQ65fMLlN1L4PUatt6ZuoaQZWz4YYBcJPbadAXD0nC6AEoSDeDOErXSMLeD2k7f6CTcNpL6UWM95j9JGuDuDhx80dJ0xnKnJ+35HYqU8XXEa4GA/zHN9ft3U6XcH9Xl/ED3BKQRuI2xpEAAAAASUVORK5CYII="}}
]
}
],
"max_tokens": 128
}'
INFO 11-27 20:01:13 logger.py:37] Received request chatcmpl-fde0bff308f34171ac7fdf1d80e1d8cc: prompt: '[gMASK]<|user|>\n描述这张图片<|assistant|>', params: SamplingParams(n=1, presence_penalty=0.0, frequency_penalty=0.0, repetition_penalty=1.0, temperature=0.7, top_p=1.0, top_k=-1, min_p=0.0, seed=None, stop=[], stop_token_ids=[], bad_words=[], include_stop_str_in_output=False, ignore_eos=False, max_tokens=128, min_tokens=0, logprobs=None, prompt_logprobs=None, skip_special_tokens=True, spaces_between_special_tokens=True, truncate_prompt_tokens=None, guided_decoding=None), prompt_token_ids: None, lora_request: None, prompt_adapter_request: None.
INFO 11-27 20:01:13 preprocess.py:215] Your model uses the legacy input pipeline instead of the new multi-modal processor. Please note that the legacy pipeline will be removed in a future release. For more details, see: vllm-project#10114
INFO 11-27 20:01:13 engine.py:285] Aborted request chatcmpl-fde0bff308f34171ac7fdf1d80e1d8cc.
INFO: 10.161.63.218:48780 - "POST /v1/chat/completions HTTP/1.1" 500 Internal Server Error
ERROR: Exception in ASGI application
Traceback (most recent call last):
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_utils.py", line 76, in collapse_excgroups
yield
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 189, in call
response_sent.set()
File "/usr/local/python3.9/lib/python3.9/site-packages/anyio/_backends/_asyncio.py", line 763, in aexit
raise BaseExceptionGroup(
exceptiongroup.ExceptionGroup: unhandled errors in a TaskGroup (1 sub-exception)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/local/python3.9/lib/python3.9/site-packages/uvicorn/protocols/http/httptools_impl.py", line 409, in run_asgi
result = await app( # type: ignore[func-returns-value]
File "/usr/local/python3.9/lib/python3.9/site-packages/uvicorn/middleware/proxy_headers.py", line 60, in call
return await self.app(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/fastapi/applications.py", line 1054, in call
await super().call(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/applications.py", line 113, in call
await self.middleware_stack(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/errors.py", line 187, in call
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/errors.py", line 165, in call
await self.app(scope, receive, _send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 189, in call
response_sent.set()
File "/usr/local/python3.9/lib/python3.9/contextlib.py", line 137, in exit
self.gen.throw(typ, value, traceback)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_utils.py", line 82, in collapse_excgroups
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 187, in call
response = await self.dispatch_func(request, call_next)
File "/workspace/vllm/vllm/entrypoints/openai/api_server.py", line 489, in add_request_id
response = await call_next(request)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 163, in call_next
raise app_exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 149, in coro
await self.app(scope, receive_or_disconnect, send_no_error)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/cors.py", line 85, in call
await self.app(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/exceptions.py", line 62, in call
await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 42, in wrapped_app
await app(scope, receive, sender)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 715, in call
await self.middleware_stack(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 735, in app
await route.handle(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 288, in handle
await self.app(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 76, in app
await wrap_app_handling_exceptions(app, request)(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 42, in wrapped_app
await app(scope, receive, sender)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 73, in app
response = await f(request)
File "/usr/local/python3.9/lib/python3.9/site-packages/fastapi/routing.py", line 301, in app
raw_response = await run_endpoint_function(
File "/usr/local/python3.9/lib/python3.9/site-packages/fastapi/routing.py", line 212, in run_endpoint_function
return await dependant.call(**values)
File "/workspace/vllm/vllm/entrypoints/openai/api_server.py", line 346, in create_chat_completion
generator = await handler.create_chat_completion(request, raw_request)
File "/workspace/vllm/vllm/entrypoints/openai/serving_chat.py", line 226, in create_chat_completion
return await self.chat_completion_full_generator(
File "/workspace/vllm/vllm/entrypoints/openai/serving_chat.py", line 586, in chat_completion_full_generator
async for res in result_generator:
File "/workspace/vllm/vllm/utils.py", line 401, in iterate_with_cancellation
item = await awaits[0]
File "/workspace/vllm/vllm/engine/multiprocessing/client.py", line 633, in _process_request
raise request_output
KeyError: 'prompt'
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: