Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: glm-4v-9b报错 #4

Open
1 task done
Sunxiaohu0406 opened this issue Nov 27, 2024 · 3 comments
Open
1 task done

[Bug]: glm-4v-9b报错 #4

Sunxiaohu0406 opened this issue Nov 27, 2024 · 3 comments
Labels
bug Something isn't working

Comments

@Sunxiaohu0406
Copy link

Your current environment

tokenizers 0.20.4
tomli 2.1.0
torch 2.5.1
torch-npu 2.5.1rc1
torchvision 0.20.1
tqdm 4.67.1
transformers 4.46.3
typing_extensions 4.12.2

Model Input Dumps

No response

🐛 Describe the bug

curl "http://10.161.63.218:50077/v1/chat/completions"
-H "Content-Type: application/json"
-d '{
"model": "GLM-4V",
"messages": [
{
"role": "user",
"content": [
{ "type": "text", "text": "描述这张图片" },
{ "type": "image_url", "image_url": { "url": "data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAACgAAAAfCAIAAAAa4xgvAAAACXBIWXMAABJ0AAASdAHeZh94AAAAEXRFWHRTb2Z0d2FyZQBTbmlwYXN0ZV0Xzt0AAAGDSURBVEiJ7ZZNboJAFMefxnMIsvQAphtkgQt7AxIJLMzEnoGkLtqEM2gIi6oLjsACmkC78hBWepHpYuwEBuTDjNSk/sJiPvnPe3nvzXSG4wdoGQwA0GtZj9JtTzjLnwlzdTWuXkK5u7oZTXzLwFrsrVykG2dlmO+XyVjZ++HeD721W1O4zGJv5UqCWDgVxJFlv9jWcjJWyMhoqtJD2NaycH1dYe1pnu6imaHKirbIDDq7N2e3ye+l5wAApBuDPmtAj86hmUnakiCS9uhRhatxEna2G2e7AQBv7YYfUaERfLm9dPLWrmW/HpJj+X6km0g3aTeIo/fP6EJh8i9tMa9UhaLgonFeV5jmxrlA5UuXqn59H0dTtY6VXDhZzGR3C/CMakkQ65fMLlN1L4PUatt6ZuoaQZWz4YYBcJPbadAXD0nC6AEoSDeDOErXSMLeD2k7f6CTcNpL6UWM95j9JGuDuDhx80dJ0xnKnJ+35HYqU8XXEa4GA/zHN9ft3U6XcH9Xl/ED3BKQRuI2xpEAAAAASUVORK5CYII="}}
]
}
],
"max_tokens": 128
}'

INFO 11-27 20:01:13 logger.py:37] Received request chatcmpl-fde0bff308f34171ac7fdf1d80e1d8cc: prompt: '[gMASK]<|user|>\n描述这张图片<|assistant|>', params: SamplingParams(n=1, presence_penalty=0.0, frequency_penalty=0.0, repetition_penalty=1.0, temperature=0.7, top_p=1.0, top_k=-1, min_p=0.0, seed=None, stop=[], stop_token_ids=[], bad_words=[], include_stop_str_in_output=False, ignore_eos=False, max_tokens=128, min_tokens=0, logprobs=None, prompt_logprobs=None, skip_special_tokens=True, spaces_between_special_tokens=True, truncate_prompt_tokens=None, guided_decoding=None), prompt_token_ids: None, lora_request: None, prompt_adapter_request: None.
INFO 11-27 20:01:13 preprocess.py:215] Your model uses the legacy input pipeline instead of the new multi-modal processor. Please note that the legacy pipeline will be removed in a future release. For more details, see: vllm-project#10114
INFO 11-27 20:01:13 engine.py:285] Aborted request chatcmpl-fde0bff308f34171ac7fdf1d80e1d8cc.
INFO: 10.161.63.218:48780 - "POST /v1/chat/completions HTTP/1.1" 500 Internal Server Error
ERROR: Exception in ASGI application
Traceback (most recent call last):
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_utils.py", line 76, in collapse_excgroups
yield
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 189, in call
response_sent.set()
File "/usr/local/python3.9/lib/python3.9/site-packages/anyio/_backends/_asyncio.py", line 763, in aexit
raise BaseExceptionGroup(
exceptiongroup.ExceptionGroup: unhandled errors in a TaskGroup (1 sub-exception)

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
File "/usr/local/python3.9/lib/python3.9/site-packages/uvicorn/protocols/http/httptools_impl.py", line 409, in run_asgi
result = await app( # type: ignore[func-returns-value]
File "/usr/local/python3.9/lib/python3.9/site-packages/uvicorn/middleware/proxy_headers.py", line 60, in call
return await self.app(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/fastapi/applications.py", line 1054, in call
await super().call(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/applications.py", line 113, in call
await self.middleware_stack(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/errors.py", line 187, in call
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/errors.py", line 165, in call
await self.app(scope, receive, _send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 189, in call
response_sent.set()
File "/usr/local/python3.9/lib/python3.9/contextlib.py", line 137, in exit
self.gen.throw(typ, value, traceback)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_utils.py", line 82, in collapse_excgroups
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 187, in call
response = await self.dispatch_func(request, call_next)
File "/workspace/vllm/vllm/entrypoints/openai/api_server.py", line 489, in add_request_id
response = await call_next(request)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 163, in call_next
raise app_exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/base.py", line 149, in coro
await self.app(scope, receive_or_disconnect, send_no_error)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/cors.py", line 85, in call
await self.app(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/middleware/exceptions.py", line 62, in call
await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 42, in wrapped_app
await app(scope, receive, sender)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 715, in call
await self.middleware_stack(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 735, in app
await route.handle(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 288, in handle
await self.app(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 76, in app
await wrap_app_handling_exceptions(app, request)(scope, receive, send)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app
raise exc
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/_exception_handler.py", line 42, in wrapped_app
await app(scope, receive, sender)
File "/usr/local/python3.9/lib/python3.9/site-packages/starlette/routing.py", line 73, in app
response = await f(request)
File "/usr/local/python3.9/lib/python3.9/site-packages/fastapi/routing.py", line 301, in app
raw_response = await run_endpoint_function(
File "/usr/local/python3.9/lib/python3.9/site-packages/fastapi/routing.py", line 212, in run_endpoint_function
return await dependant.call(**values)
File "/workspace/vllm/vllm/entrypoints/openai/api_server.py", line 346, in create_chat_completion
generator = await handler.create_chat_completion(request, raw_request)
File "/workspace/vllm/vllm/entrypoints/openai/serving_chat.py", line 226, in create_chat_completion
return await self.chat_completion_full_generator(
File "/workspace/vllm/vllm/entrypoints/openai/serving_chat.py", line 586, in chat_completion_full_generator
async for res in result_generator:
File "/workspace/vllm/vllm/utils.py", line 401, in iterate_with_cancellation
item = await awaits[0]
File "/workspace/vllm/vllm/engine/multiprocessing/client.py", line 633, in _process_request
raise request_output
KeyError: 'prompt'

Before submitting a new issue...

  • Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
@Sunxiaohu0406 Sunxiaohu0406 added the bug Something isn't working label Nov 27, 2024
@xgdyp
Copy link

xgdyp commented Dec 6, 2024

I got the same error

@wangshuai09
Copy link
Owner

You can try internVL2 and qwen2vl which are supported.

@Sunxiaohu0406
Copy link
Author

You can try internVL2 and qwen2vl which are supported.

所以目前是不支持glm-4v么,上述两个是支持的

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

3 participants