- 
          
- 
                Notifications
    You must be signed in to change notification settings 
- Fork 10.9k
[Feature] Add visionarena offline support for benchmark_throughput #14654
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: Jennifer Zhao <7443418+JenZhao@users.noreply.github.com>
| 👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run  Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add  🚀 | 
Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
[Feature] Huggingface Dataset support for benchmark_throughput
Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I left a few comments - thank you for your work!
Could you please also do a comparison run between main and this PR when running vLLM with ShareGPT? I'd like to see the differences between the two if there's any.
--------- Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
| 
 test results here https://github.com/JenZhao/vllm/pull/5#issue-2918317590 | 
Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM! I left some final comments.
Co-authored-by: Roger Wang <136131678+ywang96@users.noreply.github.com>
Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com>
…llm-project#14654) Signed-off-by: Jennifer Zhao <7443418+JenZhao@users.noreply.github.com> Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com> Co-authored-by: Jennifer Zhao <7443418+JenZhao@users.noreply.github.com> Co-authored-by: Jennifer Zhao <JenZhao@users.noreply.github.com> Co-authored-by: Roger Wang <136131678+ywang96@users.noreply.github.com> Signed-off-by: Richard Liu <ricliu@google.com>
…llm-project#14654) Signed-off-by: Jennifer Zhao <7443418+JenZhao@users.noreply.github.com> Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com> Co-authored-by: Jennifer Zhao <7443418+JenZhao@users.noreply.github.com> Co-authored-by: Jennifer Zhao <JenZhao@users.noreply.github.com> Co-authored-by: Roger Wang <136131678+ywang96@users.noreply.github.com> Signed-off-by: Louis Ulmer <ulmerlouis@gmail.com>
…llm-project#14654) Signed-off-by: Jennifer Zhao <7443418+JenZhao@users.noreply.github.com> Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com> Co-authored-by: Jennifer Zhao <7443418+JenZhao@users.noreply.github.com> Co-authored-by: Jennifer Zhao <JenZhao@users.noreply.github.com> Co-authored-by: Roger Wang <136131678+ywang96@users.noreply.github.com>
…llm-project#14654) Signed-off-by: Jennifer Zhao <7443418+JenZhao@users.noreply.github.com> Signed-off-by: Jennifer Zhao <ai.jenniferzhao@gmail.com> Co-authored-by: Jennifer Zhao <7443418+JenZhao@users.noreply.github.com> Co-authored-by: Jennifer Zhao <JenZhao@users.noreply.github.com> Co-authored-by: Roger Wang <136131678+ywang96@users.noreply.github.com> Signed-off-by: Mu Huai <tianbowen.tbw@antgroup.com>
This is a follow-up PR for #14036 to support vision arena in the
benchmark_throughputscript.With the
vllm-chatbackend, the prompt token count will now include image token counts, as we will be using the request output to perform the calculation.CLOSES #14600
hf:
lmarena-ai/vision-arena-bench-v0.1Try another model
hf:
lmms-lab/LLaVA-OneVision-Datasince we changed the token calculation for vllm and vllm-chat, here we are using sharegpt dataset to verify the change
This Branch
Main Branch
vllm-chat is not support on the main branch
python3 vllm/benchmarks/benchmark_throughput.py --model NousResearch/Hermes-3-Llama-3.1-8B --backend vllm --dataset-name sharegpt --dataset-path ~/data/vllm_benchmark_datasets/ShareGPT_V3_unfiltered_cleaned_split.json --num-prompts 1000