Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Report usage for beam search #6404

Merged
merged 3 commits into from
Jul 15, 2024
Merged

Conversation

simon-mo
Copy link
Collaborator

So we can be informed when fully removing.

@simon-mo simon-mo requested a review from WoosukKwon July 13, 2024 03:21
Copy link

👋 Hi! Thank you for contributing to the vLLM project.
Just a reminder: PRs would not trigger full CI run by default. Instead, it would only trigger fastcheck CI to run, which consists only a small and essential subset of tests to quickly catch errors with the flexibility to run extra individual tests on top (you can do this by unblocking test steps in the Buildkite run).

Full CI run is still required to merge this PR so once the PR is ready to go, please make sure to run it. If you need all test signals in between PR commits, you can trigger full CI as well.

To run full CI, you can do one of these:

  • Comment /ready on the PR
  • Add ready label to the PR
  • Enable auto-merge.

🚀

@simon-mo simon-mo added the ready ONLY add when PR is ready to merge/full CI is needed label Jul 13, 2024
@@ -184,6 +184,9 @@ def __init__(

self._verify_args()
if self.use_beam_search:
# Lazy import to avoid circular imports.
from vllm.usage.usage_lib import set_runtime_usage_data
set_runtime_usage_data("use_beam_search", True)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

IIUC, we don't track the number of beam search requests or its ratio, but track the cases that the server receives at least one beam search request, right?

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Correct. I think in the end I want to get to some understanding of "% of vLLM deployments using beam search"

@simon-mo
Copy link
Collaborator Author

Verified that the data has been received from testing.

@simon-mo simon-mo merged commit 32c9d7f into vllm-project:main Jul 15, 2024
68 of 73 checks passed
dtrifiro pushed a commit to opendatahub-io/vllm that referenced this pull request Jul 17, 2024
xjpang pushed a commit to xjpang/vllm that referenced this pull request Jul 24, 2024
Alvant pushed a commit to compressa-ai/vllm that referenced this pull request Oct 26, 2024
Signed-off-by: Alvant <alvasian@yandex.ru>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ready ONLY add when PR is ready to merge/full CI is needed
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants