Skip to content

Conversation

@ApostaC
Copy link
Collaborator

@ApostaC ApostaC commented Sep 24, 2025

Purpose

This PR puts the LMCache integration code (that does from vllm import ***) into vLLM.

The goal of this PR (and the following series) is to make sure that LMCache has a correct integration with vLLM, and the integration codes are "maintainable" by the whole vLLM community.

Test Plan

There will be a few upcoming PRs covering the testing and documentation:

  • Add unit tests for the LMCache module that does "soft failing" (i.e., gives the contributor a notification that the integration check fails, but doesn't make CI red).
  • Add detailed documentation and examples for LMCache KV cache offloading

Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Signed-off-by: ApostaC <yihua98@uchicago.edu>
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request is a significant step in improving the maintainability of the LMCache integration by vendoring its code into the vLLM repository. The changes are well-structured, primarily involving the addition of the LMCache integration code under a new directory and updating the connector to use this native implementation. I've identified a critical bug and another high-severity issue that should be addressed.

Signed-off-by: ApostaC <yihua98@uchicago.edu>
Signed-off-by: ApostaC <yihua98@uchicago.edu>
Copy link
Collaborator

@KuntaiDu KuntaiDu left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM!!!

@KuntaiDu KuntaiDu enabled auto-merge (squash) September 26, 2025 00:34
@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Sep 26, 2025
@ApostaC ApostaC disabled auto-merge September 26, 2025 00:51
@simon-mo
Copy link
Collaborator

@codex review

Copy link

@chatgpt-codex-connector chatgpt-codex-connector bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codex Review

Here are some automated review suggestions for this pull request.

ℹ️ About Codex in GitHub

Your team has set up Codex to review pull requests in this repo. Reviews are triggered when you

  • Open a pull request for review
  • Mark a draft as ready
  • Comment "@codex review".

If Codex has suggestions, it will comment; otherwise it will react with 👍.

Codex can also answer questions or update the PR. Try commenting

from lmcache.v1.offload_server.zmq_server import ZMQOffloadServer
from lmcache.v1.plugin.plugin_launcher import PluginLauncher

# Third Party
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

fix the import order

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Fixed. Just to double check, the correct import order should be something like:

  1. python native imports (os, sys, time, typing...)
  2. third party libs (torch, zmq)
  3. first party imports with absolute paths (from vllm.xxx import)
  4. type checking related imports

Pls correct me if I'm wrong.

Signed-off-by: ApostaC <yihua98@uchicago.edu>
Signed-off-by: ApostaC <yihua98@uchicago.edu>
Signed-off-by: ApostaC <yihua98@uchicago.edu>
@ApostaC
Copy link
Collaborator Author

ApostaC commented Oct 9, 2025

@simon-mo Sorry, I was travelling in the past 2 weeks.
Just addressed all the comments from you and the review bots. Would you mind taking a look again at this PR? Thanks!

@ApostaC
Copy link
Collaborator Author

ApostaC commented Oct 16, 2025

Hey @NickLucche , would you like to take a quick look at this PR? Seems like Simon is pretty busy these days.

Copy link
Collaborator

@simon-mo simon-mo left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Add a test?

Signed-off-by: ApostaC <yihua98@uchicago.edu>
Signed-off-by: ApostaC <yihua98@uchicago.edu>
Signed-off-by: ApostaC <yihua98@uchicago.edu>
Signed-off-by: ApostaC <yihua98@uchicago.edu>
Signed-off-by: ApostaC <yihua98@uchicago.edu>
@KuntaiDu KuntaiDu enabled auto-merge (squash) October 24, 2025 22:22
@KuntaiDu KuntaiDu merged commit 83f478b into vllm-project:main Oct 25, 2025
51 checks passed
kingsmad pushed a commit to kingsmad/vllm that referenced this pull request Oct 25, 2025
rohin-garg pushed a commit to rohin-garg/vllm that referenced this pull request Oct 25, 2025
0xrushi pushed a commit to 0xrushi/vllm that referenced this pull request Oct 26, 2025
…llm-project#25542)

Signed-off-by: ApostaC <yihua98@uchicago.edu>
Signed-off-by: 0xrushi <6279035+0xrushi@users.noreply.github.com>
0xrushi pushed a commit to 0xrushi/vllm that referenced this pull request Oct 26, 2025
…llm-project#25542)

Signed-off-by: ApostaC <yihua98@uchicago.edu>
Signed-off-by: 0xrushi <6279035+0xrushi@users.noreply.github.com>
@sammshen sammshen mentioned this pull request Oct 31, 2025
5 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

kv-connector ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants