-
Notifications
You must be signed in to change notification settings - Fork 713
docs: update TensorRT-LLM commit for NIXL to v1.2.0rc2 #4582
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Update the TensorRT-LLM commit reference from 'main' to 'v1.2.0rc2' in: - docs/backends/trtllm/kv-cache-transfer.md - docs/backends/trtllm/multimodal_epd.md This ensures users build with the correct TensorRT-LLM version for NIXL and EPD features. Signed-off-by: Dan Gil <dagil@nvidia.com>
|
/ok to test 620889e |
WalkthroughTwo TensorRT-LLM backend documentation files have been updated to pin the commit version from main branch to v1.2.0rc2 in build instructions and example commands. Changes
Estimated code review effort🎯 1 (Trivial) | ⏱️ ~3 minutes
Poem
Pre-merge checks✅ Passed checks (3 passed)
📜 Recent review detailsConfiguration used: Path: .coderabbit.yaml Review profile: CHILL Plan: Pro 📒 Files selected for processing (2)
🧰 Additional context used🧠 Learnings (3)📚 Learning: 2025-07-30T00:34:35.810ZApplied to files:
📚 Learning: 2025-07-22T10:22:28.972ZApplied to files:
📚 Learning: 2025-08-30T20:43:10.091ZApplied to files:
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (2)
🔇 Additional comments (2)
Tip 📝 Customizable high-level summaries are now available in beta!You can now customize how CodeRabbit generates the high-level summary in your pull requests — including its content, structure, tone, and formatting.
Example instruction:
Note: This feature is currently in beta for Pro-tier users, and pricing will be announced later. Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
Signed-off-by: Dan Gil <dagil@nvidia.com>
docs: update TensorRT-LLM commit to v1.2.0rc2
Summary
Updates the TensorRT-LLM commit reference from
maintov1.2.0rc2in the documentation for NIXL KV cache transfer and EPD (Encode-Prefill-Decode) flow features.Changes
docs/backends/trtllm/kv-cache-transfer.mdto use--tensorrtllm-commit v1.2.0rc2docs/backends/trtllm/multimodal_epd.mdto use--tensorrtllm-commit v1.2.0rc2Motivation
These features require a specific TensorRT-LLM version. Using
maincan lead to compatibility issues as the TensorRT-LLM main branch evolves. Pinning tov1.2.0rc2ensures users build with the correct tested version.Testing
Checklist
docs:prefix)-s)PR Link
Create PR from
dagil/fix-trtllm-commit-docs→mainBranch pushed to: https://github.com/ai-dynamo/dynamo/pull/new/dagil/fix-trtllm-commit-docs
Summary by CodeRabbit
✏️ Tip: You can customize this high-level summary in your review settings.