-
Notifications
You must be signed in to change notification settings - Fork 332
[Refactor] Dynamic registration of FP8 data type for compatibility with older PyTorch versions #1197
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
…nversion handling in dtypes.py
|
👋 Hi! Thank you for contributing to the TileLang project. Please remember to run We appreciate you taking this step! Our team will review your contribution, and we look forward to your awesome work! 🚀 |
WalkthroughTwo independent modifications: a CLI parameter default adjustment in a latency measurement script (dim from 512 to 256), and internal dtype mapping refactoring to support conditional FP8 dtype registration based on torch availability. Changes
Estimated code review effort🎯 2 (Simple) | ⏱️ ~12 minutes
Possibly related PRs
Poem
Pre-merge checks and finishing touches✅ Passed checks (2 passed)
✨ Finishing touches
🧪 Generate unit tests (beta)
📜 Recent review detailsConfiguration used: CodeRabbit UI Review profile: CHILL Plan: Pro 📒 Files selected for processing (2)
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (2)
🔇 Additional comments (3)
Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
|
@codex review |
|
Codex Review: Didn't find any major issues. Nice work! ℹ️ About Codex in GitHubYour team has set up Codex to review pull requests in this repo. Reviews are triggered when you
If Codex has suggestions, it will comment; otherwise it will react with 👍. Codex can also answer questions or update the PR. Try commenting "@codex address that feedback". |
…th older PyTorch versions (tile-ai#1197)
This pull request introduces a dynamic approach to handling FP8 data types in the
tilelang/language/v2/dtypes.pymodule and updates a default argument inmaint/gemm_v2/latency_mha_fwd.py. The main improvement is making FP8 dtype support conditional on their availability in the installed version of PyTorch, improving compatibility and robustness.Dynamic FP8 dtype handling and default parameter update:
Data type conversion improvements:
_dtype_cvt_baseand moving FP8-related types to a dynamic section. This ensures FP8 types are only added to_dtype_cvtif they exist in the current PyTorch installation, preventing import errors on unsupported versions. [1] [2]Parameter default update:
--dimargument inmaint/gemm_v2/latency_mha_fwd.pyfrom512to256.Summary by CodeRabbit
Chores
Improvements