Skip to content

Conversation

@fmassa
Copy link
Contributor

@fmassa fmassa commented Sep 4, 2025

This PR is in preparation to switching the flags for enabling fusing nn.Linear into a single op, as well as the new add_alias_v2 flag, as well as for merging #125

This is also helpful for identifying potential problems with the cost model for comms / compute.

This can also serve as an example of how users can customize precise shardings of their model, without having to change user-code. We might want to add helper ops so that this customization can happen directly in the model code as well, but I'm keeping it completely separate for now.

I've also removed the embedding constraint for 1d -- it is not needed anymore now that we have a cost model for memory-bound ops (like Embedding).

@fmassa fmassa requested a review from ezyang September 4, 2025 08:43
@meta-cla meta-cla bot added the CLA Signed This label is managed by the Meta Open Source bot. label Sep 4, 2025
@fmassa fmassa merged commit 273f54c into main Sep 7, 2025
5 of 6 checks passed
@fmassa fmassa deleted the fmassa/add_constraints branch September 7, 2025 05:42
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Meta Open Source bot.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants