Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add arctic model support by adding w2 to all_reduce #6856

Merged
merged 7 commits into from
Dec 18, 2024

Conversation

pi314ever
Copy link
Contributor

As title says.

Default behavior of arctic model produces shape issues with AutoTP due to the MLP layer performing w2 * act(w1*w3). However, method provided to fix Mixtral-7x8b in #5257 does not work since the MLP for Arctic is also used within a ModuleList for the MoE. This results in MLP weights hiding behind individual experts as layers #.w#, which is not caught by the fix in #5257. This adds the check directly within replace, where it can check for actual layer names for the w2 key in the model to patch with all_reduce.

Signed-off-by: Daniel Huang <daniel1.huang@intel.com>
Signed-off-by: Daniel Huang <daniel1.huang@intel.com>
@pi314ever pi314ever force-pushed the arctic-enabling-upstream branch from 2c2084b to 96eb813 Compare December 11, 2024 23:24
@pi314ever
Copy link
Contributor Author

@microsoft-github-policy-service agree company="Intel"

@loadams loadams requested review from jeffra and removed request for awan-10 December 12, 2024 00:38
@jeffra
Copy link
Collaborator

jeffra commented Dec 16, 2024

@RezaYazdaniAminabadi @sfc-gh-reyazda can you take a look?

@loadams loadams merged commit 0b25630 into deepspeedai:master Dec 18, 2024
10 of 11 checks passed
siqi654321 pushed a commit to siqi654321/DeepSpeed that referenced this pull request Feb 7, 2025
As title says. 

Default behavior of arctic model produces shape issues with AutoTP due
to the MLP layer performing `w2 * act(w1*w3)`. However, method provided
to fix Mixtral-7x8b in deepspeedai#5257 does not work since the MLP for Arctic is
also used within a ModuleList for the MoE. This results in MLP weights
hiding behind individual experts as layers `#.w#`, which is not caught
by the fix in deepspeedai#5257. This adds the check directly within replace, where
it can check for actual layer names for the `w2` key in the model to
patch with `all_reduce`.

---------

Signed-off-by: Daniel Huang <daniel1.huang@intel.com>
Co-authored-by: Olatunji Ruwase <olruwase@microsoft.com>
Co-authored-by: Logan Adams <114770087+loadams@users.noreply.github.com>
Signed-off-by: siqi <siqi@tecorigin.com>
traincheck-team pushed a commit to traincheck-team/DeepSpeed that referenced this pull request Feb 9, 2025
As title says. 

Default behavior of arctic model produces shape issues with AutoTP due
to the MLP layer performing `w2 * act(w1*w3)`. However, method provided
to fix Mixtral-7x8b in deepspeedai#5257 does not work since the MLP for Arctic is
also used within a ModuleList for the MoE. This results in MLP weights
hiding behind individual experts as layers `#.w#`, which is not caught
by the fix in deepspeedai#5257. This adds the check directly within replace, where
it can check for actual layer names for the `w2` key in the model to
patch with `all_reduce`.

---------

Signed-off-by: Daniel Huang <daniel1.huang@intel.com>
Co-authored-by: Olatunji Ruwase <olruwase@microsoft.com>
Co-authored-by: Logan Adams <114770087+loadams@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants