Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Checking if attention mask is present for ignoring pad tokens in ffn. #1188

Conversation

ShashankMosaicML
Copy link
Contributor

If attention_mask is None, that means we do not have padding tokens, and hence we do not need to unpad the input. Currently, if attention_mask is None then unpad_input(m, attention_mask) throws an error.

@ShashankMosaicML ShashankMosaicML marked this pull request as ready for review May 8, 2024 23:50
@ShashankMosaicML ShashankMosaicML enabled auto-merge (squash) May 9, 2024 00:03
@ShashankMosaicML ShashankMosaicML merged commit 0c7bc2a into mosaicml:main May 9, 2024
9 checks passed
@ShashankMosaicML ShashankMosaicML deleted the ignore_pad_tok_in_ffn_only_if_attn_mask branch May 9, 2024 00:14
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants