Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Triton SMEM] Add not-yet-landed usage of Triton SMEM feature with autotuning #72

Draft
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

plotfi
Copy link

@plotfi plotfi commented Aug 22, 2024

NOTE: This is an experiment, and a draft. Do not review.

The following change requires a private patchset that is not yet available outside of plotfi/triton#4

This patch adds usage of shared memory using the tl.local_copy and tl.gather operations for the TW (time bias) and PW (position bias) tensors for the forward pass kernel.

Autotuning is also hooked up to the usage of these shared memory operators

…totuning

The following change requires a private patchset that is not yet
available outside of plotfi/triton#4

This patch adds usage of shared memory using the tl.local_copy and
tl.gather operations for the TW (time bias) and PW (position bias)
tensors for the forward pass kernel.

Autotuning is also hooked up to the usage of these shared memory operators
@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Meta Open Source bot. label Aug 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Meta Open Source bot.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants