Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix Llama sdpa/ flash attention + adapters #722

Merged
merged 3 commits into from
Jul 20, 2024
Merged

Fix Llama sdpa/ flash attention + adapters #722

merged 3 commits into from
Jul 20, 2024

Commits on Jul 15, 2024

  1. Fix Llama attention integration

    calpt committed Jul 15, 2024
    Configuration menu
    Copy the full SHA
    ada2eeb View commit details
    Browse the repository at this point in the history

Commits on Jul 17, 2024

  1. import fix

    calpt committed Jul 17, 2024
    Configuration menu
    Copy the full SHA
    d17dbdb View commit details
    Browse the repository at this point in the history

Commits on Jul 20, 2024

  1. fix FA dtype

    calpt committed Jul 20, 2024
    Configuration menu
    Copy the full SHA
    844449b View commit details
    Browse the repository at this point in the history