Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Adds a flash attention layer. #26

Merged
merged 1 commit into from
Aug 4, 2023
Merged

Conversation

markblee
Copy link
Contributor

@markblee markblee commented Aug 4, 2023

  • Adds basic support for attention logit biases in the pallas attention op.
  • Adds a benchmarking script for fwd+bwd.
  • Adds FlashAttention layer which is (mostly) a drop-in replacement for MultiheadAttention (with caveats, see docstrings).
  • Updates licenses/acknowledgements.

@markblee markblee requested a review from ruomingp August 4, 2023 17:09
@markblee markblee merged commit 7e45e86 into apple:main Aug 4, 2023
@markblee markblee deleted the markblee/triton branch August 4, 2023 18:10
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant