-
Notifications
You must be signed in to change notification settings - Fork 80
feat: faster scaled_dot_product_attention for reactant #1452
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
646aa5e to
bf2b3e6
Compare
bf2b3e6 to
7211f5f
Compare
Contributor
Benchmark Results (ASV)
Benchmark PlotsA plot of the benchmark results have been uploaded as an artifact to the workflow run for this PR. |
1c64d78 to
7280a95
Compare
Member
Author
Comparing against
|
88c82a1 to
80da2e3
Compare
04a599e to
ecd4ca3
Compare
80da2e3 to
14cc8c7
Compare
327801c to
2cb1e29
Compare
f4b5605 to
4d2df40
Compare
4d2df40 to
d0a27ca
Compare
feat: update Lux and Qwen example to use faster sdpa fix: qwen3 model?? fix: transpose ordering feat: expose kwargs in bmm + documentation fix: repeat batch dim feat: eliminate transpose in sdpa feat: support bias in Reactant sdpa fix: reduce permutations fix: inv scale before chore: cleanup fix: zygote gradient for sdpa and batched_matmul chore: add some comments fix: throw a dimension mismatch error feat: allow a pass through feat: support gqa correctly feat: use GQA directly in Qwen3 model [Filtered to include only lib/LuxLib changes] Original commit: 88c82a1 fix: correct error messages feat: make is_causal a keyword argument test: add tests for generalized bmm fix: restrict LV uses with PermutedDimsArray docs: add attention documentation test: attention test fix: reshape test: attention test test: attention test fix: import test: fix sdpa feat: faster scaled_dot_product_attention for reactant feat: update Lux and Qwen example to use faster sdpa fix: qwen3 model?? fix: transpose ordering feat: expose kwargs in bmm + documentation fix: repeat batch dim feat: eliminate transpose in sdpa feat: support bias in Reactant sdpa fix: reduce permutations fix: inv scale before chore: cleanup fix: zygote gradient for sdpa and batched_matmul chore: add some comments fix: throw a dimension mismatch error feat: allow a pass through feat: support gqa correctly feat: use GQA directly in Qwen3 model feat: add is_causal to MHA layer feat: use implicit causal masking
d0a27ca to
c976d32
Compare
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Uh oh!
There was an error while loading. Please reload this page.