Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Sparse attention window size bug fix #60

Open
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

mklasby
Copy link

@mklasby mklasby commented Apr 3, 2025

Cherry-picked hotfix from #39 originally by @minminsun which includes critical bug fixes for sparse attention kernel window size. Original PR was closed since the files in this fork were synced with upstream. Notably, the dual-chunk attention PR in vLLM requires this fix.

@mklasby mklasby force-pushed the sparse-attn-hotfix branch from 5ac198c to 22de4f3 Compare April 3, 2025 18:26
@mklasby mklasby force-pushed the sparse-attn-hotfix branch from 22de4f3 to 5a93f56 Compare April 3, 2025 18:31
@minminsun
Copy link

Thanks a lot for cherry-picking this hotfix! @mklasby

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants