Skip to content

Add Sparse Attention Kernels on Hopper#98

Merged
interestingLSY merged 4 commits intomainfrom
open-source-h
Sep 29, 2025
Merged

Add Sparse Attention Kernels on Hopper#98
interestingLSY merged 4 commits intomainfrom
open-source-h

Conversation

@interestingLSY
Copy link
Copy Markdown
Collaborator

Add sparse attention kernels for DeepSeek-V3.2, for the Hopper architecture. This includes:

  • Sparse attention kernel for the prefill stage
  • Sparse attention kernel for the decoding stage, with FP8 KV cache and paged attention support

@interestingLSY interestingLSY merged commit 1794455 into main Sep 29, 2025
@interestingLSY interestingLSY deleted the open-source-h branch September 29, 2025 10:06
@faruknane
Copy link
Copy Markdown

@interestingLSY does it support Blackwell architecture as well?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants