Skip to content

JIA-Lab-research/Block-Approximate-Sparse-Attention

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

2 Commits
Β 
Β 
Β 
Β 

Repository files navigation

πŸš€ BA-Att: Block Approximate Sparse Attention

Efficient Long-Context Modeling in Diffusion Language Models
πŸ“„ CVPR 2026 (Findings)


πŸ”₯ Highlights

  • πŸš€ Up to 6.95Γ— speedup over FlashAttention
  • ⚑ Training-free sparse attention (no finetuning required)
  • 🧠 Maintains near full-attention performance at 50% sparsity
  • πŸŽ₯ Strong generalization across language, multimodal, and video generation

πŸ“Œ Overview

We propose Block Approximate Sparse Attention (BA-Att), a training-free block-sparse attention framework for Diffusion Language Models (DLMs).

Unlike prior works relying on fixed patterns, BA-Att:

  • Performs selection in downsampled space
  • Uses norm-based ranking to reduce approximation error
  • Applies covariance compensation for accuracy recovery

🧠 Framework Overview

πŸ“¦ Code

🚧 Code is coming soon!

We are currently cleaning and organizing the codebase.

About

[CVPR26] Official implementation of CVPR2026 Findings Paper 'Efficient Long-Context Modeling in Diffusion Language Models via Block Approximate Sparse Attention'

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors