Pinned Loading
Repositories
Showing 8 of 8 repositories
- flash-linear-attention Public
š Efficient implementations of state-of-the-art linear attention models in Torch and Triton
fla-org/flash-linear-attentionās past year of commit activity - native-sparse-attention Public
š³ Efficient Triton implementations for "Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention"
fla-org/native-sparse-attentionās past year of commit activity - flash-hybrid-attention Public
fla-org/flash-hybrid-attentionās past year of commit activity - flash-bidirectional-linear-attention Public
Triton implement of bi-directional (non-causal) linear attention
fla-org/flash-bidirectional-linear-attentionās past year of commit activity