Skip to content

Pinned Loading

  1. flash-linear-attention flash-linear-attention Public

    🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton

    Python 2.1k 128

  2. flame flame Public

    🔥 A minimal training framework for scaling FLA models

    Python 75 14

  3. native-sparse-attention native-sparse-attention Public

    🐳 Efficient Triton implementations for "Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention"

    Python 541 27

Repositories

Showing 7 of 7 repositories
  • flash-linear-attention Public

    🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton

    fla-org/flash-linear-attention’s past year of commit activity
    Python 2,088 MIT 128 23 4 Updated Mar 13, 2025
  • native-sparse-attention Public

    🐳 Efficient Triton implementations for "Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention"

    fla-org/native-sparse-attention’s past year of commit activity
    Python 541 MIT 27 8 0 Updated Mar 13, 2025
  • flame Public

    🔥 A minimal training framework for scaling FLA models

    fla-org/flame’s past year of commit activity
    Python 75 MIT 14 2 0 Updated Mar 13, 2025
  • fla-zoo Public

    Flash-Linear-Attention models beyond language

    fla-org/fla-zoo’s past year of commit activity
    Python 7 0 0 0 Updated Mar 12, 2025
  • fla-org/flash-hybrid-attention’s past year of commit activity
    7 0 0 0 Updated Mar 5, 2025
  • evals Public
    fla-org/evals’s past year of commit activity
    0 0 0 0 Updated Feb 19, 2025
  • flash-bidirectional-linear-attention Public

    Triton implement of bi-directional (non-causal) linear attention

    fla-org/flash-bidirectional-linear-attention’s past year of commit activity
    Python 43 MIT 1 0 0 Updated Feb 4, 2025