3 results found Sort:
A collection of memory efficient attention operators implemented in the Triton language.
Created
2023-10-07
47 commits to main branch, last one 8 months ago
Triton implementation of FlashAttention2 that adds Custom Masks.
Created
2024-07-20
18 commits to main branch, last one 6 months ago
Triton implement of bi-directional (non-causal) linear attention
Created
2024-11-20
32 commits to main branch, last one 14 days ago