3 results found Sort:

A collection of memory efficient attention operators implemented in the Triton language.
Created 2023-10-07
47 commits to main branch, last one 8 months ago
Triton implementation of FlashAttention2 that adds Custom Masks.
Created 2024-07-20
18 commits to main branch, last one 5 months ago
Triton implement of bi-directional (non-causal) linear attention
Created 2024-11-20
31 commits to main branch, last one 21 days ago