Historique des commits

Auteur SHA1 Message Date
  Tri Dao f1a73d0740 Run isort and black on python files il y a 1 an
  Tri Dao 5d079fdd7a [Triton] Fix benchmark_causal, mention Triton version il y a 1 an
  Tri Dao 6b5f271c6d [Triton] Avoid einops repeat by using Tensor.expand il y a 2 ans
  Tri Dao b8ccd20098 [Triton] Fix variable name from qkv to kv (h/t FrankZijlstra) il y a 2 ans
  Tri Dao 908a5b2244 Set num_warps=4 for headdim=64 in Triton fw (h/t Michael Benesty) il y a 2 ans
  Tri Dao 7479757191 Fix pipelining bug in Triton bwd with bias_type=matrix il y a 2 ans
  Tri Dao 557781933d Parallelize CUDA bwd along seqlen_k instead of seqlen_q il y a 2 ans
  Tri Dao 62025e1aff Fix more race condition in Triton bwd when there's bias il y a 2 ans
  Tri Dao ff78ea4123 Fix race condition in Triton bwd when there's bias il y a 2 ans
  Tri Dao 86862cfd7b Implement attention bias for Triton version il y a 2 ans
  Tri Dao 470010f59b Fix race condition for Triton bwd for headdim 48 and 96 il y a 2 ans
  Tri Dao aacc10fbab Fix race condition in Triton bwd for non-po2 headdims il y a 2 ans
  Tri Dao 1fb12afdfb Avoid memcpy in the Triton bwd il y a 2 ans
  Tri Dao 731f154de3 Fix race conditions in the Triton bwd for headdim=64 il y a 2 ans
  Tri Dao 9b0bc97872 Fix race condition in Triton fwd il y a 2 ans
  Tri Dao 215930bce3 Fix EVEN_M & EVEN_HEADDIM for headdim=40 in Triton bwd il y a 2 ans
  Tri Dao 4f81aff46e Add debug_barrier for all headdims in Triton bwd il y a 2 ans
  Tri Dao bedcbd6a71 Disable some autotune configs that give wrong results in Triton bwd il y a 2 ans
  Tri Dao e78d509c64 [WIP] Support all head dimensions up to 128 in the Triton bwd il y a 2 ans
  Tri Dao 008951f1d9 Support all head dimensions up to 128 in the Triton fwd il y a 2 ans
  Tri Dao b910bf14c1 Support arbitrary seqlens (both q & k) in Triton bwd il y a 2 ans
  Tri Dao dc55469355 Support arbitrary seqlen_k in Triton bwd il y a 2 ans
  Tri Dao d11341fd1a Fix Triton fwd to support seqlen not multiples of 128 il y a 2 ans
  Tri Dao b0c0db81f6 Implement FlashAttention in Triton il y a 2 ans