Commit History

Author SHA1 Message Date
  AlpinDale 32bdbd1ee4 chore: add fp8 support to `reshape_and_cache_flash` 4 months ago
  AlpinDale 9d7beaa5b9 chore: separate kv_scale into k_scale and v_scale 4 months ago
  AlpinDale 156f577f79 feat: switch from `PYBIND11_MODULE` to `TORCH_LIBRARY` (#569) 5 months ago
  AlpinDale 3bdeb3e116 fix: clang formatting for all kernels (#558) 5 months ago
  AlpinDale 251568470e initial nvidia fp8 e4m3 for kv cache 5 months ago
  AlpinDale 8b56dc4347 dict -> torch.Tensor for blocks_to_swap 5 months ago
  AlpinDale 21ce19b3ea blocks_to_copy dict -> torch.Tensor 5 months ago
  AlpinDale 2351a0e2cd feat: FlashInfer backend for decoding phase (#548) 5 months ago
  AlpinDale 9d81716bfd [v0.5.3] Release Candidate (#388) 8 months ago
  AlpinDale f8dfac6372 chore: attention refactor and upstream sync apr01 (#365) 9 months ago
  AlpinDale 9810daa699 feat: INT8 KV Cache (#298) 10 months ago
  AlpinDale 8e1cd54497 fix: do not include fp8 for rocm (#271) 10 months ago
  AlpinDale 31c95011a6 feat: FP8 E5M2 KV Cache (#226) 11 months ago
  AlpinDale 8fa608aeb7 feat: replace Ray with NCCL for control plane comms (#221) 11 months ago
  AlpinDale 15a0454172 feat: FP8 KV Cache (#185) 1 year ago
  AlpinDale b9b295d74e chore: backlogs 1 (#191) 1 year ago
  AlpinDale 7d91e9e0f2 feat: CUDA graphs (#172) 1 year ago
  AlpinDale 1334a833a4 feat: AMD ROCm support (#95) 1 year ago
  AlpinDale 3d72f05c7b feat: flattened 1D tensor -> 2D tensor (#85) 1 year ago
  AlpinDale 32844c1522 add GELU kernels and remove compile bloat 1 year ago
  AlpinDale 76b2e4a445 Merge dev branch into main (#7) 1 year ago
  AlpinDale 081545bde6 fix: various CUDA kernel tweaks 1 year ago
  AlpinDale b8f4337c5b chore: various fixes 1 year ago
  AlpinDale a409431c40 feat: draft for cuda kernels 1 year ago