Commit History

Author SHA1 Message Date
  AlpinDale 1405051912 attention: add `AttentionState` abstraction (#863) 1 month ago
  AlpinDale 7a313483f1 chore: move update_flash_attn_metadata to attn backend (#731) 3 months ago
  AlpinDale 60b702a827 chore: register custom torch ops for flash-attn and flashinfer (#724) 3 months ago
  AlpinDale 24456206a9 fix: logit softcapping in flash-attn (#688) 4 months ago
  AlpinDale 7df7b8ca53 optimization: reduce end-to-end overhead from python obj allocation (#666) 4 months ago
  AlpinDale f1d0b77c92 [0.6.0] Release Candidate (#481) 4 months ago
  AlpinDale 9d81716bfd [v0.5.3] Release Candidate (#388) 8 months ago