Commit History

Author SHA1 Message Date
  AlpinDale 50c2434267 move megatron to a top-level directory 9 months ago
  AlpinDale 41beab5dc1 add exllamav2 tensor paralell, fused MoE for GPTQ/AWQ 9 months ago
  AlpinDale 0f1399c135 feat: attention refactor part 2 9 months ago
  AlpinDale d1786645a3 fix formatting 9 months ago
  AlpinDale 0299dd41f0 fix query shape in moe models 9 months ago
  AlpinDale 688d56993a add logit scale for command-r 9 months ago
  AlpinDale f1ea36a445 add some imports 9 months ago
  AlpinDale f8dfac6372 chore: attention refactor and upstream sync apr01 (#365) 9 months ago
  AlpinDale da223153c6 feat&fix: cohere support and missing GPU blocks (#333) 10 months ago
  AlpinDale e42a78381a feat: switch from pylint to ruff (#322) 10 months ago
  AlpinDale e31c6f0b45 feat: refactor modeling logic and support more models (#274) 11 months ago
  AlpinDale 7d6ba53602 feat: fused top-k kernels for MoE (#273) 11 months ago
  AlpinDale 842912d022 feat: on-the-fly gguf conversion (#250) 11 months ago
  AlpinDale d9b65e6c5f feat: DeepSeek MoE support (#237) 11 months ago