Commit Verlauf

Autor SHA1 Nachricht Datum
  AlpinDale 4d4e767838 ci: take one of fixing lint issues vor 5 Monaten
  AlpinDale 98f9dbd734 feat: Triton Kernels for Punica (#613) vor 5 Monaten
  AlpinDale 848731f527 chore: add punica sizes for mistral nemo vor 5 Monaten
  AlpinDale d907f20908 feat: support collective comms in XLA devices, e.g. TPUs vor 5 Monaten
  AlpinDale 99680b2d23 feat: soft prompts (#589) vor 5 Monaten
  AlpinDale 5761ef8c35 feat: gemma-2 support vor 5 Monaten
  AlpinDale 0f4a9ee77b quantized lm_head (#582) vor 6 Monaten
  AlpinDale 8a6e83b52e feat: fully sharded QKVParallelLinearWithLora support vor 6 Monaten
  AlpinDale c975bba905 fix: sharded state loader with lora vor 6 Monaten
  AlpinDale 9e73559eba make use of batched rotary embedding kernels to support long context lora vor 6 Monaten
  AlpinDale e87c32bed3 feat: full tensor parallel for LoRA layers (#545) vor 6 Monaten
  AlpinDale b178ae4b4a chore: generalize linear_method to be quant_method (#540) vor 6 Monaten
  AlpinDale 9d81716bfd [v0.5.3] Release Candidate (#388) vor 9 Monaten
  AlpinDale f8dfac6372 chore: attention refactor and upstream sync apr01 (#365) vor 10 Monaten
  AlpinDale e42a78381a feat: switch from pylint to ruff (#322) vor 11 Monaten
  AlpinDale a1d8ab9f3e fix: lora on quantized models (barred gguf) (#292) vor 11 Monaten
  AlpinDale c0aac15421 feat: S-LoRA support (#222) vor 1 Jahr