Commit History

Author SHA1 Message Date
  AlpinDale 0f4a9ee77b quantized lm_head (#582) 7 months ago
  AlpinDale 8a6e83b52e feat: fully sharded QKVParallelLinearWithLora support 7 months ago
  AlpinDale c975bba905 fix: sharded state loader with lora 7 months ago
  AlpinDale 9e73559eba make use of batched rotary embedding kernels to support long context lora 8 months ago
  AlpinDale e87c32bed3 feat: full tensor parallel for LoRA layers (#545) 8 months ago
  AlpinDale b178ae4b4a chore: generalize linear_method to be quant_method (#540) 8 months ago
  AlpinDale 9d81716bfd [v0.5.3] Release Candidate (#388) 10 months ago
  AlpinDale f8dfac6372 chore: attention refactor and upstream sync apr01 (#365) 1 year ago
  AlpinDale e42a78381a feat: switch from pylint to ruff (#322) 1 year ago
  AlpinDale a1d8ab9f3e fix: lora on quantized models (barred gguf) (#292) 1 year ago
  AlpinDale c0aac15421 feat: S-LoRA support (#222) 1 year ago