Commit History

Author SHA1 Message Date
  AlpinDale 99680b2d23 feat: soft prompts (#589) 5 months ago
  AlpinDale 5761ef8c35 feat: gemma-2 support 5 months ago
  AlpinDale 0f4a9ee77b quantized lm_head (#582) 5 months ago
  AlpinDale 8a6e83b52e feat: fully sharded QKVParallelLinearWithLora support 5 months ago
  AlpinDale c975bba905 fix: sharded state loader with lora 6 months ago
  AlpinDale 9e73559eba make use of batched rotary embedding kernels to support long context lora 6 months ago
  AlpinDale e87c32bed3 feat: full tensor parallel for LoRA layers (#545) 6 months ago
  AlpinDale b178ae4b4a chore: generalize linear_method to be quant_method (#540) 6 months ago
  AlpinDale 9d81716bfd [v0.5.3] Release Candidate (#388) 8 months ago
  AlpinDale f8dfac6372 chore: attention refactor and upstream sync apr01 (#365) 10 months ago
  AlpinDale e42a78381a feat: switch from pylint to ruff (#322) 10 months ago
  AlpinDale a1d8ab9f3e fix: lora on quantized models (barred gguf) (#292) 11 months ago
  AlpinDale c0aac15421 feat: S-LoRA support (#222) 1 year ago