AlpinDale
|
99680b2d23
feat: soft prompts (#589)
|
há 5 meses atrás |
AlpinDale
|
5761ef8c35
feat: gemma-2 support
|
há 5 meses atrás |
AlpinDale
|
0f4a9ee77b
quantized lm_head (#582)
|
há 5 meses atrás |
AlpinDale
|
8a6e83b52e
feat: fully sharded QKVParallelLinearWithLora support
|
há 5 meses atrás |
AlpinDale
|
c975bba905
fix: sharded state loader with lora
|
há 6 meses atrás |
AlpinDale
|
9e73559eba
make use of batched rotary embedding kernels to support long context lora
|
há 6 meses atrás |
AlpinDale
|
e87c32bed3
feat: full tensor parallel for LoRA layers (#545)
|
há 6 meses atrás |
AlpinDale
|
b178ae4b4a
chore: generalize linear_method to be quant_method (#540)
|
há 6 meses atrás |
AlpinDale
|
9d81716bfd
[v0.5.3] Release Candidate (#388)
|
há 8 meses atrás |
AlpinDale
|
f8dfac6372
chore: attention refactor and upstream sync apr01 (#365)
|
há 10 meses atrás |
AlpinDale
|
e42a78381a
feat: switch from pylint to ruff (#322)
|
há 10 meses atrás |
AlpinDale
|
a1d8ab9f3e
fix: lora on quantized models (barred gguf) (#292)
|
há 11 meses atrás |
AlpinDale
|
c0aac15421
feat: S-LoRA support (#222)
|
há 1 ano atrás |