AlpinDale
|
9d81716bfd
[v0.5.3] Release Candidate (#388)
|
8 months ago |
AlpinDale
|
9534fcfb7b
fix: build error
|
9 months ago |
AlpinDale
|
29c241c115
fix: explicitly disallow installation on non-linux platforms (#373)
|
9 months ago |
AlpinDale
|
3d6695cfbb
feat: add approximate gelu activation kernels (#370)
|
9 months ago |
AlpinDale
|
f8dfac6372
chore: attention refactor and upstream sync apr01 (#365)
|
9 months ago |
AlpinDale
|
e120404436
Revert "feat: CMake Build System Generator (#332)"
|
9 months ago |
AlpinDale
|
ad6802690f
feat: CMake Build System Generator (#332)
|
9 months ago |
AlpinDale
|
e42a78381a
feat: switch from pylint to ruff (#322)
|
10 months ago |
AlpinDale
|
c41462cfcd
feat: exllamav2 quantization (#305)
|
10 months ago |
sgsdxzy
|
94c1543cae
fix: typo in marlin kernel path (#295)
|
10 months ago |
AlpinDale
|
e0c35bb353
feat: bitsandbytes and `--load-in{4,8}bit` support (#294)
|
10 months ago |
AlpinDale
|
705821a7fe
feat: AQLM quantization support (#293)
|
10 months ago |
AlpinDale
|
13d850334e
fix: navi support (#283)
|
10 months ago |
AlpinDale
|
7d6ba53602
feat: fused top-k kernels for MoE (#273)
|
10 months ago |
AlpinDale
|
4360684667
fix: cuda version in wheel
|
10 months ago |
AlpinDale
|
f99eb2c874
fix: hadamard tensors not included in wheel
|
11 months ago |
AlpinDale
|
d9b65e6c5f
feat: DeepSeek MoE support (#237)
|
11 months ago |
AlpinDale
|
aebd68c632
feat: backport kernels (#235)
|
11 months ago |
AlpinDale
|
5d288aa76c
feat: add fast hadamard transformation kernels (#232)
|
11 months ago |
AlpinDale
|
c3a221eb02
feat: GGUF, QuIP#, and Marlin support (#228)
|
11 months ago |
AlpinDale
|
31c95011a6
feat: FP8 E5M2 KV Cache (#226)
|
11 months ago |
AlpinDale
|
641bb0f6e9
feat: add custom allreduce kernels (#224)
|
11 months ago |
AlpinDale
|
c0aac15421
feat: S-LoRA support (#222)
|
11 months ago |
AlpinDale
|
8fa608aeb7
feat: replace Ray with NCCL for control plane comms (#221)
|
11 months ago |
AlpinDale
|
fe70c6e8d5
feat: bump cuda and pytorch (#205)
|
1 year ago |
AlpinDale
|
7e72ce0a73
feat: mixtral tensor parallelism (#193)
|
1 year ago |
AlpinDale
|
801eda0b7a
feat: support GPTQ 2, 3, and 8bit quants (#181)
|
1 year ago |
AlpinDale
|
68c2083adb
fix includes in wheel
|
1 year ago |
AlpinDale
|
62b2c4119d
feat: re-write GPTQ and refactor exllama kernels (#152)
|
1 year ago |
AlpinDale
|
1334a833a4
feat: AMD ROCm support (#95)
|
1 year ago |