AlpinDale
|
4ec08af18b
chore: update fused MoE weight loading (#700)
|
4 miesięcy temu |
AlpinDale
|
0e558e9b2f
fix: loading chameleon model with TP>1 (#695)
|
4 miesięcy temu |
AlpinDale
|
3f712cd287
feat: add progress bar for loading individual weight modules (#640)
|
4 miesięcy temu |
AlpinDale
|
f1d0b77c92
[0.6.0] Release Candidate (#481)
|
4 miesięcy temu |
AlpinDale
|
9d81716bfd
[v0.5.3] Release Candidate (#388)
|
8 miesięcy temu |
AlpinDale
|
f8dfac6372
chore: attention refactor and upstream sync apr01 (#365)
|
9 miesięcy temu |
AlpinDale
|
da223153c6
feat&fix: cohere support and missing GPU blocks (#333)
|
9 miesięcy temu |
AlpinDale
|
e42a78381a
feat: switch from pylint to ruff (#322)
|
10 miesięcy temu |
AlpinDale
|
e31c6f0b45
feat: refactor modeling logic and support more models (#274)
|
10 miesięcy temu |
AlpinDale
|
7d6ba53602
feat: fused top-k kernels for MoE (#273)
|
10 miesięcy temu |
AlpinDale
|
224b87b484
feat: add fused mixtral moe support (#238)
|
10 miesięcy temu |
AlpinDale
|
842912d022
feat: on-the-fly gguf conversion (#250)
|
11 miesięcy temu |
AlpinDale
|
d9b65e6c5f
feat: DeepSeek MoE support (#237)
|
11 miesięcy temu |
AlpinDale
|
c3a221eb02
feat: GGUF, QuIP#, and Marlin support (#228)
|
11 miesięcy temu |
AlpinDale
|
8fa608aeb7
feat: replace Ray with NCCL for control plane comms (#221)
|
11 miesięcy temu |
AlpinDale
|
97f37c1cb2
fix: use tensor parallel for quantized mixtral (#213)
|
11 miesięcy temu |
AlpinDale
|
193287b2ef
fix: mixtral unused import
|
1 rok temu |
AlpinDale
|
53d391e1f2
merge 'dev' into 'main'
|
1 rok temu |
AlpinDale
|
7e72ce0a73
feat: mixtral tensor parallelism (#193)
|
1 rok temu |
AlpinDale
|
b9b295d74e
chore: backlogs 1 (#191)
|
1 rok temu |
AlpinDale
|
f013d714c0
chore: merge dev branch into main (#177)
|
1 rok temu |
g4rg
|
fe57bb7ad2
feat: add rope scaling to mixtral (#174)
|
1 rok temu |
AlpinDale
|
7d91e9e0f2
feat: CUDA graphs (#172)
|
1 rok temu |
AlpinDale
|
725be3e0de
feat: mixtral HF with expert parallelism (#167)
|
1 rok temu |
AlpinDale
|
730357c7d5
chore: implement lazy module loader for models (#165)
|
1 rok temu |
AlpinDale
|
2755a48d51
merge dev branch into main (#153)
|
1 rok temu |
AlpinDale
|
87277c76e4
feat: Mixtral 8x7B support (#155)
|
1 rok temu |