AlpinDale
|
50c2434267
move megatron to a top-level directory
|
9 months ago |
AlpinDale
|
41beab5dc1
add exllamav2 tensor paralell, fused MoE for GPTQ/AWQ
|
9 months ago |
AlpinDale
|
0f1399c135
feat: attention refactor part 2
|
9 months ago |
AlpinDale
|
d1786645a3
fix formatting
|
9 months ago |
AlpinDale
|
0299dd41f0
fix query shape in moe models
|
9 months ago |
AlpinDale
|
688d56993a
add logit scale for command-r
|
9 months ago |
AlpinDale
|
f1ea36a445
add some imports
|
9 months ago |
AlpinDale
|
f8dfac6372
chore: attention refactor and upstream sync apr01 (#365)
|
9 months ago |
AlpinDale
|
da223153c6
feat&fix: cohere support and missing GPU blocks (#333)
|
10 months ago |
AlpinDale
|
e42a78381a
feat: switch from pylint to ruff (#322)
|
10 months ago |
AlpinDale
|
e31c6f0b45
feat: refactor modeling logic and support more models (#274)
|
11 months ago |
AlpinDale
|
7d6ba53602
feat: fused top-k kernels for MoE (#273)
|
11 months ago |
AlpinDale
|
842912d022
feat: on-the-fly gguf conversion (#250)
|
11 months ago |
AlpinDale
|
d9b65e6c5f
feat: DeepSeek MoE support (#237)
|
11 months ago |