AlpinDale
|
ba371fbbbd
feat: AWQ marlin kernels (#603)
|
5 months ago |
AlpinDale
|
19340b672e
chore: improve min_capability checking for `compressed-tensors`
|
5 months ago |
AlpinDale
|
500f3b654f
fix: support bias term in compressed-tensors quant
|
5 months ago |
AlpinDale
|
058e629f8e
chore: refactor marlin python utils
|
5 months ago |
AlpinDale
|
ddb3323f94
refactor: have w8a8 compressed tensors use `process_weights_after_load` for fp8
|
6 months ago |
AlpinDale
|
b6ff0623a6
chore: clean up branding
|
6 months ago |
AlpinDale
|
9e75007c40
chore: update w4a16 to wna16 and support w8a16
|
6 months ago |