.. |
aqlm
|
156f577f79
feat: switch from `PYBIND11_MODULE` to `TORCH_LIBRARY` (#569)
|
há 7 meses atrás |
autoquant
|
156f577f79
feat: switch from `PYBIND11_MODULE` to `TORCH_LIBRARY` (#569)
|
há 7 meses atrás |
awq
|
156f577f79
feat: switch from `PYBIND11_MODULE` to `TORCH_LIBRARY` (#569)
|
há 7 meses atrás |
compressed_tensors
|
156f577f79
feat: switch from `PYBIND11_MODULE` to `TORCH_LIBRARY` (#569)
|
há 7 meses atrás |
cutlass_w8a8
|
b03b4d4c8c
fix: compute cutlass 3.x epilogues in fp32 instead of 16
|
há 7 meses atrás |
exl2
|
156f577f79
feat: switch from `PYBIND11_MODULE` to `TORCH_LIBRARY` (#569)
|
há 7 meses atrás |
fp8
|
ad24e74a99
feat: FP8 weight-only quantization support for Ampere GPUs
|
há 6 meses atrás |
gguf
|
9d81716bfd
[v0.5.3] Release Candidate (#388)
|
há 10 meses atrás |
gptq
|
156f577f79
feat: switch from `PYBIND11_MODULE` to `TORCH_LIBRARY` (#569)
|
há 7 meses atrás |
gptq_marlin
|
156f577f79
feat: switch from `PYBIND11_MODULE` to `TORCH_LIBRARY` (#569)
|
há 7 meses atrás |
int8_kvcache
|
9810daa699
feat: INT8 KV Cache (#298)
|
há 1 ano atrás |
marlin
|
1587fab5de
fix: cuda version check for mma warning suppression
|
há 7 meses atrás |
quip
|
156f577f79
feat: switch from `PYBIND11_MODULE` to `TORCH_LIBRARY` (#569)
|
há 7 meses atrás |
squeezellm
|
156f577f79
feat: switch from `PYBIND11_MODULE` to `TORCH_LIBRARY` (#569)
|
há 7 meses atrás |
quant_ops.h
|
ad24e74a99
feat: FP8 weight-only quantization support for Ampere GPUs
|
há 6 meses atrás |