AlpinDale
|
9d7beaa5b9
chore: separate kv_scale into k_scale and v_scale
|
4 months ago |
AlpinDale
|
156f577f79
feat: switch from `PYBIND11_MODULE` to `TORCH_LIBRARY` (#569)
|
5 months ago |
AlpinDale
|
3bdeb3e116
fix: clang formatting for all kernels (#558)
|
5 months ago |
AlpinDale
|
251568470e
initial nvidia fp8 e4m3 for kv cache
|
5 months ago |
AlpinDale
|
8b56dc4347
dict -> torch.Tensor for blocks_to_swap
|
5 months ago |
AlpinDale
|
21ce19b3ea
blocks_to_copy dict -> torch.Tensor
|
5 months ago |
AlpinDale
|
2351a0e2cd
feat: FlashInfer backend for decoding phase (#548)
|
5 months ago |
AlpinDale
|
9d81716bfd
[v0.5.3] Release Candidate (#388)
|
8 months ago |
AlpinDale
|
f8dfac6372
chore: attention refactor and upstream sync apr01 (#365)
|
9 months ago |
AlpinDale
|
9810daa699
feat: INT8 KV Cache (#298)
|
10 months ago |
AlpinDale
|
8e1cd54497
fix: do not include fp8 for rocm (#271)
|
10 months ago |
AlpinDale
|
31c95011a6
feat: FP8 E5M2 KV Cache (#226)
|
11 months ago |
AlpinDale
|
8fa608aeb7
feat: replace Ray with NCCL for control plane comms (#221)
|
11 months ago |
AlpinDale
|
15a0454172
feat: FP8 KV Cache (#185)
|
1 year ago |
AlpinDale
|
b9b295d74e
chore: backlogs 1 (#191)
|
1 year ago |
AlpinDale
|
7d91e9e0f2
feat: CUDA graphs (#172)
|
1 year ago |
AlpinDale
|
1334a833a4
feat: AMD ROCm support (#95)
|
1 year ago |
AlpinDale
|
3d72f05c7b
feat: flattened 1D tensor -> 2D tensor (#85)
|
1 year ago |
AlpinDale
|
32844c1522
add GELU kernels and remove compile bloat
|
1 year ago |
AlpinDale
|
76b2e4a445
Merge dev branch into main (#7)
|
1 year ago |
AlpinDale
|
081545bde6
fix: various CUDA kernel tweaks
|
1 year ago |
AlpinDale
|
b8f4337c5b
chore: various fixes
|
1 year ago |
AlpinDale
|
a409431c40
feat: draft for cuda kernels
|
1 year ago |