.. |
__init__.py
|
5884e0b904
add bitnetforcausallm support
|
6 miesięcy temu |
arctic.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
baichuan.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
bitnet.py
|
9bbc75d2e3
wip
|
6 miesięcy temu |
bloom.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
chatglm.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
commandr.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
dbrx.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
decilm.py
|
50b7c13db0
refactor: attention selector (#552)
|
6 miesięcy temu |
deepseek.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
falcon.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
gemma.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
gpt2.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
gpt_bigcode.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
gpt_j.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
gpt_neox.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
internlm2.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
jais.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
llama.py
|
ac79d115b3
add guards for prefix caching, fp8, chunked, etc
|
6 miesięcy temu |
llama_embedding.py
|
50b7c13db0
refactor: attention selector (#552)
|
6 miesięcy temu |
llava.py
|
24a2d9c2c8
minor llava refactoring
|
6 miesięcy temu |
minicpm.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
mixtral.py
|
ac79d115b3
add guards for prefix caching, fp8, chunked, etc
|
6 miesięcy temu |
mixtral_quant.py
|
ac79d115b3
add guards for prefix caching, fp8, chunked, etc
|
6 miesięcy temu |
mpt.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
olmo.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
opt.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
orion.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
phi.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
phi3_small.py
|
696f2cd59c
add phi3_small support with blocksparse attention
|
6 miesięcy temu |
qwen.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
qwen2.py
|
ac79d115b3
add guards for prefix caching, fp8, chunked, etc
|
6 miesięcy temu |
qwen2_moe.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
stablelm.py
|
656459fd84
make fp8_e4m3 work on nvidia
|
6 miesięcy temu |
starcoder2.py
|
ac79d115b3
add guards for prefix caching, fp8, chunked, etc
|
6 miesięcy temu |
vlm_base.py
|
f970f3f3fb
add base class for VLMs
|
6 miesięcy temu |
xverse.py
|
ac79d115b3
add guards for prefix caching, fp8, chunked, etc
|
6 miesięcy temu |