.. |
__init__.py
|
d9f4c36edd
feat: Medusa speculative decoding support (#590)
|
6 meses atrás |
arctic.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
baichuan.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
bloom.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
chatglm.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
clip.py
|
c11a8bdaad
fix: calculate max number of multi-modal tokens automatically
|
6 meses atrás |
commandr.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
dbrx.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
decilm.py
|
56e0b8223c
chore: add base class for LoRA-supported models
|
7 meses atrás |
deepseek.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
deepseek_v2.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
falcon.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
gemma.py
|
05e45aeb53
fix: dtype mismatch for paligemma
|
6 meses atrás |
gemma2.py
|
5761ef8c35
feat: gemma-2 support
|
6 meses atrás |
gpt2.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
gpt_bigcode.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
gpt_j.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
gpt_neox.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
interfaces.py
|
4f7d212b70
feat: remove vision language config
|
6 meses atrás |
internlm2.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
jais.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
jamba.py
|
ebba0d9226
fix: mamba cache cuda graph padding
|
6 meses atrás |
llama.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
llama_embedding.py
|
50b7c13db0
refactor: attention selector (#552)
|
7 meses atrás |
llava.py
|
7e99578712
fix: cleanup validation and update docs for vlm
|
6 meses atrás |
llava_next.py
|
0ab35652d3
fix: llava 1.6 feature size calculation
|
6 meses atrás |
medusa.py
|
d9f4c36edd
feat: Medusa speculative decoding support (#590)
|
6 meses atrás |
minicpm.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
mixtral.py
|
9622c59f8f
chore: support 2D input shape in MoE layer
|
6 meses atrás |
mixtral_quant.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
mlp_speculator.py
|
db73f03cdc
fix: use ParallelLMHead for MLPSpeculator
|
6 meses atrás |
mpt.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
olmo.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
opt.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
orion.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
paligemma.py
|
05e45aeb53
fix: dtype mismatch for paligemma
|
6 meses atrás |
phi.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
phi3_small.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
phi3v.py
|
7e99578712
fix: cleanup validation and update docs for vlm
|
6 meses atrás |
qwen.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
qwen2.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
qwen2_moe.py
|
9622c59f8f
chore: support 2D input shape in MoE layer
|
6 meses atrás |
stablelm.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
starcoder2.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |
utils.py
|
4599c98f99
feat: dynamic image size support for VLMs
|
6 meses atrás |
xverse.py
|
0f4a9ee77b
quantized lm_head (#582)
|
6 meses atrás |