AlpinDale 93cffaf446 add flash_attn back 7 mēneši atpakaļ
..
attention 93cffaf446 add flash_attn back 7 mēneši atpakaļ
common 9e73559eba make use of batched rotary embedding kernels to support long context lora 7 mēneši atpakaļ
distributed c58589318f remove the graph mode func 7 mēneši atpakaļ
endpoints fe431bb840 check for next port if current is unavailable 7 mēneši atpakaļ
engine 9e73559eba make use of batched rotary embedding kernels to support long context lora 7 mēneši atpakaļ
executor eaa06fdd14 fix some f-strings 7 mēneši atpakaļ
kv_quant e42a78381a feat: switch from pylint to ruff (#322) 1 gadu atpakaļ
lora 9e73559eba make use of batched rotary embedding kernels to support long context lora 7 mēneši atpakaļ
modeling f970f3f3fb add base class for VLMs 7 mēneši atpakaļ
processing 9e73559eba make use of batched rotary embedding kernels to support long context lora 7 mēneši atpakaļ
quantization 8e11259e90 missing triton autoconfig for rocm flash attn 7 mēneši atpakaļ
spec_decode 236be273e5 feat: tensor parallel speculative decoding (#554) 7 mēneši atpakaļ
task_handler 9e73559eba make use of batched rotary embedding kernels to support long context lora 7 mēneši atpakaļ
transformers_utils 9e73559eba make use of batched rotary embedding kernels to support long context lora 7 mēneši atpakaļ
__init__.py be8154a8a0 feat: proper embeddings API with e5-mistral-7b support 7 mēneši atpakaļ
py.typed 1c988a48b2 fix logging and add py.typed 1 gadu atpakaļ