.. |
__init__.py
|
2bd6c92f73
fix: lora inclusion in wheels
|
11 mēneši atpakaļ |
fully_sharded_layers.py
|
e87c32bed3
feat: full tensor parallel for LoRA layers (#545)
|
5 mēneši atpakaļ |
layers.py
|
9e73559eba
make use of batched rotary embedding kernels to support long context lora
|
5 mēneši atpakaļ |
lora.py
|
e87c32bed3
feat: full tensor parallel for LoRA layers (#545)
|
5 mēneši atpakaļ |
models.py
|
9e73559eba
make use of batched rotary embedding kernels to support long context lora
|
5 mēneši atpakaļ |
punica.py
|
e87c32bed3
feat: full tensor parallel for LoRA layers (#545)
|
5 mēneši atpakaļ |
request.py
|
5b0c11d190
support pipeline parallel pynccl groups
|
5 mēneši atpakaļ |
utils.py
|
9e73559eba
make use of batched rotary embedding kernels to support long context lora
|
5 mēneši atpakaļ |
worker_manager.py
|
9e73559eba
make use of batched rotary embedding kernels to support long context lora
|
5 mēneši atpakaļ |