AlpinDale cda0e93a10 abstract away the platform for device capability il y a 7 mois
..
__init__.py 2bd6c92f73 fix: lora inclusion in wheels il y a 1 an
fully_sharded_layers.py 8a6e83b52e feat: fully sharded QKVParallelLinearWithLora support il y a 7 mois
layers.py 0f4a9ee77b quantized lm_head (#582) il y a 7 mois
lora.py 56e0b8223c chore: add base class for LoRA-supported models il y a 7 mois
models.py 0a6db357d8 fix: use safetensor keys instead of adapter_config.json to find unexpected modules il y a 7 mois
punica.py cda0e93a10 abstract away the platform for device capability il y a 7 mois
request.py 5b0c11d190 support pipeline parallel pynccl groups il y a 7 mois
utils.py 8a6e83b52e feat: fully sharded QKVParallelLinearWithLora support il y a 7 mois
worker_manager.py 25feb1d592 chore: add support for pinning lora adapters in the lru cache il y a 7 mois