-
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Paper • 2403.03507 • Published • 189 -
Let the Expert Stick to His Last: Expert-Specialized Fine-Tuning for Sparse Architectural Large Language Models
Paper • 2407.01906 • Published • 46 -
QLoRA: Efficient Finetuning of Quantized LLMs
Paper • 2305.14314 • Published • 58 -
LoRA+: Efficient Low Rank Adaptation of Large Models
Paper • 2402.12354 • Published • 7
Ruozhou He
fward
·
AI & ML interests
None yet
Organizations
None yet