Immediately after switching the page, it will work with CSR.
Please reload your browser to see how it works.

Source:https://github.com/SoraKumo001/next-streaming

⬅️ LoRA+: Efficient Low Rank Adaptation of Large Models
batterseapower 16 daysReload
The other recent improvement suggested for LoRA is DoRA: https://magazine.sebastianraschka.com/p/lora-and-dora-from-s.... It really does seem to strongly outperform LoRA - see also https://www.answer.ai/posts/2024-04-26-fsdp-qdora-llama3.htm...

cuuupid 16 daysReload
I’m struggling to understand from this paper whether the approach is better in the general sense (all cases, with wider models seeing greater benefits) or purely for wider models (with narrower models seeing detriment)?

If it’s the former this could effectively halve finetuning cost overnight which would go a significant way towards enabling a wider array of use cases for LoRA.


ironbound 16 daysReload
I've had sucess with GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection https://arxiv.org/abs/2403.03507

youssefabdelm 16 daysReload
A better name would've probably been FastLoRA or something

yau8edq12i 16 daysReload
What an unfortunate name... I initially thought this was about wireless communication. https://en.wikipedia.org/wiki/LoRa