In this article, you can download many papers.
Floating-point arithmetic for AI inference — hit or miss?
In this article, you can download many papers.
Floating-point arithmetic for AI inference — hit or miss?
LoRA (Low-Rank Adaptation)
- Paper
[Current]
[Before]
----
- YouTube
https://www.youtube.com/watch?v=dA-NhCtrrVE
https://www.youtube.com/watch?v=BJqwmDpa0wM
https://www.youtube.com/watch?v=t509sv5MT0w
----
The case for 4-bit precision: k-bit Interence Scaling Laws
Parameter-Efficient Transfer Learning for NLP
Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tuning
QLoRA: Efficient Finetuning of Quantized LLMs
https://www.youtube.com/watch?v=X4VvO3G6_vw
----
Intrinsic Dimensionality Ezplanins the Effectiveness of Language Model Fine-Tuning