Programming is Not Rocket Science, Don't let AI Write Your Code, Fight Back. And if you must use AI, find provenance, and Attribute. Long Live GNU/Linux. Full praise to SSA-Based Compilation.
Tuesday, 1 July 2025
Concept of LoRA or Low Rank Adapation in LLMs
LoRA is an approach to optimizing LLMs by reducing the "size" of the matrix of trainable parameters, as measured by "rank" of the matrix i.e. the number of linearly independent rows or columns.
No comments:
Post a Comment