Skip to content

Commit

Permalink
Update index.html
Browse files Browse the repository at this point in the history
  • Loading branch information
NamrataRShivagunde authored May 8, 2024
1 parent e1f575d commit cc8b52a
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion docs/2024/pept_relora_n_galore/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -154,7 +154,7 @@ <h2 id="intro">Parameter Efficient Pre-training (PEPT)</h2>

<h2 id="relora">ReLoRA: High-Rank Training Through Low-Rank Updates</h2>

<p>ReLoRA uses LoRA (Hu et al., 2022) decomposition technique where the pre-trained model weights are frozen and trainable rank decomposition matrices (W<sub>A</sub>, W<sub>B</sub>) are injected into each layer of the LLM. However in LoRA, the rank of the matrix is restricted by the rank r (given below), and the new trainable parameters (W<sub>A</sub> and W<sub>B</sub>) are merged back to the original matrices only after the end of the training. This suggests the potential to use PEPT techniques to pre-train LLMS. </p>
<p>ReLoRA uses LoRA (Hu et al., 2022) decomposition technique where the pre-trained model weights are frozen and trainable rank decomposition matrices (W<sub>A</sub>, W<sub>B</sub>) are injected into each attention and MLP layer of the LLM. However in LoRA, the rank of the matrix is restricted by the rank r (given below), and the new trainable parameters (W<sub>A</sub> and W<sub>B</sub>) are merged back to the original matrices only after the end of the training. </p>

<figure>
<img src="/blog/assets/images/relora-rank-property.png" />
Expand Down

0 comments on commit cc8b52a

Please sign in to comment.