AmelieSchreiber commited on
Commit
77ab0f2
1 Parent(s): f56efe3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -34,6 +34,11 @@ Note, we are only training 0.58% of the parameters, using only the query, key, a
34
  trainable params: 23682 || all params: 4075265 || trainable%: 0.5811155838945443
35
  ```
36
 
 
 
 
 
 
37
  ## Testing for Overfitting
38
 
39
  ### Checkpoint 1
 
34
  trainable params: 23682 || all params: 4075265 || trainable%: 0.5811155838945443
35
  ```
36
 
37
+ It was shown in the QLoRA paper that to obtain performance comparable or better than full finetuning, the most important hyperparameter than can
38
+ that can be adjusted is which weight matrices the LoRA adapters are applied to, with more being better. The rank and other hyperparameters
39
+ such as rank and the scaling factor alpha did not seem to matter. So, an important thing to investigate next would be to check and see if this
40
+ transfers to protein language models as well.
41
+
42
  ## Testing for Overfitting
43
 
44
  ### Checkpoint 1