AmelieSchreiber
commited on
Commit
•
77ab0f2
1
Parent(s):
f56efe3
Update README.md
Browse files
README.md
CHANGED
@@ -34,6 +34,11 @@ Note, we are only training 0.58% of the parameters, using only the query, key, a
|
|
34 |
trainable params: 23682 || all params: 4075265 || trainable%: 0.5811155838945443
|
35 |
```
|
36 |
|
|
|
|
|
|
|
|
|
|
|
37 |
## Testing for Overfitting
|
38 |
|
39 |
### Checkpoint 1
|
|
|
34 |
trainable params: 23682 || all params: 4075265 || trainable%: 0.5811155838945443
|
35 |
```
|
36 |
|
37 |
+
It was shown in the QLoRA paper that to obtain performance comparable or better than full finetuning, the most important hyperparameter than can
|
38 |
+
that can be adjusted is which weight matrices the LoRA adapters are applied to, with more being better. The rank and other hyperparameters
|
39 |
+
such as rank and the scaling factor alpha did not seem to matter. So, an important thing to investigate next would be to check and see if this
|
40 |
+
transfers to protein language models as well.
|
41 |
+
|
42 |
## Testing for Overfitting
|
43 |
|
44 |
### Checkpoint 1
|