crumb commited on
Commit
55157a6
·
1 Parent(s): db7918b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -52,4 +52,6 @@ Nearly every base model that isn't finetuned for a specific task was trained on
52
 
53
  ```
54
 
55
- The only application where I can imagine these being useful in the slightest is warm-starting very small encoder-decoder models or fitting a new scaling law that takes into account smaller models. Every model was trained on a singular GPU, either a RTX2060, RTX3060, or a T4.
 
 
 
52
 
53
  ```
54
 
55
+ The only application where I can imagine these being useful in the slightest is warm-starting very small encoder-decoder models or fitting a new scaling law that takes into account smaller models. Every model was trained on a singular GPU, either a RTX2060, RTX3060, or a T4.
56
+
57
+ I'd , uh , appreciate help in evaluating all these models probably with lm harness