File size: 1,111 Bytes
00ec26b 243f70a 00ec26b 5858480 243f70a 071281f 243f70a 071281f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 |
---
license: apache-2.0
datasets:
- ambrosfitz/ps_data_v2.2
---
### Model
A fine-tuned openllama 3B model, using primary sources from US History to provide a deeper understanding of the historical context.
Run history:
train/epoch βββββββββββ
β
β
ββββββββ<br>
train/global_step βββββββββββ
β
β
ββββββββ<br>
train/grad_norm ββββ
ββ
ββββββββββββββ<br>
train/learning_rate βββββ
β
ββββββββ
β
βββββ<br>
train/loss βββββ
β
ββββββββββββββ<br>
train/total_flos β<br>
train/train_loss β<br>
train/train_runtime β<br>
train/train_samples_per_second β<br>
train/train_steps_per_second β<br>
Run summary:
train/epoch 2.0<br>
train/global_step 20<br>
train/grad_norm 0.13779<br>
train/learning_rate 0.0<br>
train/loss 1.1365<br>
train/total_flos 4.579249185376512e+16<br>
train/train_loss 1.29891<br>
train/train_runtime 1552.5749<br>
train/train_samples_per_second 1.649<br>
train/train_steps_per_second 0.013<br> |