End of training
Browse files- README.md +1 -1
- all_results.json +1 -1
- eval_results.json +1 -1
README.md
CHANGED
@@ -26,7 +26,7 @@ print(output["generated_text"])
|
|
26 |
|
27 |
## Training procedure
|
28 |
|
29 |
-
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/shahrad_m/AIFGen-ppo-continual-test/runs/
|
30 |
|
31 |
|
32 |
This model was trained with PPO, a method introduced in [Fine-Tuning Language Models from Human Preferences](https://huggingface.co/papers/1909.08593).
|
|
|
26 |
|
27 |
## Training procedure
|
28 |
|
29 |
+
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/shahrad_m/AIFGen-ppo-continual-test/runs/ufysmsjb)
|
30 |
|
31 |
|
32 |
This model was trained with PPO, a method introduced in [Fine-Tuning Language Models from Human Preferences](https://huggingface.co/papers/1909.08593).
|
all_results.json
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
{
|
2 |
"dataset": 1,
|
3 |
-
"eval_score": 1.
|
4 |
}
|
|
|
1 |
{
|
2 |
"dataset": 1,
|
3 |
+
"eval_score": 1.718251347541809
|
4 |
}
|
eval_results.json
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
{
|
2 |
"dataset": 1,
|
3 |
-
"eval_score": 1.
|
4 |
}
|
|
|
1 |
{
|
2 |
"dataset": 1,
|
3 |
+
"eval_score": 1.718251347541809
|
4 |
}
|