Context-awareness in instruction finetuning
Collection
15 items
•
Updated
This model is a fine-tuned version of meta-llama/Meta-Llama-3-8B on the yihanwang617/WizardLM_70k_processed_indicator_unfiltered_4k dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
0.718 | 0.2225 | 200 | 0.7090 |
0.7205 | 0.4450 | 400 | 0.6897 |
0.7203 | 0.6675 | 600 | 0.6808 |
0.703 | 0.8900 | 800 | 0.6756 |
0.6759 | 1.1125 | 1000 | 0.6748 |
0.6533 | 1.3350 | 1200 | 0.6695 |
0.6458 | 1.5575 | 1400 | 0.6669 |
0.632 | 1.7800 | 1600 | 0.6655 |
Base model
meta-llama/Meta-Llama-3-8B