Trained on 3 epochs of the totally-not-an-llm/EverythingLM-data-V2-sharegpt dataset.
### HUMAN:
{prompt}
### RESPONSE:
<leave a newline for the model to answer>
note: Changed a few of the finetuning parameters this time around. I have no idea if its any good but Feel free to give it a try!
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
| Metric | Value |
|---|---|
| Avg. | 36.29 |
| ARC (25-shot) | 42.83 |
| HellaSwag (10-shot) | 73.28 |
| MMLU (5-shot) | 26.87 |
| TruthfulQA (0-shot) | 37.26 |
| Winogrande (5-shot) | 66.61 |
| GSM8K (5-shot) | 1.59 |
| DROP (3-shot) | 5.61 |
- Downloads last month
- 1,049