Adding Evaluation Results

#1
Files changed (1) hide show
  1. README.md +15 -1
README.md CHANGED
@@ -1,8 +1,8 @@
1
  ---
2
  license: apache-2.0
3
- base_model: NousResearch/Yarn-Mistral-7b-128k
4
  tags:
5
  - generated_from_trainer
 
6
  model-index:
7
  - name: unraveled-7b-dpo-lora
8
  results: []
@@ -71,3 +71,17 @@ The following hyperparameters were used during training:
71
  - Pytorch 2.1.0+cu118
72
  - Datasets 2.14.6
73
  - Tokenizers 0.14.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
 
3
  tags:
4
  - generated_from_trainer
5
+ base_model: NousResearch/Yarn-Mistral-7b-128k
6
  model-index:
7
  - name: unraveled-7b-dpo-lora
8
  results: []
 
71
  - Pytorch 2.1.0+cu118
72
  - Datasets 2.14.6
73
  - Tokenizers 0.14.1
74
+
75
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
76
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_ZoidBB__unraveled-7b-a1)
77
+
78
+ | Metric |Value|
79
+ |---------------------------------|----:|
80
+ |Avg. |56.63|
81
+ |AI2 Reasoning Challenge (25-Shot)|59.81|
82
+ |HellaSwag (10-Shot) |82.80|
83
+ |MMLU (5-Shot) |63.39|
84
+ |TruthfulQA (0-shot) |42.23|
85
+ |Winogrande (5-shot) |77.19|
86
+ |GSM8k (5-shot) |14.33|
87
+