JosephusCheung commited on
Commit
254116c
·
1 Parent(s): f45584c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -46,6 +46,11 @@ For details, please refer to the version without DPO training: [CausalLM/14B](ht
46
  | **CausalLM/14B-DPO-α** | **7.618868** |
47
  | **CausalLM/7B-DPO-α** | **7.038125** |
48
 
 
 
 
 
 
49
  It should be noted that this is not a version that continues training on CausalLM/14B & 7B, but rather an optimized version that has undergone DPO training concurrently on a previous training branch, and some detailed parameters may have changed. You will still need to download the full model.
50
 
51
  The beta branch will soon be released, employing some aggressive approaches that might be detrimental in certain tasks, in order to achieve better alignment with human preferences, aiming to meet or exceed the GPT-3.5 benchmarks. Stay tuned.
 
46
  | **CausalLM/14B-DPO-α** | **7.618868** |
47
  | **CausalLM/7B-DPO-α** | **7.038125** |
48
 
49
+ Dec 2, 2023
50
+ Rank **#2** non-base model, of its size on 🤗 Open LLM Leaderboard, outperforms ALL ~13B chat models including microsoft/Orca-2-13b.
51
+
52
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63468a143ea42ee2cb49ddd1/Df5BcU3Pxzt2oKjuzArvk.png)
53
+
54
  It should be noted that this is not a version that continues training on CausalLM/14B & 7B, but rather an optimized version that has undergone DPO training concurrently on a previous training branch, and some detailed parameters may have changed. You will still need to download the full model.
55
 
56
  The beta branch will soon be released, employing some aggressive approaches that might be detrimental in certain tasks, in order to achieve better alignment with human preferences, aiming to meet or exceed the GPT-3.5 benchmarks. Stay tuned.