Commit
·
254116c
1
Parent(s):
f45584c
Update README.md
Browse files
README.md
CHANGED
@@ -46,6 +46,11 @@ For details, please refer to the version without DPO training: [CausalLM/14B](ht
|
|
46 |
| **CausalLM/14B-DPO-α** | **7.618868** |
|
47 |
| **CausalLM/7B-DPO-α** | **7.038125** |
|
48 |
|
|
|
|
|
|
|
|
|
|
|
49 |
It should be noted that this is not a version that continues training on CausalLM/14B & 7B, but rather an optimized version that has undergone DPO training concurrently on a previous training branch, and some detailed parameters may have changed. You will still need to download the full model.
|
50 |
|
51 |
The beta branch will soon be released, employing some aggressive approaches that might be detrimental in certain tasks, in order to achieve better alignment with human preferences, aiming to meet or exceed the GPT-3.5 benchmarks. Stay tuned.
|
|
|
46 |
| **CausalLM/14B-DPO-α** | **7.618868** |
|
47 |
| **CausalLM/7B-DPO-α** | **7.038125** |
|
48 |
|
49 |
+
Dec 2, 2023
|
50 |
+
Rank **#2** non-base model, of its size on 🤗 Open LLM Leaderboard, outperforms ALL ~13B chat models including microsoft/Orca-2-13b.
|
51 |
+
|
52 |
+

|
53 |
+
|
54 |
It should be noted that this is not a version that continues training on CausalLM/14B & 7B, but rather an optimized version that has undergone DPO training concurrently on a previous training branch, and some detailed parameters may have changed. You will still need to download the full model.
|
55 |
|
56 |
The beta branch will soon be released, employing some aggressive approaches that might be detrimental in certain tasks, in order to achieve better alignment with human preferences, aiming to meet or exceed the GPT-3.5 benchmarks. Stay tuned.
|