Update README.md
Browse files
README.md
CHANGED
@@ -9,9 +9,11 @@ License: https://llama.meta.com/llama3/license
|
|
9 |
|
10 |
240421: Bellman's back!
|
11 |
This is a preliminary test run, run on mostly the same settings and dataset as bellman-mistral-instruct (But not the dpo pass). Context length is 3072.
|
12 |
-
I've done some basic testing, and it's not a total mess. Whether it's an improvement to llama-3-instruct, I'm not sure, because
|
13 |
|
14 |
-
I'll try to make a pass on full context length soon. And hopefully
|
|
|
|
|
15 |
|
16 |
|
17 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/653cd3049107029eb004f968/IDGX3d9lGe6yx-yHjsrav.png)
|
|
|
9 |
|
10 |
240421: Bellman's back!
|
11 |
This is a preliminary test run, run on mostly the same settings and dataset as bellman-mistral-instruct (But not the dpo pass). Context length is 3072.
|
12 |
+
I've done some basic testing, and it's not a total mess. Whether it's an improvement to llama-3-instruct, I'm not sure, because that's REALLY good.
|
13 |
|
14 |
+
I'll try to make a pass on full context length soon. And hopefully improve the results more.
|
15 |
+
|
16 |
+
Make sure to use the correct chat template (llama-3) for best results.
|
17 |
|
18 |
|
19 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/653cd3049107029eb004f968/IDGX3d9lGe6yx-yHjsrav.png)
|