bhenrym14 commited on
Commit
d6d4c42
1 Parent(s): 058c342

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -43,7 +43,7 @@ Previous experiments have demonstrated that orca-like datasets yield substantial
43
  | 2048 | 5.22 | 5.38 | 5.87 | 5.23 | **5.07** |
44
  | 4096 | 4.90 | 5.08 | 5.50 | 4.91 | **4.77** |
45
  | 8192 | **4.71** | 4.90 | 5.32 | Not Tested | 57.1 |
46
- | 12000 | 30 | **4.82** | 56.1 | Not Tested | Not Tested |
47
 
48
  - This model is very competitive with the Llama-1 33b extended context variants. In fact, it outperforms bhenrym14/airoboros-33b-gpt4-1.4.1-lxctx-PI-16384-fp16 everywhere <=8192 tokens.
49
  - Not presented here, but this model outperforms the base llama-2-13b on MMLU-fs with a score of 58.3. If this score ends up being be replicated on the HF LLM leaderboard, **this would place this model at 2nd or 3rd overall for MMLU among 13b models (and the #1 for extended context)**
 
43
  | 2048 | 5.22 | 5.38 | 5.87 | 5.23 | **5.07** |
44
  | 4096 | 4.90 | 5.08 | 5.50 | 4.91 | **4.77** |
45
  | 8192 | **4.71** | 4.90 | 5.32 | Not Tested | 57.1 |
46
+ | 12000 | 55 | **4.82** | 56.1 | Not Tested | Not Tested |
47
 
48
  - This model is very competitive with the Llama-1 33b extended context variants. In fact, it outperforms bhenrym14/airoboros-33b-gpt4-1.4.1-lxctx-PI-16384-fp16 everywhere <=8192 tokens.
49
  - Not presented here, but this model outperforms the base llama-2-13b on MMLU-fs with a score of 58.3. If this score ends up being be replicated on the HF LLM leaderboard, **this would place this model at 2nd or 3rd overall for MMLU among 13b models (and the #1 for extended context)**