Update README.md
Browse files
README.md
CHANGED
@@ -34,23 +34,46 @@ The lora merged model is [here](https://huggingface.co/uukuguy/Mistral-7B-OpenOr
|
|
34 |
|
35 |
The source code is [here](https://github.com/uukuguy/multi_loras)
|
36 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
## lm-evaluation-harness
|
38 |
|
39 |
[Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
40 |
|
41 |
-
| Metric | Mistral-7B-OpenOrca | Mistral-7B-OpenOrca-lora|
|
42 |
-
| --- | --- |--- |
|
43 |
-
| ARC | 64.08 | |
|
44 |
-
| HellaSwag | 83.99 | |
|
45 |
-
| MMLU | 62.24 | |
|
46 |
-
| TruthfulQA | 53.05 | |
|
47 |
-
| Average | 65.84 | |
|
48 |
|
49 |
## HumanEval
|
50 |
|
51 |
-
| Metric | Mistral-7B-OpenOrca | Mistral-7B-OpenOrca-lora|
|
52 |
-
| --- | --- | --- |
|
53 |
-
| humaneval-python | 35.976 | |
|
54 |
|
55 |
|
56 |
## Training procedure
|
|
|
34 |
|
35 |
The source code is [here](https://github.com/uukuguy/multi_loras)
|
36 |
|
37 |
+
## Mistral-7B-OpenOrca
|
38 |
+
|
39 |
+
- Extract lora model [Mistral-7B-OpenOrca-lora](https://huggingface.co/uukuguy/Mistral-7B-OpenOrca-lora) from [Mistral-7B-OpenOrca](https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca);
|
40 |
+
|
41 |
+
- Merge the base model [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) with lora model to [Mistral-7B-OpenOrca-lora-merged](https://huggingface.co/uukuguy/Mistral-7B-OpenOrca-lora-merged)
|
42 |
+
|
43 |
+
- LLM Evaluation ...
|
44 |
+
|
45 |
+
### Local Test
|
46 |
+
|
47 |
+
| | ARC_acc_norm (25-shot) | HellaSwag_acc_norm (10-shot) | MMLU_acc (5-shot) | TruthfulQA_mc2 (0-shot) | GSM8K_acc (8-shot) | Open LLM Score |
|
48 |
+
| ------ | ------ | ------ | ------ | ------ | ------ | ------ |
|
49 |
+
| Mistral-7B-OpenOrca | **71** | 83 | 61.42 | 45 | 40 | 65.11 |
|
50 |
+
| **r=64** | 67 | 84 | **64.26** | **47.32** | **41** | **65.65** |
|
51 |
+
| *r=16* | *65* | *83* | *62.84* | *46.95* | *38* | *64.45* |
|
52 |
+
|
53 |
+
### Open LLM Leaderboard
|
54 |
+
| | ARC_acc_norm (25-shot) | HellaSwag_acc_norm (10-shot) | MMLU_acc (5-shot) | TruthfulQA_mc2 (0-shot) | Open LLM Score |
|
55 |
+
| ------ | ------ | ------ | ------ | ------ | ------ |
|
56 |
+
| Mistral-7B-SlimOrca | 62.54 | 83.86 | **62.77** | **54.23** | **65.85** |
|
57 |
+
| Mistral-7B-OpenOrca | **64.08** | **83.99** | 62.24 | 53.05 | 65.84 |
|
58 |
+
|
59 |
+
|
60 |
## lm-evaluation-harness
|
61 |
|
62 |
[Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
63 |
|
64 |
+
| Metric | Mistral-7B-OpenOrca | Mistral-7B-OpenOrca-lora| Mistral-7B-OpenOrca-lora-merged |
|
65 |
+
| --- | --- |--- | --- |
|
66 |
+
| ARC | 64.08 | | |
|
67 |
+
| HellaSwag | 83.99 | | |
|
68 |
+
| MMLU | 62.24 | | |
|
69 |
+
| TruthfulQA | 53.05 | | |
|
70 |
+
| Average | 65.84 | | |
|
71 |
|
72 |
## HumanEval
|
73 |
|
74 |
+
| Metric | Mistral-7B-OpenOrca | Mistral-7B-OpenOrca-lora|Mistral-7B-OpenOrca-lora-merged |
|
75 |
+
| --- | --- | --- | --- |
|
76 |
+
| humaneval-python | 35.976 | | |
|
77 |
|
78 |
|
79 |
## Training procedure
|