Add metadata and paper link (#1)
Browse files- Add metadata and paper link (cd40ed60125a69349a23921b373d84e2a53431d7)
Co-authored-by: Niels Rogge <[email protected]>
README.md
CHANGED
@@ -9,6 +9,8 @@ datasets:
|
|
9 |
- codingsteven/Llama-3-8B-chat
|
10 |
language:
|
11 |
- zh
|
|
|
|
|
12 |
base_model:
|
13 |
- meta-llama/Llama-3.1-8B
|
14 |
model-index:
|
@@ -71,12 +73,15 @@ model-index:
|
|
71 |
value: 0.37368330167501296
|
72 |
stderr: 0.00438421288652232
|
73 |
verified: false
|
|
|
|
|
74 |
---
|
|
|
75 |
# Control-LLM-Llama3.1-8B-SynE-Concat16-Lerp
|
76 |
This is a fine-tuned model of Llama-3.1-8B for muliligual-Chinese tasks on SynE dataset by Control LLM-Concat16-Lerp.
|
77 |
|
78 |
## Linked Paper
|
79 |
-
This model is associated with the paper: [Control
|
80 |
|
81 |
## Linked Open Source code - training, eval and benchmark
|
82 |
This model is associated with the github: [Control-LLM](https://github.com/linkedin/ControlLLM).
|
@@ -110,4 +115,4 @@ The table below summarizes evaluation results across Chinese tasks and original
|
|
110 |
- **MLU**: MMLU (Massive Multitask Language Understanding)
|
111 |
- **MLUP**: MMLU Pro
|
112 |
- **O-Avg**: Original Capability - Size Weighted Average across BBH, MLU, and MLUP
|
113 |
-
- **Overall**: Combined average across all tasks
|
|
|
9 |
- codingsteven/Llama-3-8B-chat
|
10 |
language:
|
11 |
- zh
|
12 |
+
metrics:
|
13 |
+
- accuracy
|
14 |
base_model:
|
15 |
- meta-llama/Llama-3.1-8B
|
16 |
model-index:
|
|
|
73 |
value: 0.37368330167501296
|
74 |
stderr: 0.00438421288652232
|
75 |
verified: false
|
76 |
+
pipeline_tag: text-generation
|
77 |
+
library_name: transformers
|
78 |
---
|
79 |
+
|
80 |
# Control-LLM-Llama3.1-8B-SynE-Concat16-Lerp
|
81 |
This is a fine-tuned model of Llama-3.1-8B for muliligual-Chinese tasks on SynE dataset by Control LLM-Concat16-Lerp.
|
82 |
|
83 |
## Linked Paper
|
84 |
+
This model is associated with the paper: [Control LLM: Controlled Evolution for Intelligence Retention in LLM](https://huggingface.co/papers/2501.10979).
|
85 |
|
86 |
## Linked Open Source code - training, eval and benchmark
|
87 |
This model is associated with the github: [Control-LLM](https://github.com/linkedin/ControlLLM).
|
|
|
115 |
- **MLU**: MMLU (Massive Multitask Language Understanding)
|
116 |
- **MLUP**: MMLU Pro
|
117 |
- **O-Avg**: Original Capability - Size Weighted Average across BBH, MLU, and MLUP
|
118 |
+
- **Overall**: Combined average across all tasks
|