Add pipeline tag, link to paper
Browse filesThis PR adds a link to the paper [Control LLM: Controlled Evolution for Intelligence Retention in LLM](https://huggingface.co/papers/2501.10979).
It also adds the relevant `pipeline_tag` and `library_name`.
README.md
CHANGED
@@ -9,8 +9,12 @@ datasets:
|
|
9 |
- codingsteven/Llama-3-8B-chat
|
10 |
language:
|
11 |
- zh
|
|
|
|
|
12 |
base_model:
|
13 |
- meta-llama/Llama-3.1-8B
|
|
|
|
|
14 |
model-index:
|
15 |
- name: Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
|
16 |
results:
|
@@ -74,10 +78,10 @@ model-index:
|
|
74 |
---
|
75 |
|
76 |
# Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
|
77 |
-
This is a fine-tuned model of Llama-3.1-8B for muliligual-Chinese tasks on SynE dataset by Control LLM-Concat16-Dlerp.
|
78 |
|
79 |
## Linked Paper
|
80 |
-
This model is associated with the paper: [Control-LLM](https://arxiv.org/abs/
|
81 |
|
82 |
## Evaluation Results
|
83 |
Here is an overview of the evaluation results and findings:
|
@@ -108,4 +112,4 @@ The table below summarizes evaluation results across Chinese tasks and original
|
|
108 |
- **MLU**: MMLU (Massive Multitask Language Understanding)
|
109 |
- **MLUP**: MMLU Pro
|
110 |
- **O-Avg**: Original Capability - Size Weighted Average across BBH, MLU, and MLUP
|
111 |
-
- **Overall**: Combined average across all tasks
|
|
|
9 |
- codingsteven/Llama-3-8B-chat
|
10 |
language:
|
11 |
- zh
|
12 |
+
metrics:
|
13 |
+
- accuracy
|
14 |
base_model:
|
15 |
- meta-llama/Llama-3.1-8B
|
16 |
+
pipeline_tag: text-generation
|
17 |
+
library_name: transformers
|
18 |
model-index:
|
19 |
- name: Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
|
20 |
results:
|
|
|
78 |
---
|
79 |
|
80 |
# Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
|
81 |
+
This is a fine-tuned model of Llama-3.1-8B for muliligual-Chinese tasks on SynE dataset by Control LLM-Concat16-Dlerp, as described in [Control LLM: Controlled Evolution for Intelligence Retention in LLM](https://huggingface.co/papers/2501.10979).
|
82 |
|
83 |
## Linked Paper
|
84 |
+
This model is associated with the paper: [Control-LLM](https://arxiv.org/abs/2410.14745).
|
85 |
|
86 |
## Evaluation Results
|
87 |
Here is an overview of the evaluation results and findings:
|
|
|
112 |
- **MLU**: MMLU (Massive Multitask Language Understanding)
|
113 |
- **MLUP**: MMLU Pro
|
114 |
- **O-Avg**: Original Capability - Size Weighted Average across BBH, MLU, and MLUP
|
115 |
+
- **Overall**: Combined average across all tasks
|