hawei nielsr HF staff commited on
Commit
dce1a7b
·
verified ·
1 Parent(s): 47df0a8

Add pipeline tag, link to paper (#1)

Browse files

- Add pipeline tag, link to paper (029547c60db5337eca28ca90c83c2ab4c95f27cf)


Co-authored-by: Niels Rogge <[email protected]>

Files changed (1) hide show
  1. README.md +7 -3
README.md CHANGED
@@ -9,8 +9,12 @@ datasets:
9
  - codingsteven/Llama-3-8B-chat
10
  language:
11
  - zh
 
 
12
  base_model:
13
  - meta-llama/Llama-3.1-8B
 
 
14
  model-index:
15
  - name: Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
16
  results:
@@ -74,10 +78,10 @@ model-index:
74
  ---
75
 
76
  # Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
77
- This is a fine-tuned model of Llama-3.1-8B for muliligual-Chinese tasks on SynE dataset by Control LLM-Concat16-Dlerp.
78
 
79
  ## Linked Paper
80
- This model is associated with the paper: [Control-LLM](https://arxiv.org/abs/2501.10979).
81
 
82
  ## Linked Open Source code - training, eval and benchmark
83
  This model is associated with the github: [Control-LLM](https://github.com/linkedin/ControlLLM).
@@ -111,4 +115,4 @@ The table below summarizes evaluation results across Chinese tasks and original
111
  - **MLU**: MMLU (Massive Multitask Language Understanding)
112
  - **MLUP**: MMLU Pro
113
  - **O-Avg**: Original Capability - Size Weighted Average across BBH, MLU, and MLUP
114
- - **Overall**: Combined average across all tasks
 
9
  - codingsteven/Llama-3-8B-chat
10
  language:
11
  - zh
12
+ metrics:
13
+ - accuracy
14
  base_model:
15
  - meta-llama/Llama-3.1-8B
16
+ pipeline_tag: text-generation
17
+ library_name: transformers
18
  model-index:
19
  - name: Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
20
  results:
 
78
  ---
79
 
80
  # Control-LLM-Llama3.1-8B-SynE-Concat16-Dlerp
81
+ This is a fine-tuned model of Llama-3.1-8B for muliligual-Chinese tasks on SynE dataset by Control LLM-Concat16-Dlerp, as described in [Control LLM: Controlled Evolution for Intelligence Retention in LLM](https://huggingface.co/papers/2501.10979).
82
 
83
  ## Linked Paper
84
+ This model is associated with the paper: [Control-LLM](https://arxiv.org/abs/2410.14745).
85
 
86
  ## Linked Open Source code - training, eval and benchmark
87
  This model is associated with the github: [Control-LLM](https://github.com/linkedin/ControlLLM).
 
115
  - **MLU**: MMLU (Massive Multitask Language Understanding)
116
  - **MLUP**: MMLU Pro
117
  - **O-Avg**: Original Capability - Size Weighted Average across BBH, MLU, and MLUP
118
+ - **Overall**: Combined average across all tasks