hawei nielsr HF staff commited on
Commit
6c8b4d8
·
verified ·
1 Parent(s): 59f5f87

Add metadata and paper link (#1)

Browse files

- Add metadata and paper link (cd40ed60125a69349a23921b373d84e2a53431d7)


Co-authored-by: Niels Rogge <[email protected]>

Files changed (1) hide show
  1. README.md +7 -2
README.md CHANGED
@@ -9,6 +9,8 @@ datasets:
9
  - codingsteven/Llama-3-8B-chat
10
  language:
11
  - zh
 
 
12
  base_model:
13
  - meta-llama/Llama-3.1-8B
14
  model-index:
@@ -71,12 +73,15 @@ model-index:
71
  value: 0.37368330167501296
72
  stderr: 0.00438421288652232
73
  verified: false
 
 
74
  ---
 
75
  # Control-LLM-Llama3.1-8B-SynE-Concat16-Lerp
76
  This is a fine-tuned model of Llama-3.1-8B for muliligual-Chinese tasks on SynE dataset by Control LLM-Concat16-Lerp.
77
 
78
  ## Linked Paper
79
- This model is associated with the paper: [Control-LLM](https://arxiv.org/abs/2501.10979).
80
 
81
  ## Linked Open Source code - training, eval and benchmark
82
  This model is associated with the github: [Control-LLM](https://github.com/linkedin/ControlLLM).
@@ -110,4 +115,4 @@ The table below summarizes evaluation results across Chinese tasks and original
110
  - **MLU**: MMLU (Massive Multitask Language Understanding)
111
  - **MLUP**: MMLU Pro
112
  - **O-Avg**: Original Capability - Size Weighted Average across BBH, MLU, and MLUP
113
- - **Overall**: Combined average across all tasks
 
9
  - codingsteven/Llama-3-8B-chat
10
  language:
11
  - zh
12
+ metrics:
13
+ - accuracy
14
  base_model:
15
  - meta-llama/Llama-3.1-8B
16
  model-index:
 
73
  value: 0.37368330167501296
74
  stderr: 0.00438421288652232
75
  verified: false
76
+ pipeline_tag: text-generation
77
+ library_name: transformers
78
  ---
79
+
80
  # Control-LLM-Llama3.1-8B-SynE-Concat16-Lerp
81
  This is a fine-tuned model of Llama-3.1-8B for muliligual-Chinese tasks on SynE dataset by Control LLM-Concat16-Lerp.
82
 
83
  ## Linked Paper
84
+ This model is associated with the paper: [Control LLM: Controlled Evolution for Intelligence Retention in LLM](https://huggingface.co/papers/2501.10979).
85
 
86
  ## Linked Open Source code - training, eval and benchmark
87
  This model is associated with the github: [Control-LLM](https://github.com/linkedin/ControlLLM).
 
115
  - **MLU**: MMLU (Massive Multitask Language Understanding)
116
  - **MLUP**: MMLU Pro
117
  - **O-Avg**: Original Capability - Size Weighted Average across BBH, MLU, and MLUP
118
+ - **Overall**: Combined average across all tasks