jianqing666 commited on
Commit
1c8cc96
·
verified ·
1 Parent(s): 69b0f09

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -12
README.md CHANGED
@@ -2,21 +2,19 @@
2
  license: apache-2.0
3
  language:
4
  - ar
5
- - zh
6
- - en
7
  ---
8
 
9
- # <b>MgGPT-8B-Chat</b>
10
- MgGPT-8B-Chat is a fully fine-tuned generative text model collection, particularly focused on the Arabic language domain.
11
- This is the repository for the 8B-chat model, developed based on [MgGPT-8B](https://huggingface.co/MgGPT/MgGPT-8B).
12
 
13
  ---
14
  ## Model Details
15
- We have released the MgGPT family of large language models, which is a collection of fully fine-tuned generative text models based on LlaMA2(MgGPT-7B, MgGPT-13B), LlaMA3(MgGPT-8B, MgGPT-70B), Qwen2(MgGPT-32B). Our models include two main categories: MgGPT and MgGPT-chat. MgGPT-chat is an optimized version specifically designed for dialogue applications. It is worth mentioning that our models have demonstrated superior performance compared to all currently available open-source Arabic dialogue models in multiple benchmark tests. Furthermore, in our human evaluations, our models have shown comparable satisfaction levels to some closed-source models, such as ChatGPT, in the Arabic language.
16
  <!-- ## Model Developers -->
17
  <!-- We are from the King Abdullah University of Science and Technology (KAUST), the Chinese University of Hong Kong, Shenzhen (CUHKSZ) and the Shenzhen Research Institute of Big Data (SRIBD). -->
18
  ## Variations
19
- MgGPT families come in a range of parameter sizes —— 7B, 8B, 13B, 32B and 70B, each size of model has a base category and a -chat category.
20
  <!-- ## Paper -->
21
  <!-- The paper can be accessed at [link](https://huggingface.co/FreedomIntelligence/AceGPT-v2-70B-Chat/blob/main/Alignment_at_Pre_training__a_Case_Study_of_Aligning_LLMs_in_Arabic.pdf). -->
22
  ## Input
@@ -27,11 +25,11 @@ Models output text only.
27
 
28
  | Model | Avg. | [ArabicMMLU]((https://github.com/mbzuai-nlp/ArabicMMLU)) | [ArabicMMLU]((https://github.com/mbzuai-nlp/ArabicMMLU)) | ARC | BoolQ | EXAMs | ACVA (clean) | ACVA (all) |
29
  |------------------|--------|--------|--------|-------|--------|--------|--------------|------------|
30
- | MgGPT-7B-chat | 59.90 | 45.77 | 56.62 | 60.49 | 72.45 | 43.69 | 69.46 | 70.86 |
31
- | **MgGPT-8B-chat** | 66.69 | 54.45 | 62.21 | 72.44 | 71.65 | 52.98 | 76.54 | 76.55 |
32
- | MgGPT-13B-chat | 63.42 | 47.33 | 61.70 | 63.99 | 69.33 | 48.37 | 76.90 | 76.37 |
33
- | MgGPT-32B-chat | 70.63 | 57.12 | 68.70 | 78.07 | 77.22 | 52.89 | 81.36 | 79.03 |
34
- | MgGPT-70B-chat | 73.99 | 64.26 | 72.50 | 85.53 | 82.66 | 56.99 | 78.61 | 77.38 |
35
  | Jais-30B-v3-chat | 57.84 | 35.68 | 62.36 | 51.02 | 76.30 | 32.24 | 73.63 | 73.66 |
36
  | GPT-3.5 | 62.44 | 46.07 | 57.72 | 60.24 | 76.12 | 45.63 | 74.45 | 76.88 |
37
  | GPT-4 | 75.78 | 65.04 | 72.50 | 85.67 | 85.99 | 57.76 | 84.06 | 79.43 |
 
2
  license: apache-2.0
3
  language:
4
  - ar
 
 
5
  ---
6
 
7
+ # <b>MgGPT0-8B-Chat</b>
8
+ MgGPT0-8B-Chat is a fully fine-tuned generative text model collection, particularly focused on the Arabic language domain.
9
+ This is the repository for the 8B-chat model, developed based on [MgGPT0-8B](https://huggingface.co/MgGPT/MgGPT0-8B).
10
 
11
  ---
12
  ## Model Details
13
+ We have released the MgGPT0 family of large language models, which is a collection of fully fine-tuned generative text models based on LlaMA2(MgGPT0-7B, MgGPT0-13B), LlaMA3(MgGPT0-8B, MgGPT0-70B), Qwen2(MgGPT0-32B). Our models include two main categories: MgGPT0 and MgGPT0-chat. MgGPT0-chat is an optimized version specifically designed for dialogue applications. It is worth mentioning that our models have demonstrated superior performance compared to all currently available open-source Arabic dialogue models in multiple benchmark tests. Furthermore, in our human evaluations, our models have shown comparable satisfaction levels to some closed-source models, such as ChatGPT, in the Arabic language.
14
  <!-- ## Model Developers -->
15
  <!-- We are from the King Abdullah University of Science and Technology (KAUST), the Chinese University of Hong Kong, Shenzhen (CUHKSZ) and the Shenzhen Research Institute of Big Data (SRIBD). -->
16
  ## Variations
17
+ MgGPT0 families come in a range of parameter sizes —— 7B, 8B, 13B, 32B and 70B, each size of model has a base category and a -chat category.
18
  <!-- ## Paper -->
19
  <!-- The paper can be accessed at [link](https://huggingface.co/FreedomIntelligence/AceGPT-v2-70B-Chat/blob/main/Alignment_at_Pre_training__a_Case_Study_of_Aligning_LLMs_in_Arabic.pdf). -->
20
  ## Input
 
25
 
26
  | Model | Avg. | [ArabicMMLU]((https://github.com/mbzuai-nlp/ArabicMMLU)) | [ArabicMMLU]((https://github.com/mbzuai-nlp/ArabicMMLU)) | ARC | BoolQ | EXAMs | ACVA (clean) | ACVA (all) |
27
  |------------------|--------|--------|--------|-------|--------|--------|--------------|------------|
28
+ | MgGPT0-7B-chat | 59.90 | 45.77 | 56.62 | 60.49 | 72.45 | 43.69 | 69.46 | 70.86 |
29
+ | **MgGPT0-8B-chat** | 66.69 | 54.45 | 62.21 | 72.44 | 71.65 | 52.98 | 76.54 | 76.55 |
30
+ | MgGPT0-13B-chat | 63.42 | 47.33 | 61.70 | 63.99 | 69.33 | 48.37 | 76.90 | 76.37 |
31
+ | MgGPT0-32B-chat | 70.63 | 57.12 | 68.70 | 78.07 | 77.22 | 52.89 | 81.36 | 79.03 |
32
+ | MgGPT0-70B-chat | 73.99 | 64.26 | 72.50 | 85.53 | 82.66 | 56.99 | 78.61 | 77.38 |
33
  | Jais-30B-v3-chat | 57.84 | 35.68 | 62.36 | 51.02 | 76.30 | 32.24 | 73.63 | 73.66 |
34
  | GPT-3.5 | 62.44 | 46.07 | 57.72 | 60.24 | 76.12 | 45.63 | 74.45 | 76.88 |
35
  | GPT-4 | 75.78 | 65.04 | 72.50 | 85.67 | 85.99 | 57.76 | 84.06 | 79.43 |