Update README.md
Browse files
README.md
CHANGED
@@ -30,11 +30,11 @@ We introduce two innovative techniques: Gating Logit Normalization, which enhanc
|
|
30 |
Skywork-MoE demonstrates comparable or superior performance to models with more parameters or more activated parameters, such as Grok-1, DBRX, Mistral 8*22, and Deepseek-V2.
|
31 |
|
32 |
# News and Updates
|
33 |
-
* 2024.6.3 We release the **Skywork-MoE-
|
34 |
|
35 |
# Table of contents
|
36 |
|
37 |
-
|
38 |
- [👨💻Benchmark Results](#Benchmark-Results)
|
39 |
- [🏆Demonstration of Hugging Face Model Inference](#Demonstration-of-HuggingFace-Model-Inference)
|
40 |
- [📕Demonstration of vLLM Model Inference](#Demonstration-of-vLLM-Model-Inference)
|
@@ -42,15 +42,8 @@ Skywork-MoE demonstrates comparable or superior performance to models with more
|
|
42 |
- [🤝Contact Us and Citation](#Contact-Us-and-Citation)
|
43 |
|
44 |
|
45 |
-
# Download URL
|
46 |
-
|
47 |
-
| | HuggingFace Model | ModelScope Model | Wisemodel Model |
|
48 |
-
|:-------:|:-----------:|:-----------------------------:|:-----------------------------:|
|
49 |
-
| **Skywork-MoE-base** | 🤗 [Skywork-MoE-base](https://github.com/SkyworkAI/Skywork-MoE) | 🤖[Skywork-MoE-base](https://www.modelscope.cn/models/skywork/Skywork-MoE-base) | 👾[Skywork-MoE-base](https://wisemodel.cn/models/Skywork/Skywork-MoE-base) |
|
50 |
-
| **Skywork-MoE-Base-FP8** | 🤗 [Skywork-MoE-Base-FP8](https://github.com/SkyworkAI/Skywork-MoE) | 🤖 | 👾 |
|
51 |
-
|
52 |
# Benchmark Results
|
53 |
-
We evaluated Skywork-MoE-
|
54 |
<img src="misc/skywork_moe_base_evaluation.png" alt="Image" width="600" height="280">
|
55 |
|
56 |
|
@@ -58,9 +51,9 @@ We evaluated Skywork-MoE-base model on various popular benchmarks, including C-E
|
|
58 |
|
59 |
## Quickstart with vLLM
|
60 |
|
61 |
-
We provide a method to quickly deploy the Skywork-
|
62 |
|
63 |
-
Under fp8 precision you can run Skywork-
|
64 |
|
65 |
You can get the source code in [`vllm`](https://github.com/SkyworkAI/vllm)
|
66 |
|
@@ -128,7 +121,7 @@ docker run \
|
|
128 |
registry.cn-wulanchabu.aliyuncs.com/triple-mu/skywork-moe-vllm:v1
|
129 |
```
|
130 |
|
131 |
-
Now, you can run the Skywork
|
132 |
|
133 |
### Text Completion
|
134 |
|
|
|
30 |
Skywork-MoE demonstrates comparable or superior performance to models with more parameters or more activated parameters, such as Grok-1, DBRX, Mistral 8*22, and Deepseek-V2.
|
31 |
|
32 |
# News and Updates
|
33 |
+
* 2024.6.3 We release the **Skywork-MoE-Base** model.
|
34 |
|
35 |
# Table of contents
|
36 |
|
37 |
+
|
38 |
- [👨💻Benchmark Results](#Benchmark-Results)
|
39 |
- [🏆Demonstration of Hugging Face Model Inference](#Demonstration-of-HuggingFace-Model-Inference)
|
40 |
- [📕Demonstration of vLLM Model Inference](#Demonstration-of-vLLM-Model-Inference)
|
|
|
42 |
- [🤝Contact Us and Citation](#Contact-Us-and-Citation)
|
43 |
|
44 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
# Benchmark Results
|
46 |
+
We evaluated Skywork-MoE-Base model on various popular benchmarks, including C-Eval, MMLU, CMMLU, GSM8K, MATH and HumanEval.
|
47 |
<img src="misc/skywork_moe_base_evaluation.png" alt="Image" width="600" height="280">
|
48 |
|
49 |
|
|
|
51 |
|
52 |
## Quickstart with vLLM
|
53 |
|
54 |
+
We provide a method to quickly deploy the Skywork-MoE-Base model based on vllm.
|
55 |
|
56 |
+
Under fp8 precision you can run Skywork-MoE-Base with just only 8*4090.
|
57 |
|
58 |
You can get the source code in [`vllm`](https://github.com/SkyworkAI/vllm)
|
59 |
|
|
|
121 |
registry.cn-wulanchabu.aliyuncs.com/triple-mu/skywork-moe-vllm:v1
|
122 |
```
|
123 |
|
124 |
+
Now, you can run the Skywork MoE model for fun!
|
125 |
|
126 |
### Text Completion
|
127 |
|