Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ license_link: https://huggingface.co/Qwen/Qwen1.5-72B-Chat/blob/main/LICENSE
|
|
5 |
---
|
6 |
**--Qwen1.5-55B-Chat-Cut--**
|
7 |
|
8 |
-
|
9 |
|
10 |
**I remove 20 layers of qwen1.5-72B-Chat and got a 55B model with 60 layers. It still works well, but with some degradation (compared to the 72B model, it appears slightly more confused and the logic is somewhat more disordered, but there seems to be no discernible loss in writing ability). I also tried removing more layers; when the qwen1.5-72B-Chat model is cut down to 40B+, it becomes noticeably extremely chaotic and foolish but still functions. However, when cut down to 30B+, it completely collapses and only generates meaningless gibberish.**
|
11 |
|
|
|
5 |
---
|
6 |
**--Qwen1.5-55B-Chat-Cut--**
|
7 |
|
8 |
+
**-It's a experimental model made by mergekit-**
|
9 |
|
10 |
**I remove 20 layers of qwen1.5-72B-Chat and got a 55B model with 60 layers. It still works well, but with some degradation (compared to the 72B model, it appears slightly more confused and the logic is somewhat more disordered, but there seems to be no discernible loss in writing ability). I also tried removing more layers; when the qwen1.5-72B-Chat model is cut down to 40B+, it becomes noticeably extremely chaotic and foolish but still functions. However, when cut down to 30B+, it completely collapses and only generates meaningless gibberish.**
|
11 |
|