cataluna84
commited on
Commit
•
1fae59e
1
Parent(s):
fec90af
Initial Commit
Browse files
README.md
CHANGED
@@ -7,4 +7,14 @@ sdk: static
|
|
7 |
pinned: false
|
8 |
---
|
9 |
|
10 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
pinned: false
|
8 |
---
|
9 |
|
10 |
+
Multilingual language models are typically large, requiring significant computational resources.
|
11 |
+
|
12 |
+
Can we create multilingual models that maintain performance comparable to their larger models while reducing size, latency and inference speeds?
|
13 |
+
|
14 |
+
Potential Techniques:
|
15 |
+
- Pruning
|
16 |
+
- SparseGPT
|
17 |
+
- ShortGPT
|
18 |
+
- Knowledge Distillation
|
19 |
+
- Quantization
|
20 |
+
|