JonahYixMAD
commited on
Commit
•
8edf197
1
Parent(s):
84907ce
Update README.md
Browse files
README.md
CHANGED
@@ -51,4 +51,13 @@ outputs = model.generate(**inputs, do_sample=True, max_new_tokens=256)
|
|
51 |
print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
|
52 |
```
|
53 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
54 |
For additional xMADified models, access to fine-tuning, and general questions, please contact us at [email protected] and join our waiting list.
|
|
|
51 |
print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
|
52 |
```
|
53 |
|
54 |
+
Other xMADified models and their GPU memory requirements are listed below.
|
55 |
+
|
56 |
+
Model | GPU Memory Requirement
|
57 |
+
--- | ---
|
58 |
+
Llama-3.2-3B-Instruct-xMADai-4bit | 6.5 GB → 3.5 GB
|
59 |
+
Llama-3.2-1B-Instruct-xMADai-4bit | 2.5 → 2 GB
|
60 |
+
Llama-3.1-405B-Instruct-xMADai-4bit | 258.14 GB → 250 GB
|
61 |
+
Llama-3.1-8B-Instruct-xMADai-4bit | 16 → 7 GB
|
62 |
+
|
63 |
For additional xMADified models, access to fine-tuning, and general questions, please contact us at [email protected] and join our waiting list.
|