JonahYixMAD commited on
Commit
ef2d27e
β€’
1 Parent(s): 8edf197

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -3
README.md CHANGED
@@ -51,13 +51,11 @@ outputs = model.generate(**inputs, do_sample=True, max_new_tokens=256)
51
  print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
52
  ```
53
 
54
- Other xMADified models and their GPU memory requirements are listed below.
55
-
56
  Model | GPU Memory Requirement
57
  --- | ---
58
  Llama-3.2-3B-Instruct-xMADai-4bit | 6.5 GB β†’ 3.5 GB
59
  Llama-3.2-1B-Instruct-xMADai-4bit | 2.5 β†’ 2 GB
60
- Llama-3.1-405B-Instruct-xMADai-4bit | 258.14 GB β†’ 250 GB
61
  Llama-3.1-8B-Instruct-xMADai-4bit | 16 β†’ 7 GB
62
 
63
  For additional xMADified models, access to fine-tuning, and general questions, please contact us at [email protected] and join our waiting list.
 
51
  print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
52
  ```
53
 
 
 
54
  Model | GPU Memory Requirement
55
  --- | ---
56
  Llama-3.2-3B-Instruct-xMADai-4bit | 6.5 GB β†’ 3.5 GB
57
  Llama-3.2-1B-Instruct-xMADai-4bit | 2.5 β†’ 2 GB
58
+ Llama-3.1-405B-Instruct-xMADai-4bit | 800 GB (16 H100s) β†’ 250 GB (8 V100)
59
  Llama-3.1-8B-Instruct-xMADai-4bit | 16 β†’ 7 GB
60
 
61
  For additional xMADified models, access to fine-tuning, and general questions, please contact us at [email protected] and join our waiting list.