joaoalvarenga commited on
Commit
fb4c1b6
1 Parent(s): cf7c1ed

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -57,7 +57,7 @@ Heavily inspired by [Hivemind's GPT-J-6B with 8-bit weights](https://huggingface
57
 
58
  Here, we also apply [LoRA (Low Rank Adapters)](https://arxiv.org/abs/2106.09685) to reduce model size. The original version takes \~353GB memory, this version takes **\~180GB**.
59
 
60
- Our main objective is to generate a model compressed enough to be deployed in a traditional Kubernetes cluster.
61
  ### How to use
62
 
63
  This model can be used by adapting Bloom original implementation. This is an adaptation from [Hivemind's GPT-J 8-bit](https://nbviewer.org/urls/huggingface.co/hivemind/gpt-j-6B-8bit/raw/main/convert-gpt-j.ipynb):
 
57
 
58
  Here, we also apply [LoRA (Low Rank Adapters)](https://arxiv.org/abs/2106.09685) to reduce model size. The original version takes \~353GB memory, this version takes **\~180GB**.
59
 
60
+ Our main goal is to generate a model compressed enough to be deployed in a traditional Kubernetes cluster.
61
  ### How to use
62
 
63
  This model can be used by adapting Bloom original implementation. This is an adaptation from [Hivemind's GPT-J 8-bit](https://nbviewer.org/urls/huggingface.co/hivemind/gpt-j-6B-8bit/raw/main/convert-gpt-j.ipynb):