Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -32,7 +32,7 @@ The context size of the model is 4096 tokens. We use the GPTNeoX tokenizer with
32
 
33
  The model has been trained on the [MosaicML](https://www.mosaicml.com/) platform on 128 H100-80GB GPUs using their [LLM Foundry](https://github.com/mosaicml/llm-foundry) and [Composer](https://github.com/mosaicml/composer) training library built on top of PyTorch.
34
 
35
- ## Dependancies
36
  You will need to install the latest versions of the following dependencies:
37
  ```
38
  einops
 
32
 
33
  The model has been trained on the [MosaicML](https://www.mosaicml.com/) platform on 128 H100-80GB GPUs using their [LLM Foundry](https://github.com/mosaicml/llm-foundry) and [Composer](https://github.com/mosaicml/composer) training library built on top of PyTorch.
34
 
35
+ ## Dependencies
36
  You will need to install the latest versions of the following dependencies:
37
  ```
38
  einops