dropout05's picture
Update README.md
ff0308f
---
license: apache-2.0
---
**Don't use this model for any applied task. It too small to be practically useful. It is just a part of a weird research project.**
An extremely small version of T5 with these parameters
```python
"d_ff": 1024,
"d_kv": 64,
"d_model": 256,
"num_heads": 4,
"num_layers": 1, # yes, just one layer
```
The model was pre-trained on `realnewslike` subset of C4 for 1 epoch with sequence length `64`. Corresponding WandB run: [click](https://wandb.ai/guitaricet/t5-lm/runs/2yvuxsfz?workspace=user-guitaricet).