ambrosfitz's picture
Upload README.md with huggingface_hub
708a0d1 verified
# FLAN-T5-Base pre-trained on Historical Text Completion
This model is a pre-trained version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) fine-tuned on historical text completion tasks.
## Model description
This model has been trained to complete historical texts, enhancing its understanding of historical contexts and language.
## Intended uses & limitations
This model is intended for further fine-tuning on specific historical NLP tasks or for generating historically-aware text completions.
## Training and evaluation data
The model was trained on a subset of the dataset [ambrosfitz/just_history_xl_masked](https://huggingface.co/datasets/ambrosfitz/just_history_xl_masked),
limited by available GPU memory.
## Training procedure
The model was trained using the following hyperparameters:
- Number of epochs: 3
- Batch size: 2
- Learning rate: 5e-05
- Gradient Accumulation Steps: 8
- Weight Decay: 0.01
## Results
Evaluation results: {'eval_loss': nan, 'eval_runtime': 0.7866, 'eval_samples_per_second': 38.139, 'eval_steps_per_second': 19.069, 'epoch': 2.909090909090909}