File size: 1,118 Bytes
77b656c
708a0d1
77b656c
708a0d1
77b656c
708a0d1
77b656c
708a0d1
77b656c
708a0d1
77b656c
708a0d1
77b656c
708a0d1
77b656c
708a0d1
 
77b656c
708a0d1
77b656c
708a0d1
 
 
 
 
 
77b656c
708a0d1
77b656c
708a0d1
77b656c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
# FLAN-T5-Base pre-trained on Historical Text Completion

This model is a pre-trained version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) fine-tuned on historical text completion tasks.

## Model description

This model has been trained to complete historical texts, enhancing its understanding of historical contexts and language.

## Intended uses & limitations

This model is intended for further fine-tuning on specific historical NLP tasks or for generating historically-aware text completions.

## Training and evaluation data

The model was trained on a subset of the dataset [ambrosfitz/just_history_xl_masked](https://huggingface.co/datasets/ambrosfitz/just_history_xl_masked), 
limited by available GPU memory.

## Training procedure

The model was trained using the following hyperparameters:
- Number of epochs: 3
- Batch size: 2
- Learning rate: 5e-05
- Gradient Accumulation Steps: 8
- Weight Decay: 0.01

## Results

Evaluation results: {'eval_loss': nan, 'eval_runtime': 0.7866, 'eval_samples_per_second': 38.139, 'eval_steps_per_second': 19.069, 'epoch': 2.909090909090909}