metadata
datasets:
- DEplain/DEplain-APA-sent
language:
- de
metrics:
- sari
- bleu
- bertscore
library_name: transformers
base_model: google/mT5-large
pipeline_tag: text2text-generation
Model Card for mT5-large-trimmed_deplain-apa
Finetuned mT5-Model for German sentence-level text-simplification.
Model Details
Model Description
- Model type: Encoder-Decoder-Transformer
- Language(s) (NLP): German
- Finetuned from model: google/mT5-large
- Task: Text-Simplification
Training Details
Training Data
DEplain/DEplain-APA-sent
Stodden et al. (2023): arXiv:2305.18939
Training Procedure
Parameter-efficient Fine-Tuning with LoRA. Vocabulary trimmed to 32.000 most frequent tokens for German.
Training Hyperparameters
- Batch Size: 16
- Epochs: 1
- Learning Rate: 0,001
- Optimizer: Adafactor
LoRA Hyperparameters
- R: 32
- Alpha: 64
- Dropout:
- Target modules: all linear layers