File size: 1,820 Bytes
5ab419d 5dfb023 a77a2fe 5dfb023 4a5e650 5ab419d 5dfb023 302a420 5dfb023 87fca81 5dfb023 a77a2fe 5dfb023 a77a2fe 4c897dc a77a2fe 5dfb023 a77a2fe 5dfb023 a77a2fe 5dfb023 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 |
---
license: apache-2.0
datasets:
- opus_books
- iwslt2017
language:
- en
- nl
pipeline_tag: text2text-generation
tags:
- translation
metrics:
- bleu
- chrf
- chrf++
widget:
- text: ">>en<< Was het leuk?"
---
# Model Card for mt5-small nl-en translation
The mt5-small nl-en translation model is a finetuned version of [google/mt5-small](https://huggingface.co/google/mt5-small).
It was finetuned on 237k rows of the [iwslt2017](https://huggingface.co/datasets/iwslt2017/viewer/iwslt2017-en-nl) dataset and roughly 38k rows of the [opus_books](https://huggingface.co/datasets/opus_books/viewer/en-nl) dataset. The model was trained in multiple phases with different epochs & batch sizes.
## How to use
**Install dependencies**
```bash
pip install transformers
pip install sentencepiece
pip install protobuf
```
You can use the following code for model inference. This model was finetuned to work with an identifier when prompted that needs to be present for the best results.
```Python
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer, GenerationConfig
# load tokenizer and model
tokenizer = AutoTokenizer.from_pretrained("Michielo/mt5-small_nl-en_translation")
model = AutoModelForSeq2SeqLM.from_pretrained("Michielo/mt5-small_nl-en_translation")
# tokenize input
inputs = tokenizer(">>en<< Your Dutch text here", return_tensors="pt")
# calculate the output
outputs = model.generate(**inputs, generation_config=generation_config)
# decode and print
print(tokenizer.batch_decode(outputs, skip_special_tokens=True))
```
## Benchmarks
| Benchmark | Score |
|--------------|:-----:|
| BLEU | 51.92% |
| chr-F | 67.90% |
| chr-F++ | 67.62% |
## License
This project is licensed under the Apache License 2.0 - see the [LICENSE](https://www.apache.org/licenses/LICENSE-2.0) file for details. |