WelfCrozzo
commited on
Commit
·
6bada3c
1
Parent(s):
65772b6
Update README.md
Browse files
README.md
CHANGED
@@ -28,6 +28,8 @@ widget:
|
|
28 |
|
29 |
# T5 for belarusian language
|
30 |
|
|
|
|
|
31 |
This model is based on T5-small with sequence length equal 128 tokens. Model trained from scratch on RTX 3090 24GB.
|
32 |
|
33 |
# Supported tasks:
|
@@ -59,4 +61,7 @@ x = tokenizer.encode('<extra_id_1>да зорак праз цяжкасці', re
|
|
59 |
result = model.generate(x, return_dict_in_generate=True, output_scores=True,max_length=128)
|
60 |
print(tokenizer.decode(result["sequences"][0]))
|
61 |
```
|
62 |
-
</details>
|
|
|
|
|
|
|
|
28 |
|
29 |
# T5 for belarusian language
|
30 |
|
31 |
+
![model image](https://camo.githubusercontent.com/623b4dea0b653f2ad3f36c71ebfe749a677ac0a1/68747470733a2f2f6d69726f2e6d656469756d2e636f6d2f6d61782f343030362f312a44304a31674e51663876727255704b657944387750412e706e67)
|
32 |
+
|
33 |
This model is based on T5-small with sequence length equal 128 tokens. Model trained from scratch on RTX 3090 24GB.
|
34 |
|
35 |
# Supported tasks:
|
|
|
61 |
result = model.generate(x, return_dict_in_generate=True, output_scores=True,max_length=128)
|
62 |
print(tokenizer.decode(result["sequences"][0]))
|
63 |
```
|
64 |
+
</details>
|
65 |
+
|
66 |
+
# References
|
67 |
+
- [Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer](https://jmlr.org/papers/volume21/20-074/20-074.pdf)
|