Update README.md
Browse files
README.md
CHANGED
@@ -2,11 +2,13 @@
|
|
2 |
language:
|
3 |
- gl
|
4 |
licence:
|
5 |
-
-
|
6 |
tags:
|
7 |
- galician
|
8 |
- FLOR
|
9 |
- bloom
|
|
|
|
|
10 |
---
|
11 |
|
12 |
# FLOR-1.3B-GL
|
@@ -33,7 +35,6 @@ tags:
|
|
33 |
- [Copyright](#copyright)
|
34 |
- [License](#license)
|
35 |
- [Funding](#funding)
|
36 |
-
- [Disclaimer](#disclaimer)
|
37 |
|
38 |
</details>
|
39 |
|
@@ -91,24 +92,7 @@ The language adaptation technique used to train FLOR-1.3B-GL is based in the use
|
|
91 |
|
92 |
### Training data
|
93 |
|
94 |
-
|
95 |
-
It consists of 26B tokens of several corpora gathered from web crawlings and public domain data.
|
96 |
-
|
97 |
-
| Dataset | Language | Words (per-epoch) | Epochs |
|
98 |
-
|---------------------|----------|--------------------|--------------|
|
99 |
-
| Wikipedia | en | 2169.97M | 1.428144485 |
|
100 |
-
| C4_es | es | 53709.80M | 0.1049686196 |
|
101 |
-
| Biomedical | es | 455.03M | 0.7140722425 |
|
102 |
-
| Legal | es | 995.70M | 0.7140722425 |
|
103 |
-
| Wikipedia | es | 693.60M | 1.428144485 |
|
104 |
-
| Gutenberg | es | 53.18M | 0.7140722425 |
|
105 |
-
| C4_ca | ca | 2826.00M | 2.142216727 |
|
106 |
-
| Biomedical | ca | 11.80M | 1.428144485 |
|
107 |
-
| RacoCatalà Noticias | ca | 17.16M | 2.142216727 |
|
108 |
-
| RacoCatalà Forums | ca | 333.73M | 2.142216727 |
|
109 |
-
| CaWaC | ca | 57.79M | 2.142216727 |
|
110 |
-
| Wikipedia | ca | 228.01M | 3.570361212 |
|
111 |
-
| Vilaweb | ca | 50.34M | 2.142216727 |
|
112 |
|
113 |
|
114 |
### Training hyperparameters
|
|
|
2 |
language:
|
3 |
- gl
|
4 |
licence:
|
5 |
+
- MIT
|
6 |
tags:
|
7 |
- galician
|
8 |
- FLOR
|
9 |
- bloom
|
10 |
+
license: mit
|
11 |
+
pipeline_tag: text-generation
|
12 |
---
|
13 |
|
14 |
# FLOR-1.3B-GL
|
|
|
35 |
- [Copyright](#copyright)
|
36 |
- [License](#license)
|
37 |
- [Funding](#funding)
|
|
|
38 |
|
39 |
</details>
|
40 |
|
|
|
92 |
|
93 |
### Training data
|
94 |
|
95 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
96 |
|
97 |
|
98 |
### Training hyperparameters
|