Update README.md
Browse files
README.md
CHANGED
@@ -26,10 +26,10 @@ It has different versions that were trained for different variants of Portuguese
|
|
26 |
namely the European variant from Portugal (PT-PT) and the American variant from Brazil (PT-BR),
|
27 |
and it is distributed free of charge and under a most permissible license.
|
28 |
|
29 |
-
**Albertina PT-
|
30 |
and to the best of our knowledge, at the time of its initial distribution,
|
31 |
-
it
|
32 |
-
that had been made publicly available and distributed for reuse.
|
33 |
|
34 |
It was developped by a joint team from the University of Lisbon and the University of Porto, Portugal.
|
35 |
For further details, check the respective publication:
|
@@ -55,16 +55,21 @@ Please use the above cannonical reference when using or citing this model.
|
|
55 |
|
56 |
# Model Description
|
57 |
|
58 |
-
**This model card is for Albertina-PT-
|
59 |
|
60 |
-
This model is distributed free of charge
|
|
|
61 |
|
62 |
|
63 |
<br>
|
64 |
|
65 |
# Training Data
|
66 |
|
67 |
-
|
|
|
|
|
|
|
|
|
68 |
|
69 |
- [OSCAR](https://huggingface.co/datasets/oscar-corpus/OSCAR-2301): the OSCAR data set includes documents in more than one hundred languages, including Portuguese, and it is widely used in the literature. It is the result of a selection performed over the [Common Crawl](https://commoncrawl.org/) data set, crawled from the Web, that retains only pages whose metadata indicates permission to be crawled, that performs deduplication, and that removes some boilerplate, among other filters. Given that it does not discriminate between the Portuguese variants, we performed extra filtering by retaining only documents whose meta-data indicate the Internet country code top-level domain of Portugal. We used the January 2023 version of OSCAR, which is based on the November/December 2022 version of Common Crawl.
|
70 |
- [DCEP](https://joint-research-centre.ec.europa.eu/language-technology-resources/dcep-digital-corpus-european-parliament_en): the Digital Corpus of the European Parliament is a multilingual corpus including documents in all official EU languages published on the European Parliament's official website. We retained its European Portuguese portion.
|
@@ -72,8 +77,6 @@ This model is distributed free of charge under the [MIT](https://choosealicense.
|
|
72 |
- [ParlamentoPT](https://www.parlamento.pt/): the ParlamentoPT is a data set we obtained by gathering the publicly available documents with the transcription of the debates in the Portuguese Parliament.
|
73 |
|
74 |
|
75 |
-
**Albertina PT-BR**, in turn, was trained over the [BrWac](https://huggingface.co/datasets/brwac) data set.
|
76 |
-
|
77 |
|
78 |
## Preprocessing
|
79 |
|
@@ -85,12 +88,6 @@ We skipped the default filtering of stopwords since it would disrupt the syntact
|
|
85 |
|
86 |
As codebase, we resorted to the [DeBERTa V2 XLarge](https://huggingface.co/microsoft/deberta-v2-xlarge), for English.
|
87 |
|
88 |
-
To train **Albertina-PT-PT**, the data set was tokenized with the original DeBERTa tokenizer with a 128 token sequence truncation and dynamic padding.
|
89 |
-
The model was trained using the maximum available memory capacity resulting in a batch size of 832 samples (52 samples per GPU and applying gradient accumulation in order to approximate the batch size of the PT-BR model).
|
90 |
-
Similarly to the PT-BR variant above, we opted for a learning rate of 1e-5 with linear decay and 10k warm-up steps.
|
91 |
-
However, since the number of training examples is approximately twice of that in the PT-BR variant, we reduced the number of training epochs to half and completed only 25 epochs, which resulted in approximately 245k steps.
|
92 |
-
The model was trained for 3 days on a2-highgpu-8gb Google Cloud A2 VMs with 8 GPUs, 96 vCPUs and 680 GB of RAM.
|
93 |
-
|
94 |
To train **Albertina-PT-BR** the BrWac data set was tokenized with the original DeBERTA tokenizer with a 128 token sequence truncation and dynamic padding.
|
95 |
The model was trained using the maximum available memory capacity resulting in a batch size of 896 samples (56 samples per GPU without gradient accumulation steps).
|
96 |
We chose a learning rate of 1e-5 with linear decay and 10k warm-up steps based on the results of exploratory experiments.
|
@@ -98,6 +95,13 @@ In total, around 200k training steps were taken across 50 epochs.
|
|
98 |
The model was trained for 1 day and 11 hours on a2-megagpu-16gb Google Cloud A2 VMs with 16 GPUs, 96 vCPUs and 1.360 GB of RAM.
|
99 |
|
100 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
101 |
<br>
|
102 |
|
103 |
# Evaluation
|
|
|
26 |
namely the European variant from Portugal (PT-PT) and the American variant from Brazil (PT-BR),
|
27 |
and it is distributed free of charge and under a most permissible license.
|
28 |
|
29 |
+
**Albertina PT-BR** is the version for **American Portuguese from Brazil**,
|
30 |
and to the best of our knowledge, at the time of its initial distribution,
|
31 |
+
it set a new state of the art for this language and variant
|
32 |
+
for a model that had been made publicly available and distributed for reuse.
|
33 |
|
34 |
It was developped by a joint team from the University of Lisbon and the University of Porto, Portugal.
|
35 |
For further details, check the respective publication:
|
|
|
55 |
|
56 |
# Model Description
|
57 |
|
58 |
+
**This model card is for Albertina-PT-BR**, with 900M parameters, 24 layers and a hidden size of 1536.
|
59 |
|
60 |
+
This model is distributed free of charge with the license granted to the [BrWac](https://huggingface.co/datasets/brwac) data set on which it was trained,
|
61 |
+
a license for non-commercial use.
|
62 |
|
63 |
|
64 |
<br>
|
65 |
|
66 |
# Training Data
|
67 |
|
68 |
+
|
69 |
+
**Albertina PT-BR** was trained over the [BrWac](https://huggingface.co/datasets/brwac) data set.
|
70 |
+
|
71 |
+
|
72 |
+
**Albertina PT-PT**, in turn, was trained over a data set that resulted from gathering some openly available corpora of European Portuguese from the following sources:
|
73 |
|
74 |
- [OSCAR](https://huggingface.co/datasets/oscar-corpus/OSCAR-2301): the OSCAR data set includes documents in more than one hundred languages, including Portuguese, and it is widely used in the literature. It is the result of a selection performed over the [Common Crawl](https://commoncrawl.org/) data set, crawled from the Web, that retains only pages whose metadata indicates permission to be crawled, that performs deduplication, and that removes some boilerplate, among other filters. Given that it does not discriminate between the Portuguese variants, we performed extra filtering by retaining only documents whose meta-data indicate the Internet country code top-level domain of Portugal. We used the January 2023 version of OSCAR, which is based on the November/December 2022 version of Common Crawl.
|
75 |
- [DCEP](https://joint-research-centre.ec.europa.eu/language-technology-resources/dcep-digital-corpus-european-parliament_en): the Digital Corpus of the European Parliament is a multilingual corpus including documents in all official EU languages published on the European Parliament's official website. We retained its European Portuguese portion.
|
|
|
77 |
- [ParlamentoPT](https://www.parlamento.pt/): the ParlamentoPT is a data set we obtained by gathering the publicly available documents with the transcription of the debates in the Portuguese Parliament.
|
78 |
|
79 |
|
|
|
|
|
80 |
|
81 |
## Preprocessing
|
82 |
|
|
|
88 |
|
89 |
As codebase, we resorted to the [DeBERTa V2 XLarge](https://huggingface.co/microsoft/deberta-v2-xlarge), for English.
|
90 |
|
|
|
|
|
|
|
|
|
|
|
|
|
91 |
To train **Albertina-PT-BR** the BrWac data set was tokenized with the original DeBERTA tokenizer with a 128 token sequence truncation and dynamic padding.
|
92 |
The model was trained using the maximum available memory capacity resulting in a batch size of 896 samples (56 samples per GPU without gradient accumulation steps).
|
93 |
We chose a learning rate of 1e-5 with linear decay and 10k warm-up steps based on the results of exploratory experiments.
|
|
|
95 |
The model was trained for 1 day and 11 hours on a2-megagpu-16gb Google Cloud A2 VMs with 16 GPUs, 96 vCPUs and 1.360 GB of RAM.
|
96 |
|
97 |
|
98 |
+
To train **Albertina-PT-PT**, the data set was tokenized with the original DeBERTa tokenizer with a 128 token sequence truncation and dynamic padding.
|
99 |
+
The model was trained using the maximum available memory capacity resulting in a batch size of 832 samples (52 samples per GPU and applying gradient accumulation in order to approximate the batch size of the PT-BR model).
|
100 |
+
Similarly to the PT-BR variant above, we opted for a learning rate of 1e-5 with linear decay and 10k warm-up steps.
|
101 |
+
However, since the number of training examples is approximately twice of that in the PT-BR variant, we reduced the number of training epochs to half and completed only 25 epochs, which resulted in approximately 245k steps.
|
102 |
+
The model was trained for 3 days on a2-highgpu-8gb Google Cloud A2 VMs with 8 GPUs, 96 vCPUs and 680 GB of RAM.
|
103 |
+
|
104 |
+
|
105 |
<br>
|
106 |
|
107 |
# Evaluation
|