Commit
·
47353cb
1
Parent(s):
b351c4a
Update README.md
Browse files
README.md
CHANGED
@@ -23,7 +23,7 @@ pipeline_tag: text-to-image
|
|
23 |
In order to improve the [RoBERTa-large-bne](https://huggingface.co/PlanTL-GOB-ES/roberta-large-bne) encoder performance, this model has been trained using the generated corpus ([in this respository](https://huggingface.co/oeg/RoBERTa-CelebA-Sp/))
|
24 |
and following the strategy of using a Siamese network together with the loss function of cosine similarity. The following steps were followed:
|
25 |
- Define [sentence-transformer](https://www.sbert.net/) and _torch_ libraries for the implementation of the encoder.
|
26 |
-
- Divide the training corpus into two parts, training with 249,
|
27 |
- Load training / validation data for the model. Two lists are generated for the storage of the information and, in each of them,
|
28 |
the entries are composed of a pair of descriptive sentences and their similarity value.
|
29 |
- Implement [RoBERTa-large-bne](https://huggingface.co/PlanTL-GOB-ES/roberta-large-bne) as a baseline model for transformer training.
|
|
|
23 |
In order to improve the [RoBERTa-large-bne](https://huggingface.co/PlanTL-GOB-ES/roberta-large-bne) encoder performance, this model has been trained using the generated corpus ([in this respository](https://huggingface.co/oeg/RoBERTa-CelebA-Sp/))
|
24 |
and following the strategy of using a Siamese network together with the loss function of cosine similarity. The following steps were followed:
|
25 |
- Define [sentence-transformer](https://www.sbert.net/) and _torch_ libraries for the implementation of the encoder.
|
26 |
+
- Divide the training corpus into two parts, training with 249,000 sentences and validation with 1,000 sentences.
|
27 |
- Load training / validation data for the model. Two lists are generated for the storage of the information and, in each of them,
|
28 |
the entries are composed of a pair of descriptive sentences and their similarity value.
|
29 |
- Implement [RoBERTa-large-bne](https://huggingface.co/PlanTL-GOB-ES/roberta-large-bne) as a baseline model for transformer training.
|