jarodrigues
commited on
Commit
•
ce02612
1
Parent(s):
88fcd72
Update README.md
Browse files
README.md
CHANGED
@@ -106,6 +106,10 @@ These datasets were machine translated into Portuguese and from the [extraGLUE](
|
|
106 |
Furthermore, instruction templates have been manually crafted for each task.
|
107 |
These take the various fields in the dataset and arrange them into prompts, which were collected into the [extraGLUE-instruct](https://huggingface.co/datasets/PORTULAN/extraglue-instruct) dataset.
|
108 |
|
|
|
|
|
|
|
|
|
109 |
# Training Details
|
110 |
|
111 |
We applied supervised fine-tuning with a causal language modeling training objective following a zero-out technique during the fine-tuning process.
|
@@ -120,8 +124,7 @@ In other words, each example occupies the full input sequence length.
|
|
120 |
# Evaluation
|
121 |
|
122 |
For testing, we reserved the translated datasets MRPC (similarity) and RTE (inference), from GLUE, and COPA (reasoning/qa), from SuperGLUE, which were taken as representatives of three major types of tasks, and were not seen during training.
|
123 |
-
|
124 |
-
This involves repurposing the tasks in various ways, such as generation of answers from MultiRC, question generation from BoolQ, and other relevant modifications.
|
125 |
|
126 |
|
127 |
| Model | MRPC (F1) | RTE (F1) | COPA (F1) |
|
|
|
106 |
Furthermore, instruction templates have been manually crafted for each task.
|
107 |
These take the various fields in the dataset and arrange them into prompts, which were collected into the [extraGLUE-instruct](https://huggingface.co/datasets/PORTULAN/extraglue-instruct) dataset.
|
108 |
|
109 |
+
We also employ data augmentation techniques to enhance the size and diversity of our dataset.
|
110 |
+
This involves repurposing the tasks in various ways, such as generation of answers from MultiRC, question generation from BoolQ, and other relevant modifications.
|
111 |
+
|
112 |
+
|
113 |
# Training Details
|
114 |
|
115 |
We applied supervised fine-tuning with a causal language modeling training objective following a zero-out technique during the fine-tuning process.
|
|
|
124 |
# Evaluation
|
125 |
|
126 |
For testing, we reserved the translated datasets MRPC (similarity) and RTE (inference), from GLUE, and COPA (reasoning/qa), from SuperGLUE, which were taken as representatives of three major types of tasks, and were not seen during training.
|
127 |
+
|
|
|
128 |
|
129 |
|
130 |
| Model | MRPC (F1) | RTE (F1) | COPA (F1) |
|