jarodrigues commited on
Commit
ce02612
1 Parent(s): 88fcd72

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -2
README.md CHANGED
@@ -106,6 +106,10 @@ These datasets were machine translated into Portuguese and from the [extraGLUE](
106
  Furthermore, instruction templates have been manually crafted for each task.
107
  These take the various fields in the dataset and arrange them into prompts, which were collected into the [extraGLUE-instruct](https://huggingface.co/datasets/PORTULAN/extraglue-instruct) dataset.
108
 
 
 
 
 
109
  # Training Details
110
 
111
  We applied supervised fine-tuning with a causal language modeling training objective following a zero-out technique during the fine-tuning process.
@@ -120,8 +124,7 @@ In other words, each example occupies the full input sequence length.
120
  # Evaluation
121
 
122
  For testing, we reserved the translated datasets MRPC (similarity) and RTE (inference), from GLUE, and COPA (reasoning/qa), from SuperGLUE, which were taken as representatives of three major types of tasks, and were not seen during training.
123
- We also employ data augmentation techniques to enhance the size and diversity of our dataset.
124
- This involves repurposing the tasks in various ways, such as generation of answers from MultiRC, question generation from BoolQ, and other relevant modifications.
125
 
126
 
127
  | Model | MRPC (F1) | RTE (F1) | COPA (F1) |
 
106
  Furthermore, instruction templates have been manually crafted for each task.
107
  These take the various fields in the dataset and arrange them into prompts, which were collected into the [extraGLUE-instruct](https://huggingface.co/datasets/PORTULAN/extraglue-instruct) dataset.
108
 
109
+ We also employ data augmentation techniques to enhance the size and diversity of our dataset.
110
+ This involves repurposing the tasks in various ways, such as generation of answers from MultiRC, question generation from BoolQ, and other relevant modifications.
111
+
112
+
113
  # Training Details
114
 
115
  We applied supervised fine-tuning with a causal language modeling training objective following a zero-out technique during the fine-tuning process.
 
124
  # Evaluation
125
 
126
  For testing, we reserved the translated datasets MRPC (similarity) and RTE (inference), from GLUE, and COPA (reasoning/qa), from SuperGLUE, which were taken as representatives of three major types of tasks, and were not seen during training.
127
+
 
128
 
129
 
130
  | Model | MRPC (F1) | RTE (F1) | COPA (F1) |