Datasets:
init
Browse files
training_scripts/finetune_t5.py
CHANGED
@@ -137,7 +137,7 @@ def train(
|
|
137 |
|
138 |
if not skip_train:
|
139 |
lr = [1e-6, 1e-5, 1e-4] if lr is None else lr
|
140 |
-
batch = [
|
141 |
epoch = [3, 5] if not epoch else epoch
|
142 |
eval_batch_size = min(batch) if not eval_batch_size else eval_batch_size
|
143 |
for n, (lr_tmp, batch_tmp, epoch_tmp) in enumerate(product(lr, batch, epoch)):
|
|
|
137 |
|
138 |
if not skip_train:
|
139 |
lr = [1e-6, 1e-5, 1e-4] if lr is None else lr
|
140 |
+
batch = [32] if not batch else batch
|
141 |
epoch = [3, 5] if not epoch else epoch
|
142 |
eval_batch_size = min(batch) if not eval_batch_size else eval_batch_size
|
143 |
for n, (lr_tmp, batch_tmp, epoch_tmp) in enumerate(product(lr, batch, epoch)):
|
training_scripts/script.sh
CHANGED
@@ -1,6 +1,8 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
python finetune_t5.py --dataset-name
|
4 |
-
python finetune_t5.py --dataset-name
|
|
|
|
|
5 |
|
6 |
|
|
|
1 |
+
|
2 |
+
|
3 |
+
python finetune_t5.py --dataset-name ja --low-cpu-mem-usage --model-alias mt5-small-tweet-topic-multi-ja --model-organization cardiffnlp
|
4 |
+
python finetune_t5.py --dataset-name gr --low-cpu-mem-usage --model-alias mt5-small-tweet-topic-multi-gr --model-organization cardiffnlp
|
5 |
+
python finetune_t5.py --dataset-name es --low-cpu-mem-usage --model-alias mt5-small-tweet-topic-multi-es --model-organization cardiffnlp
|
6 |
+
python finetune_t5.py --dataset-name en --low-cpu-mem-usage --model-alias mt5-small-tweet-topic-multi-en --model-organization cardiffnlp
|
7 |
|
8 |
|