File size: 511 Bytes
e3c7b5a
 
 
64a6414
 
e3c7b5a
64a6414
 
e3c7b5a
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
train:
  ver: m-1-2
  desc: just overfitting the model, but on the entire PIE dataset.
  bart: facebook/bart-base
  lr: 0.0001
  literal2idiomatic_ver: d-1-2
  max_epochs: 100
  batch_size: 100
  shuffle: true

# for building & uploading datasets or others
upload:
  idioms:
    ver: d-1-2
    description: the set of idioms in the traning set of literal2idiomatic:d-1-2
  literal2idiomatic:
    ver: d-1-2
    description: PIE data split into train & test set (80 / 20 split)
    train_ratio: 0.8
    seed: 104