OpenSLU / app.yaml
LightChen2333's picture
Upload app.yaml
89c300b
raw
history blame
2.21 kB
device: "NVIDIA GeForce RTX 2080 Ti"
host: 127.0.0.1
port: 7860
is_push_to_public: false
save-path: save/stack/outputs.jsonl
page-size: 2
base:
name: "OpenSLUv1"
train: false
test: false
device: cpu
ckpt_path: null
seed: 42
best_key: EMA
epoch_num: 300
batch_size: 16
eval_by_epoch: true
model_dir: save/stack
template: application.html
accelerator:
use_accelerator: false
dataset:
dataset_name: atis
metric:
- intent_acc
- slot_f1
- EMA
tokenizer:
_tokenizer_name_: word_tokenizer
_padding_side_: right
_align_mode_: fast
_to_lower_case_: true
add_special_tokens: false
max_length: 512
optimizer:
_model_target_: torch.optim.Adam
_model_partial_: true
lr: 0.001
weight_decay: 1e-6
scheduler:
_model_target_: transformers.get_scheduler
_model_partial_: true
name : "linear"
num_warmup_steps: 0
model:
_model_target_: model.OpenSLUModel
encoder:
_model_target_: model.encoder.AutoEncoder
encoder_name: self-attention-lstm
embedding:
embedding_dim: 256
dropout_rate: 0.55
lstm:
layer_num: 1
bidirectional: true
output_dim: 256
dropout_rate: 0.5
attention:
hidden_dim: 1024
output_dim: 128
dropout_rate: 0.6
return_with_input: true
return_sentence_level_hidden: false
decoder:
_model_target_: model.decoder.StackPropagationDecoder
interaction:
_model_target_: model.decoder.interaction.StackInteraction
differentiable: false
intent_classifier:
_model_target_: model.decoder.classifier.AutoregressiveLSTMClassifier
layer_num: 1
bidirectional: false
force_ratio: 0.9
hidden_dim: 64
embedding_dim: 8
ignore_index: -100
dropout_rate: 0.5
mode: "token-level-intent"
use_multi: false
return_sentence_level: true
slot_classifier:
_model_target_: model.decoder.classifier.AutoregressiveLSTMClassifier
layer_num: 1
bidirectional: false
force_ratio: 0.9
hidden_dim: 64
embedding_dim: 32
ignore_index: -100
dropout_rate: 0.55
mode: "slot"
use_multi: false
return_sentence_level: false