Titouan P
commited on
Commit
·
4301d51
1
Parent(s):
c806d04
update model
Browse files- config.json +10 -1
config.json
CHANGED
@@ -6,6 +6,8 @@
|
|
6 |
],
|
7 |
"attention_dropout": 0.1,
|
8 |
"bos_token_id": 1,
|
|
|
|
|
9 |
"conv_bias": true,
|
10 |
"conv_dim": [
|
11 |
512,
|
@@ -36,12 +38,14 @@
|
|
36 |
],
|
37 |
"ctc_loss_reduction": "sum",
|
38 |
"ctc_zero_infinity": false,
|
|
|
39 |
"do_stable_layer_norm": true,
|
40 |
"eos_token_id": 2,
|
41 |
"feat_extract_activation": "gelu",
|
42 |
"feat_extract_dropout": 0.0,
|
43 |
"feat_extract_norm": "layer",
|
44 |
"feat_proj_dropout": 0.1,
|
|
|
45 |
"final_dropout": 0.0,
|
46 |
"gradient_checkpointing": false,
|
47 |
"hidden_act": "gelu",
|
@@ -65,11 +69,16 @@
|
|
65 |
"mask_time_selection": "static",
|
66 |
"model_type": "wav2vec2",
|
67 |
"num_attention_heads": 16,
|
|
|
|
|
68 |
"num_conv_pos_embedding_groups": 16,
|
69 |
"num_conv_pos_embeddings": 128,
|
70 |
"num_feat_extract_layers": 7,
|
71 |
"num_hidden_layers": 24,
|
|
|
72 |
"pad_token_id": 0,
|
73 |
-
"
|
|
|
|
|
74 |
"vocab_size": 32
|
75 |
}
|
|
|
6 |
],
|
7 |
"attention_dropout": 0.1,
|
8 |
"bos_token_id": 1,
|
9 |
+
"codevector_dim": 384,
|
10 |
+
"contrastive_logits_temperature": 0.1,
|
11 |
"conv_bias": true,
|
12 |
"conv_dim": [
|
13 |
512,
|
|
|
38 |
],
|
39 |
"ctc_loss_reduction": "sum",
|
40 |
"ctc_zero_infinity": false,
|
41 |
+
"diversity_loss_weight": 0.1,
|
42 |
"do_stable_layer_norm": true,
|
43 |
"eos_token_id": 2,
|
44 |
"feat_extract_activation": "gelu",
|
45 |
"feat_extract_dropout": 0.0,
|
46 |
"feat_extract_norm": "layer",
|
47 |
"feat_proj_dropout": 0.1,
|
48 |
+
"feat_quantizer_dropout": 0.0,
|
49 |
"final_dropout": 0.0,
|
50 |
"gradient_checkpointing": false,
|
51 |
"hidden_act": "gelu",
|
|
|
69 |
"mask_time_selection": "static",
|
70 |
"model_type": "wav2vec2",
|
71 |
"num_attention_heads": 16,
|
72 |
+
"num_codevector_groups": 2,
|
73 |
+
"num_codevectors_per_group": 320,
|
74 |
"num_conv_pos_embedding_groups": 16,
|
75 |
"num_conv_pos_embeddings": 128,
|
76 |
"num_feat_extract_layers": 7,
|
77 |
"num_hidden_layers": 24,
|
78 |
+
"num_negatives": 100,
|
79 |
"pad_token_id": 0,
|
80 |
+
"proj_codevector_dim": 256,
|
81 |
+
"torch_dtype": "float32",
|
82 |
+
"transformers_version": "4.9.2",
|
83 |
"vocab_size": 32
|
84 |
}
|