Update safety_checker/config.json
Browse files
safety_checker/config.json
CHANGED
|
@@ -14,7 +14,6 @@
|
|
| 14 |
"architectures": null,
|
| 15 |
"attention_dropout": 0.0,
|
| 16 |
"bad_words_ids": null,
|
| 17 |
-
"begin_suppress_tokens": null,
|
| 18 |
"bos_token_id": 0,
|
| 19 |
"chunk_size_feed_forward": 0,
|
| 20 |
"cross_attention_hidden_size": null,
|
|
@@ -68,7 +67,6 @@
|
|
| 68 |
"return_dict": true,
|
| 69 |
"return_dict_in_generate": false,
|
| 70 |
"sep_token_id": null,
|
| 71 |
-
"suppress_tokens": null,
|
| 72 |
"task_specific_params": null,
|
| 73 |
"temperature": 1.0,
|
| 74 |
"tf_legacy_loss": false,
|
|
@@ -79,7 +77,7 @@
|
|
| 79 |
"top_p": 1.0,
|
| 80 |
"torch_dtype": null,
|
| 81 |
"torchscript": false,
|
| 82 |
-
"transformers_version": "4.
|
| 83 |
"typical_p": 1.0,
|
| 84 |
"use_bfloat16": false,
|
| 85 |
"vocab_size": 49408
|
|
@@ -98,7 +96,6 @@
|
|
| 98 |
"architectures": null,
|
| 99 |
"attention_dropout": 0.0,
|
| 100 |
"bad_words_ids": null,
|
| 101 |
-
"begin_suppress_tokens": null,
|
| 102 |
"bos_token_id": null,
|
| 103 |
"chunk_size_feed_forward": 0,
|
| 104 |
"cross_attention_hidden_size": null,
|
|
@@ -154,7 +151,6 @@
|
|
| 154 |
"return_dict": true,
|
| 155 |
"return_dict_in_generate": false,
|
| 156 |
"sep_token_id": null,
|
| 157 |
-
"suppress_tokens": null,
|
| 158 |
"task_specific_params": null,
|
| 159 |
"temperature": 1.0,
|
| 160 |
"tf_legacy_loss": false,
|
|
@@ -165,7 +161,7 @@
|
|
| 165 |
"top_p": 1.0,
|
| 166 |
"torch_dtype": null,
|
| 167 |
"torchscript": false,
|
| 168 |
-
"transformers_version": "4.
|
| 169 |
"typical_p": 1.0,
|
| 170 |
"use_bfloat16": false
|
| 171 |
},
|
|
@@ -176,4 +172,4 @@
|
|
| 176 |
"num_hidden_layers": 24,
|
| 177 |
"patch_size": 14
|
| 178 |
}
|
| 179 |
-
}
|
|
|
|
| 14 |
"architectures": null,
|
| 15 |
"attention_dropout": 0.0,
|
| 16 |
"bad_words_ids": null,
|
|
|
|
| 17 |
"bos_token_id": 0,
|
| 18 |
"chunk_size_feed_forward": 0,
|
| 19 |
"cross_attention_hidden_size": null,
|
|
|
|
| 67 |
"return_dict": true,
|
| 68 |
"return_dict_in_generate": false,
|
| 69 |
"sep_token_id": null,
|
|
|
|
| 70 |
"task_specific_params": null,
|
| 71 |
"temperature": 1.0,
|
| 72 |
"tf_legacy_loss": false,
|
|
|
|
| 77 |
"top_p": 1.0,
|
| 78 |
"torch_dtype": null,
|
| 79 |
"torchscript": false,
|
| 80 |
+
"transformers_version": "4.22.0.dev0",
|
| 81 |
"typical_p": 1.0,
|
| 82 |
"use_bfloat16": false,
|
| 83 |
"vocab_size": 49408
|
|
|
|
| 96 |
"architectures": null,
|
| 97 |
"attention_dropout": 0.0,
|
| 98 |
"bad_words_ids": null,
|
|
|
|
| 99 |
"bos_token_id": null,
|
| 100 |
"chunk_size_feed_forward": 0,
|
| 101 |
"cross_attention_hidden_size": null,
|
|
|
|
| 151 |
"return_dict": true,
|
| 152 |
"return_dict_in_generate": false,
|
| 153 |
"sep_token_id": null,
|
|
|
|
| 154 |
"task_specific_params": null,
|
| 155 |
"temperature": 1.0,
|
| 156 |
"tf_legacy_loss": false,
|
|
|
|
| 161 |
"top_p": 1.0,
|
| 162 |
"torch_dtype": null,
|
| 163 |
"torchscript": false,
|
| 164 |
+
"transformers_version": "4.22.0.dev0",
|
| 165 |
"typical_p": 1.0,
|
| 166 |
"use_bfloat16": false
|
| 167 |
},
|
|
|
|
| 172 |
"num_hidden_layers": 24,
|
| 173 |
"patch_size": 14
|
| 174 |
}
|
| 175 |
+
}
|