add onnx file + quantized model file
Browse files- model.onnx +3 -0
- model_quantized.onnx +3 -0
- ort_config.json +33 -0
model.onnx
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fa24fc03bbfd34bc3db3a25512196c71d7eac76e13a011efd81c4db275450235
|
3 |
+
size 47172693
|
model_quantized.onnx
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2daec03215e030a45dad9b835e2d0e79dd99455cda100245785ac99841a82f42
|
3 |
+
size 40805606
|
ort_config.json
ADDED
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"one_external_file": true,
|
3 |
+
"opset": null,
|
4 |
+
"optimization": {},
|
5 |
+
"quantization": {
|
6 |
+
"activations_dtype": "QUInt8",
|
7 |
+
"activations_symmetric": false,
|
8 |
+
"format": "QOperator",
|
9 |
+
"is_static": false,
|
10 |
+
"mode": "IntegerOps",
|
11 |
+
"nodes_to_exclude": [],
|
12 |
+
"nodes_to_quantize": [],
|
13 |
+
"operators_to_quantize": [
|
14 |
+
"Conv",
|
15 |
+
"MatMul",
|
16 |
+
"Attention",
|
17 |
+
"LSTM",
|
18 |
+
"Gather",
|
19 |
+
"Transpose",
|
20 |
+
"EmbedLayerNormalization"
|
21 |
+
],
|
22 |
+
"per_channel": false,
|
23 |
+
"qdq_add_pair_to_weight": false,
|
24 |
+
"qdq_dedicated_pair": false,
|
25 |
+
"qdq_op_type_per_channel_support_to_axis": {
|
26 |
+
"MatMul": 1
|
27 |
+
},
|
28 |
+
"reduce_range": false,
|
29 |
+
"weights_dtype": "QInt8",
|
30 |
+
"weights_symmetric": true
|
31 |
+
},
|
32 |
+
"use_external_data_format": false
|
33 |
+
}
|