quantized models
#3
by
Sio-Nii
- opened
there is several quantized model files with same int8 size
model.onnx + model.onnx.data
model.ort
model.with_runtime_opt.ort
and another model.onnx + model.onnx.data in fp16fullonnxsdquantized folder
what difference between them?
which one we need use?
You use ort models for android