I have a .pb model, which I want to use as a custom MLKit model. MLKit only supports .tflite models, but even after I use toco
to get the TensorFlow Lite model, the file size is too large for Firebase (95 MB and only 40 MB allowed).
Is there a way to quantize the graph and then convert to TFLite or quantize a .tflite graph?
When I do the former, I get the following error message: Unsupported TensorFlow op: Dequantize) for which the quantized form is not yet implemented. Sorry, and patches welcome (that's a relatively fun patch to write, mostly providing the actual quantized arithmetic code for this op).