You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
TensorFlow Lite now supports converting weights to 16-bit floating point values during model conversion from TensorFlow to TensorFlow Lite's flat buffer format. This results in a 2x reduction in model size.
However, this will insert new dequantize for ops (like conv2d) used for dequantize fp16 weight to fp32. Like this:
TVM doesn't support this behavior. List the things we mainly should to do:
TensorFlow Lite now supports converting weights to 16-bit floating point values during model conversion from TensorFlow to TensorFlow Lite's flat buffer format. This results in a 2x reduction in model size.
However, this will insert new
dequantize
for ops (likeconv2d
) used fordequantize
fp16 weight to fp32. Like this:TVM doesn't support this behavior. List the things we mainly should to do:
dequantize
to support fp16 to fp32Related issue:#5774
The text was updated successfully, but these errors were encountered: