-
Notifications
You must be signed in to change notification settings - Fork 51
Yolo-v3-tiny-tf model with INT-8 precision gives bad inferences #116
Comments
If you change the precision you need to specify it in the pipeline as per Referencing Models in Pipeline Definitions. See updated pipeline definition entry for gvadetect.
|
@whbruce These are the logs on the pipeline server, I could see that it is using the INT-8 models
over the inferencing side keeps incorrectly predicting bench continuously for uri=https://github.com/intel-iot-devkit/sample-videos/raw/master/bottle-detection.mp4?raw=true
|
Please run OMZ object detection demo to get a baseline for model accuracy. |
@whbruce These are the same model that i used for the pipeline server (setup) intel@intel-WL10:~/workspace/open_model_zoo/demos/object_detection_demo/python$ python3 object_detection_demo.py -d CPU -i bottle.mp4 -m /home/intel/workspace/pipeline-server/models/object_detection/yolo-v3-tiny-tf/FP32-INT8/yolo-v3-tiny-tf.xml -at yolo --labels /home/intel/workspace/pipeline-server/models/object_detection/yolo-v3-tiny-tf/coco-80cl.txt |
Using the default-algorithm for INT8 quantization (instead of using the "accuracy-aware" algorithm) can come with a slightly reduced accuracy... The sample "object_detection_demo.py" has a default-value
|
Hi,
I am able to use it with the omz object detection sample, However it seems to fail on the pipeline server, could you please suggest any solution for this issue |
Can you provide more details about the problem you are seeing, please? |
yes, @brmarkus |
Hi
We are working on integrating yolo-v3-tiny-tf int-8 IR model into the dlstreamer pipeline following the documentation provided for changing model
We were able to integrate yolo-v3-tiny-tf IR model (non quantized) and test it. but we failed to get proper inference with the INT-8 models for the same. these converted INT-8 model were validated using open model zoo sample for object detection and it was giving a proper inference.
The steps follow for conversion of the yolo-v3-tiny-tf to int8 model are provided below:
This Quantization document is based on yolo_v3_tiny_tf model
Requirements
Steps for Quantization
Step 1: Obtain the OMZ model (yolo_v3_tiny_tf)
This step downloads the frozen model and converts them to it's appropriate IR representation.
Step 2 : Obtain the DataSet for Optimization
For this Model COCO 2017 Validation dataset was selected
Step 3: Create a Json <quantization_spec.json> (Optional instead of pot arguments)
Note: Use FP16 to convert to FP16-INT8
Step 4: Use post optimization tool of the openvino to finish the process
This step converts the FP32/FP16 models to its FP32-INT8/FP16-INT8 models
the INT8 models will be available in "yolov3_int8" directory
Step 5 : Validation
Test the converted model with open model zoo demo object detection sample.
For integrating the pipeline server the steps followed are as per the document
copy the downloaded and converted model under
Directory structure looks something like this under yolo-v3-tiny-tf
Created new pipeline
Edited the template of pipeline.json
Ran the pipeline server
with this we were able to the inferencing with FP16, FP32 but we were not able to do inference with FP32-INT8 IR model
Could you please let us know what steps we are missing to integrate Quantized model ?
Thanks
The text was updated successfully, but these errors were encountered: