You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I had the same issue and address it with a quick fix.
I dont have enough background in NN to fix it properly.
TLDR; make sure the output of transform method has the same dimensions as your output video.
Context and debug
The issue come from the shape reduction and augmentation in the src/transform.py file that lead to a different format.
For example with a video of 640x338 you will output frames of 640x340, if you debug the network layer by layer
conv1 shape = (4, 338, 640, 32)
conv2 shape = (4, 169, 320, 64)
conv3 shape = (4, 85, 160, 128)
resid1 shape = (4, 85, 160, 128)
resid2 shape = (4, 85, 160, 128)
resid3 shape = (4, 85, 160, 128)
resid4 shape = (4, 85, 160, 128)
resid5 shape = (4, 85, 160, 128)
conv_t1 shape = (4, 170, 320, 64)
conv_t2 shape = (4, 340, 640, 32)
You can see the shape of each layer and notice the difference between the first and the last. (640x338 vs 640x340)
338 --> 169 --> 85 (instead of 84.5) --> 170 --> 340
In the evaluate.py file, when you write the video, the output file is written with frames of same size as the original clip, nethertheless the NN will output 640x340 images and this will slightly move your video frame by frame (in my example, 2 row at a time)
Quick fix :
Compute the video output size based on the preds shape instead of original clip
when i transforming video, the result video is frame malposition,see below two pic
is there other partner get it?
The text was updated successfully, but these errors were encountered: