Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

key points extraction in textureless region #3

Open
zhaoxin111 opened this issue Mar 4, 2021 · 4 comments
Open

key points extraction in textureless region #3

zhaoxin111 opened this issue Mar 4, 2021 · 4 comments

Comments

@zhaoxin111
Copy link

Excellent job, but I have a question. When there is a large area of sky in the picture with little texture, can the key point detection model based on CNN be able to effectively extract the key points?

@Annbless
Copy link
Owner

Sorry for the late reply. I am afraid that the CNN-based keypoint detection model cannot extract enough keypoints in such areas without texture and edges. I visualized the keypoints extracted in the case you mentioned. The CNN model can extract keypoints on the edges of clouds, but not in texture-free regions such as the sky. However, this does not affect the final performance of the video stabilization, because the distortion of the texture-free regions is insignificant.

@zhaoxin111
Copy link
Author

Thank you for your reply, I noticed that in the ablation study, replacing the FAST keypoint detector and the KLT tracker with RFNet and PWCNet respectively, the score of stability, distortion and cropping is evey close to the DUT results. Is that mean the keypoint detector and tracker are the key resulting in good performence in DUT? Will traditional method like MeshFlow also use deep learning-based feature extractors and trackers also have a significant performance improvement?

@Annbless
Copy link
Owner

Not exactly. Please refer to Figures 3 and 5 in the paper. The distortion metric only describes the global artifacts as the frame ratio changes, but it does not measure the artifacts that affect visual quality. The deep learning-based motion estimation and trajectory smoothing module proposed in this paper is specifically designed to correct such artifacts and thus polish the visual experience. Moreover, this is the reason why we use additional metrics to measure the performance of the motion estimation module. Does this answer solve your problem?

@zhaoxin111
Copy link
Author

Your answer basically answered my doubts. I will read your paper again, and there should be new gains.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants