Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About previous sota pdc-net+ #44

Open
Moreland-cas opened this issue Sep 23, 2023 · 5 comments
Open

About previous sota pdc-net+ #44

Moreland-cas opened this issue Sep 23, 2023 · 5 comments

Comments

@Moreland-cas
Copy link

Why does PDC-net perform reasonably well in pck but much worse in two-view geometry estimation?
pdc-net pck:
image
image
image
Is it because the confidence it predicts is learned self-supervised? Also is it fair to compare DKM with pdc-net without retraining it fully supervised?
If I understand it wrong, please point it out, Thank you! : )

@Parskatt
Copy link
Owner

Hi!

The PCK in pdcnet is on a different subset of megadepth and therefore not directly comparable.

As for the confidence estimate, yes it helps, but if you qualitatively look at the results you will see pdcnet has way more failure cases (not involving certainty).

If you want a network-to-network comparison of pdcnet and DKM/RoMa I suggest you use the training set and strategy of DKM with pdcnet architecture. I'm not sure exactly how big the gain would be, but if you look at some of our ablations I think we have some pretty big improvements.

The point of our papers are not to say that we are better than pdcnet. Rather we want to show that dense matching is competitive for two view geometry.

@Parskatt
Copy link
Owner

Ah, another point is that pdcnet does actually train on megadepth (just different strategy), so I think the conparison is pretty fair.

@Moreland-cas
Copy link
Author

Thanks for your quick reply, I believe there are certainly some improvements to the architecture of DKM.
There are two more questions that are bothering me, could you please help me out?

  1. In the process of computing the fundamental matrix using OpenCV's cv2.findFundamentalMatrix, the algorithm typically uses a subset of points, such as 8-point correspondences, along with a RANSAC approach to filter out outliers. Given that the remaining points only contribute implicitly to the estimation of F, would it improve accuracy to re-estimate F using the inliers identified by the initial computation?
  2. Is it possible to employ the dense warp in a differentiable manner to compute F, thereby enabling the training of all point correspondences directly?

@Parskatt
Copy link
Owner

  1. Yes. Typically you would use a few LM optimization steps to improve. This is included in most opencv ransac, not sure about the most basic one. If using LO-RANSAC it would even be done for the scoring of the model using a local set.

  2. Yes. There are works in this direction and Ive been meaning to try it out myself as well.

@Moreland-cas
Copy link
Author

  1. Yes. Typically you would use a few LM optimization steps to improve. This is included in most opencv ransac, not sure about the most basic one. If using LO-RANSAC it would even be done for the scoring of the model using a local set.
  2. Yes. There are works in this direction and Ive been meaning to try it out myself as well.

Thanks again :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants