Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reproduction of the paper results #7

Open
cornerfarmer opened this issue Sep 2, 2020 · 8 comments
Open

Reproduction of the paper results #7

cornerfarmer opened this issue Sep 2, 2020 · 8 comments

Comments

@cornerfarmer
Copy link
Contributor

I am not able to reproduce the results on 7-Scenes listed in the paper.
In my eyes the code should be correct, but the model is totally overfitting on the training data.
While the training loss goes towards zero, the performance of model on the test data stagnates (e.q. mean median translational deviation is always >0.20m)

Is anybody else experiencing similar problems?

@zhuixunforever
Copy link

Do you reproduce it now? I have the same problem. @cornerfarmer

@cornerfarmer
Copy link
Contributor Author

No, I have given up on that. Let me know, if you get it to work.

@zhuixunforever
Copy link

Well, I didn't succeed either. Frustrated. Do you have torchE, or do you use nn.BatchNorm2d instead of SyncBatchNorm2d ?

@cornerfarmer
Copy link
Contributor Author

cornerfarmer commented Jun 2, 2021

I trained on one GPU, but I don't see why this would be a problem

@westlife35
Copy link

@cornerfarmer Do you reproduce it now? or some other new open source method like this (retrieval+relative pose, such as relocnet,camnet) ?

@cornerfarmer
Copy link
Contributor Author

Hey @westlife35,

I have not tried it anymore, so I have no news here.
Regarding other methods: You can take a look at my approach which also works in this way, but is also scene-agnostic:
https://github.com/DLR-RM/ExReNet

@westlife35
Copy link

@cornerfarmer hi, I have a question about this code:
In Train/segdata.py, line 71:
def quaternion_t(t1, t2):
return t2-t1
why directly use "t2-t1" as relative translation? why not use the code like this T1-R1*(R2)^(-1)*T2 ? which consider the rotation of images. I think this may be the reason of cannot reproduction the results of the paper

@huhulike
Copy link

@westlife35 Hi,I think the author used t2-t1 because the internal parameters of the two cameras are the same by default, so there is no scaling between the camera coordinate systems.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants