Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is it the refiner rather than the self-supervised training on real data? #27

Open
phquentin opened this issue Mar 14, 2024 · 1 comment

Comments

@phquentin
Copy link

phquentin commented Mar 14, 2024

Hi there,

First of all, congratulations on your great work and making it available on github!

I have a question regarding the ablation study of your paper on Occluded Linemod and the results of Table 3.
If I understand the results correctly (please correct me if I'm wrong), the row OURS(LB) + Dref is the performance of the baseline algorithm only with the additional refiner in the teacher-student training paradigm. These results show that this addition already achieves an average recall of 62.1% and the other branches "only" add 2.6% to reach the top performance of 64.7%.

So could it be that the main performance gain simply comes from the refiner? That is, the refiner's capabilities are transferred to the GDR net in this way, and therefore the additional self-supervised learning from the other branches from the unannotated real data is actually minimal?

In other words, if we were to compare the performance of your best version with just the GDR-net and a downstream refiner (both trained on synthetic data only), would we get similar results?

If I understand correctly, this is what happened in the results in Table 7. The results suggest that the difference between GDR-Net with a downstream refiner and your self-supervised method is not really significant.

Would be nice to hear your opinion on that, as such an interpretation could influence further research :). Thx in advance!

Best regards,

Philipp

@wangg12
Copy link
Member

wangg12 commented Mar 15, 2024

Hi, thanks for the comments. In my view, "the additional self-supervised learning from the other branches from the unannotated real data is actually minimal" may not be correct. Without this self-supervised learning paradigm, we cannot gain any knowledge from the unlabeled real images. Besides, after self-training, we don't need the tedious and maybe much slower iterative refining process. The refiner and the self-supervised learning both play important roles in the performance boost of the base GDR-Net.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants