Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

what is the attributes in the large-kernel CNN #7

Open
ccx1997 opened this issue Mar 15, 2022 · 2 comments
Open

what is the attributes in the large-kernel CNN #7

ccx1997 opened this issue Mar 15, 2022 · 2 comments

Comments

@ccx1997
Copy link

ccx1997 commented Mar 15, 2022

Great analysis! I wonder the attributes of large-kernel CNN. In your paper, the basic 3x3 resnet is fully explored. 3x3 conv extracts detailed local patterns, thus may contribute to the high pass filtering. However, recent works investigate the effect of larger kernel. The attribute of 3x3 resnet might change, and similar to ViT?

@ccx1997
Copy link
Author

ccx1997 commented Mar 15, 2022

[1] A ConvNet for the 2020s.
[2] Scaling Up Your Kernels to 31x31: Revisiting Large Kernel Design in CNNs.

@xxxnell
Copy link
Owner

xxxnell commented Mar 15, 2022

Thank you for your support and insightful question!

In our observation, the attributes of Conv depend primarily on the architecture or the group size (e.g., depthwise-separable Conv) rather than the kernel size. For example, some Conv blocks at the end of stages of ConvNeXt behave like a low-pass filter in terms of Fourier analysis:

image

This figure provides ∆ log amplitude of ConvNeXt at high-frequency. ConvNext layers at the end of stages reduce high-frequency information, but ResNet layers do not. We leave a detailed investigation for future work.

∆ log amplitude of ResNet-50 at high-frequency image

Cf. MSAs generally behave like low-pass filters and reduce feature map variance. Convs, on the contrary, behave like high-pass filters and increase the feature map variance. In terms of feature map variance, all ConvNeXt blocks diversify feature maps (i.e., they always increase the variance).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants