Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CNNCifar.weight_keys #8

Open
KiteFlyKid opened this issue Mar 6, 2021 · 4 comments
Open

CNNCifar.weight_keys #8

KiteFlyKid opened this issue Mar 6, 2021 · 4 comments

Comments

@KiteFlyKid
Copy link

Nice Work!
But in Net.py CNNCifar.weight_keys, why are fc layers ahead of conv?

       self.weight_keys = [['fc1.weight', 'fc1.bias'],
                            ['fc2.weight', 'fc2.bias'],
                            ['fc3.weight', 'fc3.bias'],
                            ['conv2.weight', 'conv2.bias'],
                            ['conv1.weight', 'conv1.bias'],
                            ]

If nothing wrong with my understanding, I suppose it should be like this because you do conv first then feedfoward

        self.weight_keys = [['conv1.weight', 'conv1.bias'],
                            ['conv2.weight', 'conv2.bias'],
                            ['fc1.weight', 'fc1.bias'],
                            ['fc2.weight', 'fc2.bias'],
                            ['fc3.weight', 'fc3.bias'],
                            ]
@mengcz13
Copy link

I think this is for keeping conv layers global: the logic of codes is to select the last N layers as global layers.

@CarlBye
Copy link

CarlBye commented Oct 3, 2021

I think this is for keeping conv layers global: the logic of codes is to select the last N layers as global layers.

I think local part extract high level, compact features like feature extractor and global part acts like classifier. If the codes select conv layers as global part, does it meet the original meaning of the paper ? Or am i misunderstand the paper?

@AlphaPav
Copy link

I have the same question here. Is this problem solved?

@KiteFlyKid
Copy link
Author

You can try to reproduce their experiments but as I tried to modify the code the results are different from what they put in the paper (I did this last year so I may forget the details).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants