Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Densenets supermask #2

Open
bhack opened this issue Jul 3, 2020 · 10 comments
Open

Densenets supermask #2

bhack opened this issue Jul 3, 2020 · 10 comments

Comments

@bhack
Copy link

bhack commented Jul 3, 2020

Have you never tried to find supermask over densenets?

@mitchellnw
Copy link
Collaborator

This seems like more of a question for

@bhack
Copy link
Author

bhack commented Jul 3, 2020

I was interested in your specific context 😉 and the comments and FAQ section in https://mitchellnw.github.io/blog/2020/supsup/ was poiting to this repo 😸

@bhack
Copy link
Author

bhack commented Jul 3, 2020

P.s. I got this vague idea reading the conclusions of https://arxiv.org/abs/2006.12156.

If he is wondering about skip connections why not about dense connections?

@mitchellnw
Copy link
Collaborator

Oops! Sorry about that :)

We tried skip-connections with resnets here which worked well.

I believe dense-connections have not been explored with supermasks and it seems like a really interesting direction!

@bhack
Copy link
Author

bhack commented Jul 3, 2020

Yes I know but I meant in the mentioned work the conclusion was more related to their strong claim that subnetworks "only needs a logarithmic factor (in all variables but depth) number of neurons per weight of the target subnetwork".

So the open question was more about the impact of convolutional and batch norm layers, skip-connections, (densenet like connections?)
and LSTMs on the number of required sampled neurons to maintain a good accuracy.

@bhack
Copy link
Author

bhack commented Jul 3, 2020

I also meant that this claim could has an interesting impact in your continual learning specific setup.
If you can free-up "more resources" it is useful when you need to expand on new task.

@mitchellnw
Copy link
Collaborator

Thanks, that could definitely help!

@bhack
Copy link
Author

bhack commented Jul 3, 2020

@mitchellnw
Copy link
Collaborator

Thank you, we have seen this but haven't taken a close look! Hopefully we can soon it seems awesome

@bhack
Copy link
Author

bhack commented Jul 3, 2020

Other then densenets another interesting direction are Transformers. Some early exploring efforts were made in:

https://arxiv.org/abs/2005.00561
https://arxiv.org/abs/2005.03454

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants