Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Rigging the Lottery: Making All Tickets Winners #4

Open
raulincze opened this issue Oct 13, 2020 · 0 comments
Open

Rigging the Lottery: Making All Tickets Winners #4

raulincze opened this issue Oct 13, 2020 · 0 comments

Comments

@raulincze
Copy link

Stumbled on this here: https://www.reddit.com/r/MachineLearning/comments/j0xr24/d_machine_learning_wayr_what_are_you_reading_week/g71offh/?utm_source=reddit&utm_medium=web2x&context=3

You'll really see a patter in what I'm reading lately...

Typical way to get a sparse model is to train a dense model and drop connections (may or may not include multiple iterations.). But this approach have two main limitations.

  1. The maximum size of sparse model is limited to the largest dense model that can be trained.
  2. Large amount of computation is performed for parameters that are zero valued or that will be zero during inference.
    In this new approach, called "RigL", they randomly initialize a sparse neural network, and at regularly spaced intervals it removes a fraction of connections based on their magnitudes and activates new ones using gradient information.

Link to the paper: https://arxiv.org/pdf/1911.11134.pdf

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant