Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

No speed-up in my implementation too #7

Open
LSC527 opened this issue May 19, 2022 · 3 comments
Open

No speed-up in my implementation too #7

LSC527 opened this issue May 19, 2022 · 3 comments

Comments

@LSC527
Copy link

LSC527 commented May 19, 2022

I implemented this papaer with torch.autograd.forward_ad. However, fwd gradient showed no speed-up compared to fwd+bwd.

@DavideTr8
Copy link
Collaborator

It would be interesting for us to see your implementation as well. If you want, you can make a PR to our repo with your code. So we can have multiple implementations available.

@LittleWork123
Copy link

I ran the code from the repository, but I couldn't replicate the results mentioned in the paper, especially regarding the CNN. I used the hyperparameter settings specified in the paper.

May I inquire if there are alternative parameter settings available?

image

@DavideTr8
Copy link
Collaborator

DavideTr8 commented Jan 20, 2024

Hi, unfortunately we weren't able to reproduce the same results too.
The hyperparameters we used are the same reported in the paper, but we don't now if alternative hyperparameters settings are available.

We believe that the difference between our implementation and the official one are due to the fact that they did not use functorch

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants