-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Regarding the issue with the new version #14
Comments
yes, the improvement is brought by using the Snake activation function instead of ReLU. |
I guess it's often the case that we get hindsight about little simple things to tweak for further improvement. |
And, I noticed that you have implemented some functionalities in both the encoder and decoder, and you've removed the loss updates related to zero padding. I don't quite understand why this was done. |
hmm.. I've noticed that the training on the FordA dataset generally resulted in unsatisfactory validation loss (which is shown in your figure -- reconstruction on a test sample). Have you tried it on other datasets? Additionally, the same model with the ReLU activation results in poorer optimization (i.e., slower training and poorer convergence). I think you can try it out yourself. I've personally experimented quite extensively to confirm the positive effect of the current VQVAE. I found it more stable with faster and better convergence. |
I've simplified the code by leaving the essentials only so that end users can read and use it more easily. |
In the first stage of the new version, the reconstructed signal appears smoother compared to the old version. The GT and reconstructed signals are almost overlapping, but is this form correct?
The text was updated successfully, but these errors were encountered: