-
Notifications
You must be signed in to change notification settings - Fork 96
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pretraining #5
Comments
Hi I am also interested in your pre-training code. I did pre-training based on your description in your paper. However, with pre-training, gamma output will always assign the same class to all data points. |
Also, why you assign weights from one previous layer in pretrained AE to the layers in VaDE as below: |
Also having trouble replicating the results. Using the pretrained weights provided works fine, except for HAR dataset. But using pretraining code from DEC-keras, which achieves good results for AE+kmeans and DEC, does not make the VaDE model work. Also, in the code for HAR dataset, it specifies the random state for GMM, which shouldn't be done. Removing the random state specification and repeat many times, the performance is significantly lower than the result reported. Is the author using the pretrain code from original DEC code? If not, could you provide it? |
@michelleowen |
They use VAE or AAE to pretrain the model. You need to constraint the latent space with KL divergence in the loss (Or use discriminator in AAE). |
@wangmn93 Could you elaborate more on the VAE pretraining? How to control the range of latent space? By setting coefficient on the KL divergence term? Also, it seems that their provided pretrain weights only have autoencoder weight, but not enc_sigma weight. It would even better if you could share your code for the pretraining. Thanks. |
Actually, i found that sometimes you can get high accuracy(around 94%)when
you just use autoencoder for pretrain training instead of vae, which means
you do not need to resitict the range of latent space. But the whole algo
is sensitive to initialization (both ae and kmean). In short, you can not
gurantee to get 94% on average. If you want to reproduce 94% acc,use their
pretrained weight. Or pretrain with ae and then use kmean in the latent
space to test the acc if it is more than 80% or higher, you may get 94% on
VaDE.
eelxpeng <[email protected]> 于 2018年5月21日周一 08:01写道:
… @wangmn93 <https://github.com/wangmn93> Could you elaborate more on the
VAE pretraining? How to control the range of latent space? By setting
coefficient on the KL divergence term? Also, it seems that their provided
pretrain weights only have autoencoder weight, but not enc_sigma weight.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#5 (comment)>, or mute
the thread
<https://github.com/notifications/unsubscribe-auth/AS2TVB_5GRHYOJsLnW3YZweMv5vTEh-oks5t0koPgaJpZM4PoKtX>
.
|
@wangmn93 Thank you for your reply. I actually tried many possible initializations, including ae, sdae, vae, with all kinds of random initialization. However, I haven't got one work. Could you share one code that at least sometimes works? I am trying to find out the reason of the instability, and good initialization method to make things work robustly. Your help would be much appreciated. |
i use the pretrain of dec https://github.com/XifengGuo/DEC-keras
eelxpeng <[email protected]> 于 2018年5月22日周二 09:06写道:
… @wangmn93 <https://github.com/wangmn93> Thank you for your reply. I
actually tried many possible initializations, including ae, sdae, vae, with
all kinds of random initialization. However, I haven't got one work. Could
you share one that at least sometimes works? I am trying to find out the
reason of the instability, and good initialization method to make things
work robustly. Your help would be much appreciated.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#5 (comment)>, or mute
the thread
<https://github.com/notifications/unsubscribe-auth/AS2TVM2HCF-HmqZImPW0vveb7_wiHkPvks5t06rXgaJpZM4PoKtX>
.
|
@eelxpeng Did you make any progress on this problem? Did you get the DEC-Keras pre-training method to work? I could get the AE pre-training on DEC-Keras to reach ~86% ... However, once I plug that in to VaDE, accuracy drops dramatically to ~57%. Not really sure what is going on wrong there. |
@wangmn93 Did you train some other datasets ?I could get 85%+ on MNIST using VaDE. But when i train the new dataset, the acc is only about 20% . |
no, i only test on MNIST
Zizi6947 <[email protected]> 于 2018年12月17日周一 下午12:35写道:
… @wangmn93 <https://github.com/wangmn93> Did you train some other datasets
?I could get 85%+ on MNIST using VaDE. But when i train the new dataset,
the acc is only about 20% .
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#5 (comment)>, or mute
the thread
<https://github.com/notifications/unsubscribe-auth/AS2TVBi86jXQaY6iPcMeuP7IF6n8mAzcks5u5x8qgaJpZM4PoKtX>
.
|
@Zizi6947 @devyhia @michelleowen |
Hi,
My team and I are trying to duplicate the results of your paper, but cannot. Would it be possible to gain access to the code that pretrains the data? That would help us a lot. Thank you.
The text was updated successfully, but these errors were encountered: