-
-
Notifications
You must be signed in to change notification settings - Fork 35
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to run #12
Comments
Hello @mognc , |
I didn't changed any parameter and yes I used config/celebhq_text_cond.yaml |
Okay, then you can you check what is the name of checkpoint file that was created in the |
But there should be a celebhq folder after you run the autoencoder. I am assuming you ran |
Sorry for troubling you, I missed that block. Its downloading different models atm and I hope my error will be resolved now. Thanks for helping. I am new to this image generation thing so I make silly mistakes. |
No problem at all @mognc :) Will keep the issue open for now, and you can close it once you have successfully ran the text conditional training. |
After running train_vqvae , did you run the train_ddpm_cond ? Did that fail ? |
After running this command "!python -m tools.train_vqvae --config config/celebhq_text_cond.yaml" it displayed the training is completed. Then i ran "!python -m tools.sample_ddpm_text_cond --config config/celebhq_text_cond.yaml" which failed. I also have attached pics above for your reference. |
Yes but the train vqvae is only for Stage I. This will only train the autoencoder and not the diffusion model. |
Hey there, my model got in running state yesterday but the output was not that I desired. And I am bit confused what changes I have to make to improve them. I have 100 images dataset with 100 captions file. The dimensions of pics are 600*800. I have made dataset and yaml file similar to "celebhq.yaml" and celeb_dataset.py file. I also have modified the targets to my own files. I want to generate similar pics as my dataset but with variations in them. I figured out to achieve this task I will be using Training text conditional LDM method and will be creating similar file as "celebhq_text_cond.yaml" file right?. But now I am unsure what additional changes I have to make in my new three files I have created that will help me achieve my goal like can you pin out parameters I have to change like no of epochs I have to train and etc. Also I have enables save latents in those file as you mentioned in readme file to speed up the training process. |
How many epochs/steps did you train the autoencoder for ? and could you add some output examples of autoencoder. The first thing I would suggest is have more images, maybe 2K images to start with. |
Well I didn't change epochs or samples. At the moment I don't have access to those outputs as Collab erase all the data after session terminates. But the ldm stage was not producing correct output it was just blur pic. |
If you didnt change any parameters then that means the autoencoder ran for only 20 epochs and the discriminator didnt even start because the config has the start of discriminator at 15000 steps. For the conditioning, if all you have are texts of type ' on walls' where obj can be one of K things, then you can use class conditioning with K classes rather than text conditioning. |
Ok I will change that parameter. And I will try conditioning too, but I just want to make sure like my dataset is simple it don't include type of variations I want like snow or dust. This will not be a problem right ? |
Also the ldm epochs are set at 100 epochs but this was for celebhq dataset with 30000 images. |
"but I just want to make sure like my dataset is simple it don't include type of variations I want like snow or dust" |
Yes, I don't have pics of variations I want. |
But if the model has never seen what 'snow' looks like anytime during the training, it will not be able to generate 'snow on walls' right ? |
Well my friend used some pre trained models and those were producing results, so I am not sure of this model how it works. So should I add simple pics of snow, dust and other variations and merge them with walls dataset? |
Yes pre-trained model would work because that has seen what 'snow' looks like. But this model will be trained from scratch. |
Well I will stick with this repo and will add variations pic along with captions and will merge all dataset together. Thanks for removing all confusions really appreciate your time. |
This part of the readme is just saying that the dataset class must return a tuple of image tensor and a dictionary of conditional inputs. |
This is my config file which I edited: diffusion_params: ldm_params: autoencoder_params: train_params: |
I think it would benefit by training the autoencoder more. Specifically two changes:
Basically train for longer and start discriminator after your autoencoder generates the best reconstructions it can. The disc_start is the number of steps after which discriminator should start. |
Yes the ddpm_ckpt_text_cond_clip.pth is overwritten every time you run the training. |
Hello there, I am trying to run conditional text part and have followed all the instructions but at the end I am facing following error. Screenshot attached below. It says "Model checkpoint celebhq/ddpm_ckpt_text_cond_clip.pth not found"
The text was updated successfully, but these errors were encountered: