You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am evaluating your released model on Birds dataset. Particularly, I want to get the number claimed in your paper to make sure that everything I have done is correct. However, in your paper, it is claimed IS=52.53±0.45 and FID=11.25. But I got IS=43.20±0.54 and FID=22.08. I think I might have made some mistakes in details.
What I did
I generate 30K 128x128 child images with your released model on Birds, wherein 150 images for each child category.
(1) I compute the IS with your released finetuned inception model. The generated images are resized to 299x299 normalized to be within [-1, 1] before fed into the network. Mean and std are computed over 10 splits.
(2) I compute the FID with default inception model using
Do you use the finetuned inception model for computing FID?
Should I use the images that are cropped with 1.5x bounding boxes from original images to compute FID?
Should I first resize the real images to 128x128 and then feed them into the inception network (which automatically resizes input to 299x299) to compute FID?
Is the number I got lies in the normal variation? I suppose the quality of generated images may be different for different times of generation.
B.T.W. I am also curious about the results of LR-GAN you got in Table1. Do you train LR-GAN on the original CUB images or on the cropped images?
B.T.W. I am using pytorch==1.3.0. Not sure if there is any version issue.
The text was updated successfully, but these errors were encountered:
UPDATE:
I first resize the real images to 128x128 and then feed them into the inception model to compute FID. This gives me FID=12.39. I think this is quite close to the number claimed in the paper. However, I still can't get a higher IS score.
Hi, apologies for the delayed response. The FID will depend on how exactly you're generating the 30k images. Are you iterating through all the child codes, and generating the images with parent/background code fixed?
And yes, you should use 1.5 times cropped real images for computing the FID.
The inception model used for FID is the model pre-trained on ImageNet. We use the fine-tuned version only for computing the Inception score, where the model should be finetuned on all the 200 categories (for birds).
As for the variation, I believe FID could still be in that range, but IS does appear to be lower than usual.
Double check the image generation process, and see if that improves the score.
Hi, nice work!
I am evaluating your released model on Birds dataset. Particularly, I want to get the number claimed in your paper to make sure that everything I have done is correct. However, in your paper, it is claimed IS=52.53±0.45 and FID=11.25. But I got IS=43.20±0.54 and FID=22.08. I think I might have made some mistakes in details.
What I did
I generate 30K 128x128 child images with your released model on Birds, wherein 150 images for each child category.
(1) I compute the IS with your released finetuned inception model. The generated images are resized to 299x299 normalized to be within [-1, 1] before fed into the network. Mean and std are computed over 10 splits.
(2) I compute the FID with default inception model using
Note that /path/to/real/images are original CUB images without any cropping or resizing.
Evaluation codes
IS: https://github.com/sbarratt/inception-score-pytorch
FID: https://github.com/mseitzer/pytorch-fid
Questions
B.T.W. I am also curious about the results of LR-GAN you got in Table1. Do you train LR-GAN on the original CUB images or on the cropped images?
B.T.W. I am using pytorch==1.3.0. Not sure if there is any version issue.
The text was updated successfully, but these errors were encountered: