Description
Hi, I'm trying to reproduce the result but don't seem to able to get the same training set and testing set as you do. I followed the other issue #3 and the provided code code/celebA_to_torch.py but got something that's a bit off:
Specifically, I downloaded img_align_celeba from the website and used their provided list_eval_partition.txt to separate training and testing sets. There are 202,599 images, and the file lists 0-162,770 as train, 162,771-182,637 as val and 182,638-202,599 as test. I passed the path of the train and test folders to code/celebA_to_torch.py and get a training set that's not the same size as you used (I noticed that your notebook has the training checkpoint of size torch.Size([198650, 1, 80, 80]). Mine ends up being somewhat smaller (~ 162,769) and also probably in the wrong order (as a result I can't reproduce the curve for the train input/output PSNR). Is there a different cutoff you used for train/test set of the celebA images? Or is there any missteps in the procedures I did above? (alternatively, if there's a link to download the checkpoints used)
I also noticed that code/celebA_to_torch.py uses s=.125 for the load_CelebA_dataset function, but I have to use s=.5 to get the 80x80 images (since the original images from img_align_celeba are 178 × 218, if I used 0.125 it seems to give me 20x20 images). I just want to double check that this is expected and not because I did not download the right dataset.
Thanks for the clarifications!