Skip to content

Preprocessing for the celebA dataset #4

Open
@jeffjiang1204

Description

@jeffjiang1204

Hi, I'm trying to reproduce the result but don't seem to able to get the same training set and testing set as you do. I followed the other issue #3 and the provided code code/celebA_to_torch.py but got something that's a bit off:

Specifically, I downloaded img_align_celeba from the website and used their provided list_eval_partition.txt to separate training and testing sets. There are 202,599 images, and the file lists 0-162,770 as train, 162,771-182,637 as val and 182,638-202,599 as test. I passed the path of the train and test folders to code/celebA_to_torch.py and get a training set that's not the same size as you used (I noticed that your notebook has the training checkpoint of size torch.Size([198650, 1, 80, 80]). Mine ends up being somewhat smaller (~ 162,769) and also probably in the wrong order (as a result I can't reproduce the curve for the train input/output PSNR). Is there a different cutoff you used for train/test set of the celebA images? Or is there any missteps in the procedures I did above? (alternatively, if there's a link to download the checkpoints used)

I also noticed that code/celebA_to_torch.py uses s=.125 for the load_CelebA_dataset function, but I have to use s=.5 to get the 80x80 images (since the original images from img_align_celeba are 178 × 218, if I used 0.125 it seems to give me 20x20 images). I just want to double check that this is expected and not because I did not download the right dataset.

Thanks for the clarifications!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions