This is the official webpage of the paper "ChildPredictor: A Child Face Prediction Framework with Disentangled Learning", accepted to IEEE TMM, 2022
Arxiv: https://arxiv.org/abs/2204.09962
IEEE Xplore: https://ieeexplore.ieee.org/document/9749880
🚀 🚀 🚀 News:
-
May. 19, 2022: We release the data of the FF-Database, please see section 1.1 for the terms of use.
-
Apr. 15, 2022: We release the pre-trained models with validation samples for ChildPredictor.
-
Mar. 31, 2022: The paper is accepted by the IEEE Transactions on Multimedia.
-
Feb. 8, 2022: We release the code for ChildPredictor. We are considerring to release the original data of the collected FF-Database.
If you would like to download the FF-Database data, please fill out an agreement to the FF-Database Terms of Use and send it to us at [email protected].
Please use the institutional email instead of anonymous addresses such as Gmail, QQMail, and hotmail. Make sure your email address is the same as it in the FF-Database Terms of Use.
The data collection pipeline is shown as follows:
Some families are shown as follows:
The generated results on the collected FF-Database:
The generated results on other datasets:
The disentangled learning analysis is as:
The ablation study is as:
Some files are not included in the current implementation since they are too large. The network architectures can be found in the code
folder.
code
│
└───baby_model_pool
│ └───attgan
│ │ │ attgan_without_claloss_baby.pth
│ │ │ attgan_without_ganloss_celeba_baby.pth
│ │ │ attgan_without_ganloss_claloss_celeba_baby.pth
│ │ │ ...
│ └───inverse
│ │ │ Inverse_ProGAN_GAN_ACGAN_start-with-code.pth
│ │ │ Inverse_ProGAN_GAN_MSGAN_ACGAN_start-with-code.pth
│ │ │ Inverse_ProGAN_GAN_MSGAN_ACGAN_start-with-image.pth
│ │ │ ...
│ └───mapping
│ │ └───Mapping_Xencoder_full_ProGAN_GAN_MSGAN_ACGAN_deepArch_multi-gt_v4
│ │ │ │ MappingNet_Batchsize_32_Epoch_298.pth
│ │ └───Mapping_Xencoder_full_ProGAN_GAN_deepArch_multi-gt_v4
│ │ │ │ MappingNet_Batchsize_32_Epoch_298.pth
│ │ └───Mapping_Xencoder_wo-class_ProGAN_GAN_MSGAN_deepArch_multi-gt_v4
│ │ │ │ MappingNet_Batchsize_32_Epoch_298.pth
│ │ │ ...
│ └───ProGAN-ckp
│ │ │ ProGAN_pt_mixtureData_GAN.pth
│ │ │ ProGAN_pt_mixtureData_GAN_ACGAN.pth
│ │ │ ProGAN_pt_mixtureData_GAN_MSGAN.pth
│ │ │ ProGAN_pt_mixtureData_GAN_MSGAN_ACGAN.pth
│ │ │ ...
│
└───babyinverse (Ey)
│ │ ...
|
└───babymapping_1219 (T)
│ │ ...
│
└───Datasets
│ │ ...
│
└───ProGAN (Gy)
│ │ ...
│
└───AttGAN (please refer to AttGAN official webpage)
│ │ ...
│
The implementation is based on CUDA 9.0 and PyTorch 1.1.0. The following packages are needed to be installed:
pytorch==1.1.0
torchvision==0.3.0
tensorflow-gpumkdir ./babymapping_1219/Models/pretrain
mv ./baby_model_pool/ProGAN-ckp/* ./babymapping_1219/Models/pretrain/
tensorboardx
pyyaml
tqdm
easydict
First, download the pre-trained models from this link. It should be a large zip file with size of approximately 3.9 Gb.
After you have already downloaded the pre-trained models, enter code
folder and unzip all the models under the ./code/baby_model_pool folder:
cd code
mkdir baby_model_pool
cd baby_model_pool
unzip Onedrive_baby_model_pool.zip
cd ..
Then, you need to move all the ProGAN pre-trained models under another path:
mkdir ./babymapping_1219/Models/pretrain
mv ./baby_model_pool/ProGAN-ckp/* ./babymapping_1219/Models/pretrain/
Next, you can test some validation samples (we have already put some examples under the code/babymapping_1219 folder):
cd babymapping_1219
python main.py
If you want to change the input images, see lines 38-39 of validation.yaml: https://github.com/zhaoyuzhi/ChildPredictor/blob/main/code/babymapping_1219/yaml/yaml/validation.yaml#L38-L39
Currently, we do not release the full codes for training due to privacy issue.
Please refer to code_FFDatabase_collection.
-
Zaman, Ishtiak and Crandall, David. Genetic-GAN: Synthesizing Images Between Two Domains by Genetic Crossover. European Conference on Computer Vision Workshops, 312--326, 2020.
-
Gao, Pengyu and Robinson, Joseph and Zhu, Jiaxuan and Xia, Chao and Shao, MIng and Xia, Siyu. DNA-Net: Age and Gender Aware Kin Face Synthesizer. IEEE International Conference on Multimedia and Expo (ICME), 2021.
-
Robinson, Joseph Peter and Khan, Zaid and Yin, Yu and Shao, Ming and Fu, Yun. Families in wild multimedia (FIW MM): A multimodal database for recognizing kinship. IEEE Transactions on Multimedia, 2021.
If you find this work useful for your research, please cite:
@article{zhao2022childpredictor,
title={ChildPredictor: A Child Face Prediction Framework with Disentangled Learning},
author={Zhao, Yuzhi and Po, Lai-Man and Wang, Xuehui and Yan, Qiong and Shen, Wei and Zhang, Yujia and Liu, Wei and Wong Chun-Kit and Pang, Chiu-Sing and Ou, Weifeng and Yu, Wing-Yin and Liu, Buhua},
journal={IEEE Transactions on Multimedia},
year={2022}
}
Please contact [email protected] for further questions.