Disentangled Representation Learning for Realistic and Diverse Child Face Prediction from Parent Images
DOI:
https://doi.org/10.62051/ijcsit.v3n3.09Keywords:
Child face prediction, Generative adversarial networks (GANs), Disentangled representation, Factor-based map-ping, Family-focused dataset, Genetic factors, Facial attributesAbstract
Predicting a child’s facial appearance from their parents’ photos is a challenging task with potential applications in various fields, including kinship verification, age progression, and forensic investigations. Existing methods often struggle to balance the need for accurate genetic representation with the generation of diverse and realistic child faces. We propose a novel approach that leverages a Generative Adversarial Net- work (GAN) framework with factor-based disentanglement and mapping, trained exclusively on a family-focused dataset. Our model explicitly separates and represents distinct facial factors: genetic (inherited traits), external (changeable attributes), and variety (individual differences). By focusing on genetic factors and employing a dedicated mapping module to learn parent-to- child genetic relationships, we aim to achieve higher accuracy and realism compared to traditional style-based or direct mapping methods. Comprehensive experiments on a large-scale Family Face Database demonstrate that our model outperforms existing state-of-the-art approaches in generating realistic and diverse child face images. The predicted faces not only capture the nuanced resemblance between parents and children but also exhibit a wide range of individual variations, aligning with real- world observations. Additionally, our method addresses ethical concerns by focusing on heritable traits and utilizing family- specific data, promoting privacy and minimizing potential biases. This work opens up new possibilities for child face prediction, offering a more accurate and ethically sound approach for future research and applications.
Downloads
References
[1] P S Chandran, N Byju, R Deepak, K Nishakumari, P Devanand, and P Sasi. Missing child identification system using deep learning and multiclass svm. In Proc. IEEE RAICS, pages 113–116, 2018.
[2] Yunjey Choi, Minje Choi, Munyoung Kim, Jung-Woo Ha, Sunghun Kim, and Jaegul Choo. Stargan: Unified generative adversarial networks for multi-domain image- to-image translation. In Proceedings of the IEEE confer- ence on computer vision and pattern recognition, pages 8789–8797, 2018.
[3] Jiankang Deng, Jia Guo, Niannan Xue, and Stefanos Zafeiriou. Arcface: Additive angular margin loss for deep face recognition. In Proc. CVPR, pages 4690–4699, 2019.
[4] Hamdi Dibeklioglu, Albert Ali Salah, and Theo Gevers. Like father, like son: Facial expression dynamics for kinship verification. In Proc. ICCV, pages 1497–1504, 2013.
[5] Hajar Emami, Majid Moradi Aliabadi, Ming Dong, and Ratna Babu Chinnam. Spa-gan: Spatial attention gan for image-to-image translation. IEEE Transactions on Multimedia, 23:391–401, 2020.
[6] Pengcheng Gao, Julian Robinson, Jian Zhu, Chunyong Xia, Ming Shao, and Shu Xia. Dna-net: Age and gender aware kin face synthesizer. In Proc. ICME, pages 1–6, 2021.
[7] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Proc. NeurIPS, pages 2672–2680, 2014.
[8] Ishaan Gulrajani, Faruk Ahmed, Martin Arjovsky, Vin- cent Dumoulin, and Aaron C Courville. Improved training of wasserstein gans. In Proc. NeurIPS, pages 5767–5777, 2017.
[9] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Proc. NeurIPS, pages 6626–6637, 2017.
[10] Xun Huang and Serge Belongie. Arbitrary style transfer in real-time with adaptive instance normalization. In Proceedings of the IEEE international conference on computer vision, pages 1501–1510, 2017.
[11] Zhi Huang, Jian Zhang, and Hongming Shan. When age- invariant face recognition meets face age synthesis: A multi-task learning framework. In Proc. CVPR, pages 7282–7291, 2021.
[12] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A Efros. Image-to-image translation with conditional adver- sarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1125– 1134, 2017.
[13] Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehti- nen. Progressive growing of gans for improved quality, stability, and variation. arXiv preprint arXiv:1710.10196, 2017.
[14] Tero Karras, Samuli Laine, and Timo Aila. A style- based generator architecture for generative adversarial networks. In Proc. CVPR, pages 4401–4410, 2019.
[15] Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and Timo Aila. Analyzing and improv- ing the image quality of stylegan. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 8110–8119, 2020.
Downloads
Published
Issue
Section
License

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.







