×

Imperfect imaGANation: implications of GANs exacerbating biases on facial data augmentation and snapchat face lenses. (English) Zbl 07482904

Summary: In this paper, we show that popular Generative Adversarial Network (GAN) variants exacerbate biases along the axes of gender and skin tone in the generated data. The use of synthetic data generated by GANs is widely used for a variety of tasks ranging from data augmentation to stylizing images. While practitioners celebrate this method as an economical way to obtain synthetic data to train data-hungry machine learning models or provide new features to users of mobile applications, it is unclear whether they recognize the perils of such techniques when applied to real world datasets biased along latent dimensions. Although one expects GANs to replicate the distribution of the original data, in real-world settings with limited data and finite network capacity, GANs suffer from mode collapse. First, we show readily-accessible GAN variants such as DCGANs ‘imagine’ faces of synthetic engineering professors that have masculine facial features and fair skin tones. When using popular GAN architectures that attempt to address mode-collapse, we observe that these variants either provide a false sense of security or suffer from other inherent limitations due to their design choice. Second, we show that a conditional GAN variant transforms input images of female and nonwhite faces to have more masculine features and lighter skin when asked to generate faces of engineering professors. Worse yet, prevalent filters on Snapchat end up consistently lightening the skin tones in people of color when trying to make face images appear more feminine. Thus, our study is meant to serve as a cautionary tale for practitioners and educate them about the side-effect of bias amplification when applying GAN-based techniques.

MSC:

68Txx Artificial intelligence
PDF BibTeX XML Cite
Full Text: DOI arXiv

References:

[1] Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y., Generative adversarial nets, (Advances in Neural Information Processing Systems (2014)), 2672-2680
[2] O’Gorman, L.; Kasturi, R., Document Image Analysis, vol. 39 (1995), IEEE Computer Society Press: IEEE Computer Society Press Los Alamitos
[3] Bloice, M. D.; Roth, P. M.; Holzinger, A., Biomedical image augmentation using augmentor, Bioinformatics, 35, 21, 4522-4524 (2019)
[4] Teich, D. A., Synthetic data is a tool for improving training and accuracy of deep learning systems (2019)
[5] Nisselson, E., Deep learning with synthetic data will democratize the tech industry (2018)
[6] Arora, S.; Risteski, A.; Zhang, Y., Do GANs learn the distribution? Some theory and empirics, (6th International Conference on Learning Representations. 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018. 6th International Conference on Learning Representations. 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, OpenReview.net (2018))
[7] Tolstikhin, I. O.; Gelly, S.; Bousquet, O.; Simon-Gabriel, C.-J.; Schölkopf, B., AdaGAN: boosting generative models, (Advances in Neural Information Processing Systems (2017)), 5424-5433
[8] Karras, T.; Aila, T.; Laine, S.; Lehtinen, J., Progressive growing of GANs for improved quality, stability, and variation, (6th International Conference on Learning Representations. 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018. 6th International Conference on Learning Representations. 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, OpenReview.net (2018))
[9] Yu, N.; Li, K.; Zhou, P.; Malik, J.; Davis, L.; Fritz, M., Inclusive gan: improving data and minority coverage in generative models, (European Conference on Computer Vision (2020), Springer), 377-393
[10] Mulaudzi, S., Let’s be honest: snapchat filters are a little racist (2017)
[11] Jagota, V., Why do all the snapchat filters try to make you look white? (2016)
[12] BetStyle, Hold up: does snapchat have a problem with brown skin? (2016)
[13] Crawford, K.; Paglen, T., Excavating AI (2019)
[14] Arora, S.; Ge, R.; Liang, Y.; Ma, T.; Zhang, Y., Generalization and equilibrium in Generative Adversarial Nets (GANs), (Proceedings of the 34th International Conference on Machine Learning, vol. 70 (2017)), 224-232
[15] Che, T.; Li, Y.; Jacob, A. P.; Bengio, Y.; Li, W., Mode regularized generative adversarial networks, (5th International Conference on Learning Representations. 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017. 5th International Conference on Learning Representations. 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings, OpenReview.net (2017))
[16] Zhao, S.; Ren, H.; Yuan, A.; Song, J.; Goodman, N.; Ermon, S., Bias and generalization in deep generative models: an empirical study, (Advances in Neural Information Processing Systems (2018)), 10792-10801
[17] Grnarova, P.; Levy, K. Y.; Lucchi, A.; Perraudin, N.; Hofmann, T.; Krause, A., Evaluating GANs via duality, arXiv preprint
[18] Goodfellow, I., NIPS 2016 tutorial: generative adversarial networks, arXiv preprint
[19] Hoang, Q.; Nguyen, T. D.; Le, T.; Phung, D. Q., MGAN: training generative adversarial nets with multiple generators, (6th International Conference on Learning Representations. 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3. 6th International Conference on Learning Representations. 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, Conference Track Proceedings, OpenReview.net (2018))
[20] Zhang, H.; Cissé, M.; Dauphin, Y. N.; Lopez-Paz, D., Mixup: beyond empirical risk minimization, (6th International Conference on Learning Representations. 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018. 6th International Conference on Learning Representations. 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, OpenReview.net (2018))
[21] He, H.; Wang, H.; Lee, G.; Tian, Y., ProbGAN: towards probabilistic GAN with theoretical guarantees, (7th International Conference on Learning Representations. 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019, OpenReview.net (2019))
[22] Metz, L.; Poole, B.; Pfau, D.; Sohl-Dickstein, J., Unrolled generative adversarial networks, (5th International Conference on Learning Representations. 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017. 5th International Conference on Learning Representations. 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings (2017), ICLR)
[23] Srivastava, A.; Valkov, L.; Russell, C.; Gutmann, M. U.; Sutton, C., VEEGAN: reducing mode collapse in GANs using implicit variational learning, (Advances in Neural Information Processing Systems (2017)), 3308-3318
[24] Arjovsky, M.; Chintala, S.; Bottou, L., Wasserstein generative adversarial networks, (International Conference on Machine Learning, PMLR (2017)), 214-223
[25] Miyato, T.; Kataoka, T.; Koyama, M.; Yoshida, Y., Spectral normalization for generative adversarial networks, (6th International Conference on Learning Representations. 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018. 6th International Conference on Learning Representations. 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, OpenReview.net (2018))
[26] Mishra, D.; Jayendran, A.; Srivastava, V.; Chaudhury, S., Mode matching in GANs through latent space learning and inversion, arXiv preprint
[27] Odena, A.; Olah, C.; Shlens, J., Conditional image synthesis with auxiliary classifier GANs, (International Conference on Machine Learning (2017)), 2642-2651
[28] Mariani, G.; Scheidegger, F.; Istrate, R.; Bekas, C.; Malossi, A. C.I., BAGAN: data augmentation with balancing GAN, CoRR
[29] Ali-Gombe, A.; Elyan MFC-GAN, E., Class-imbalanced dataset classification using multiple fake class generative adversarial network, Neurocomputing, 361, 212-221 (2019)
[30] Radford, A.; Metz, L.; Chintala, S., Unsupervised representation learning with deep convolutional generative adversarial networks, (Bengio, Y.; LeCun, Y., 4th International Conference on Learning Representations. 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016. 4th International Conference on Learning Representations. 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings (2016))
[31] Liu, Z.; Luo, P.; Wang, X.; Tang, X., Deep learning face attributes in the wild, (Proceedings of International Conference on Computer Vision. Proceedings of International Conference on Computer Vision, ICCV (2015))
[32] A. Krizhevsky, Learning Multiple Layers of Features from Tiny Images, University of Toronto.
[33] Yu, F.; Zhang, Y.; Song, S.; Seff, A.; Xiao, J., LSUN: construction of a large-scale image dataset using deep learning with humans in the loop, CoRR
[34] Goodfellow, I. J., NIPS 2016 tutorial: generative adversarial networks, CoRR
[35] Frid-Adar, M.; Klang, E.; Amitai, M.; Goldberger, J.; Greenspan, H., Synthetic data augmentation using GAN for improved liver lesion classification, (Biomedical Imaging (ISBI 2018), 2018 IEEE 15th International Symposium on (2018), IEEE), 289-293
[36] Guibas, J. T.; Virdi, T. S.; Li, P. S., Synthetic medical images from dual generative adversarial networks, CoRR
[37] (2015), carpedm20, DCGAN implementation in Tensorflow
[38] Rätsch, G.; Onoda, T.; Müller, K.-R., Soft margins for AdaBoost, Mach. Learn., 42, 3, 287-320 (2001) · Zbl 0969.68128
[39] Lala, S.; Shady, M.; Belyaeva, A.; Liu, M., Evaluation of mode collapse in generative adversarial networks, high performance extreme computing
[40] Karras, T.; Laine, S.; Aila, T., A style-based generator architecture for generative adversarial networks, (Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019)), 4401-4410
[41] Karras, T.; Laine, S.; Aittala, M.; Hellsten, J.; Lehtinen, J.; Aila, T., Analyzing and improving the image quality of StyleGAN, (Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2020)), 8110-8119
[42] Li, K.; Malik, J., Implicit maximum likelihood estimation, arXiv preprint
[43] Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A. A., Image-to-image translation with conditional adversarial networks, (Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2017)), 1125-1134
[44] Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A. A., Unpaired image-to-image translation using cycle-consistent adversarial networks, (Proceedings of the IEEE International Conference on Computer Vision (2017)), 2223-2232
[45] junyanz, CycleGAN and pix2pix in PyTorch (2018)
[46] junyanz, CycleGAN (2017)
[47] Hiasa, Y.; Otake, Y.; Takao, M.; Matsuoka, T.; Takashima, K.; Carass, A.; Prince, J.; Sugano, N.; Sato, Y., Cross-Modality Image Synthesis from Unpaired Data Using CycleGAN: Effects of Gradient Consistency Loss and Training Data Size, 31-41 (2018), Springer Verlag
[48] Sandfort, V.; Yan, K.; Pickhardt, P. J.; Summers, R. M., Data augmentation using generative adversarial networks (CycleGAN) to improve generalizability in CT segmentation tasks, Sci. Rep., 9, 1, 1-9 (2019)
[49] Huang, S.-W.; Lin, C.-T.; Chen, S.-P.; Wu, Y.-Y.; Hsu, P.-H.; Lai, S.-H., AugGAN: cross domain adaptation with GAN-based data augmentation, (Proceedings of the European Conference on Computer Vision. Proceedings of the European Conference on Computer Vision, ECCV (2018)), 718-731
[50] U.S.N., Best Engineering Schools (2019)
[51] Dalal, N.; Triggs, B., Histograms of oriented gradients for human detection, (2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), vol. 1 (2005)), 886-893, vol. 1
[52] Ma, S.; Fu, J.; Wen Chen, C.; Mei, T., DA-GAN: instance-level image translation by deep attention generative adversarial networks, (Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018)), 5657-5666
[53] Center, P. R., Social media fact sheet (June 2019)
[54] Li, E. Yanjia, Gender Swap and CycleGAN in TensorFlow 2.0 - Ethan Yanjia Li (2020)
[55] Magazine, P., The dark implications of facial swap filter technology (Sep 2019)
[56] Jang, E., Fun with snapchat’s gender swapping filter (Jan 1970)
[57] (2019), r/MachineLearning - [D] Is the new Snapchat gender filter GAN-based?
[58] Baeza-Yates, R., Data and algorithmic bias in the web, (Proceedings of the 8th ACM Conference on Web Science (2016)), 1
[59] Omnicore, Snapchat statistics (Jan 2020)
[60] Hunter, M., The persistent problem of colorism: skin tone, status, and inequality, Sociol. Compass, 1, 1, 237-254 (2007)
[61] Nittle, N. K., The harmful effects of colorism (Mar 2019)
[62] Mehrabi, N.; Morstatter, F.; Saxena, N.; Lerman, K.; Galstyan, A., A survey on bias and fairness in machine learning, ACM Comput. Surv., 54, 6, 1-35 (2021)
[63] Howard, A.; Borenstein, J., The ugly truth about ourselves and our robot creations: the problem of bias and social inequity, Sci. Eng. Ethics, 24, 1521-1536 (2018)
[64] N. Kodali, J. Hays, J.D. Abernethy, Z. Kira, On Convergence and Stability of GANs, Artificial Intelligence, arXiv.
[65] Mishra, Deepak; Prathosh, A. P.; Jayendran, Aravind; Srivastava, Varun; Chaudhury, Santanu, Mode matching in GANs through latent space learning and inversion (2019)
[66] Hymas, C., AI used for first time in job interviews in UK to find best applicants (Sep 2019)
[67] Bahrampour, T., Can your face reveal how long you’ll live? New technology may provide the answer (Jul 2014)
[68] Kaur, H.; Marco, T., A New York school district is bringing in facial recognition software. Rights groups say it could spell trouble for students (May 2019)
[69] Harwell, D., Oregon became a testing ground for Amazon’s facial-recognition policing. But what if Rekognition gets it wrong? (Apr 2019)
[70] Gebru, T., Oxford Handbook on AI Ethics book chapter on race and gender, arXiv preprint
[71] Angwin, J.; Larson, J., Machine Bias (2016)
[72] Menon, S.; Damian, A.; Hu, S.; Ravi, N.; Rudin, C., PULSE: self-supervised photo upsampling via latent space exploration of generative models, (Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2020)), 2437-2445
[73] Afchar, D.; Nozick, V.; Yamagishi, J.; Echizen, I., MesoNet: a compact facial video forgery detection network, (2018 IEEE International Workshop on Information Forensics and Security. 2018 IEEE International Workshop on Information Forensics and Security, WIFS (2018)), 1-7
[74] Güera, D.; Delp, E. J., Deepfake video detection using recurrent neural networks, (2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance. 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance, AVSS (2018), IEEE), 1-6
[75] Li, Y.; Yang, X.; Sun, P.; Qi, H.; Lyu, S., Celeb-DF: a large-scale challenging dataset for DeepFake forensics, (2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR (2020)), 3204-3213
[76] Rössler, A.; Cozzolino, D.; Verdoliva, L.; Riess, C.; Thies, J.; Niessner, M., FaceForensics++: learning to detect manipulated facial images, (2019 IEEE/CVF International Conference on Computer Vision. 2019 IEEE/CVF International Conference on Computer Vision, ICCV (2019)), 1-11
[77] Korshunov, P.; Marcel, S., DeepFakes: a new threat to face recognition? Assessment and detection, CoRR
[78] Dolhansky, B.; Howes, R.; Pflaum, B.; Baram, N.; Ferrer, C. C., The deepfake detection challenge (DFDC) preview dataset (2019)
This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. It attempts to reflect the references listed in the original paper as accurately as possible without claiming the completeness or perfect precision of the matching.