[1910.09399] A Survey and Taxonomy of Adversarial Neural Networks for Text-to-Image Synthesis

Text-to-image synthesis refers to computational methods which translate human written textual descriptions, in the form of keywords or sentences, into images with similar semantic meaning to the text. In earlier research, image synthesis relied mainly on word to image correlation analysis combined with supervised methods to find best alignment of the visual content matching to the text. Recent progress in deep learning (DL) has brought a new set of unsupervised deep learning methods, particularly deep generative models which are able to generate realistic visual images using suitably trained neural network models. In this paper, we review the most recent development in the text-to-image synthesis research domain. Our survey first introduces image synthesis and its challenges, and then reviews key concepts such as generative adversarial networks (GANs) and deep convolutional encoder-decoder neural networks (DCNN). After that, we propose a taxonomy to summarize GAN based text-to-image sy

5 mentions: @arxiv_cs_cv_pr
Date: 2019/10/26 03:49

Related Entries

Read more [1910.07192] Animating Landscape: Self-Supervised Learning of Decoupled Motion and Appearance for Si...
0 users, 6 mentions 2019/10/17 02:19
Read more [2002.05509] Replacing Mobile Camera ISP with a Single Deep Learning Modelcontact arXivarXiv Twitter
0 users, 9 mentions 2020/02/17 06:51
Read more [2003.03040] DeLTra: Deep Light Transport for Projector-Camera Systemscontact arXivarXiv Twitter
0 users, 4 mentions 2020/03/09 02:20
Read more [2003.02874] Optimizing JPEG Quantization for Classification Networkscontact arXivarXiv Twitter
0 users, 4 mentions 2020/03/09 21:51
Read more [2003.10333] Neural Contours: Learning to Draw Lines from 3D Shapescontact arXivarXiv Twitter
0 users, 13 mentions 2020/03/29 02:21