Abstract
Generative adversarial network, in short GAN, is a new convolution neural network (CNN) based framework with the great potential to determine high dimensional data from its feedback. It is a generative model built using two CNN blocks named generator and discriminator. GAN is a recent and trending innovation in CNN with evident progress in applications like computer vision, cyber security, medical and many more. This paper presents a complete overview of GAN with its structure, variants, application and current existing work. Our primary focus is to review the growth of GAN in the computer vision domain, specifically on image enhancement techniques. In this paper, the review is carried out in a funnel approach, starting with a broad view of GAN in all domains and then narrowing down to GAN in computer vision and, finally, GAN in image enhancement. Since GAN has cleverly acquired its position in various disciplines, we are showing a comparative analysis of GAN v/s ML v/s MATLAB computer vision methods concerning image enhancement techniques in existing work. The primary objective of the paper is to showcase the systematic literature survey and execute a comparative analysis of GAN with various existing research works in different domains and understand how GAN is a better approach compared to existing models using PRISMA guidelines. In this paper, we have also studied the current GAN model for image enhancement techniques and compared it with other methods concerning PSNR and SSIM.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
Generative adversarial networks are the sub-class of the generative model, with the competence to produce/verify a new set of data. A generative adversarial network was introduced in 2014 by researchers Ian J GoodFellow et al. [28] in his research paper published in IEEE Journal.
Most neural networks aim to learn from the limited data set, which usually faces misclassification and overfitting problems. The GAN model is a powerful architecture with a component of self-generate, self-learning and competence to overcome the limitation of traditional networks.
According to GoodFellow et al. [28] research paper published in 2014, GAN and its structure are described as a two-player min–max game or Nash Equilibrium with the function value V(D,G). The detailed mathematical description is given by Good Fellow is shown in Formula 1.
In 2015, a new variant of GAN was proposed by ABC, and this work has become a basic approach for all upcoming variants of GAN. In this work, the GAN is mainly broken down into two modules, Generator G(A) and Discriminator D(A). Here, the generator generates the data, similar to the training dataset, and the discriminator is a network trying to identify the real and generated data. The GAN model work on the principles of game probability. The theory is to generate a random variable (A) whose properties are similar to the actual variable. Specifically, the generation of the random variable experiment is repeated N number of times until it gets the actual variable value which is known as Probability(P). And the possible outcome of this is known as sample space represented by Ω. Overall we claim it as probability distribution function P(A) where the probability of all outcomes can generate the result R as shown in Formula 2.
P: Ω → Z
Hence, we can say the summation of all probability can give an actual variable, i.e. ΣAЄΩP(A) = 1. A simple real-time example of GAN is two people playing Guess the number in the mind game. R. Chang et al. 2023 [143] and Z Pan et al. 2019 [105] are some of the experimental works that supported the above hypothesis. The simple GAN model concerning Game probability is shown in Fig. 1.
1.1 Basic Modules of GAN
GAN deep learning module is mainly made of two adversarial network modules Generator and a Discriminator.
Generator: It is an unsupervised model in GAN that generates new values in input distribution based on the summary of real input variable distribution. The generator reads fixed-length random vectors based on the Gaussian distribution concept; after training, the generator forms compressed data distribution corresponding to multi-dimensional vector space. The architecture of the generator is shown in Fig. 2.
Discriminator: Discriminator is a supervised GAN model that uses input and general variables based on the class label. The discriminator inputs value from real and generated dataset and predict a binary label 0 and 1, classifying the received data as fake or the same, respectively. The architecture of the discriminator is shown in Fig. 3.
1.2 Applications of GAN
A generative adversarial network is a trending neural network model with several fascinating applications in various domains. The usage of the GAN model in many applications has shown a drastic change in the result and system accuracy. In this study, we have discussed some of the well-known applications of GAN as follows.
1.2.1 Application of GAN in Cyber and Network Security
The various anomalies in the security system are damaging the system and our privacy. The new GAN approach is vital in improving cyber security and building a safer system environment that protects against various attacks. GAN is one of the latest ideas in self-driving cars to enhance their safety and protection during navigation and collection of specific sensor data. These days, applying GAN in cyber security has become one of the exciting fields among researchers. A large set of research works can be observed using the GAN approach in the cyber security area.
The GAN model can be practiced to detect various cyber intrusions like distributed denial of service attacks, botnet attacks etc. [1]. To detect cyber-physical system attacks, FID-GAN, an unsupervised intrusion detection system, is designed [2]. Many imbalanced data set problems during intrusion detection are solved by using simple GAN and GAN with Earth-Mover distance in [6, 7]. To enhance the accuracy of GAN model, the labelled sample set is expanded by using an advanced binary classification model [3]. In Yixuan Wimu et al., a mining approach is presented based on the fuzzy rough set, CNN and GAN to enhance intrusion detection based on feature extraction [4, 5]. GAN and modified versions of GAN, like PAC-GAN, have notably contributed to detecting malware and standard packets in cyber security [8, 9].
Overall, GAN can be used in most of the studies related to threat detection [10,11,12], false data injection attacks imbalanced data problems etc., in the cyber and network security domain.
1.2.2 Application of GAN in Healthcare Industry
GAN is one of the fascinating inventions of AI that has contributed to most of the domains in today's research environment. Most of the SURPRISING and splendid tasks of human and AI bots are the work of GAN. The healthcare industry is one of the majorly benefited fields of GAN. Radiology images like CT, MRI, ultrasound, radiography, and elastography resolution can be enhanced by GAN. The small data set problem during the training phase is one of the major issues addressed in the healthcare domain by GAN.
To understand the role of GAN in healthcare, we have gone through different research works. The major work was observed in enhancing image clarity. In Yuhui Ma et al., [13] a versatile novel approach, Still-GAN, is introduced to enhance low and high-quality images. Lesion Focused Multiscale in [14] and enhancement of low-resolution counterparts of CT images by the GAN-Circle approach [15] are a few other enhancement techniques noted. To enhance and generate a high-resolution 3D medical image, hierarchical amortized GAN is used in research work presented in [16].
The other notable application of GAN is image generation and synthesis. Chikato Yamasoba et al. [17] presented an approach to generate different modality images using DCGAN and Cycle GAN. In [18], a one more approach where DC-GAN is used for medical data synthesis, and generating MR images using GAN is observed [19]. Strategies like GAN augmentation for liver lesion classification [20], fund-GAN approach to augment fundus image for retinal image classification [21], pseudo-3D cycle GAN lumbar spine data synthesis [22] and 3D multi-conditional GAN for image augmentation in lung module classification some more work reviewed in image augmentation [23]. Finally, we noticed a few more applications like medical image segmentation by using MS-GAN [24], U-net Based GAN [26], image fusion on GAN [25] and tumour classification [27]. In conclusion, GAN has become a boon and advantage for the growth of the medical field.
1.2.3 Application of GAN in Computer Vision
In this survey, we have considered some of the applications of GAN, which have made revolutionary improvements in computer vision. The application of GAN in computer vision can be classified into the generation of image datasets, super-resolution, creating human face photographs, image-to-image translation, generating realistic pictures, face frontal view generation and generating new human poses.
Generating image datasets is an approach to creating new plausible images from existing images. Firstly, this approach was designed by Ian Goodfellow et al. in 2014 [28]. In this paper, the author has generated a likely image from the MINIST data set. The MINIST dataset combines CIFAR-10 small objects and the Toronto face database. In 2015 [29], Alec Radford et al. designed an approach to stabilize GAN. This approach was beneficial to overcome with small dataset overfit problem in CNN and ML.
To enhance the image resolution, SRGAN is one of the well-known approaches used widely. In this approach, the generated image has a higher pixel resolution; some of the known works using SRGAN were conducted in 2016 by Christin Leidg et al. [30] and in 2017 by Huang Bin et al. [31]. In 2018, Subeesh et al. [32] presented an approach to creating a high-resolution image for photographs using the SR network.
The GAN model can also be applied to generate pictures of human faces. In 2017, Tero Karras et al. [33] published a work where celebrity faces are generated from input samples, and the generated output is quite similar. Later many works were published using Tero Karras et al. work as a base paper.
The image-to-Image translation is a vital application of image translation research using GAN. The first paper on image translation was published in 2016 by Philip Isola et al. [34]. The work was proposed on conditional adversarial Network and pix2pix approach. In 2018, Andrew Brock et al. [35] proposed a work to generate realistic photographs using bigGAN. It is noticed the generated images are very similar to the old photos with better accuracy. Face frontal view generation by GAN came to light in 2017 by Rui Hang et al. [36]. The global and local GAN is used in this paper. The face photos taken from various angle is used to generate the different frontal view and human poses.
To analyze the growth and advancement of GAN in various fields, we have queried across the different journals with a keyword "GAN" and "Generative Adversarial Network" with a filter of publication year from 2016 to 2023. This search aims to give a detailed, comprehensive overview for researchers and practitioners where we can answer the following research questions based on the growth of GAN, as shown in Table 1. In Table 2, CONF: Conference, JOR: Journal, EAA: Early Access Article, MAG: Magazine, BOK: Book, RA: Review Article, RSA: Research Article, BOC: Book Chapter, COP: Conference Proceeding, RWE: Reference Work Entry and RW: Reference Work.
After analyzing research questions, we understood that the progress of GAN in various domains is increasing exponentially, especially in computer vision, as observed in RQ5 in Table 2. This paper aims to analyze and understand current practices, approaches and ground truth of GAN in computer vision and image enhancement techniques. Our contribution to this paper is as follows:
-
A detailed literature survey on GAN and its variants is carried out. The detailed report on the technique and the current tool is outlined by framing the research questions.
-
A detailed review of existing work in image enhancement techniques in GAN is discussed. Depth analysis of evaluation metrics, datasets, methodology and tools of various methods are explained in detail by carrying out a systematic literature review.
-
We highlighted some of the gaps and challenges in the spectrum of image enhancement techniques using GAN, which can be helpful for future research work.
Overall, this paper is structured as follows, in Sect. 2, the detailed review process is presented by defining the research question. In Sect. 3, variants of GAN in computer vision and outcome of research questions are outlined; Sect. 4, gaps and challenges are discussed, and in Sect. 5 conclusion.
2 Taxonomy of Systematic Literature Review
To perform a detailed and systematic literature survey, we have referred few benchmark review works proposed by Bugen et al. [37], B Kitchenham et al. [38] and M. A Barbar et al. [39] in the area of software engineering. Throughout this paper, we have taken up their approaches to design our review and manifested our survey into three significant steps planning, conducting, and reporting, as shown in Fig. 4.
2.1 Planning
The primary aim of this stage is to give sufficient information and give a systematic path for the conduction and reporting stage. This phase consists of three steps.
-
Identifying the need for a Survey
Before a systematic survey, the research scholar must understand how important the survey is. The researcher should undergo existing survey work available, and we have read a good count of work to perform this step.
-
Formulate Research Question
A well-structured research question will help to understand the identified study in a proper direction. We have drawn all possible research questions in this phase to match our study.
-
Review Protocol
Generally, protocols are the critical element in most of the literature survey. Analyzing the described research question, planned strategy, and background context meet the designed survey or not is executed in this step. In this study, we have followed a hierarchical approach to review protocol.
2.2 Conducting
Conducting is the next step after the planning. In this phase, there are four steps.
-
Search Strategy
It is a predefined approach that aims to find possible primary research papers related to our work. In this step, we designed a search technique based on a specific keyword, a synonym of a keyword or a constructed string using possible keywords.
-
Selection of Study on Criteria Basis
Various challenges are encountered during the literature selection process, like language, author, journal etc. The presented work follows a well-defined protocol to decrease bias and ensure fairness.
-
Study Quality Assessment
This process's primary goal is to ensure the quality and relevance of selected papers from the previous steps. Here, we have fixed a set of quality metrics to appraise the quality of this study.
-
Data Extraction and Monitoring
In this phase, the source and form used to collect the required data for the study are designed. We have carefully selected the necessary references and entities in our research and well-recorded them.
2.3 Reporting
In this phase, all the extracted and analyzed data is summarized well. This phase consists of two steps.
-
Data Synthesis
In this step, data synthesis and summarization are achieved using a graphical and tabular approach, which is more suitable for understanding.
-
Reporting Finding
In this stage, the synthesized data is reported in the proper channel that can target research scholars and evidence.
2.4 Implementation of Systematic Literature Review
2.4.1 Identifying the Need for a Survey
To identify the importance of the study, we tried to analyse the current research trend, especially in GAN. We have searched various journals, and it is observed there has been a steady growth in the count of papers published over the years, as shown in Table 2.
2.4.2 Formulate Research Question
Picking a research question is an essential first step to define the overall purpose of the specific study. In this paper, we have established stable research questions (RQ) to guide researchers, increase confidence in the domain and understand the recent exercise and trend of GAN in computer vision. The established RQs and SRQs are given in Table 3.
2.4.3 Review Protocol
After defining the RQs, the research questions are sent to the research guide, research supervisor and co-supervisor to check the depth and correctness of the RQ. The research guide has also evaluated the protocol design of this study. After reviewing the protocol from the supervisor, we proceeded further in our research.
2.4.4 Search Strategy
We have started our research with the intent to compile as many studies and work related to our research domain. In this phase of the collection, we included all possible keywords and also phrases that match the keywords. The possible keyword used is shown in Table 4.
To collect the study papers, we looked into several journal repositories. However, many digital journals are available these days; the selected journals for this paper are listed below.
-
Web of Science
-
IEEE digital library
-
ACM digital library
-
Springer
-
Semantic Scholar
This search is restrained to the period of 2014 to 2023, including journals, conferences and archives.
2.4.5 Selection of Study on Criteria Basis
In selecting the relevant work after the search and collection process, we established two inclusion criteria to pick the most relevant study, as listed below.
-
The keyword should be part of the abstract, keyword and title.
-
Few papers have worked in GAN and do not involve the keyword in the abstract, title and keywords. We have gone through the complete article to complete the selection process in such cases.
To skip some studies that do not support the objective and aim of the study, we have defined three exclusion criteria as follows.
-
Studies which are not in English.
-
GAN papers related to healthcare, cyber security, networks and other domains unrelated to computer vision.
-
Conference proceedings are not considered for the study.
The detailed inclusion process is shown using the PRISMA approach in Fig. 5.
2.4.6 Study Quality Assessment
After the selection process, accessing quality proof is crucial to conduct a proper systematic review. The result obtained from the survey should be firm and avoid all sorts of bias. This paper uses the criteria stated in research work [40] to analyse the quality assessment.
2.4.7 Data Extraction and Monitoring
In this phase, we will extract the data required for the study. After going through six journal repositories to answer the defined RQs, we have set some rules and minimal entities required from each paper. In this paper, we extracted author details, publication details, journal details, dataset, features, methods, and metrics used.
2.4.8 Data Synthesis and Reporting
The data synthesis and reporting is the last phase of the systematic review, where the findings from the data extraction stage are segregated and presented as a supportive definition for RQs. In this phase, we have used graphs and tables to visualize the summarized data.
3 Outcomes
3.1 RQ-1: What are the Well-known Variants of GAN?
3.1.1 Deep Convolutional Generative Adversarial Networks (DCGAN)
The DCGAN layer model was proposed by Radford et al. in 2015, in which they presented two CNN models, namely discriminator and generator with a convolution transpose layer as shown in Fig. 6.
The principal aim of DCGAN is to support unsupervised learning using stride and transposed convolution for downsampling and upsampling[66].
The essence of DCGAN is as follows:
-
Eliminates all hidden layers.
-
Max pooling layers are replaced with the stride convolution layer and fractional stride convolution layer in the discriminator and generator, respectively.
-
Batch normalisation is used, except for the generator's output layer and the discriminator's input layer.
-
Leaky ReLu is applied in all layers of the discriminator.
-
ReLu is used in the generator except in the output layer. In the generator output layer, tanh is applied.
In this paper, some of the work based on DCGAN are presented. In the survey process, our foremost aim is to identify the methodology, model and application where DCGAN can be applied. In [41], Yurika Sagawa et al. presented a model for facial image generation using attributes and labels by DCGAN, and a few more works are noticed where researchers' primary motivation was to generate a facial image using DCGAN in [44, 46, 53, 58, 61].
The DCGAN gives a higher contribution in data augmentation to enhance any target CNN model's accuracy by increasing the dataset's size or building a training model, as seen in [52, 59]. However, the most noticeable work of DCGAN is in creating and performing analysis of Anime Characters [61, 63]. It is noticed using the DCGAN with the CNN model or some well-known algorithm like self-learning [58], SVM [46] etc., will give better accuracy. The detailed study of DCGAN is outlined in Table 5.
3.1.1.1 SRQ-1.2: What Are the Applications of DCGAN?
Based on the applications of DCGAN in computer vision, we noticed the higher contribution of DCGAN is marked in image generation and data augmentation. Considering all 25 works together, we observed five papers specially used for face image synthesis, six on data augmentation, two on anime character generation, four on resolution enhancement, and eight on data generation. Table 5 illustrates a detailed study of 25 research papers on DCGAN; based on this table, Fig. 7 outlines a list of DCGAN applications. Hence it concludes DCGAN works fine in situations of image generation.
3.1.2 Conditional Generative Adversarial Networks (CGAN)
Conditional GAN (CGAN) is a novel approach and a well-known variant of GAN designed to train generative models. The first glance of CGAN was in 2015, presented by Mehdi Mirza et al. [67].
The primary function of conditional GAN is to learn samples from distribution instead of sampling from marginal distribution. In conditional GAN sampling is based on additional auxiliary information like labels and data. The detailed architecture is given in Fig. 8. Based on Fig. 8 the 2-player min–max function v(G, D) given in [29] can be redefined for CGAN as shown below.
Here D(x|y) is the discriminator with x input and y label, and G(x|y) is the generator with noise vector and y label.
Generally, the major applications of CGAN are video generation, face generation, Image-to-Image Synthesis and Text to Image Synthesis. When we queried IEEE digital library with the keyword CGAN and filtered from 2019 to 2023, 24 publication topics were listed; in the extracted list, image classification, feature extraction, and medical image processing are the top 3 publication topics for CGAN. In this study, we have received 34 papers on CGAN by restricting our subject to CGAN in computer vision and image processing. The detailed outline of the studied research papers is given in Table 6.
In the survey phase, we came across various works; among these, image processing in the medical field using CGAN has many notable results. In [68], Changhee Han et al. used 3D Multi conditional GAN to augment a small fragmented CT image dataset. Similar works are observed in Ke Xu et al. [69] and Meng Li et al. [70], presenting a novel approach of CGAN named MCRGAN with the capacity to generate pseudo-CT images under limited training dataset conditions and transform-based architecture CGAN called MedViTGAN for augmentation of synthetic histopathology image. In the medical field, one more application of CGAN is image segmentation. In [71, 72], we noticed the application of CGAN in improving lesion contrast of MR images and retinal vessel segmentation. Image denoising by Zhao Yang et al. [73],[74]and Miao Tian et al. [75], Image synthesis by Huan Yang et al. [76], Zhaohui Liang et al. [77] and Yulin Yang et al. [78] are some of the noticed works of CGAN in image processing for the medical field.
Apart from medical image processing, we have studied the application of CGAN in the computer vision domain. In Jeongik Cho et al. [79], CGAN increases hyperparameters and reduces training speed. The designed approach uses multiple GANs, sharing all the hidden layers. In [80], the work presented by Tetsuya Ishikawa et al. illustrated a method to augment training data using CGAN. Few works in computer vision addressed problems like large model size and high interface time [81], and in [82], Felipe Coelho Silva et al. demonstrated a semi-automatic frame for manga art colourization. The other application of CGAN is in quality reconstruction, Art font, image generation, video games, rejuvenation of face image etc. In Table 6, we have given a comparative analysis of all our studies in CGAN based on parameters like purpose, model and outcome.
3.1.2.1 SRQ-1.2: What are the Applications of CGAN?
After studying 34 research works on CGAN in computer vision, we recorded Image to Image Synthesis is one of the well-noted applications. Considering the application and purpose of all 34 works, a detailed pictorial view is given in the graph of Fig. 9. From Fig. 9, we can conclude Image to Image Synthesis, Image Enhancement and Text to Image Synthesis are some of the applications where CGAN can definitely be used.
3.1.3 Cycle Generative Adversarial Networks (CYCLEGAN)
CycleGAN is another noteworthy variant of GAN presented in 2017 by Jun-Yan Zhu et al. [102]. The principal objective of the model is to map the images without paired data using the mapping function G(x- > y) and an adversarial loss function.
The image generates from the first generator, G(x), is similar to y, that is, G(x- > y) = > y = G(x). Moreover, in this approach using inverse mapping, y will learn from x that is F(y- > x) = > x = F(y). It can be said F(G(x)) = x and G(F(y)) = y using inverse mapping and cycle consistency loss. The pictorial representation of the Cycle GAN methodology is given in Fig. 10.
During the training process, Cycle GAN focuses more on the training dataset and follows a few practices as follows.
-
The training set paired image {xi,yi} where all xi in a dataset has yi as its counterpart.
-
The training set paired image {xi,yi} where every xi in the dataset dont have any match with yi.
To get a broad view of CycleGAN and its methodology, we have surveyed more than 25 research papers. The significant observation is that CycleGAN is majorly used for Image Synthesis, especially in the medical field. In Taesung Kwon et al. [103] and Jawook Gu et al. [123], image synthesis is used for denoising low-dose CT images. CycleGAN is also used for augmentation purposes in the classification of Melanoma medical images when a limited labelled dataset is available for training purposes[104]. ECG restoration [104] and fundus image enhancement in diabetic retinopathy classification [112] are the other recognized applications of CycleGAN in medical image processing. Moving apart, if we consider the general application of CycleGAN in computer vision, SAR to optical image registration [106, 120], NIR to RGB image [116] and VIS to NIR image [117] are the maximal noted research works. Along with this, image colourization, denoising and image enhancement in low light and night images are the few other works observed. A detailed study of Cycle GAN is given in Table 7.
3.1.3.1 SRQ-1.2: What are the Applications Of CycleGAN?
Based on the research and problem addressed in the state of art methods from Table 7, we collected some of the following basic observations. Firstly, CycleGAN is majorly used in Image Synthesis for unpaired data in various domains. Secondly, using CycleGAN, training time and memory consumption can be reduced. At last, CycleGAN is also helpful for converting any existing supervised method to an unsupervised one. The detailed usage of CycleGAN is given in Fig. 11.
3.1.3.2 Style Generative Adversarial Networks (STYLEGAN)
StyleGAN is a variant of GAN introduced by Tero Karras et al. in 2019 [134]. It is the first variant of GAN focused on the advancement and improvement of the generator, then the discriminator. This model is built with two networks, namely the mapping network and the synthesis network. The StyleGAN inputs the latent space vector directly into the mapping network, which comprises eight fully connected layers. The output of the mapping network is later sent to the synthesis network architecture consisting of 18 convolution layers and an AdaIN style network.
The synthesis network produces 4 × 4 to 1024X1024-sized images in every layer. Gaussian noise is added to the activation map before sending the images into the AdaIN method. And this is the primary reason that StyleGAN can produce high-resolution images. The comphrehensive architecture of StyleGAN is shown in Fig. 12.
The significant changes and updation in the StyleGAN compared to other GAN architecture are as follows.
-
Tuning and bilinear upsampling are added.
-
Gaussian noise is added in each block.
-
Mapping and Synthesis networks are added.
-
Latent vector input is not added to the generator.
Since StyleGAN was introduced in 2019, we got only a few research work on this model related to computer vision. The survey shows that most of the work collected from the paper is on the enhancement of image quality and advancement of StyleGAN. Dongsik Yoon et al. [135] started with the objective of generating diverse face images using available static faces. A similar work is observed in Shao Xiaofeng et al. [150], where the author develops the image using StyleGAN with ResNet using the FFHQ dataset. The idea of single-dimension pluralistic face image generation is taken to 3D pluralistic image generation in [136], where they worked on fixed styleGAN and RigNet with the 3DMM model. StyleGAN can also be used for classification, as demonstrated in [137], Face generation from the masked image in [138] and [151]. StyleGAN is widely used in fashion [154] and painting [145] [155] for better-quality images. The detailed study on StyleGAN is outlined in Table 8.
3.1.3.3 SRQ-1.2: What are the Applications of STYLEGAN?
To understand the application of StyleGAN in computer vision, we have been through 20 research papers. As we observed, StyleGAN in computer vision is widely used to address quality enhancement problems in generated images. Another major application of StyleGAN, as per the literature study, is Image Synthesis. For a better understanding of applications of StyleGAN in computer vision, we have plotted the graph as shown in Fig. 13.
3.1.4 Super Resolution Generative Adversarial Networks (SRGAN)
Super Resolution GAN is a well-known GAN variant to convert images with low resolution to high-resolution. This model was proposed by Twitter researchers in 2017. SRGAN model mainly subsist of three networks, namely generator, discriminator and VGG16 network, which is built using perceptual loss function.
The generator network consists of a convolution layer, PReLU layer and k3n64S1 strands with skip connection. And the discriminator network consists of a convolution layer, Leaky ReLU layer and k3n64S1 strands. The simple training network of SRGAN is illustrated in Fig. 14.
Super Resolution GAN is mainly used for creating photo-realistic images by using down-sampled images. In this study, we have been through some existing works to understand the role of SRGAN in removing the artefacts in low-resolution images. SRGAN can be used across various domains using computer vision techniques. In Yudai Nagano et al. [156], SRGAN creates a high-resolution food image. The author has mainly focused on inducing noises like jpg, blur etc. Junchao et al. [167], in this work the author used SRGAN for textile image reconstruction to get better accuracy than bilinear. In the survey, we observed most of the SRGAN works are based on facial resolution enhancement in the face image. In Hao Dou et al. [157], Minjie et al. [160], and Hai Nguyen Truong et al. [166], the SRGAN is used for facial resolution enhancement using orthogonal projection, wavelet transform and total variation loss, respectively. The SRGAN can be used to enhance the CT images [161] and fundus images [163] in medical image processing. The researcher Nai Feng Zhang et al. [174] have used SRGAN to deblur distant pedestrians. and Yong Hun Kim et al. [158] used SRGAN to restore old documents. The detailed study on SRGAN is outlined in Table 9.
3.1.4.1 SRQ-1.2: What are the Applications of SRGAN?
After analyzing several research works on SRGAN, we noted that image resolution enhancement, especially facial, medical image, textile, and pedestrian images, are the main areas in which SRGAN is used. SRGAN can also be used for image segmentation, classification and restoration purposes. The detailed use of SRGAN in various domains is shown in Fig. 15.
3.2 SRQ-1.1: What are the Frameworks Available to Work with GAN?
Generative Adversarial Network (GAN) is successfully used for image synthesis, data augmentation, image restoration and many more. Practising GAN on primary python IDE or any framework is challenging and lengthy. To minimize the complexity these days, we have various tools in the market to support GAN. In this section, we have discussed available GAN tools, their features and applications that simplify the usage of GAN.
-
GAN LAB
It is a visual interactive experiment tool to train GAN with a 2D data distribution model and visualize the internal working system. The GAN lab is built on TensorFlow. js and UI on GPU accelerated deep learning library. Using the GAN Lab, model learning visualization and improving fake samples is much easier.
Some of the features of GAN LAB are:
-
Slow motion code
-
Adjustment of the interactive hyperparameter is possible
-
User-defined data distribution is possible.
-
VeGANs
VeGANs is a python library with PyTorch framework for GAN. This library is mainly designed for developers willing to develop their own generator and discriminator network.
-
TORCH-GAN
Torch-GAN is a PyTorch framework for GAN. This framework is a collection of building blocks of GAN which gives customization for popular GAN datasets. Torch-GAN library offers provision for adding a new plugin for loss function and architecture, as well as the option to visualize various logging backgrounds.
-
HYPERGAN
HyperGAN is a framework with a user interface and API. Building the GAN model on HyperGAN makes the training process more straightforward. In HyperGAN, replacing part of GAN with JSON file or creating a new GAN is way easier than in other frameworks.
-
IMAGINAIRE
Imaginaire is an invention of NVIDIA; also a PyTorch-based GAN library integrating all the NVIDIA image and video synthesis projects. This library has several functionalities with six algorithms like Pix2PixHD, FUNIT, MUNIT, UNIT, COCO-FUNIT and SPADE.
-
MIMICRY
Mimicry is a lightweight PyTorch library to monitor GAN's loss and probability curves. This library is supported by the Tensor board, which is helpful in the performance comparison of multiple GAN models.
-
GAN TOOLKIT
GAN toolkit is a flexible library by IBM based on No code approach. This library helps the user to work with config files and command line arguments. It is an open-source library that allows multiple libraries like Keras, PyTorch and Tensor flow.
-
TFGAN
TFGAN is a light weighted library used for the evaluation of GAN. This library comprises many GAN operations, normalization techniques, losses etc. TFGAN can be used in Google TPU and GPU and is also compatible with Tensorflow2. For the self-study of GAN, TFGAN is the best tool.
-
PyGAN
PyGAN is a library in Python to implement models like CGAN, GAN, adversarial autoencoder and energy-based GAN. This library is mainly used for semi-supervised learning.
-
STUDIOGAN
StudioGAN is a library for GAN on PyTorch Framework for both conditional and unconditional image generation. StudioGAN has an inbuilt benchmark for CIFARIO, TinyImage Net and ImageNet. This library has a unique feature that performs better for low memory consumption.
3.3 RQ-2: What are the Well-known Approaches for Image Enhancement Techniques Using GAN?
Image Enhancement is a technique of manipulating digital pixel value so that the resultant images are more suitable for visualization and further analysis. The general idea of image enhancement is to process the given image and make it more convenient for the specific application.
Image enhancement can be executed in different ways; it can be the sharpening of image features such as boundaries, edges etc. It can also be removing noise, increasing an image's brightness or changing contrast. It is said that image enhancement can't improve the inherent content of data, but it can enhance the dynamic range of chosen features.
There are numerous techniques for image enhancement in computer vision. And Fig. 16 shows a general approach or hierarchy to carry out image enhancement.
To understand the methodology used for image enhancement using GAN, we studied many research papers on different variants of GAN for image enhancement techniques. Some researchers worked on the enhancement of face images and their features [31, 175, 184, 200], and some papers mainly concentrated on computer vision in the medical field. In [178] [180, 190], the author focused on enhancing the clarity of the fundus image for better recognition of the iris. And in [76, 128, 191, 195, 202, 204], the author's principal objective was enhancing X-Ray, MRI and CT Scan images. The research in image enhancement is not only restricted to image processing in the medical field; it has also shown a comprehensive improvement in enhancing low light, low luminance and underwater images. In Table 10, we have illustrated all the studied research work in detail based on their methodologies.
3.4 RQ-2.1: Which are the Datasets Typically Used in Image Enhancement by GAN?
We observed various datasets were used in numerous studies related to image enhancement using GAN variants for testing and training purposes. Generally, the datasets are publicly available on the internet; in some cases, datasets are private, self-created, and acquired. It is found these dataset has made an incredible advancement in image enhancement using GAN. In turn, because of these datasets, most of the GAN variants can achieve their desired outcome. In Table 11, all the datasets used in different research papers related to image enhancement in GAN are displayed respectively to their variants.
3.5 RQ-2.2: What are the Models Used in Image Enhancement Techniques Using GAN?
This section illustrates various GAN variants used in image enhancement. Based on studies and considering all the GAN variants used for image enhancement, we have outlined Table 12 in this paper. While presenting the summary table, we considered noise removal, clarity enhancement, blurriness removal, contrast enhancement and brightness enhancement as image enhancement techniques. In this paper, we assessed 69 reports to study image enhancement using GAN. Based on 69 articles, Table 12 is drawn, listing all the variants of GAN used for image enhancement, the number of studies in each category and the percentage of studies in each category (PSC). Using Table 12, we can reveal SRGAN is the most used GAN in the image enhancement approach.
3.6 RQ-2.3: What are the Metrics Used to Evaluate Image Enhancement Using GAN?
This section of the paper showcases various measurement metrics used in calculating, analysing and assessing the performance of the model used for image enhancement in GAN. Table 13 defines multiple metrics and performance units in all the studies on image enhancement methods. It also gives the proper explanation and description of each measurement metric and the number of studies related to each metric. Based on Table 13, it can be concluded PSNR and SSIM are often used measurement metrics to evaluate image enhancement studies across various GAN models.
3.7 RQ-3: Whether GAN is a Better Approach for Image Enhancement? How is Image Enhancement Performance in GAN, MATLAB, and Other Platforms for Image Enhancement?
To analyse how the GAN model is efficient for image enhancement compared to the other existing techniques, we split our analysis based on three categories: (i) Image enhancement using the GAN model (ii) Image enhancement using machine learning (iii) Image enhancement using MATLAB.
In this review work, we considered a maximum of ten sample existing studies from each category [204] [205] [206] [207] [208] [209] [210] [211] [212] [213] [214] [215] [216] [217] [218] [219]. And the PSNR and SSIM performance metrics are used for comparative analysis. We recorded minimum and maximum PSNR and SSIM observed from the collected sample study from each category as given in Table 14. Overall, in this section, Table 14 and Fig. 17 present the gist of the comparative analysis. By analyzing Table 14, we can say the GAN model is a better approach for image enhancement.
4 Limitations and Challenges
Please make sure that the paper you submit is final and complete, that any copyright This section lists some of the challenges, limitations and gaps noticed during the study. The observed gaps are as follows.
-
Minimal work is proposed to enhance and restore the image by extracting the original features of the image.
-
Using the GAN model for training purposes can increase the output, but it is noticed the model will become very unstable so that in each iteration result gets varied.
-
One more notable observation in numerous image enhancement works is that handling high-frequency and low-frequency features in images using the same model doesn’t give effective results.
-
Combining GAN with the extra deep neural network can increase the accuracy of output, but a rapid increase in training time is observed.
-
It is noticed no single GAN model is designed to address all possible noise in the image during the image enhancement technique.
5 Conclusion
The presented SLR illustrates the study of various state-of-the-art methods on GAN, variants on GAN and image enhancement techniques using GAN. This research gives a detailed view of the existing work of GAN published from 2018 to 2023. Throughout this paper, we answered all the possible questions on GAN by discussing its history, application, variants, limitations, image enhancement approaches, and conducted a comparative and summarizing examination of distinctions with other existing works. The overall summary of this study is as follows.
-
The GAN model is widely used in many domains like machine designing, architecture, medicine, construction, computer vision etc.
-
Linear growth is observed in research publications related to GAN. And in 2019–2020, a rapid increase in the publication count was seen.
-
Every GAN model has its own specialization approach; for example, the DCGAN can be mainly used in data augmentation like this; the detailed explanation of every variant of GAN is given in section III.
-
The SRGAN model holds a significant role in image enhancement.
-
PSNR and SSIM have widely used performance metrics for image enhancements.
-
The experimental result demonstrates that GAN is a practical approach and outperforms as a better model for image enhancement than other techniques.
With the rapid progress in technology and multimedia, GAN still needs to address many challenges. And this study gives a route map and valuable basic details for the research community in developing compelling research works on GAN.
Data Availability Statement
Data sharing is not applicable to this article, as no datasets were generated or analyzed during the course of the current study.
References
Smith TF, Waterman MS (1981) Identification of common molecular subsequences. J Mol Biol 147:195–197. https://doi.org/10.1016/0022-2836(81)90087-5
Shi A Cyber attacks detection based on generative adversarial networks, 2021 2nd Asia conference on computers and communications (ACCC), 2021, pp 111–114, https://doi.org/10.1109/ACCC54619.2021.00025.
Freitas de Araujo-Filho P, Kaddoum G, Campelo DR, Gondim Santos A, Macêdo D, Zanchettin C (2021) Intrusion detection for cyber–physical systems using generative adversarial networks in fog environment, In: IEEE internet of things journal, 8(8): 6247–6256. https://doi.org/10.1109/JIOT.2020.3024800.
Liao D, Huang S, Tan Y, Bai G (2020) Network intrusion detection method based on GAN model. Int Conf Comput Commun Netw Secur (CCNS) 2020:153–156. https://doi.org/10.1109/CCNS50731.2020.00041
Wu Y, Nie L, Wang S, Ning Z, Li S Intelligent intrusion detection for internet of things security: a deep convolutional generative adversarial network-enabled approach, In: IEEE Internet of Things Journal. https://doi.org/10.1109/JIOT.2021.3112159.
Piplai A, Chukkapalli SSL, Joshi A NAttack (2020) Adversarial Attacks to bypass a GAN based classifier trained to detect Network intrusion. In: 2020 IEEE 6th Intl Conference on Big Data Security on Cloud (BigDataSecurity), IEEE Intl Conference on High Performance and Smart Computing, (HPSC) and IEEE Intl Conference on Intelligent Data and Security (IDS), , pp 49–54, https://doi.org/10.1109/BigDataSecurity-HPSC-IDS49724.2020.00020.
Yilmaz I, Masum R, Siraj A Addressing imbalanced data problem with generative adversarial network for intrusion detection. In: 2020 IEEE 21st International conference on information reuse and integration for data science (IRI), 2020, pp. 25–30, https://doi.org/10.1109/IRI49571.2020.00012.
Hao X et al. "Producing More with Less: A GAN-based network attack detection approach for imbalanced data. In: 2021 IEEE 24th international conference on computer supported cooperative work in design (CSCWD), 2021, pp. 384–390. https://doi.org/10.1109/CSCWD49262.2021.9437863.
Liu Z, Li S, Zhang Y, Yun X, Cheng Z (2020) Efficient malware originated traffic classification by using generative adversarial networks. IEEE Symp Comput Commun (ISCC) 2020:1–7. https://doi.org/10.1109/ISCC50000.2020.9219561
Cheng A PAC-GAN: packet generation of network traffic using generative adversarial networks. In: 2019 IEEE 10th annual information technology, electronics and mobile communication conference (IEMCON), 2019, pp. 0728-0734, https://doi.org/10.1109/IEMCON.2019.8936224
Zhu D, Huang X, Li N, Sun H, Liu M, Liu J (2022) RAP-Net: a resource access pattern network for insider threat detection. Int Joint Conf Neural Netw (IJCNN) 2022:1–8. https://doi.org/10.1109/IJCNN55064.2022.9892183
Yun X, Huang J, Wang Y, Zang T, Zhou Y, Zhang Y (2020) Khaos: an adversarial neural network DGA with high anti-detection ability. IEEE Trans Inf Forensics Secur 15:2225–2240. https://doi.org/10.1109/TIFS.2019.2960647
Ma Y et al (Dec.2021) Structure and illumination constrained GAN for medical image enhancement. IEEE Trans Med Imaging 40(12):3955–3967. https://doi.org/10.1109/TMI.2021.3101937
Zhu J, Yang G, Lio P How can we make gan perform better in single medical image super-resolution? A lesion focused multi-scale approach. In: 2019 IEEE 16th international symposium on biomedical imaging (ISBI 2019), 2019, pp. 1669–1673, doi: https://doi.org/10.1109/ISBI.2019.8759517.
You C et al (2020) CT Super-resolution GAN constrained by the identical, residual, and cycle learning ensemble (GAN-CIRCLE). IEEE Trans Med Imaging 39(1):188–203. https://doi.org/10.1109/TMI.2019.2922960
Sun L, Chen J, Xu Y, Gong M, Yu K, Batmanghelich K (2022) Hierarchical amortized GAN for 3D high resolution medical image synthesis. IEEE J Biomed Health Inform 26(8):3966–3975. https://doi.org/10.1109/JBHI.2022.3172976
Yamasoba C, Tozaki T, Senda M (2021) Generation and evaluation of different modality of medical image based on GAN, In: 2021 IEEE nuclear science symposium and medical imaging conference (NSS/MIC), pp. 1–3. https://doi.org/10.1109/NSS/MIC44867.2021.9875577.
Divya S, Suresh LP, John A Medical MR Image Synthesis using DCGAN, In: 2022 First international conference on electrical, electronics, information and communication technologies (ICEEICT), 2022, pp. 01–04. https://doi.org/10.1109/ICEEICT53079.2022.9768647.
Han C et al. (2018) GAN-based synthetic brain MR image generation, In: 2018 IEEE 15th international symposium on biomedical imaging (ISBI 2018), pp. 734–738, https://doi.org/10.1109/ISBI.2018.8363678.
Frid-Adar M, Klang E, Amitai M, Goldberger J, Greenspan H (2018) Synthetic data augmentation using GAN for improved liver lesion classification, In: 2018 IEEE 15th international symposium on biomedical imaging (ISBI 2018), pp 289–293, https://doi.org/10.1109/ISBI.2018.8363576.
Shenkut D, Bhagavatula V (2022) Fundus GAN—GAN-based fundus image synthesis for training retinal image classifiers, In: 2022 44th Annual international conference of the IEEE engineering in medicine & biology society (EMBC), pp. 2185–2189. https://doi.org/10.1109/EMBC48229.2022.9871771.
. Oulbacha R, Kadoury S (2020) MRI to CT synthesis of the lumbar spine from a pseudo-3D cycle GAN. In: 2020 IEEE 17th international symposium on biomedical imaging (ISBI), pp 1784–1787, https://doi.org/10.1109/ISBI45749.2020.9098421.
Han C et al. Synthesizing diverse lung nodules wherever massively: 3d multi-conditional GAN-based CT image augmentation for object detection, In: 2019 international conference on 3D vision (3DV), 2019, pp. 729–737, https://doi.org/10.1109/3DV.2019.00085.
Zhang C et al (2018) MS-GAN: GAN-based semantic segmentation of multiple sclerosis lesions in brain magnetic resonance imaging. Digital Image Comput (DICTA) 2018:1–8. https://doi.org/10.1109/DICTA.2018.8615771
Wang L, Chang C, Hao B, Liu C (2020) Multi-modal medical image fusion based on GAN and the shift-invariant shearlet transform. IEEE Int Conf Bioinform Biomed (BIBM) 2020:2538–2543. https://doi.org/10.1109/BIBM49941.2020.9313288
Lou Z, Huo W, Le K, Tian X (2020) Whole heart auto segmentation of cardiac CT images using U-net based GAN, In: 2020 13th international congress on image and signal processing, BioMedical engineering and informatics (CISP-BMEI), pp. 192–196. https://doi.org/10.1109/CISP-BMEI51763.2020.9263532.
Yerukalareddy DR, Pavlovskiy E (2021) Brain Tumor classification based on MR images using GAN as a pre-trained model, In: 2021 IEEE ural-siberian conference on computational technologies in cognitive science, genomics and biomedicine (CSGB), pp 380–384, https://doi.org/10.1109/CSGB53040.2021.9496036.
Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, David Warde-Farley, Sherjil Ozair, et al. (2014) Generative Adversarial Networks.
Radford A, Metz L, Chintala S (2015) Unsupervised representation learning with deep convolutional generative adversarial networks
Ledig C, Theis L, Huszar F, Caballero J, Cunningham A, Acosta A, Aitken A, Tejani A, Totz J, Wang Z, Shi W (2017) Photo-realistic single image super-resolution using a generative adversarial network. 105–114. https://doi.org/10.1109/CVPR.2017.19
Bin H, Chen W, Xingming W, Chun-Liang L (2017) High-quality face image SR using conditional generative adversarial networks.
Vasu S, Madam N, Rajagopalan AN Analyzing perception-distortion tradeoff using enhanced perceptual super-resolution network, 2018.
Karras T, Aila T, Laine S, Lehtinen J Progressive growing of GANs for improved quality, stability, and variation, international conference on learning representations,2018.
Isola P, Zhu J-Y, Zhou T, Efros A (2017) Image-to-image translation with conditional adversarial networks. 5967–5976. https://doi.org/10.1109/CVPR.2017.632.
Brock A, Lim T, Ritchie J, Weston N (2016) Neural photo editing with introspective adversarial networks.
Huang R, Zhang S, Li T, He R (2017) Beyond face rotation: global and local perception GAN for photorealistic and identity preserving frontal view synthesis. 2458–2467. https://doi.org/10.1109/ICCV.2017.267.
Budgen D, Brereton P Performing systematic literature reviews in software engineering, In: Proc. 28th Int. Conf. Softw. Eng., New York, NY, USA, May 2006, pp. 1051–1052, https://doi.org/10.1145/1134285.1134500.
Kitchenham B, Charters S (2007) Guidelines for performing systematic literature reviews in software engineering. Softw. Eng. Group; Keele Univ., Durham University Joint, Durham, U.K., Tech. Rep. EBSE-2007- 01.
Babar MA, Zhang H (2009) Systematic literature reviews in software engineering: Preliminary results from interviews with researchers. In: Proc. 3rd Int. Symp. Empirical Softw. Eng. Meas., Lake Buena Vista, FL, USA, pp. 346–355, https://doi.org/10.1109/ESEM.2009.5314235.
Do SE, Rothermel G (2005) Supporting controlled experimentation with testing techniques: an infrastructure and its potential impact. Empirical Softw Eng 10(4):405–435
Sagawa Y, Hagiwara M (2018) Face image generation system using attribute information with DCGANs. In: 2nd international conference on machine learning and soft computing (ICMLSC '18). Association for Computing Machinery, New York, NY, USA, 109–113.
Zhong J, Gao J, Chen R, Li J (2019) Digital recognition of street view house numbers based on DCGAN. In: 2nd international conference on image and graphics processing (ICIGP '19). association for computing machinery, New York, NY, USA, 19–22.
Kim DD, Tanseef Shahid M, Kim Y, Lee WJ, Song HC, Piccialli F, Choi KN Generating Pedestrian Training Dataset using DCGAN. In: 3rd international conference on advances in image processing (ICAIP 2019). Association for computing machinery, New York, NY, USA, 1–4, 2020.
Liu B, Lv J, Fan X, Luo J, Zou T Application of an improved DCGAN for image generation. Mobile information systems, 2022, Article ID 9005552, 14 pages, 2022.
Mahmoud MAB, Guo P (2019) A novel method for traffic sign recognition based on DCGAN and MLP With PILAE algorithm. IEEE Access 7:74602–74611. https://doi.org/10.1109/ACCESS.2019
Sun L, Liang K, Song Y, Wang Y (2021) An improved CNN-based apple appearance quality classification method with small samples. IEEE Access 9:68054–68065
Yang X, Xu P, Xue Y, Jin H (2021) Contextual feature constrained semantic face completion with paired discriminator. IEEE Access 9:42100–42110
Xu J et al. (2021) Point AE-DCGAN: A deep learning model for 3D point cloud lossy geometry compression, 2021 Data Compression Conference (DCC), Snowbird, UT, USA, pp. 379–379.
Wu Q, Chen Y, Meng J (2020) DCGAN-based data augmentation for tomato leaf disease identification. IEEE Access 8:98716–98728
Mostofa M, Mohamadi S, Dawson J, Nasrabadi NM (2021) Deep GAN-based cross-spectral cross-resolution iris recognition. IEEE Trans Biometrics Behav Identity Sci 3(4):443–463
Kumari N, Anwar S, Bhattacharjee V (2021) DCGAN based pre-trained model for image reconstruction using ImageNet. 2021 Seventh International conference on Bio Signals, Images, and Instrumentation (ICBSII), Chennai, India.
Li W, Jia T, Chen Q, Wu Y, Wang J, Huang J (2020) Omnidirectional Ring Structured Light Noise Filtering Based On DCGAN Network And Autoencoder. IN: 2020 international conference on culture-oriented science & technology (ICCST), Beijing, China.
Wu K, Yu Y, Zhang X, Li J, Zhang Q (2020) Application of face data augmentation based on rotate-and-render-DCGAN in campus security. In: 2020 IEEE 3rd international conference of safe production and informatization (IICSPI), Chongqing City, China.
Yu S-N, Wang S-W, Chang YP (2022) Improving distinguishability of photoplethysmography in emotion recognition using deep convolutional generative adversarial networks. IEEE Access 10:119630–119640
Lv T, Wen C, Zhang J, Chen Y (2020) A face recognition algorithm based on CNN with ELBP and DCGAN. In: 2020 international symposium on computer engineering and intelligent communications (ISCEIC), Guangzhou, China.
Wang K, Liu MZ (2020) Object Recognition at Night Scene Based on DCGAN and Faster R-CNN. IEEE Access 8:193168–193182
Sun Q, Guo J, Liu Y (2022) Face image synthesis from facial parts. J Image Video Proc.
Shawi RE, Kilanava K, Sakr S (2022) An interpretable semi-supervised framework for patch-based classification of breast cancer. Sci Rep 12:16734
Nasrin SS, Rasel RI HennaGAN: Henna Art Design Generation using Deep Convolutional Generative Adversarial Network (DCGAN)," 2020 IEEE International Women in Engineering (WIE) Conference on Electrical and Computer Engineering (WIECON-ECE), Bhubaneswar, India, 2020.
Dewi C, Chen R-C, Liu Y-T, Jiang X, Hartomo KD (2021) Yolo V4 for advanced traffic sign recognition with synthetic training data generated by various GAN. IEEE Access 9:97228–97242
Li Z, Wan Q Generating anime characters and experimental analysis based on DCGAN Model. In: 2021 2nd international conference on intelligent computing and human-computer interaction (ICHCI), Shenyang, China, 2021.
Yang X, Xu P, Jin H, Zhang J (2021) Semantic face completion based on DCGAN with dual-discriminator. In: 2021 7th annual international conference on network and information systems for Computers (ICNISC), Guiyang, China.
Jiang Y (2021) Performance analysis anime character generation based on DCGAN Model. In: 2021 2nd international conference on intelligent computing and human-computer interaction (ICHCI), Shenyang, China.
Qiao M, Chen M, Liang X, Guo Y (2021) Flower Data set Expansion Based on DCGAN and ResNet Classification Algorithm Based on Transfer Learning, In: 2021 6th international symposium on computer and information processing technology (ISCIPT), Changsha, China.
Tingfei W, Jingpeng G, Zhiye J (2021) ISAR Image Inpainting Algorithm Based on DCGAN, 2021 International Symposium on Antennas and Propagation (ISAP), Taipei, Taiwan.
Alec R, Luke M, Soumith C (2015) Unsupervised representation learning with deep convolutional generative adversarial networks.
Mirza M, Osindero S (2014) Conditional Generative Adversarial Nets.
Han C et al. Synthesizing Diverse Lung Nodules Wherever Massively: 3D Multi-Conditional GAN-Based CT Image Augmentation for Object Detection, 2019 international conference on 3D vision (3DV), Quebec City, QC, Canada, 2019.
Xu K et al (2019) Multichannel residual conditional GAN-leveraged abdominal pseudo-CT generation via Dixon MR images. IEEE Access 7:163823–163830
Li M, Li C, Hobson P, Jennings T, Lovell BC MedViTGAN: End-to-End Conditional GAN for Histopathology Image Augmentation with Vision Transformers, In: 2022 26th international conference on pattern recognition (ICPR), Montreal, QC, Canada, 2022.
Hamghalam M, Wang T, Qin J, Lei B Transforming intensity distribution of brain lesions via conditional gans for segmentation. In: 2020 IEEE 17th international symposium on biomedical imaging (ISBI), Iowa City, IA, USA, 2020.
Rammy SA, Anwar SJ, Abrar M, Zhang W (2019) Conditional Patch-based Generative Adversarial Network for Retinal Vessel Segmentation. In: 2019 22nd international multitopic conference (INMIC), Islamabad, Pakistan.
Yang Z, Yan C, Chen H (2021) Unpaired low-dose CT denoising using conditional GAN with structural loss, in: 2021 international conference on wireless communications and smart grid (ICWCSG), Hangzhou, China.
de Almeida MB, Alves Pereira LF, Ren TI, Cavalcanti GDC, Sijbers J (2021) The Gated Recurrent Conditional Generative Adversarial Network (GRC-GAN): application to denoising of low-dose CT images, In: 2021 34th SIBGRAPI conference on graphics, patterns and images (SIBGRAPI), Gramado, Rio Grande do Sul, Brazil.
Tian M, Song K (2021) Boosting magnetic resonance image denoising with generative adversarial networks. IEEE Access 9:62266–62275
Yang H, Xia K, Anqi B, Qian P (2019) Khosravi Abdomen MRI synthesis based on conditional gan, 2019 international conference on computational science and computational intelligence (CSCI), Las Vegas, NV, USA.
. Liang Z, Huang JX, Li J, Chan S Enhancing Automated COVID-19 Chest X-ray Diagnosis by image-to-image GAN translation. In: 2020 IEEE international conference on bioinformatics and biomedicine (BIBM), Seoul, Korea (South), 2020.
Yang Y et al. (2022) Synthesizing contrast-enhanced computed tomography images with an improved conditional generative adversarial network. In: 2022 44th annual international conference of the IEEE engineering in medicine & biology society (EMBC), Glasgow, Scotland, United Kingdom.
Cho J, Yoon K (2020) Conditional activation GAN: improved auxiliary classifier GAN. IEEE Access 8:216729–216740
Ishikawa T, Stent S (2022) Boosting supervised learning in small data regimes with conditional GAN augmentation. In: 2022 IEEE international conference on image processing (ICIP), Bordeaux, France.
Li M, Lin J, Ding Y, Liu Z, Zhu J-Y, Han S (2022) GAN compression: efficient architectures for interactive conditional GANs, In: IEEE transactions on pattern analysis and machine intelligence, 44(12): 9331–9346.
Silva FC, André Lima de Castro P, Júnior HR, Cordeiro Marujo E (2019) Mangan: assisting colorization of manga characters concept art using conditional GAN. In: 2019 IEEE international conference on image processing (ICIP), Taipei, Taiwan.
. Gu AN, Jafari MH, Luong C, Tsang T, Abolmaesumi P (2021) Echocardiogram view conversion using clinically constrained conditional GAN. In: 2021 IEEE 18th international symposium on biomedical imaging (ISBI), Nice, France.
Tao L, Li X, Fisher J, Levin CS Application of conditional adversarial networks for automatic generation of MR-based attenuation Map in PET/MR, In: 2018 IEEE nuclear science symposium and medical imaging conference proceedings (NSS/MIC), Sydney, NSW, Australia, 2018
Ann K, Jang Y, Shim H, Chang H-J (2021) Multi-scale conditional generative adversarial network for small-sized lung nodules using class activation region influence maximization. IEEE Access 9:139426–139437
Deng C, Chen Z, Wang R, Su W, Qu Y (2021) Modality-shared MRI image translation based on conditional GAN. IN: 2021 IEEE international conference on bioinformatics and biomedicine (BIBM), Houston, TX, USA.
Liao K, Lin C, Zhao Y, Gabbouj M (2020) DR-GAN: automatic radial distortion rectification using conditional GAN in real-time. IEEE Trans Circuits Syst Video Technol 30(3):725–733
Mahmud T, Billah M, Roy-Chowdhury AK (2018) Multi-view frame reconstruction with conditional GAN. In: 2018 IEEE Global Conference on Signal and Information Processing (GlobalSIP), Anaheim, CA, USA.
Vo DM, Sugimoto A, Nakayama H, PPCD-GAN: Progressive pruning and class-aware distillation for large-scale conditional gans compression, In: 2022 IEEE/CVF winter conference on applications of computer vision (WACV), Waikoloa, HI, USA, 2022.
Rodriguez Torrado R, Khalifa A, Cerny Green M, Justesen N, Risi S, Togelius J (2020) Bootstrapping Conditional GANs for Video Game Level Generation. In: 2020 IEEE conference on games (CoG), Osaka, Japan.
Yuan Y, Ito Y, Nakano K (2020) Art Font Image generation with conditional generative adversarial networks. In: 2020 eighth international symposium on computing and networking workshops (CANDARW), Naha, Japan.
Heo M, Choe Y (2019) Single-image reflection removal using conditional GANs. In: 2019 international conference on electronics, information, and communication (ICEIC), Auckland, New Zealand.
Murali S, Rajati MR, Suryadevara S (2019) Image Generation and style transfer using conditional generative adversarial networks. In: 2019 18th IEEE international conference on machine learning and applications (ICMLA), Boca Raton, FL, USA.
Song J, Zhang J, Gao L, Zhao Z, Shen HT (2022) AgeGAN++: face aging and rejuvenation with dual conditional GANs. IEEE Trans Multimedia 24:791–804
Liu X, Gao Z, Chen BM (Sept. 2020) MLFcGAN: multilevel feature fusion-based conditional gan for underwater image color correction. IEEE Geosci Remote Sens Lett 17(9):1488–1492
Li Y et al. (2019) StoryGAN: a sequential conditional GAN for story visualization. In: 2019 IEEE/CVF conference on computer vision and pattern recognition (CVPR), Long Beach, CA, USA.
Chen M, Li C, Li K, Zhang H, He X (2018) Double Encoder Conditional GAN for Facial Expression Synthesis. IN: 2018 37th Chinese Control Conference (CCC), Wuhan, China.
Nguyen V, Vicente TFY, Zhao M, Hoai M, Samaras D (2017) Shadow Detection with Conditional Generative Adversarial Networks. IN: 2017 IEEE international conference on computer vision (ICCV), Venice, Italy
Silva MC, Bianchi AGC, Ribeiro SP, Oliveira RAR (2021) Leaf shape reconstruction and damage estimation using a U-net-based conditional GAN. In: Proceedings of the 36th annual ACM symposium on applied computing (SAC '21). Association for computing machinery, New York, NY, USA.
Raipurkar P, Pal R, Raman S (2021) HDR-cGAN: single LDR to HDR image translation using conditional GAN. In: Proceedings of the twelfth indian conference on computer vision, graphics and image processing (ICVGIP '21). Association for Computing Machinery, New York, NY, USA
Zhang Y, Li Y, Zhou X, Kong X, Luo J (2020) Curb-GAN: Conditional Urban Traffic Estimation through Spatio-Temporal Generative Adversarial Networks. In: proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining (KDD '20). Association for Computing Machinery, New York, NY, USA, 842–852.
Jun-Yan Z, Taesung P, PhillP I, Alexei E (2017) Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks.
Kwon T, Ye JC (2021) Cycle-Free CycleGAN using invertible generator for unsupervised low-dose CT denoising. IEEE Trans Comput Imaging 7:1354–1368
Liu X, Meng G, Xiang S, Pan C (2018) Semantic image synthesis via conditional cycle-generative adversarial networks. In: 2018 24th international conference on pattern recognition (ICPR), Beijing, China.
Zhaoqing P, Weijie Y, Xiaokai Y, Asifullah K, Feng Y, Yuhui Z, Recent progress on generative adversarial networks (GANs): A Survey. IEEE Access. PP. 1–1. https://doi.org/10.1109/ACCESS.2019.
Huang X, Wen L, Ding J SAR and Optical image registration method based on improved CycleGAN. In: 2019 6th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Xiamen, China, 2019.
Chen H, Guan M, Li H (2021) ArCycleGAN: improved CycleGAN for style transferring of fruit images. IEEE Access 9:46776–46787
Kiranyaz S et al (Dec.2022) Blind ECG restoration by operational cycle-GANs. IEEE Trans Biomed Eng 69(12):3572–3581
Liu J et al (2022) CycleGAN Clinical Image Augmentation Based on Mask Self-Attention Mechanism. IEEE Access 10:105942–105953
Yoon HS, Park KR (2020) CycleGAN-based deblurring for gaze tracking in vehicle environments. IEEE Access 8:137418–137437
Zhang F, Gao H, Lai Y (2020) Detail-preserving CycleGAN-AdaIN framework for image-to-Ink painting translation. IEEE Access 8:132002–132011
You Q, Wan C, Sun J, Shen J, Ye H, Yu Q (2019) Fundus Image Enhancement Method Based on CycleGAN. IN: 2019 41st annual international conference of the IEEE engineering in medicine and biology society (EMBC), Berlin, Germany,.
Kim YH, Nam SH, Park KR (2021) Enhanced cycle generative adversarial network for generating face images of untrained races and ages for age estimation. IEEE Access 9:6087–6112
Zhu H, Liu D (2021) FakeSafe: human level steganography techniques by disinformation mapping using cycle-consistent adversarial network. IEEE Access 9:159364–159370
Shao X, Wei C, Shen Y, Wang Z (2021) Feature enhancement based on CycleGAN for nighttime vehicle detection. IEEE Access 9:849–859
Dou H, Chen C, Hu X, Peng S (2019) Asymmetric Cyclegan for Unpaired NIR-to-RGB Face Image Translation. In: ICASSP 2019–2019 IEEE international conference on acoustics, speech and signal processing (ICASSP), Brighton, UK.
Xiao X, Kuang P, Gu XI, He M (2019.) Pedestrian image generation with target pose based on the improved cyclegan. In: 2019 16th international computer conference on wavelet active media technology and information processing, Chengdu, China
Wang H, Zhang H, Yu L, Wang L, Yang X (2020) Facial Feature Embedded Cyclegan For Vis-Nir Translation. IN: ICASSP 2020—2020 IEEE international conference on acoustics, speech and signal processing (ICASSP), Barcelona, Spain.
Chen Y, Zhu Y, Chang Y (2020) CycleGAN based data augmentation for melanoma images classification. In proceedings of the 2020 3rd international conference on artificial intelligence and pattern recognition (AIPR '20). Association for computing machinery, New York, NY, USA, 115–119. https://doi.org/10.1145/3430199.3430217.
Hwang J, Yu C, Shin Y (2020) SAR-to-Optical image translation using SSIM and perceptual loss based cycle-consistent GAN. In: 2020 international conference on information and communication technology convergence (ICTC), Jeju, Korea (South).
Oh G, Sim B, Ye JC (2020) Unsupervised learning for compressed sensing MRI using Cyclegan. 2020 IEEE 17th international symposium on biomedical imaging (ISBI), Iowa City, IA, USA.
Cho SW, Baek NR, Koo JH, Arsalan M, Park KR (2020) Semantic segmentation with low light images by modified CycleGAN-based image enhancement. IEEE Access 8:93561–93585
Gu J, Ye JC (2021) AdaIN-based tunable CycleGAN for efficient unsupervised low-dose CT denoising. IEEE Trans Comput Imaging 7:73–85
He X, Ji W, Xie J (2022) Unsupervised haze removal for aerial imagery based on asymmetric contrastive CycleGAN. IEEE Access 10:67316–67328
Lv X, Zhang X (2019) Generating Chinese classical landscape paintings based on cycle-consistent adversarial networks. In: 2019 6th international conference on systems and informatics (ICSAI), Shanghai, China.
Sanaat A, Shiri I, Arabi H, Mainta I, Nkoulou R, Zaidi H (2020) Whole-body PET image synthesis from low-dose images using cycle-consistent generative adversarial networks, 2020 IEEE nuclear science symposium and medical imaging conference (NSS/MIC). MA, USA, Boston
Ji G, Wang Z, Zhou L, Xia Y, Zhong S, Gong S (2021) SAR Image colorization using multidomain cycle-consistency generative adversarial network. IEEE Geosci Remote Sens Lett 18(2):296–300
Parihar AS, Anand P, Sharma A (2021) Grover A UndarkGAN: low-light image enhancement with cycle-consistent adversarial networks. In: 2021 5th international conference on computer, communication and signal processing (ICCCSP), Chennai, India.
Su T, Pan F, Cao J (2021) ES-CycleGAN: An Improved CycleGAN for VI-to-IR Translation, 2021 40th chinese control conference (CCC), Shanghai, China.
Yang M, He J (2021) Image Style Transfer Based on DPN-CycleGAN. In: 2021 4th international conference on pattern recognition and artificial intelligence (PRAI), Yibin, China.
Güzel S, Yavuz S (2022) Infrared Image Generation From RGB Images Using CycleGAN. IN: 2022 international conference on innovations in intelligent systems and applications (INISTA), Biarritz, France.
Takahashi A, Koda Y, Ito K, Aoki T Domain Transformation of Fingerprint Images Using CycleGAN. In: 2019 IEEE 8th global conference on consumer electronics (GCCE), Osaka, Japan, 2019.
Lee J-H, Kim K, Kim J-H (2021) Design of CycleGAN model for SAR image colorization. In: 2021 IEEE VTS 17th Asia pacific wireless communications symposium (APWCS), Osaka, Japan.
Karras T, Laine S, Aila T (2019) A Style-Based generator architecture for generative adversarial networks. In: 2019 IEEE/CVF conference on computer vision and pattern recognition (CVPR), Long Beach, CA, USA.
Yoon D, Kwak J-G, Li Y, Han D, Ko H (2022) DIFAI: diverse facial inpainting using StyleGAN inversion. In: 2022 IEEE international conference on image processing (ICIP), Bordeaux, France.
Tewari A et al. (2020) StyleRig: rigging StyleGAN for 3D control over portrait images. In: 2020 IEEE/CVF conference on computer vision and pattern recognition (CVPR), Seattle, WA, USA.
Zhao C, Shuai R, Ma L, Liu W, Hu D, Wu M (2021) Dermoscopy image classification based on StyleGAN and DenseNet201. IEEE Access 9:8659–8679
Varkarakis V, Bazrafkan S, Corcoran P Re-training StyleGAN—a first step towards building large, scalable synthetic facial datasets. In: 2020 31st irish signals and systems conference (ISSC), Letterkenny, Ireland, 2020.
Khodadadeh S, Ghadar S, Motiian S, Lin WA, Bölöni L, Kalarot R (2022) Latent to latent: a learned mapper for identity preserving editing of multiple face attributes in StyleGAN-generated images. In: 2022 IEEE/CVF winter conference on applications of computer vision (WACV), Waikoloa, HI, USA.
Salem Hussin SH, Yildirim R (2021) StyleGAN-LSRO method for person re-identification. In: IEEE Access, vol. 9, pp. 13857–13869.
Hermosilla G, Tapia D-IH, Allende-Cid H, Castro GF, Vera E Thermal Face Generation Using StyleGAN. In: IEEE Access 9: 80511–80523.
Shi Y, Aggarwal D, Jain AK (2021) Lifting 2D StyleGAN for 3D-aware face generation. In: 2021 IEEE/CVF conference on computer vision and pattern recognition (CVPR), Nashville, TN, USA,.
Chang R, Zhang B, Zhang Y, Gao S, Zhao S, Rao Y, Zhai X, Wang T, Yang Y (2023) FFA-GAN: a generative adversarial network based on feature fusion attention for intelligent safety monitoring. Advances in Multimedia, 2023, Article ID 5588547, 13 pages, https://doi.org/10.1155/2023/5588547.
Hariharan B, Nalina KSIPSE, WBNR, Senthil Prakash PN Hybrid deep convolutional generative adversarial networks (DCGANS) and style generative adversarial network (STYLEGANS) algorithms to improve image quality. In: 2022 3rd international conference on electronics and sustainable communication systems (ICESC), Coimbatore, India, 2022.
Liao S, Arakawa K (2021) Interactive poster design system for movies with StyleGAN. In: 2021 international symposium on intelligent signal processing and communication systems (ISPACS), Hualien City, Taiwan.
Cohen Hochberg D, Greenspan H, Giryes R (2022) A self supervised StyleGAN for image annotation and classification with extremely limited labels. In: IEEE transactions on medical imaging, 41(12): 3509–3519.
Ma P, Xue T (2022) Embedding Chinese Face painting into the StyleGAN latent space. In: 2021 ninth international conference on advanced cloud and big data (CBD), Xi'an, China.
. Richardson E et al. (2021) Encoding in Style: a StyleGAN Encoder for Image-to-Image Translation. In: 2021 IEEE/CVF conference on computer vision and pattern recognition (CVPR), Nashville, TN, USA.
Tan W, Wen B, Chen C, Zeng Z, Yang X (2021) Systematic analysis of circular artifacts for stylegan. In: 2021 IEEE international conference on image processing (ICIP), Anchorage, AK, USA.
Xiaofeng S, Zhenping Q, Fei D, Shiqi S, Yuxuan Z, Libo H (2021) Pluralistic Face Image Completion Based on StyleGAN. In: 2021 IEEE 23rd Int Conf on High Performance Computing & Communications; 7th Int Conf on Data Science & Systems; 19th Int Conf on Smart City; 7th Int Conf on Dependability in Sensor, Cloud & Big Data Systems & Application (HPCC/DSS/SmartCity/DependSys), Haikou, Hainan, China.
Su W, Ye H, Chen S-Y, Gao L, Fu H (2022) DrawingInStyles: Portrait Image Generation and Editing with Spatially Conditioned StyleGAN. In: IEEE transactions on visualization and computer graphics.
Wei T et al (2022) E2Style: improve the efficiency and effectiveness of StyleGAN inversion. IEEE Trans Image Process 31:3267–3280
Choi I, Park S, Park J (2022) Generating and modifying high resolution fashion model image using StyleGAN. In: 2022 13th international conference on information and communication technology convergence (ICTC), Jeju Island, Korea, Republic of.
Rohilla R, Bansal H, Kediyal G, Dhamija A (2022) Portraits Editing using StyleGANs & Transfer Learning. In: 2022 6th international conference on computing methodologies and communication (ICCMC), Erode, India.
Christian L, Lucas T, Ferenc H, Jose C, Cunningham, Acosta A, Aitken A, Tejani A, Totz A, Wang J, Shi Z, Wenzhe (2017) Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network.
Nagano Y, Kikuta Y (2018) SRGAN for super-resolving low-resolution food images, In: Proceedings of the joint workshop on multimedia for cooking and eating activities and multimedia assisted dietary management, association for computing machinery, New York, NY, USA.
Dou H, Chen C, Hu X, Xuan Z, Hu Z, Peng S (2020) PCA-SRGAN: incremental orthogonal projection discrimination for face super-resolution. In: Proceedings of the 28th ACM international conference on multimedia (MM '20). Association for Computing Machinery, New York, NY, USA.
Jun Kim Y, Hazra D, Byun Y, Ahn K-J Old Document restoration using super resolution gan and semantic image inpainting, In: Proceedings of the International Workshop on Artificial Intelligence and Education, Association for Computing Machinery, New York, NY, USA,2019
Liu B, Chen J (2021) A super resolution algorithm based on attention mechanism and SRGAN network. IEEE Access 9:139138–139145
Cao M, Liu Z, Huang X, Shen Z, Research for face image super-resolution reconstruction based on wavelet transform and SRGAN. In: 2021 IEEE 5th advanced information technology, electronic and automation control conference (IAEAC), Chongqing, China, 2021.
Jiang X, Xu Y, Wei P, Zhou Z (2020) CT Image super resolution based on improved SRGAN. In: 2020 5th international conference on computer and communication systems (ICCCS), Shanghai, China.
Xu M-C, Yin F, Liu C-L (2020) SRR-GAN: Super-Resolution based recognition with GAN for low-resolved text images. In: 2020 17th international conference on frontiers in handwriting recognition (ICFHR), Dortmund, Germany.
Dehzangi O, Gheshlaghi SH, Amireskandari A, Nasrabadi NM, Rezai A (2021) OCT image segmentation using neural architecture search and SRGAN. In: 2020 25th international conference on pattern recognition (ICPR), Milan, Italy.
Wang L, Zheng M, Du W, Wei M, Li L (2018) Super-resolution SAR image reconstruction via generative adversarial network. In: 2018 12th international symposium on antennas, propagation and EM Theory (ISAPE), Hangzhou, China.
Adil M, Mamoon S, Zakir A, Manzoor MA, Lian Z (2020) Multi scale-adaptive super-resolution person re-identification using GAN. IEEE Access 8:177351–177362
Nguyen-Truong H, Nguyen KNA, Cao S (2020) SRGAN with total variation loss in face super-resolution. IN: 2020 7th NAFOSTED conference on information and computer science (NICS), Ho Chi Minh City, Vietnam.
Li J, Wu L, Wang S, Wu W, Song F, Zheng G (2019) Super resolution image reconstruction of textile based on SRGAN. In: 2019 IEEE international conference on smart internet of things (SmartIoT), Tianjin, China.
Huang J-H, Wang H-K, Liao Z-W (2020) HFD-SRGAN: super-resolution generative adversarial network with high-frequency discriminator. In: 2020 IEEE international conference on systems, man, and cybernetics (SMC), Toronto, ON, Canada.
Ambudkar S, Raj R, Billa K, Hukumchand R (2022) Super-Resolution for Cross-Sensor Optical Remote Sensing Images, IGARSS 2022–2022 IEEE international geoscience and remote sensing symposium. Kuala Lumpur, Malaysia
Zhang Z, Jin L, Gao T (2022) research on underwater image enhancement algorithm based on SRGAN, In: 2022 international conference on cyber-physical social intelligence (ICCSI), Nanjing, China.
Kim D-h, Lee J-w, Park S-h (2022) A Study on Model Compression Methods for SRGAN. In: 2022 International conference on electronics, information, and communication (ICEIC), Jeju, Korea, Republic of.
Fu Y, Zheng C, Yuan L, Chen H, Nie J (2021) Small object detection in complex large scale spatial image by concatenating SRGAN and Multi-Task WGAN. In: 2021 7th international conference on big data computing and communications (BigCom), Deqing, China.
Nandhini PS, Srinath P, Veeramanikandan P (2022) Detection of glaucoma using convolutional neural network (CNN) with super resolution generative adversarial network (SRGAN). In: 2022 3rd international conference on smart electronics and communication (ICOSEC), Trichy, India.
Zhang N (2022) Distant blurred pedestrian detection: based on SSD network with SRGAN image super resolution. In: 2022 3rd international conference on computer vision, image and deep learning & international conference on computer engineering and applications (CVIDL & ICCEA), Changchun, China.
Ullah M, Hamza A, Ahmad Taj I, Tahir M (2021) Low resolution face recognition using enhanced SRGAN generated images. In: 2021 16th international conference on emerging technologies (ICET), Islamabad, Pakistan.
Rakotonirina NC, Rasoanaivo A (2020) ESRGAN+ : Further Improving Enhanced Super-Resolution Generative Adversarial Network. In: ICASSP 2020—2020 IEEE international conference on acoustics, speech and signal processing (ICASSP), Barcelona, Spain.
Zheng W-F, Xie W-L (2020) A Comic Head images generation algorithm based on improved deep convolutional generative adversarial networks. IN: 2020 3rd international conference on advanced electronic materials, computers and software engineering (AEMCSE), Shenzhen, China.
Santosh KC, Ghosh S, Bose M (2021) Ret-GAN: retinal image enhancement using generative adversarial networks, 2021 IEEE 34th International Symposium on Computer-Based Medical Systems (CBMS). Aveiro, Portugal
Panetta K, Kezebou L, Oludare V, Agaian S (2022) Comprehensive underwater object tracking benchmark dataset and underwater image enhancement with GAN. IEEE J Oceanic Eng 47(1):59–75
Lee MB, Kim YH, Park KR (2019) Conditional generative adversarial network- based data augmentation for enhancement of iris recognition accuracy. IEEE Access 7:122134–122152
Zhao K, Lu T, Wang Y, Wang Y, Nie X (2020) Single image dehazing based on enhanced generative adversarial network. In: 2020 5th International conference on control, robotics and cybernetics (CRC), Wuhan, China.
Zhao B, Li W, Gong W (2019) Deep pyramid generative adversarial network with local and nonlocal similarity features for natural motion image deblurring. IEEE Access 7:185893–185907
Cai Y et al (2019) Dualattn-GAN: text to image synthesis with dual attentional generative adversarial network. IEEE Access 7:183706–183716
Kim J-H, Won CS Emotion enhancement for facial images using GAN. In: 2020 IEEE international conference on consumer electronics—Asia (ICCE-Asia), Seoul, Korea (South), 2020.
Mao Q, Wang S, Wang S, Zhang X, Ma S (2018) Enhanced image decoding via edge-preserving generative adversarial networks. In: 2018 IEEE international conference on multimedia and expo (ICME), San Diego, CA, USA.
Bhargav K, Ambika S, Deepak S, Sudha S (2020) Imagenation—a DCGAN based method for Image reconstruction from fMRI. In: 2020 Fifth international conference on research in computational intelligence and communication networks (ICRCICN), Bangalore, India.
Gu F, Zhang H, Wang C, Wu F (2019) SAR image super-resolution based on noise-free generative adversarial network. In: IGARSS 2019—2019 IEEE international geoscience and remote sensing symposium, Yokohama, Japan.
Xu B, Zhou D, Li W (2022) Image enhancement algorithm based on GAN neural network. IEEE Access 10:36766–36777
Liu Y, Qiao Y, Li Y (2021) Single image super resolution reconstruction based on the combination of residual encoding-decoding structure and GAN. In: 2021 4th international conference on artificial intelligence and pattern recognition (AIPR 2021). Association for Computing Machinery, New York, NY, USA.
Pham QTM, Shin J (2021) Generative adversarial networks for retinal image enhancement with pathological information. In: 2021 15th international conference on ubiquitous information management and communication (IMCOM), Seoul, Korea (South),.
Jammes-Floreani M, Laine AF, Angelini ED (2021) Enhanced-Quality Gan (EQ-GAN) on Lung CT Scans: Toward Truth and Potential Hallucinations. In: 2021 IEEE 18th international symposium on biomedical imaging (ISBI), Nice, France.
Zhou Y et al (2022) Speckle noise reduction for OCT images based on image style transfer and conditional GAN. IEEE J Biomed Health Inform 26(1):139–150
Wang R, Jiang B, Yang C, Li Q, Zhang B (2022) MAGAN: unsupervised low-light image enhancement guided by mixed-attention. Big Data Min Anal 5(2):110–119
Weligampola H et al. (2020) A Retinex based GAN pipeline to utilize paired and unpaired datasets for enhancing low light images. In: 2020 moratuwa engineering research conference (MERCon), Moratuwa, Sri Lanka.
Kim W et al (2018) Contrast CT image generation model using CT image of PET/CT. 2018 IEEE nuclear science symposium and medical imaging conference proceedings (NSS/MIC). NSW, Australia, Sydney
Avanaki NJ, Zadtootaghaj S, Barman N, Schmidt S, Martini MG, Möller S (2020) Quality enhancement of gaming content using generative adversarial networks. In: 2020 Twelfth international conference on quality of multimedia experience (QoMEX), Athlone, Ireland.
Lee D, Lee S, Lee H, Lee K, Lee H-J (2019) Resolution-preserving generative adversarial networks for image enhancement. IEEE Access 7:110344–110357
Hall J, Bocanegra MG, Haddad RJ, Optimizing generative adversarial networks for low-resolution image enhancement. In: 2020 SoutheastCon, Raleigh, NC, USA, 2020.
Wang Z, Hall J, Haddad RJ (2021) Improving Pneumonia diagnosis accuracy via systematic convolutional neural network-based image enhancement, SoutheastCon 2021, Atlanta, GA, USA.
Howard M, Suharjito (2020) Low resolution face recognition using generative adversarial network (GAN).
Pang Y, Xie J, Li X (2019) Visual haze removal by a unified generative adversarial network. IEEE Trans Circuits Syst Video Technol 29(11):3211–3221
Joshi OS, Joshi AD, Sawant ST (2022) Enhancing Two dimensional magnetic resonance image using generative adversarial network. In: 2022 IEEE 9th Uttar Pradesh section international conference on electrical, electronics and computer engineering (UPCON), Prayagraj, India.
Qu Y, Ou Y, Xiong R (2019) Low illumination enhancement for object detection in self-driving. In: 2019 IEEE international conference on robotics and biomimetics (ROBIO), Dali, China.
Fang C, Lv C, Cai F, Liu H, Wang J, Shuai M (2022) Low light image enhancement for color images combined with sky region segmentation. In: 2022 international conference on machine learning and knowledge engineering (MLKE), Guilin, China.
Zhang M, Jia L (2021) Chromaticity preserving contrast enhancement for low illumination images. In: 2021 3rd international conference on applied machine learning (ICAML), Changsha, China.
Kurniadi FI, Retno Dewi D, Putri VK (2019) Comparison several hue preserving image enhancement methods from under exposure image. In: 2019 5th international conference on science and technology (ICST), Yogyakarta, Indonesia.
Wang W, Wu X, Yuan X, Gao Z (2020) An experiment-based review of low-light image enhancement methods. IEEE Access 8:87884–87917
Seo J, Park H (2019) Object recognition in very low resolution images using deep collaborative learning. IEEE Access 7:134071–134082
Zamir SW et al. (2023) Learning Enriched Features for Fast Image Restoration and Enhancement. In: IEEE transactions on pattern analysis and machine intelligence, 45(2): 1934–1948.
Akita K, Hayama M, Kyutoku H, Ukita N (2021) AVM Image Quality Enhancement by Synthetic Image Learning for Supervised Deblurring. In: 2021 17th international conference on machine vision and applications (MVA), Aichi, Japan.
Abomaali M, Abosinnee AS, Malik RQ, Jaafar AA (2022) Online back-lit image enhancement based on learning restoration. In: 2022 5th international conference on engineering technology and its applications (IICETA), Al-Najaf, Iraq.
Chen L, Zhao C, Wang Y, Huang X (2022) Dust Image Enhancement Algorithm Based on Feature Transformation. In: 2022 Asia conference on algorithms, computing and machine learning (CACML), Hangzhou, China.
Mallikeswari B, Sripriya P (2019) A powerful and efficient method for enhancing images of different types. In: 2019 international conference on machine learning, big data, cloud and parallel computing (COMITCon), Faridabad, India.
Poudalov AD, Piltsov MV, Mazur VG (2018) Algorithmization of the Spatial Method of Image Enhancement Based on Aligning the Brightness Histograms. In: 2018 XIV international scientific-technical conference on actual problems of electronics instrument engineering (APEIE), Novosibirsk, Russia.
Sahu YK, Pandey C, Biswas P, Khan MR, Rathore S (2020) Minimum Time Delay and More Efficient Image Filtering Brain Tumour Detection with the help of MATLAB. In: 2020 international conference on communication and signal processing (ICCSP), Chennai, India.
Li T, Zhang H (2017) Digital image enhancement system based on MATLAB GUI. In: 2017 8th IEEE international conference on software engineering and service science (ICSESS), Beijing, China.
Grover S, Sharma A (2015) A Hybrid Method for Contrast Enhancement with Edge Preservation of Generalized Images.
Dorothy R, Joany RM, Joseph R, Rajendran PS, Susai J St (2015) Image enhancement by Histogram equalization. In: International journal of nano corrosion science and engineering.
Goel R (2021) The implementation of image enhancement techniques using matlab. In: Proceedings of the international conference on innovative computing & communication (ICICC)
Author information
Authors and Affiliations
Contributions
The authors have made equal contributions to this work, and they have collectively reviewed and approved the final manuscript.
Corresponding author
Ethics declarations
Competing Interests
This study has no conflicts of interest and is not funded by any Organization/Institution. All authors have participated in conception and design, anal-lysis and interpretation of the data, drafting the article or revising it critically for important intellectual content, and approval of the final version.
Ethical and Informed Consent for Data Used
Not Applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Nayak, A.A., Venugopala, P.S. & Ashwini, B. A Systematic Review on Generative Adversarial Network (GAN): Challenges and Future Directions. Arch Computat Methods Eng (2024). https://doi.org/10.1007/s11831-024-10119-1
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11831-024-10119-1