Underwater Image Enhancement Using GAN
Underwater Image Enhancement Using GAN
Abstract:- The process of enhancing the distorted In this project, we investigate the use of GANs for
underwater images to clear image is known as enhanc- ing underwater images. Our goal is to develop a
Underwater image enhance- ment. Distorted images are GAN-based method that can produce high-quality enhanced
the raw underwater images that taken from the deep images from a dataset of degraded underwater images. The
portion of ocean, river etc by using different cameras. In proposed method involves training a GAN on a dataset of
general underwater images are mainly used in degraded and high- quality underwater images to learn the
underwater robotics, ocean pasture and environmental statistical properties of the data. In this paper [2] Evaluate
monitor- ing, ocean exploration etc. The underwater the effectiveness of our method by comparing the quality of
image enhancement process is done by using underwater the generated images with the original degraded images and
image dataset which includes the distorted images (raw with other state-of- the-art methods. We will also analyze the
underwater images) and the corre- sponding enhanced performance of our method in terms of its ability to preserve
underwater images. Currently used image enhancement important features and details in the images.
methods cannot provide sufficient satisfaction to the
underwater image enhancement. So proposed a new The contributions of this project are two fold. First, we
method by using Generative Adversarial Network (GAN), pro- pose a GAN-based method for enhancing underwater
which tries to produce more images from the dataset. images that can produce high-quality images with improved
contrast, color, and sharpness. Second, we present an
Keywords:- Generative Adversarial Network(GAN), Under- evaluation of the effectiveness of our method in comparison
Water Image Enhancement. to other state-of- the-art methods. The results of this project
have the potential to advance the field of underwater imaging
I. INTRODUCTION and provide a valuable tool for underwater exploration and
research.
Underwater imaging is an important task for various
appli- cations, including marine biology, oceanography, and Table I: Abbreviations and Acronyms
under- water exploration. However, the quality of underwater
images is often degraded due to the effects of water on GAN : Generative adversarial network
light, such as scattering, absorption, and attenuation. This UW-GAN : Underwater generative
degradation can result in poor contrast, color distortion, and adversarial network
blurring, which can make it difficult to extract useful UWC-Net : Underwater coarse-level
information from the images. In this paper [1] traditional generative network
image enhancement methods have limited effectiveness in UWF-Net : Underwater fine-level network
improving the quality of underwater images. In Figure 1.1 CNN : Convolutional neural network
shows that the general overview of the underwater image PAM : Parallel attention module
enhancement. Here we can see that the left image set is the ALM : Adaptive learning module
distorted (poor quality) underwater images and the right
EUVP : Enhancing underwater visual
image set is the corresponding enhanced (good quality)
perception
underwater images.
SSIM : Structural similiarity
PSNR : Peak signal-to-noise ratio
In recent years, deep learning methods, particularly
Genera- tive Adversarial Networks (GANs), have shown
great promise in image enhancement tasks. GANs are a type II. LITERATURE SURVEY
of neural network that can learn to generate realistic images
by training a generator network to produce images that are A. Underwater Image Enhancement
The image enhancement is done to improve an image
similar to a dataset of real images, and a discriminator
suitability for a given task, such as making it more pleasing
network to differentiate between real and generated images.
GANs have been successfully applied to tasks such as super- to the eye. To put it another way, the main goal of image
resolution, and style transfer. enhance- ment is to modify a given image so that the finished
product is better suited than the original image for a given
application. In this paper [2] propose a conditional generative
adversarial network-based technique for improving real time
underwater images. Similarly, [3] present an improved
multiscale dense generative adversarial network for the
process of underwater image enhancement. Create an difficulties, [7] offer LANet, an adaptive learning attention
objective function that monitors the adversarial training on the network for underwater image enhancement based on
dataset and examines the global content, colour, local texture, supervised learning. A multiscale fu- sion module is first
and style information to assess the perceived image quality. suggested to merge various spatial data. Then, create a
Additionally, propose a sizable dataset with paired and brandnew parallel attention module (PAM) that combines
unpaired underwater image collections of ( ”bad” and pixel and channel attention with a focus on lighted elements
”excellent” quality) collected with seven distinct cameras and more important colour information. The shallow
under a range of visibility conditions during mar- itime trips information can then be retained by an adaptive learning
and experiments using human-robot collaboration [EUVP]. module (ALM), which can then learn crucial feature
Perform Both qualitative and quantitative analyses show that information.
The suggested model can improve underwater image quality
through paired and unpaired training. The effects of absorption and scattering frequently result
in the degradation of underwater image quality. When utilised
The task of predicting underwater single-image depth is for analysis and display, degraded underwater photos have
challenging because there aren’t any large-scale datasets of some restrictions. For instance, low contrast and colour cast
underwater depth image data and the difficulties aren’t well- in underwater photos reduce accuracy of marine biological
posed. An precise depth estimate is needed for a single recog- nition rate and underwater item detection rate. A
underwater image in applications like marine engineering contrast en- hancement algorithm an underwater image
and underwater robots. In this paper [4] from an underwater dehazing algorithm [8] are part of a systematic underwater
single image, propose an end to-end underwater generative image enhancement technique that is proposed to get around
adversarial network (UW-GAN) for depth estimation . those restrictions. The vision, colour, and a natural undersea
Firstly, by using using the underwater coarse-level generative image appearance are restored using an effective underwater
net- work (UWC-Net) a coarse-level depth map is estimated. image dehazing approach. The lowest information loss
The estimated coarse-level depth map and the input picture principle serves as its foundation. An effective contrast
are concatenated as input into the underwater fine-level enhancement technique is suggested that enhances the
network (UWF- Net), which computes a fine-level depth contrast and brightness of underwater images based on a
map. Also, suggest a method for creating synthetic histogram distribution prior type. Two improved output
underwater images for huge databases. For the performance versions can be produced using the suggested procedure. One
analysis of the proposed network, both real-world and variant, ideal for display, has fairly authentic colours and a
artificial underwater datasets are used. natural appearance. The other version, which has greater
brightness and contrast, can be utilised to glean more
The major causes of the quality decline in photographs important data and reveal more specifics.
taken under cloudy settings are 1) various meteorological
factors and 2) the attenuation of reflected light. These Underwater image improvement has received a lot of in-
elements significantly alter the hue and visibility of the terest recently in both underwater vision and image pro-
acquired photos. To tackle these problems, [5] suggest an cessing. Enhancement of underwater images is a tough task
end-to-end trainable image de-hazing network [LIGHT-Net]. because, the complex underwater environment and lighting
Haze reduction and colour constancy modules make up the circumstances. Underwater images are typically wavelength-
proposed LIGHT- Net. The colour constancy module among dependent absorption causes degradation and scattering [9],
them removes the colour tint that the weather supplied to the including backward scattering and forward scattering.
hazy image. The suggested haze reduction module, which
uses an inception- residual block as its building block, aims B. Underwater Object Tracking
to lessen the impact of haze as well as to enhance visibility Applications including deep ocean exploration,
in the foggy image. In contrast to conventional feature underwater robot navigation, marine life monitoring [10], and
concatenation, the haze reduction module proposed in this homeland and maritime security all rely heavily on
research uses dense feature sharing to efficiently distribute underwater object tracking. These applications call for
the features discovered in the network’s first layers. effective and precise vision-based underwater marine
analytics, including methods for image augmentation, picture
The underwater image typically has low contrast, colour quality evaluation, and target tracking. Understanding marine
distortion, and fuzzy features because of the light’s atten- image/video analytics faces huge difficulties due to the
uation and dispersion in the water. By taking into account excessive noise and poor light con- ditions. Because of this,
underwater imaging specifics, an unique two-stage computer vision tasks like detection, recognition, and
underwater image convolutional neural network (CNN) based tracking are substantially more difficult in underwater
on structural decomposition (UWCNN-SD) [6] for situations than in open-air environments.
underwater picture en- hancement is suggested.
In this paper [5] numerous sophisticated tracking
Due to light’s dispersion and absorption as it moves methods have been presented as a result of the availability of
through water, underwater photographs have colour casts and big annotated datasets. Depending on how tracking features
inade- quate illumination. Tasks requiring vision of are gathered and applied, the methodologies used in the
underwater, such as detection and recognition, may be majority of these trackers can be divided into different groups.
hampered by these issues. To address these degradation KCF, HCF, DCF, CFNet, STRCF, BACF, and fDSST are
examples of trackers that fall within the category of kernelized feature-based tech- nique. The foundation of older trackers
correlation filters. Modern trackers including MDNet, like Struck is local and global feature extraction.
SiamFC, CCOT, and ECO have primarily used the CNN
However, GANs also have their limitations, such as the Figure 3 shows the generator network that consist
need for large amounts of high-quality training data and the convolu- tional layer, batch normalization layer, Leaky
diffi- culty of training stable and robust models. Nevertheless, ReLU, decon- volutional layer and tanh function. This
GANs are a powerful and promising deep learning technique convolutional layer is used for the feature extractionby using
that has shown impressive results in a wide range of kernal or filter. Batch normalization is used for to standardize
applications. each input and output of each layer. Leaky ReLU is an
activation function. Tanh function is an activation function
B. Framework Overview that produce [0, 1] result.
The new underwater images (duplicate) are generated
us- ing a Generative Adversarial Network (GAN). The Here is a General Overview of How a Generator in a GAN
proposed framework includes a generator and Works:
discriminator, as seen in Figure 2. This architecture shows
that the input taken by the generator from the dataset and Input: The generator receives random noise vectors as
generate it’s duplicates of images. After that the input, which are typically drawn from a simple probabil-
discriminator takes the output of generator as it’s input for the ity distribution like a Gaussian distribution. These noise
discrimination of whether the image is real or fake. The output vectors are low-dimensional representations that capture
of discriminator indicates that true or false (1 or 0) the variability of the data.
corresponding to real or fake respectively.
C. Generator Network
There are many modules for feature extraction that have
been created. In the paper [12] includes the popular architec-
ture of inception looks for the structure of optimal local sparse
inside a structure of network. Even so, at the conclusion of
the block, these various scale characteristics concatenate in a
straightforward manner, contributing in part to the underuti-
lization of feature maps. Following the U-Net principles [13],
our generator network has been modified from UGAN. The
encoder and decoder networks make up this system.
Transformation: The generator takes the input noise vec- D. Discriminator Network
tors and applies a series of mathematical transformations Figure 4 shows the network of discriminator. The pro-
through its neural network layers. These layers gradually posed discriminator network includes four layers. It consist
increase the complexity and dimensionality of the input of convolutional layer, batch normalization, Leaky ReLU and
to produce a higher-dimensional output. sigmoid function. Convolutional layer is used for the feature
Output: The final layer of the generator typically em- extraction by using the kernal or filter map. And the batch
ploys an activation function, such as a sigmoid or a normalization is used for to standardize each input and output
hyperbolic tangent, to squash the output values into the of each layer. The Leaky ReLU is an activation function. The
desired range. The generated output is meant to resemble sigmoid activation function is the last layer of discriminator
the real data samples from the training set. which predict the output as true or false.
Training: During the training phase of the GAN, the
generator’s objective is to generate synthetic samples that Here is a General Overview of How a Discriminator in a
can deceive the discriminator, another component of the GAN Works:
GAN architecture (discussed further below). The
generator’s weights are updated based on the feedback it Input: The discriminator takes as input either real data
receives from the discriminator. samples or generated data samples from the generator.
Table 2: Quantitative Comparison with Euvp Dataset used metrics in the quantitative evaluation are SSIM and
Model PSNR SSIM PSNR. Table II shows the both PSNR and SSIM values
FUnIE-GAN 21.92 0.88 obtained from different methods. Based on this analysis, it is
FUnIE-GAN-UP 21.36 0.81 obvious that our method does not demonstrates good
UGAN 19.59 0.66 performance compared to the other previously used methods,
Ours 23.92 0.92 as evidenced by the PSNR and SSIM value.
Where, µx(µy) denotes the mean, and σ2 (σ2) is The PSNR approximates the reconstruction quality of a
the variance of x(y); whereas σxy denotes the cross-correlation gen- erated image x compared to its ground truth y based on
between x and y. Additionally, c1 = (255 × 0.01)2 and c2 their Mean Squared Error (MSE). PSNR is expressed in
= (255 ×0.03)2 are constants that ensure numeric stability. decibels (dB) and provides a quantitative assessment of image
fidelity. Higher PSNR values indicate lower perceptual
SSIM takes into account both low-level pixel-wise differences between the generated and reference images,
compar- isons and higher-level structural comparisons. It implying higher image quality.
considers how well the local structures, textures, and edges of
the generated image match those of the ground truth (real) VI. CONCLUSION
image. Higher SSIM scores indicate higher visual similarity.
In this paper focused on exploring the application of
PSNR: PSNR [16] is a metric commonly used to mea- Gen- erative Adversarial Networks (GANs) for underwater
sure the quality of reconstructed or generated images by image enhancement. The objective was to improve the visual
comparing them to a reference image. It measures the quality and clarity of underwater images, which often suffer
ratio of the peak power of the signal [17] (the maximum from poor visibility, color distortion, and low contrast.
possible value) to the noise power (the difference between Through the implementation and experimentation with GAN-
the generated and reference images). based ap- proaches, significant advancements have been
made in under- water image enhancement.