A_Perception-Aware_Decomposition_and_Fusion_Framework_for_Underwater_Image_Enhancement
A_Perception-Aware_Decomposition_and_Fusion_Framework_for_Underwater_Image_Enhancement
3, MARCH 2023
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
KANG et al.: PERCEPTION-AWARE DECOMPOSITION AND FUSION FRAMEWORK FOR UIE 989
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
990 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 33, NO. 3, MARCH 2023
Fig. 2. Flowchart of our proposed SPDF-based UIE approach. Given an input underwater image I , SPDF first involves a pre-processing stage to produce a
contrast-corrected version Ictr and a detail-sharpened version Isharp of the white-balanced image Iwb . Then, SPD is applied on Ictr and Isharp separately to
obtain their corresponding three conceptually independent components: mean intensity (L), contrast (C), and structure (S). Afterwards, the mean intensity (L),
contrast (C), and structure (S) are fused separately with different schemes. Finally, the fused mean intensity, contrast, and structure, are used to reconstruct
an enhanced underwater image Iout as output.
achieve a systematic visual quality improvement of underwa- on each of the input separately, producing three conceptually
ter images; 3) We demonstrate the superiority of SPDF by independent components: mean intensity (L), contrast (C), and
comparing it with 11 state-of-the-art UIE algorithms including structure (S) for each input. Afterwards, the corresponding
seven traditional and four most recently deep learning-based components of the two inputs are respectively fused with
approaches on two benchmark datasets. In addition, the effec- different strategies by considering the properties of human
tiveness of SPDF is also verified on another two relevant tasks, visual system and the characteristics of underwater image
i.e., low-light image enhancement (LIE) and single image degradations. Concretely, the mean intensity component L
dehazing (SID). which mainly accounts for the low-frequency information is
The rest of this paper is organized as follows. Section II fused with weights determined by a statistical prior derived
presents the proposed SPDF-based UIE approach. Section III from high-quality underwater images; the structure component
presents the experimental results and qualitative and quantita- S which mainly accounts for the high-frequency information
tivel performance comparisons. Finally, conclusion are drawn is fused with a Laplacian pyramid scheme to effectively
in Section IV. eliminate potential artifacts due to the sharp transitions caused
by noise or other unwanted high-frequency artifacts in the
II. P ROPOSED SPDF A PPROACH two inputs; and the contrast component C which mainly
A. Algorithm Overview accounts for the local contrast perception is fused with a
simple maximum operation. Since the SPD is completely
The flow chart of our proposed SPDF approach is depicted
invertible, the three components after fusion are naturally
in Fig. 2. Our framework is built on a fusion pipeline where
integrated to reconstruct a final enhanced underwater image
two inputs are derived by correcting the contrast and sharpen-
with appealing visual quality. In what follows, we will describe
ing the details of a white-balanced version of a single under-
the proposed SPDF approach with details.
water image, respectively. The white balance is to compensate
for the color casts caused by wavelength-dependent absorp-
tion of colors in water medium. The two initially enhanced B. Pre-Proccessing
versions (i.e., contrast-corrected and detail-sharpened) of the Unlike terrestrial (in-air) images, underwater images usually
white-balanced image generally contain complementary infor- suffer from serious color deviations (the hue is biased toward
mation that is useful for producing a high-quality underwater blue or green) due to special imaging and light propagation
image. conditions. Such color deviations seriously affect the visual
By taking the contrast-corrected and detail-sharpened quality of underwater images. Therefore, the pre-processing
images as two inputs, an invertible SPD operation is performed stage first involves a white-balance operation which targets
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
KANG et al.: PERCEPTION-AWARE DECOMPOSITION AND FUSION FRAMEWORK FOR UIE 991
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
992 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 33, NO. 3, MARCH 2023
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
KANG et al.: PERCEPTION-AWARE DECOMPOSITION AND FUSION FRAMEWORK FOR UIE 993
Fig. 5. Illustration of the pixel value intensity distributions of high-quality underwater images. From left to right: red channel, green channel, and blue
channel.
Fig. 6. Pipeline of the multi-scale fusion scheme where the the number of pyramid levels equals to three.
2) Fusion of Contrast: We then deal with the fusion of However, a naive weighted combination of Sctr and Sshar p
contrast, i.e., Cctr and Cshar p . Generally, the visibility of a using the contrast components Cctr and Cshar p as weights may
local image patch is highly related with the magnitude of easily result in unnatural artifacts such as halos in the fused
contrast, i.e., a higher contrast results in a better visibility. structure component. To alleviate this problem, we adopt a
Considering that the two input images may have different multi-scale fusion strategy where each structure component
contrasts, the one that has the higher contrast between them Sκ is decomposed into a Laplacian pyramid while the contrast
would provide better visibility. Based on this observation, the component Cκ is decomposed using a Gaussian pyramid. Both
expected contrast of the fused image patch is determined by pyramids have the same number of levels, and the mixing of
the higher contrast between Cctr and Cshar p : the Laplacian inputs with the Gaussian contrast components
is performed independently at each level l:
Ĉ = max Cctr , Cshar p , (11)
Ŝl = Gl {Ĉctr · 1} · Ll {Sctr }+Gl {Ĉshar p · 1} · Ll {Sshar p },
where max{A, B} represents the selection of the larger value (12)
between A and B. Note that the above processes are applied Cctr
to the three color channels separately. Ĉctr = , (13)
Cctr + Cshar p
3) Fusion of Structure: We finally deal with the fusion of Cctr
structure, i.e., Sctr and Sshar p . Since the structure component Ĉshar p = , (14)
Cctr + Cshar p
Sκ mainly accounts for high-frequency information such as
sharp edges and details, the weights for fusing the structure where Gl and Ll represent the l-th level of the Gaussian
components should be closely related to the visibility of high- and Laplacian pyramid, respectively. The final fused structure
frequency information. As we have stated before, the decom- component Ŝ is obtained by summing Ŝl at all L levels, after
posed contrast component is a good indicator of visibility. appropriate upsampling:
Thus, the contrast component Cκ serves as another input when
fusing the structure component Sκ . Ŝ =↑ ↑ ↑ (Ŝ L ) + Ŝ L−1 + · · · + Ŝ1 , (15)
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
994 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 33, NO. 3, MARCH 2023
Algorithm 1 SPDF-Based UIE Approach UIE algorithms on two benchmark datasets. Finally, we extend
Input: Original underwater image I our proposed SPDF pipeline to another two relevant tasks, i.e.,
1: Perform white balancing to obtain the white-balanced LIE and SID.
image Iwb
2: Perform contrast correction on Iwb to obtain the A. Datasets
contrast-corrected image Ictr We test different UIE approaches on two datasets:
3: Perform detail sharpening on Iwb to obtain the UIEB [20] and RUIE [32]. The UIEB dataset contains
detail-sharpened image Ishar p 890 real-world underwater images. The images suffer from
4: for each image patch P in I do diverse degradations including greenish/blueish color casts,
5: Extract its co-located patches Pctr and Pshar p from Ictr different degrees of contrast reduction, and different degrees
and Ishar p , respectively of water turbidity. In addition, there are a large number of
6: Perform SPD on Pctr and Pshar p separately to obtain organisms and objects in the underwater images. Therefore,
{Lctr , Cctr , Sctr } and {Lshar p , Cshar p , Sshar p } the UIEB dataset is suitable for verifying the robustness
7: Fuse Lctr and Lshar p , Cctr and Cshar p , Sctr and Sshar p , of UIE approaches. In order to further verify the effec-
to obtain L̂, Ĉ, Ŝ, respectively tiveness of our proposed SPDF approach, we also use the
8: Reconstruct the fused patch by inverting the decomposi- RUIE dataset for testing. The real-world underwater images
tion: P̂ = L̂ + Ĉ · Ŝ in RUIE are different from those in UIEB as it usually
9: end for contains few or vey small-scale objects, and also the under-
10: Aggregate the fused patches into Iout water creature is more severe, making it a good comple-
Output: Enhanced underwater image Iout mentary benchmark dataset for performance comparison of
UIE methods.
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
KANG et al.: PERCEPTION-AWARE DECOMPOSITION AND FUSION FRAMEWORK FOR UIE 995
Fig. 7. Visual comparisons on a typical real underwater image with obvious color deviation, reduced contrast, and detail loss. The compared methods
are Rayleigh [8], RGHS [9], Fusion [10], UDCP [11], BL-TM [12], Haze-Line [13], Bayesian [14], UWCNN [19], Water-Net [20], Two-Branch [21], and
TOPAL [33].
effectively removes the blueish tone and improves the contrast Haze-Line [13], and Bayesian [14]. Among them, Fusion [10]
and detail visibility without obvious extra artifacts and the and Water-Net [20] achieve relatively better color correction
over-enhancement issue. Although the Fusion [10] method also performance but they still have some other quality defects.
applied a fusion framework to enhance underwater images, For example, the details in some local dark regions of the
the details in some dark regions are still under-enhanced results by Fusion [10] are not visible and the color tones of
and the visibility is also unsatisfactory, which demonstrate the results by Water-Net [20] are still not quite realistic as
the advantage of our specially designed SPD method for they seem to become slightly blueish especially in the far-
fusing the mean intensity, contrast, and structure components away regions. By contrast, our proposed SPDF approach is
separately. able to effectively remove the color deviations and improve
We then show the enhanced results of several underwater the visibility of details in some local dark regions.
images sampled from the UIEB dataset in the left side of All the visual comparisons demonstrate that our SPDF not
Fig. 8. The sampled underwater images are either with obvious only produces visually pleasing results but also generalizes
color deviation or poor visibility due to low contrast and well to different underwater scenes.
blurry details, as shown in the first row. As presented, some
of the compared UIE methods even introduce artificial colors,
D. Quantitative Comparisons
such as Rayleigh [8] (the third image), UWCNN [19] (the
second and fifth images), and Two-Branch [21] (the forth 1) Quantitative Evaluation Metrics: We employ three
image). In terms of color deviation removal, most compared different no-reference underwater image quality evaluation
methods fail to recover realistic color appearance, such as metrics, i.e., NUIQ [34], UCIQE [35], and UIQM [36],
RGHS [9], BL-TM [12], Haze-Line [13], Water-Net [20], Two- to quantitatively compare different UIE methods. A higher
Branch [21], and TOPAL [33]. In addition, some of the com- NUIQ, UCIQE or UIQM score indicates a better visual quality.
pared methods also suffer from the under-/over-enhancement Note that all existing underwater image quality evaluation
and over-saturation problems, such as Rayleigh [8] (the first, metrics are not sufficiently accurate, i.e., the scores of NUIQ,
second, and fifth images), Fusion [10] (the second image), UCIQE, and UIQM cannot accurately reflect the visual quality
UDCP [11] (the first, third, and fifth images), BL-TM [12] of enhanced underwater images in some cases. In our study,
(the second and fifth images), Bayesian [14] (the first and we only provide the scores of NUIQ, UCIQE, and UIQM as
third images), UWCNN [19] (the first, third, and fifth images), the reference for the following research. In addition, we also
Water-Net [20] (the first image), Two-Branch [21] (the third provide the scores of BRISQUE [37] as the reference though
image), and TOPAL [33] (the first image). By contrast, the it was not originally devised for underwater images. A lower
proposed SPDF approach not only recovers more realistic BRISQUE score indicates a better image quality. Overall,
color but also more effectively enhances contrast and details, we use four metrics including NUIQ, UCIQE, UIQM, and
which are credited to the fusion pipeline and the specially BRISQUE, to evaluate the visual quality of different results.
designed SPD for separate fusion and reconstruction. The quantitative comparisons of different UIE algorithms
We finally show the the enhanced results of several under- in terms of NUIQ, UCIQE, UIQM, and BRISQUE, are
water images sampled from the RUIE database in the right side shown in Table I. From this table, we have the following
of Fig. 8. As presented, the input underwater images suffer observations. First, the proposed SPDF approach has the best
from serious color deviations (either greenish aor blueish) and NUIQ, UIQM, and BRISQUE scores on the RUIE dataset and
low contrast. The compared methods cannot well enhance the best NUIQ and BRISQUE scores on the UIEB dataset.
these images since the greenish and blueish tones are still It also ranks the second place in terms of UIQM on the
preserved in some enhanced results, such as RGHS [9], UIEB dataset. These results indicate that our proposed SPDF
UDCP [11], BL-TM [12], Haze-Line [13], Bayesian [14], approach performs generally well on different datasets and
Two-Branch [12], and TOPAL [33]. Even worse, some in terms of different metrics. Second, the Rayleigh method
compared methods introduce artificial colors that affect the has achieved the highest UCIQE values on both datasets.
human perception, such as Rayleigh [8], UWCNN [19], The reason is that UCIQE tends to produce higher scores
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
996 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 33, NO. 3, MARCH 2023
Fig. 8. Visual comparisons on several real underwater images sampled from the UIEB and RUIE datasets. The compared methods are Rayleigh [8], RGHS [9],
Fusion [10], UDCP [11], BL-TM [12], Haze-Line [13], Bayesian [14], UWCNN [19], Water-Net [20], Two-Branch [21], and TOPAL [33].
TABLE I
Q UANTITATIVE C OMPARISONS IN T ERMS OF NUIQ, UCIQE, UIQM, AND BRISQUE. T HE B EST
P ERFORMER I S H IGHLIGHTED IN R ED U NDER E ACH C ASE
for the underwater images with sufficient red channel infor- of underwater images is not well addressed in UCIQE. Some
mation. As we can observe from Fig. 7 and Fig. 8, the examples are given in Fig. 9. As presented, although the results
results enhanced by the Rayleigh method are easily subject of BL-TM and Rayleigh are much worse than the results
to color artifacts with excessive reddish tone. Although the obtained by our proposed SPDF method, their UCIQE scores
attenuation of the red channel is the heaviest among the are much higher, which accordingly demonstrates the inaccu-
three channels and the compensation of red channel is of rate prediction of UCIQE. Overall, the superiority of SPDF
great importance for UIE especially for color casts removal, has been well verified by different quantitative evaluation
a proper treatment of the red channel for quality evaluation metrics.
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
KANG et al.: PERCEPTION-AWARE DECOMPOSITION AND FUSION FRAMEWORK FOR UIE 997
Fig. 11. Percentage of positive votes (i.e., “+1”) for each UIE algorithm out
of the total number of votes (including both “+1” and “−1”) that this UIE
algorithm is involved.
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
998 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 33, NO. 3, MARCH 2023
Fig. 14. Running times and NUIQ scores with different parameter values.
(a) Patch sizes and (b) Pyramid levels.
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
KANG et al.: PERCEPTION-AWARE DECOMPOSITION AND FUSION FRAMEWORK FOR UIE 999
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
1000 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 33, NO. 3, MARCH 2023
Fig. 17. Visual results enhanced by different LIE methods. The compared methods from left to right are RRD-Net [41], GLAD-Net [42], Retinex-Net [43],
KIND [44], RUAS [45], RCTNet [46], UTVNet [47], and the proposed SPDF.
Fig. 18. Visual results enhanced by different SID methods. The compared methods are DCP [51], CAP [52], AOD-Net [53], Dehaze-Net [54], NLD [55],
NLBF [56], and USID-Net [57].
Note that we do not intend to rigorously demonstrate the performed separately without any interactions and information
state-of-the-art performance of our SPDF in LIE and SID, loss. Comprehensive qualitative and quantitative comparisons
but try to deliver the message that our SPDF pipeline has a on two benchmark datasets have demonstrated the superior-
good applicability and potential usage in some other relevant ity of SPDF against several state-of-the-art UIE algorithms.
applications. A simple extention to the LIE and SID tasks also verified the
IV. C ONCLUSION good applicability of SPDF and its potential usage in some
other relevant applications.
This paper has presented an novel UIE method based on
SPDF. The key insight is to enhance the raw underwater image
R EFERENCES
by fusing two complementary images derived from the input in
a perception-aware and conceptually independent image space. [1] S. Anwar and C. Li, “Diving deeper into underwater image enhance-
ment: A survey,” Signal Process., Image Commun., vol. 89, Nov. 2020,
Specifically, we perform SPD to represent each to-be-fused Art. no. 115978.
image with mean intensity, contrast, and structure, and then [2] M. Yang, J. Hu, C. Li, G. Rohde, Y. Du, and K. Hu, “An in-depth survey
fuse each component with different schemes by considering of underwater image enhancement and restoration,” IEEE Access, vol. 7,
pp. 123638–123657, 2019.
the properties of human visual system and the characteristics [3] C. S. Tan, G. Seet, A. Sluzek, and D. He, “A novel application of range-
of underwater degradations. The main advantage of SPDF is gated underwater laser imaging system (ULIS) in near-target turbid
that we perform the fusion of two complementary images in medium,” Opt. Lasers Eng., vol. 43, no. 9, pp. 995–1009, Sep. 2005.
[4] Z. Murez, T. Treibitz, R. Ramamoorthi, and D. J. Kriegman, “Photo-
a perception-consistent and conceptually independent image metric stereo in a scattering medium,” in Proc. IEEE Int. Conf. Comput.
space and the fusions of different components can be Vis. (ICCV), Oct. 2015, pp. 3415–3423.
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
KANG et al.: PERCEPTION-AWARE DECOMPOSITION AND FUSION FRAMEWORK FOR UIE 1001
[5] C. Ancuti, C. O. Ancuti, T. Haber, and P. Bekaert, “Enhancing under- [28] D. Akkaynak and T. Treibitz, “A revised underwater image formation
water images and videos by fusion,” in Proc. IEEE Conf. Comput. Vis. model,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit.,
Pattern Recognit., Jun. 2012, pp. 81–88. Jun. 2018, pp. 6723–6732.
[6] X. Fu, P. Zhuang, Y. Huang, Y. Liao, X.-P. Zhang, and X. Ding, [29] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image
“A retinex-based enhancing approach for single underwater image,” in quality assessment: From error visibility to structural similarity,” IEEE
Proc. IEEE Int. Conf. Image Process. (ICIP), Oct. 2014, pp. 4572–4576. Trans. Image Process., vol. 13, no. 4, pp. 600–612, Apr. 2004.
[7] X. Fu, Z. Fan, M. Ling, Y. Huang, and X. Ding, “Two-step approach [30] Y. Fang, K. Ma, Z. Wang, W. Lin, Z. Fang, and G. Zhai, “No-reference
for single underwater image enhancement,” in Proc. Int. Symp. Intell. quality assessment of contrast-distorted images based on natural scene
Signal Process. Commun. Syst. (ISPACS), Nov. 2017, pp. 789–794. statistics,” IEEE Signal Process. Lett., vol. 22, no. 7, pp. 838–842,
[8] A. S. A. Ghani and N. A. M. Isa, “Underwater image quality enhance- Jul. 2015.
ment through composition of dual-intensity images and Rayleigh- [31] A. S. A. Ghani and N. A. M. Isa, “Underwater image quality enhance-
stretching,” in Proc. IEEE 4th Int. Conf. Consum. Electron. Berlin ment through integrated color model with Rayleigh distribution,” Appl.
(ICCE-Berlin), Sep. 2014, pp. 219–220. Soft Comput., vol. 27, pp. 219–230, Feb. 2015.
[9] D. Huang, W. Yan, S. Wei, J. Sequeira, and S. Mavromatis, “Shallow- [32] R. Liu, X. Fan, M. Zhu, M. Hou, and Z. Luo, “Real-world underwater
water image enhancement using relative global histogram stretching enhancement: Challenges, benchmarks, and solutions under natural
based on adaptive parameter acquisition,” in MultiMedia Modeling. light,” IEEE Trans. Circuits Syst. Video Technol., vol. 30, no. 12,
Cham, Switzerland: Springer, 2018. pp. 4861–4875, Dec. 2020.
[10] C. O. Ancuti, C. Ancuti, C. De Vleeschouwer, and P. Bekaert, “Color [33] Z. Jiang, Z. Li, S. Yang, X. Fan, and R. Liu, “Target oriented perceptual
balance and fusion for underwater image enhancement,” IEEE Trans. adversarial fusion network for underwater image enhancement,” IEEE
Image Process., vol. 27, no. 1, pp. 379–393, Jan. 2018. Trans. Circuits Syst. Video Technol., early access, May 13, 2022, doi:
[11] P. L. J. Drews, Jr., E. R. Nascimento, S. S. C. Botelho, and 10.1109/TCSVT.2022.3174817.
M. F. M. Campos, “Underwater depth estimation and image restoration [34] Q. Jiang, Y. Gu, C. Li, R. Cong, and F. Shao, “Underwater image
based on single images,” IEEE Comput. Graph. Appl., vol. 36, no. 2, enhancement quality evaluation: Benchmark dataset and objective
pp. 24–35, Mar./Apr. 2016. metric,” IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 9,
[12] W. Song, Y. Wang, D. Huang, A. Liotta, and C. Perra, “Enhancement pp. 5959–5974, Sep. 2022.
of underwater images with statistical model of background light and [35] M. Yang and A. Sowmya, “An underwater color image quality evaluation
optimization of transmission map,” IEEE Trans. Broadcast., vol. 66, metric,” IEEE Trans. Image Process., vol. 24, no. 12, pp. 6062–6071,
no. 1, pp. 153–169, Mar. 2020. Dec. 2015.
[13] D. Berman, D. Levy, S. Avidan, and T. Treibitz, “Underwater single [36] K. Panetta, C. Gao, and S. Agaian, “Human-visual-system-inspired
image color restoration using haze-lines and a new quantitative dataset,” underwater image quality measures,” IEEE J. Ocean. Eng., vol. 41, no. 3,
IEEE Trans. Pattern Anal. Mach. Intell., vol. 43, no. 8, pp. 2822–2837, pp. 541–551, Jul. 2015.
Aug. 2021. [37] A. Mittal, A. K. Moorthy, and A. C. Bovik, “No-reference image quality
[14] P. Zhuang, C. Li, and J. Wu, “Bayesian retinex underwater assessment in the spatial domain,” IEEE Trans. Image Process., vol. 21,
image enhancement,” Eng. Appl. Artif. Intell., vol. 101, May 2021, no. 12, pp. 4695–4708, Dec. 2012.
Art. no. 104171. [38] Methodology for the Subjective Assessment of the Quality of Television
[15] C.-Y. Li, J.-C. Guo, R.-M. Cong, Y.-W. Pang, and B. Wang, “Underwater Pictures, document ITU-R BT.500-14, 2019.
image enhancement by dehazing with minimum information loss and [39] J.-S. Lee, “On designing paired comparison experiments for subjective
histogram distribution prior,” IEEE Trans. Image Process., vol. 25, multimedia quality assessment,” IEEE Trans. Multimedia, vol. 16, no. 2,
no. 12, pp. 5664–5677, Dec. 2016. pp. 564–571, Feb. 2014.
[16] Y. Wang, H. Liu, and L.-P. Chau, “Single underwater image restoration [40] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look
using adaptive attenuation-curve prior,” IEEE Trans. Circuits Syst. I, once: Unified, real-time object detection,” in Proc. IEEE Conf. Comput.
Reg. Papers, vol. 65, no. 3, pp. 992–1002, Mar. 2018. Vis. Pattern Recognit. (CVPR), Jun. 2016, pp. 779–788.
[17] Y.-T. Peng and P. C. Cosman, “Underwater image restoration based on [41] A. Zhu, L. Zhang, Y. Shen, Y. Ma, S. Zhao, and Y. Zhou, “Zero-shot
image blurriness and light absorption,” IEEE Trans. Image Process., restoration of underexposed images via robust retinex decomposition,”
vol. 26, no. 4, pp. 1579–1594, Apr. 2017. in Proc. IEEE Int. Conf. Multimedia Expo (ICME), Jul. 2020, pp. 1–6.
[18] Z. Liang, X. Ding, Y. Wang, X. Yan, and X. Fu, “GUDCP: Generaliza- [42] W. Wang, C. Wei, W. Yang, and J. Liu, “GLADNet: Low-light enhance-
tion of underwater dark channel prior for underwater image restoration,” ment network with global awareness,” in Proc. 13th IEEE Int. Conf.
IEEE Trans. Circuits Syst. Video Technol., vol. 32, no. 7, pp. 4879–4884, Autom. Face Gesture Recognit. (FG), May 2018, pp. 751–755.
Jul. 2022. [43] C. Wei, W. Wang, W. Yang, and J. Liu, “Deep retinex decomposition
[19] C. Li, S. Anwar, and F. Porikli, “Underwater scene prior inspired deep for low-light enhancement,” 2018, arXiv:1808.04560.
underwater image and video enhancement,” Pattern Recognit., vol. 98, [44] Y. Zhang, J. Zhang, and X. Guo, “Kindling the darkness: A practical
Feb. 2020, Art. no. 107038. low-light image enhancer,” in Proc. 27th ACM Int. Conf. Multimedia,
[20] C. Li et al., “An underwater image enhancement benchmark dataset and 2019, pp. 1632–1640.
beyond,” IEEE Trans. Image Process., vol. 29, pp. 4376–4389, 2020. [45] R. Liu, L. Ma, J. Zhang, X. Fan, and Z. Luo, “Retinex-inspired
[21] J. Hu, Q. Jiang, R. Cong, W. Gao, and F. Shao, “Two-branch deep neural unrolling with cooperative prior architecture search for low-light image
network for underwater image enhancement in HSV color space,” IEEE enhancement,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit.
Signal Process. Lett., vol. 28, pp. 2152–2156, 2021. (CVPR), Jun. 2021, pp. 10561–10570.
[22] J. Li, K. A. Skinner, R. M. Eustice, and M. Johnson-Roberson, [46] H. Kim, S.-M. Choi, C.-S. Kim, and Y. J. Koh, “Representative color
“WaterGAN: Unsupervised generative network to enable real-time color transform for image enhancement,” in Proc. IEEE/CVF Int. Conf.
correction of monocular underwater images,” IEEE Robot. Autom. Lett., Comput. Vis. (ICCV), Oct. 2021, pp. 4459–4468.
vol. 3, no. 1, pp. 387–394, Jan. 2018. [47] C. Zheng, D. Shi, and W. Shi, “Adaptive unfolding total variation
[23] Y. Guo, H. Li, and P. Zhuang, “Underwater image enhancement using a network for low-light image enhancement,” in Proc. IEEE/CVF Int.
multiscale dense generative adversarial network,” IEEE J. Ocean. Eng., Conf. Comput. Vis. (ICCV), Oct. 2021, pp. 4439–4448.
vol. 45, no. 3, pp. 862–870, Jul. 2020. [48] Z. Ying, G. Li, and W. Gao, “A bio-inspired multi-exposure fusion
[24] X. Ye et al., “Deep joint depth estimation and color correction from framework for low-light image enhancement,” 2017, arXiv:1711.00591.
monocular underwater images based on unsupervised adaptation net- [49] M. D. Grossberg and S. K. Nayar, “Modeling the space of camera
works,” IEEE Trans. Circuits Syst. Video Technol., vol. 30, no. 11, response functions,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 26,
pp. 3995–4008, Nov. 2019. no. 10, pp. 1272–1282, Oct. 2004.
[25] X. Fu and X. Cao, “Underwater image enhancement with global– [50] J. Liang et al., “Recurrent exposure generation for low-light face
local networks and compressed-histogram equalization,” Signal Process., detection,” IEEE Trans. Multimedia, vol. 24, pp. 1609–1621, 2022.
Image Commun., vol. 86, Aug. 2020, Art. no. 115892. [51] K. He, J. Sun, and X. Tang, “Single image haze removal using dark
[26] C. Li, S. Anwar, J. Hou, R. Cong, C. Guo, and W. Ren, “Underwater channel prior,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 33, no. 12,
image enhancement via medium transmission-guided multi-color space pp. 2341–2353, Dec. 2011.
embedding,” IEEE Trans. Image Process., vol. 30, pp. 4985–5000, 2021. [52] Q. Zhu, J. Mai, and L. Shao, “A fast single image haze removal algorithm
[27] B. McGlamery, “A computer model for underwater camera systems,” using color attenuation prior,” IEEE Trans. Image Process., vol. 24,
Proc. SPIE, vol. 208, pp. 221–231, Mar. 1979. no. 11, pp. 3522–3533, Nov. 2015.
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.
1002 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 33, NO. 3, MARCH 2023
[53] B. Li, X. Peng, Z. Wang, J. Xu, and D. Feng, “AOD-Net: All-in- Chongyi Li (Member, IEEE) received the Ph.D.
one dehazing network,” in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), degree from the School of Electrical and Informa-
Oct. 2017, pp. 4780–4788. tion Engineering, Tianjin University, Tianjin, China,
[54] B. L. Cai, X. M. Xu, K. Jia, C. M. Qing, and D. C. Tao, “DehazeNet: in June 2018. From 2016 to 2017, he was a Joint-
An end-to-end system for single image haze removal,” IEEE Trans. Training Ph.D. Student with The Australian National
Image Process., vol. 25, no. 11, pp. 5187–5198, Aug. 2016. University, Australia. He is currently a Research
[55] D. Berman, T. Treibitz, and S. Avidan, “Non-local image dehazing,” in Assistant Professor with the School of Computer
Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2016, Science and Engineering, Nanyang Technological
pp. 1674–1682. University, Singapore. His current research interests
[56] S. C. Raikwar and S. Tapaswi, “Lower bound on transmission using include image processing, computer vision, and deep
non-linear bounding function in single image dehazing,” IEEE Trans. learning.
Image Process., vol. 29, pp. 4832–4847, 2020.
[57] J. Li, Y. Li, L. Zhuo, L. Kuang, and T. Yu, “USID-Net: Unsu-
pervised single image dehazing network via disentangled represen-
tations,” IEEE Trans. Multimedia, early access, Mar. 30, 2022, doi:
10.1109/TMM.2022.3163554. Wenqi Ren (Member, IEEE) received the Ph.D.
[58] J.-L. Yin, Y.-C. Huang, B.-H. Chen, and S.-Z. Ye, “Color transferred degree from Tianjin University, Tianjin, China,
convolutional neural networks for image dehazing,” IEEE Trans. Circuits in 2017. From 2015 to 2016, he was supported by
Syst. Video Technol., vol. 30, no. 11, pp. 3957–3967, Nov. 2020. the China Scholarship Council and working with
[59] S. Zhang, Y. Wu, Y. Zhao, Z. Cheng, and W. Ren, “Color-constrained Prof. Ming-Husan Yang as a Joint-Training Ph.D.
dehazing model,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recog- Student with the Electrical Engineering and Com-
nit. Workshops (CVPRW), Jun. 2020, pp. 3799–3807. puter Science Department, University of California
[60] A. Dudhane, K. M. Biradar, P. W. Patil, P. Hambarde, and S. Murala, at Merced. He is currently an Associate Professor
“Varicolored image de-hazing,” in Proc. IEEE/CVF Conf. Comput. Vis. with the School of Cyber Science and Technology,
Pattern Recognit. (CVPR), Jun. 2020, pp. 4563–4572. Sun Yat-sen University, Shenzhen Campus, Shen-
[61] C. O. Ancuti, C. Ancuti, R. Timofte, and C. D. Vleeschouwer, “I-HAZE: zhen, China. His research interests include image
A dehazing benchmark with real hazy and haze-free indoor images,” in processing and related high-level vision problems.
Proc. Int. Conf. Adv. Concepts Intell. Vis. Syst., 2018, pp. 620–631.
[62] C. O. Ancuti, C. Ancuti, R. Timofte, and C. De Vleeschouwer,
“O-HAZE: A dehazing benchmark with real hazy and haze-free outdoor
images,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit.
Workshops (CVPRW), Jun. 2018, pp. 754–762. Hantao Liu (Senior Member, IEEE) received the
Ph.D. degree from the Delft University of Technol-
ogy, Delft, The Netherlands, in 2011. He is currently
an Associate Professor with the School of Computer
Science and Informatics, Cardiff University, Cardiff,
Yaozu Kang received the bachelor’s degree in U.K. He is the Chair of the IEEE Multimedia Com-
communication engineering from Ningbo Univer- munications Technical Committee, Interest Group on
sity, Ningbo, China, in 2019, where he is currently Quality of Experience for Multimedia Communica-
pursuing the master’s degree. His research inter- tions. He serves as an Associate Editor for IEEE
ests include underwater image processing and deep T RANSACTIONS ON C IRCUITS AND S YSTEMS
learning. FOR V IDEO T ECHNOLOGY and IEEE S IGNAL
P ROCESSING L ETTERS .
Authorized licensed use limited to: Tsinghua University. Downloaded on January 16,2024 at 12:59:32 UTC from IEEE Xplore. Restrictions apply.