Underwater Image Enhancement
Underwater Image Enhancement
net/publication/51899044
CITATIONS READS
1,120 7,158
2 authors:
All content following this page was uploaded by John Y. Chiang on 04 January 2014.
A. Underwater Model
The background light in an underwater image can be used
light attenuation [10], employing histogram equalization in both
to approximate the true in-scattering term in the full radiative
RGB and HSI color spaces to balance the luminance distribu-
transport equation to achieve the following simplified hazy
tions of color [11], and dynamically mixing the illumination of
image formation model [14], [15]:
an object in a distance-dependent way by using a controllable
multicolor light source to compensate color loss [12]. Despite
red, green, blue
the improved color balance, these methods are ineffective in re-
(1)
moving the image blurriness caused by light scattering. A sys-
tematic approach is needed to take all the factors concerning where is a point in the underwater scene, is the image
light scattering, color change, and possible presence of artificial captured by the camera, is the scene radiance at point
light source into consideration. , is the residual energy ratio of after reflecting
The algorithm for wavelength compensation and image de- from point in the underwater scene and reaching the camera,
hazing (WCID) proposed in this paper combines techniques is the homogeneous background light, and is the light
of WCID to remove distortions caused by light scattering and wavelength. Note that the residual energy ratio is a
color change. Dark-channel prior [13], an existing scene-depth function of both wavelength and the object–camera distance
derivation method, is used first to estimate the distances of the . summarizes the overall effects for both light scat-
scene objects to the camera. The low intensities in the dark tering and color change suffered by light with wavelength
channel are mainly due to three factors: 1) shadows, e.g., the traveling the underwater distance . The direct attenuation
shadows of creatures, plankton, plants, or rocks in seabed im- term describes the decay of scene radiance in the
ages; 2) colorful objects or surfaces, e.g., green plants, red or water [16]. The residual energy ratio can be represented
yellow sands, and colorful rocks/minerals, deficient in certain alternatively as the energy of a light beam with wavelength
color channels; and 3) dark objects or surfaces, e.g., dark crea- before and after traveling distance within the water
tures and stone [8]. Based on the depth map derived, the fore- and , respectively, as follows:
ground and background areas within the image are segmented.
The light intensities of foreground and background are then Nrer (2)
compared to determine whether an artificial light source is em-
ployed during the image acquiring process. If an artificial light
source is detected, the luminance introduced by the auxiliary where the normalized residual energy ratio Nrer corresponds
lighting is removed from the foreground area to avoid overcom- to the ratio of residual to initial energy for every unit of distance
pensation in the stages followed. Next, the dehazing algorithm propagated and is the medium extinction coefficient [15].
and wavelength compensation are utilized to remove the haze The normalized residual energy ratio Nrer depends on the
effect and color change along the underwater propagation path light wavelength transmitted [17], as illustrated in Fig. 3, where
to the camera. The residual energy ratio among different color red light possesses longer wavelength and lower frequency and
channels in the background light is employed to estimate the thereby attenuates faster than the blue counterpart. This results
water depth within an underwater scene. Energy compensation in the bluish tone prevalent in underwater images [18].
for each color channel is carried out subsequently to adjust the Other than the wavelength of light transmitted, the normal-
bluish tone to a natural color. With WCID, expensive optical in- ized residual energy ratio Nrer is also affected by water
struments or stereo image pairs are no longer required. WCID salinity and concentration of phytoplankton [17]. In light of this
can effectively enhance visibility and restore the color balance observation, oceanic water is further classified into three cate-
of underwater images, rendering high visual clarity and color gories. Type-I waters represent extremely clear oceanic waters.
fidelity. Most clear coastal waters with a higher level of attenuation
1758 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 4, APRIL 2012
Nrer
belong to the Type-II class. Turbid upwelling coastal waters are red, green, blue (5)
listed as Type III. Water types I, II, and III roughly correspond
to oligo-, meso-, and eutrophic waters [19]. For every meter where the background light represents the part of the object
of ocean type I that a light beam passes through, the values of reflected light and ambient light scattered toward the
normalized residual energy ratio Nrer in red (700 m) light, camera by particles in the water.
green (520 m) light, and blue (440 m) light are 82%, 95%, The background light increases as the object is placed
and 97.5%, respectively. Based on the water type considered, farther away from the camera [13], [20]–[23]. Alternatively, the
the normalized residual energy ratio Nrer can be adjusted residual energy ratio in the above equation can be rep-
based on that of Ocean Type I as follows: resented in terms of normalized residual energy ratio Nrer
using (2) as follows:
if m red ,
Nrer Nrer
Nrer if m green ,
if m blue . Nrer red, green, blue (6)
(3)
Equation (6) incorporates light scattering during the course
of propagation from object to the camera , and the wave-
II. UNDERWATER IMAGE FORMATION MODEL
length attenuation along both the surface–object path
The proposed WCID algorithm proceeds in a direction in- and object–camera route , Once the scene depth, i.e., ob-
verse to the underwater image formation path discussed above, ject–camera distance , is known through the dark-channel
as depicted in Fig. 4. First, consider the possible presence and prior, the value of residual energy ratio Nrer after wave-
influence of the artificial light source . Next, remove the light length attenuation can be calculated; thus, the direct attenuation
scattering and color change that occurred along the course of term Nrer is derivable through a
propagation from the object to the camera. Finally, com- dehazing procedure. The surface–object distance is cal-
pensate the disparities of wavelength attenuation for traversing culated by comparing the residual energy ratio of different color
the water depth to the top of the image and fine-tune the en- channels. Given the water depth , the amount of reflecting
ergy loss by deriving a more precise depth value for every point light , i.e., free of light scattering and color
within an image. change, from point illuminated by airlight is determined.
Fig. 2 illustrates an underwater image formation model. Ho- Moreover, the artificial light source is often provided to
mogeneous skylight entering above into the water is the major overcome insufficient lighting commonly encountered in an un-
source of illumination in an underwater environment. Incident derwater photographic environment. The luminance contributed
light traverses from the surface of water reaching the image by the artificial light source has to be removed before the de-
scene, covering a range from depth through , where hazing and wavelength compensation operations to avoid over-
corresponds to the image depth range. During the course of compensation. When the artificial light source is detected, the
propagation, light with different wavelengths is subjected to light emitted has first to travel distance before reaching
varying degrees of attenuation. The color change of ambient point . The residual energy after the course of propagation is
lighting makes the underwater environment tinted with a bluish Nrer . The total amount of light impinges on point
hue. As airlight incident from air to the water reaches the un- is therefore the summation of ambient lighting and the
derwater scene point with depth , i.e., attenuated artificial light Nrer . The total amount of
CHIANG AND CHEN: UNDERWATER IMAGE ENHANCEMENT BY WAVELENGTH COMPENSATION AND DEHAZING 1759
incident light Nrer Nrer is re- If point belongs to a part of the foreground object, the value
flected with reflectivity and bounces back distance of the dark channel is very small, i.e., . Taking the
before reaching the camera. During both forward and backward min operation in the local patch on the hazy image
courses of propagation pertinent to , color change occurs. in (7), we have
Accordingly, (6) can be further modified as the hazy image for-
mation equation Nrer
Fig. 6. (a) Depth map obtained after refining with image matting. Blowups of
(b) Frame I and (c) Frame II. When compared with Fig. 5(b) and (c), the depth
Fig. 5. (a) Depth map obtained by estimating , which is the distance be- map after refinement reduces the mosaic effect and captures the contours of
tween the object and the camera using dark-channel prior. Blowups of (b) Frame objects more accurately.
I and (c) Frame II. Visible mosaic artifacts are observed due to the block-based
operation of dark-channel prior.
hazy image . The brightest pixel value among all local minima
corresponds to the background light as follows:
(16)
foreground if
area - type (18)
background if
Nrer Nrer
Fig. 10. Underwater image obtained after the removal of the artificial light is
red, green, blue (19) shown in the right panel of the split screen, whereas the original one is in the left
panel. The influence of an artificial light depends on the object–camera distance.
The closer the object, the more significant the impact is.
Fig. 9 shows the distribution of the luminance of an artificial
light source present and the reflectance of red, green, and blue
channels in Fig. 7.
After deriving the luminance contributed by the artificial light The split screens of Fig. 10(a) and (b) show the results (see
source and reflectivity , red, green, blue , at right panel) after eliminating the artificial lighting detected in
point , the influence caused by the artificial lighting can be Figs. 1 and 7 (see left panel), respectively. Due to the size of
removed by subtraction from (7) as follows: scene area covered, the amount of artificial light received in
Fig. 7 is more concentrated and larger than that of Fig. 1.
Nrer Nrer C. Compensation of Light Scattering and Color Change Along
Nrer Nrer the Object–Camera Path
After removing the artificial light source and deriving dis-
Nrer red, green, blue tance between an object and the camera, the haze can be
(20) removed by subtracting the in-scattering term
1762 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 4, APRIL 2012
Fig. 12. Underwater image after removing light scattering and color change by
considering the artificial light source , the object–camera distance , and
the depth from the water surface to top of image . As the depths of the top and
bottom of the image are different, visible color change distortion still exists at
the lower portion of the image.
Fig. 13. Underwater image obtained after processing with WCID. The depth
of the underwater scene ranges from 5 m at the top of the image to 8 m at the
bottom portion.
Fig. 15. (a) The color board, serving as a ground truth, of the image taken
above water in Fig. 14(a). Color boards obtained after processing Fig. 14(b) with
(b) WCID, (c) dark-channel-based dehazing algorithm, (d) chromatism-based
dehazing algorithm, and (e) histogram equalization. The dotted white rectangles
correspond to the effective color areas for computing SNR values, as listed in
Table I.
TABLE I
QUANTITATIVE PERFORMANCE EVALUATION: SNR VALUES FOR THE
Fig. 14. Diver holding a board with six patches of designated colors, (a) before EFFECTIVE COLOR AREAS IN FIG. 15(b)–(e) (WATER DEPTH 5 m) AND
diving, diving at a depth of (b) 5 m, and (c) 15 m. FIG. 16(a)–(d) (WATER DEPTH 10 m) BY EMPLOYING THE EFFECTIVE COLOR
AREA IN FIG. 15(a) AS THE GROUND-TRUTH IMAGE ARE LISTED
Once the underwater depths for all pixels are obtained, the re- Other than taking the artificial light source , the ob-
stored energy of the underwater image after haze removal ject–camera distance , and the water depth into
and calibration of color change in (24) can be modified to better consideration, Fig. 13 demonstrates the result of fine tuning
compensate the energy loss during light propagation along dif- the amount of wavelength compensation by deriving the water
ferent water depths by dividing (22) with Nrer instead depth of every image pixel , where the Nrer values
of Nrer used for red, green, and blue light are 82%, 95%, and 97.5%,
respectively. In comparison with Fig. 12, the color change
suffered at the lower part of the image is greatly reduced. Color
Nrer balance is restored for the whole image, rather than just the top
red, green, blue (26) portion of the frame.
1764 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 4, APRIL 2012
Fig. 17. Image obtained after processing with the (a) WCID. Fig. 13 is included here for ease of comparison. (b) Dark-channel-based dehazing algorithm. (c) Chro-
matism-based dehazing algorithm. (d) Histogram equalization.
Fig. 18. (a) Image obtained after processing with WCID. The underwater depth of the photographic scene ranges from 17 to 18 m. The brightness emitted by
the artificial light source is estimated to be 100. (b) Image obtained after processing with the dark-channel-based algorithm, and (c) chromatism-based dehazing
algorithm, respectively. (d) Image obtained by histogram equalization, haze effects, and color change still exist.
III. EXPERIMENTAL RESULTS color patches and videos downloaded from Youtube website.
Both results demonstrate superior haze removing and color bal-
The performance of the proposed WCID algorithm is evalu- ancing capabilities of the proposed WCID over traditional de-
ated both objectively and subjectively by utilizing ground-truth hazing and histogram equalization methods.
CHIANG AND CHEN: UNDERWATER IMAGE ENHANCEMENT BY WAVELENGTH COMPENSATION AND DEHAZING 1765
Fig. 19. (a) Underwater image extracted from the Youtube website [26]. (b) Image obtained after processing with WCID. The underwater depth of the photo-
graphic scene ranges from 18 to 22 m. (c)–(e) Image obtained after processing with dark-channel-based and chromatism-based dehazing algorithms and histogram
equalization, respectively.
A. Objective Performance Evaluation match those of the ground-truth image. The areas occluded by
the diver’s fingers and the lower portion of the color board,
Pictures taken for a diver holding a board with six patches which is not within the scope of the image taken in Fig. 14(b),
of designated colors before diving (ground truth), and are exluded. The dotted white rectangles in Figs. 15 and 16
diving at a depth of 5 and 15 m, respectively, are shown correspond to the effective color areas employed in deriving
in Fig. 14(a)–(c). The colors processed by the WCID proposed, the SNR values. As shown in Table I, the WCID proposed
by dark-channel dehazing [8], by chromatism-based dehazing obtains the highest SNR value in both depths. In addition, the
[24], and histogram equalization algorithms [11] are compared performance of WCID is the most robust through different
with the ground-truth ones taken above water, as shown in water depths due to the incorporation of wavelength compen-
Figs. 15(a)–(e) and 16(a)–;(d). Due to the difficulty in main- sation, image dehazing, and artificial lighting removal. On the
taining a consistent spatial relationship between the color board other hand, the SNR value obtained by histogram equalization
and the camera, size and orientation variances exist between drops significantly as the depth increases. This is due to the fact
pictures taken underwater. In calculating the SNR values, the that the dynamic range of the red color component decreases
size and orientation of the color board are first adjusted to substantially with the depth due to wavelength attenuation. An
1766 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 4, APRIL 2012
Fig. 20. (a) Underwater shipwreck image extracted from the Youtube website [27]. (b) Image obtained after processing with WCID. The underwater depth of
the photographic scene ranges from 3 to 8 m. (c)–(e) Image obtained after processing with dark-channel-based and chromatism-based dehazing algorithms and
histogram equalization, respectively.
attempt, without taking dehazing or wavelength compensa- dark-channel-based algorithm and a chromatism-based de-
tion into consideration, blindly expanding the range of color hazing algorithm, respectively. Although the contrast of the
components will adversely accompany with a huge amount of image is increased, the color change appears even more promi-
error. The relatively low value of the SNR can be attributed to nent as the attenuated energy is not compensated individually
the presence of water bubbles, the misalignment in matching based on different wavelengths. Fig. 17(d) illustrates the image
color patches with different dimensions and orientations, and after histogram equlaization processing, from which the haze
water permeated through the seam between the green color effect and color change still remain. In comparison with results
patch and the board, making it distinct from the counterpart in processed by these conventional techniques, the image pro-
the ground-truth image. cessed by the WCID method, as shown in Fig. 13, is effectively
haze free and color balanced. We give the image the original
B. Subjective Performance Evaluation color and clarity that it would have had if it were not taken
Fig. 1 is extracted from an underwater video on the Youtube underwater.
website filmed by the Bubble Vision Company. [26]. This Fig. 7 is another image from the same video sequence used
AVI-formatted video is 350 s long with a resolution of 720p. to demonstrate the effect of the presence of a strong artificial
Fig. 17(b) and (c) shows the result after processing with a light source. The result after WCID processing, i.e., Fig. 18(a),
CHIANG AND CHEN: UNDERWATER IMAGE ENHANCEMENT BY WAVELENGTH COMPENSATION AND DEHAZING 1767
Fig. 21. (a) Relatively large white shiny regions are marked by the red lines. (b) The marked regions are mistakenly classified as background by the dark-channel
prior. (c) Depth map refinement through graph-cut-based -expansion. Most of the misjudged areas are corrected.
is superior to those by the dehazing algorithms or histogram reasonable estimate of the actual value. In addition, a calibra-
equalization, as shown in Fig. 18(b)–(d), in terms of both color tion procedure might be performed first by drivers before an
balance and clarity. image-capturing session by taking a test picture at known water
Two more sets of test images and processing results are depth and underwater propagation distance to fine-tune the
shown in Figs. 19 and 20 to facilitate further comparison. rate of light energy loss. In addition, the artificial lighting is
Fig. 19(a) is included for its variety in terms of subjects’ scene assumed to be a point source emitting uniform omnidirectional
depth, encompassing close to far range, whereas Fig. 20(a) is light beams across all wavelength spectrums. This is different
included for the uniqueness of subject, i.e., a shipwreck, on the from the linear or surface light source with a strong beam
sea floor. A consistent dehazing effect and color balance are directionality and nonuniform color characteristic commonly
obtained throught the processing of the WCID proposed. More
encountered in underwater photography. The precise estima-
test videos can be downloaded from ftp://140.117.168.28.
tion of the luminance distribution of the light source is also
IV. CONCLUSION demanding. If the geographic location in taking the underwater
footage and the characteristics of the light source employed are
The WCID algorithm proposed in this paper can effectively
known a priori, even better results in haze removal and color
restore image color balance and remove haze. To the best
balance can be reached.
of our knowledge, no existing techniques can handle light
scattering and color change distortions suffered by underwater Another source for causing compensation errors is the esti-
images simultaneously. The experimental results demonstrate mation of the scene depth by the dark-channel prior, as
superior haze removing and color balancing capabilities of commonly encountered in depth derivation by utilizing a single
the proposed WCID over traditional dehazing and histogram image. Relatively large white shiny regions of a foreground
equalization methods. However, the salinity and the amount of object might be misjudged as far away ones. For example, in
suspended particles in ocean water vary with time, location, Fig. 21(a), the area marked by red lines are mistaken as back-
and season, making accurate measurement of the rate of light ground, as shown in Fig. 21(b). Two approaches are suggested
energy loss Nrer difficult. Errors in the rate of light energy to alleviate the above situation [25]. One is enlarging the size of
loss will affect the precision of both the water depth and the local patch formulated in (9). Another is depth refine-
the underwater propagation distance derived. Constant ment by utilizing spatial and temporal correlation within and be-
monitoring and long-term tabulation of the rate of light energy tween video frames. The depth map refined through the graph-
loss according to time, location, and season might provide a cut-based -expansion method [25] is shown in Fig. 21(c).
1768 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 4, APRIL 2012
Nrer Nrer
Nrer
Nrer Nrer
Nrer