A survey 2019
A survey 2019
PII: S1566-2535(17)30758-3
DOI: https://doi.org/10.1016/j.inffus.2018.07.010
Reference: INFFUS 1002
Please cite this article as: Bikash Meher , Sanjay Agrawal , Rutuparna Panda , Ajith Abraham ,
A survey on region based image fusion methods , Information Fusion (2018), doi:
https://doi.org/10.1016/j.inffus.2018.07.010
This is a PDF file of an unedited manuscript that has been accepted for publication. As a service
to our customers we are providing this early version of the manuscript. The manuscript will undergo
copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please
note that during the production process errors may be discovered which could affect the content, and
all legal disclaimers that apply to the journal pertain.
ACCEPTED MANUSCRIPT
Research Highlights
A state-of-the-art survey on region based image fusion is presented.
A first hand classification of region based fusion methods is fostered.
A comprehensive comparison among the existing methods is highlighted.
A detailed analysis with encouraging results is presented.
This survey may attract researchers to explore the domain of region based fusion.
T
IP
CR
US
AN
M
ED
PT
CE
AC
1
ACCEPTED MANUSCRIPT
1
Bikash Meher, 1Sanjay Agrawal, 1Rutuparna Panda, 2Ajith Abraham
1
Department of Electronics & Telecommunication Engineering
VSS University of Technology, Burla, India
Email-id: [email protected]; [email protected]
T
2
Machine Intelligence Research Labs, Washington, USA
IP
Email-id: [email protected]
CR
Corresponding author: Rutuparna Panda, email-id: [email protected]
US
AN
Abstract
Image fusion has been emerging as an important area of research. It has attracted
many applications such as surveillance, photography, medical diagnosis, etc. Image fusion
M
techniques are developed at three levels: pixel, feature and decision. Region based image
fusion is one of the methods of feature level. It possesses certain advantages – less sensitive
ED
to noise, more robust and avoids misregistration. This paper presents a review of region based
fusion approaches. A first hand classification of region based fusion methods is carried out. A
PT
This may attract researchers to further explore the research in this direction.
2
ACCEPTED MANUSCRIPT
1. Introduction
Image fusion is the method of merging information from many images of the same
scene taken from various sensors, different positions or different time. The fused image
retains all the complementary and redundant information of the input images that are very
useful for human visual perception and image processing task. The objective of image fusion
is to fuse the details of the important information extracted from the two or multiple images.
In order to meet these objectives, the fused result should meet the following requirements: (a)
T
the fused image should retain the most complementary and significant information of the
IP
input images, (b) the fusion technique should not generate any synthetic information which
may divert the human observer or the advance image processing application, (c) it must avoid
CR
imperfect states, for instance, misregistration and noise [1]. It is observed from the literature
that image fusion approaches are classified into two types, spatial based and transform based.
n 1
On the other hand, in transform based techniques, the input images to be combined
are transformed from the space domain to some other domain by applying appropriate
ED
transforms such as wavelets or pyramids. A suitable fusion rule is utilized to fuse the
transformed images. The inverse transform is utilized to reconstruct the original image. The
PT
fused image is represented mathematically as I F T 1 T I1 , T I 2 ,..., T I N , where T
1
CE
3
ACCEPTED MANUSCRIPT
Complementary Information
Sensor 1 Sensor 2
Redundant Information
T
Fig. 1. Image fusion example.
IP
As discussed above, the fused image will retain both the complementary and the redundant
information from the input images.
CR
On the basis of levels of abstraction, image fusion algorithms are applied at three
stages: pixel, feature and decision level. In pixel level, images are combined directly using
US
individual pixels to form the fusion decision. A comprehensive survey on pixel level image
fusion is found in [2-6]. The feature level image fusion is achieved by region based fusion
AN
scheme [7-9]. In region based image fusion, initially an image is partitioned into a set of
regions. The different features of these regions are extracted. The appropriate features from
all the source images are merged to get the fused image. They are less responsive to noise and
M
more robust. Further, the feature information is used for considering more intelligent
semantic fusion rules. A first hand survey on region based image fusion is presented in
ED
Section 2. Decision level image fusion techniques are based on the outputs of initial object
detection and classification task [10-12]. Usually, a preliminary decision from the feature
PT
based image fusion serves as the input to the decision level fusion. A generic classification of
image fusion methods is shown in Fig. 2.
CE
Image
Fusion
Methods
AC
4
ACCEPTED MANUSCRIPT
together for better analysis and diagnosis of the diseases. Similarly, in the remote sensing
applications, multispectral (MS) images, which have low resolution and high spectral density
are fused with panchromatic (PAN) images possessing high resolution and low spectral
density, to obtain the spectral contents of the MS image with high spatial resolution. In
surveillance applications, the different images (infrared, visible, near infrared) are taken from
different sensors and fused for detection or for night vision. In photography applications, the
multifocus images, multi-exposure images, etc. are fused to get the image which is better
perceived to the human vision and computer processing. Some of the examples of the
different image fusion applications are illustrated in Fig. 3. The methods used are described
T
in [13-16]. The following methods are used in Fig. 3 – a) deep convolutional neural network
IP
based fusion [13], b) optimum spectrum mask based fusion [14], c) non-subsampled
CR
contourlet transform and intensity-hue-saturation transform based fusion [15], and d) fuzzy
integral based fusion [16]. Some of the applications of image fusion are carried out in real
time. This has inspired the researchers to develop more effective techniques for image fusion.
US
The number of articles published has increased rapidly since 2010 [5].
AN
M
(a)
AC
(b)
5
ACCEPTED MANUSCRIPT
IR Visible Fused image
(c)
T
A survey on different image fusion methods is done by many researchers. A
IP
classification of image fusion methods based on multi-scale decomposition technique is
described in [17]. A review of image fusion techniques in remote sensing is presented in [18-
CR
20]. Solanky and Katiyar [18] focussed on pixel based fusion methods in remote sensing.
Ghassemian [19] has carried out the survey on pixel, feature, and decision level with more
US
emphasis on pixel level method. Vivone et al. [20] presented a comparison among different
remote sensing image fusion algorithms, particularly focussing on multiresolution analysis
AN
and component substitution methods. Wang et al. [21] presented a review of image fusion
methods based on pulse coupled neural network (PCNN). An overview of multimodal
medical image fusion is described in [22,23]. The authors in [22] have used different image
M
rules and compared the results using different image quality assessment parameters.
Most of these methods describe the image fusion techniques based on pixels.
PT
However, it has several limitations: (i) blurring effects, (ii) high sensitivity to noise and (iii)
misregistration. These limitations may be reduced by deploying the region based image
CE
fusion techniques. These techniques have the capability to utilize intelligent fusion rules. As
far as our knowledge is concerned, a survey on region based image fusion methods is not
AC
available in the literature. This has motivated us to carry out a survey on different region
based image fusion methods. In this context, we present a survey on image fusion methods
based on the regions. Mainly, the paper focusses on different approaches and ideas of the
existing region based fusion practices. The results of the different techniques are summarized.
An elaborate discussion is presented at the end, comparing the various fusion methods. This
may clearly set a path for more investigations into region based fusion techniques. This may
open doors to new unsolved problems in the domain of image fusion.
6
ACCEPTED MANUSCRIPT
The remainder of the manuscript has been structured as follows. Section 2 explains
the different region based image fusion methods. Section 3 presents the performance
evaluation parameters. An elaborate discussion on a comparison of various methods is
presented in Section 4. Finally, Section 5 draws the conclusion giving a brief summary and
critique of the findings.
It is observed from the literature that the feature level image fusion technique can be
further classified into machine learning, region based and similarity matching to content
T
based. In the machine learning method, the features are extracted and a suitable classifier is
IP
used for fusion. In region based method, the input images are divided into different regions
using some segmentation techniques. The features are extracted from the regions and suitable
CR
fusion rules are used to get the fused image. The similarity matching to content based image
retrieval technique uses the visual contents of an image, for instance, shape, texture, colour,
US
and spatial layout to denote the index. The relevant indices are combined to get the fused
image. A first hand classification of region based image fusion methods is proposed here. It is
AN
classified as shown in Fig. 4.
Feature Level
M
ED
Similarity matching to
Machine Learning Region Based
Content Based
PT
The region based image fusion is carried out in three different approaches – (i) the
region partition approach is based on partitioning the source images into distinct regions by
using standard segmentation methods. The characteristics of the regions are used to get the
fused image; (ii) the statistical and estimation based approach partitions the source images
into regions using advanced region segmentation algorithms followed by a statistical image
formation model. A joint region map is developed by analysing the region map of each
source image to produce the fused image; (iii) the focus region detection and saliency map
7
ACCEPTED MANUSCRIPT
based approach aims at the separation of the significant foreground object from the
background leading to perceptually coherent regions. The advantages and disadvantages of
these three approaches are mentioned later in this section.
A generic block diagram of the region based image fusion scheme is presented in Fig
5. The region based image fusion procedure reads two or more input images. The images are
segmented into different regions using various segmentation algorithms. The various features
like edge, texture, contour, etc. are extracted from each of the regions using suitable feature
extraction techniques. The features are joined using relevant fusion rules to get the fused
image
T
IP
Input
image A
CR
Feature
Feature Fused
Combination
Segmentation Extraction
using Fusion image
rule
US
Input
image B
Table 1
Classification of various region based image fusion algorithms.
Algorithm Method References
ED
8
ACCEPTED MANUSCRIPT
estimation based energy evaluation model [61,62]
algorithm bivariate alpha-stable (BαS) [63]
non expectation maximization (NEM) and bootstrap
sampling [64]
Focus region lifting stationary wavelet transform (LSWT) [70]
detection and NSCT and LSWT [71]
saliency map based quaternion wavelet transform (QWT) and normalized cut [72]
algorithm NSCT and Focus area detection [73]
Surface area based [74]
shearlet and graph based visual system (GBVS) [75]
NSCT and Frequency Tuned (FT) [76]
T
In the region partition based algorithms, the first step is to partition the input images
into distinct regions by employing standard segmentation techniques. By considering the
IP
characteristics of the regions, fusion rules are employed to get the fused image. In general,
CR
the different approaches of multiresolution, ICA based transform, optimization etc. are
incorporated with the segmentation algorithms. Some flow diagrams of the existing methods
US
are shown. The flow diagram of region partition algorithm based method explained in [49] is
depicted in Fig. 6.
AN
Region
segmentation
M
Composite
Region based BEMD Fused Image
Fusion Rules representation
Region
segmentation
CE
AC
9
ACCEPTED MANUSCRIPT
general framework for pixel and region based image fusion utilizing multiresolution
approach. The segmentation stage is improved by performing a single segmentation from all
the source images in a multiresolution fashion. The images are fused following the additive
or weighted combination fusion rule. However, the author has not optimized the performance
and investigated the effect of different parameters on the fusion process. Further, the regions
are combined based on simple region property like average activity. The DWT lacks shift
invariance and directionality property.
To overcome these problems, Lewis et al. reviewed a lot of pixel level fusion
algorithms (using averaging, pyramids, DWT, DT-CWT). The authors compared their
T
findings with a new region based technique [28]. The authors used DT-CWT for
IP
segmentation of the features of the input images to develop a region map. The properties of
CR
every region are determined, and the fusion is performed region by region utilizing region
based approach in the wavelet domain. However, the authors pointed towards the
improvement of higher level region based fusion rules. The quality of the fused image may
US
deteriorate due to the inverse DT-CWT transform. To improve the fusion results, Zaveri and
Zaveri [29] used highboost filter concept with DWT for the fusion. Their proposed technique
AN
overcomes the shift variance problem; as inverse wavelet transform is not needed in the
algorithm. The highboost filter is used to get accurate segmented image. The segmented
image is utilized for obtaining the regions from the input images. The extracted regions are
M
fused using the fusion rule i.e. mean max standard deviation and spatial frequency. However,
they concluded that, to enhance the robustness of the method, complex fusion rules may be
ED
developed.
Many researchers have also used the NSCT approach to image fusion. A region based
PT
image fusion procedure for infrared (IR) and visible image using NSCT is suggested by Guo
et al. [30]. The features of the target region are used to segment the IR image. The input
CE
images are divided using NSCT. The target region and the background regions are merged
using different fusion rules. The inverse NSCT is used to get the fused image. However, the
AC
fusion process involved has a high computational complexity. Zheng et al. used NSCT and
fuzzy logic for the fusion of IR and visible image [31]. Firstly, the required input images are
segmented using the live wire segmentation method followed by decomposition using NSCT
transform. The fuzzy fusion method is used in the low frequency domain and region based
rule is used in high frequency domain. The inverse NSCT is applied to get the fused image.
To effectively analyse the geometric structure of remote images with computationally
efficient approach , a SIST and regional selection algorithm is suggested by Luo et al. [32] for
remote sensing image fusion application. The feature vectors from MS and PAN images are
10
ACCEPTED MANUSCRIPT
divided into regions using fuzzy c-means (FCM) clustering. Based on the regional similarity,
an adaptive multi-strategy fusion rule of high frequency band is suggested. Finally, the fused
image is found by taking the inverse SIST and inverse entropy component analysis.
The local features of an image are usually not considered for fusion. To resolve this,
Wang et al. suggested a fusion procedure based on the DWFT and regional characteristics
[33]. The transform coefficients are found from the two input images using the DWFT.
Taking the mean of the transform coefficients, an average image is obtained which represents
the approximate features of the input images. The average image is segmented based on the
region features to get the region coordinates. The coefficients and the region coordinates are
T
mapped. The fused image is produced by combining the coefficients of each region using
IP
suitable fusion rules. However, it needs a relatively long time for processing due to the image
CR
segmentation process involved.
Researchers have extensively used ICA for image fusion [34-42]. Mitianoudis and
Stathaki [34] used ICA for region based image fusion. The authors investigated the
US
effectiveness of a transform utilizing ICA and topographic ICA bases in image fusion. The
fused image in ICA domain is obtained by utilizing new pixel and region based fusion rules.
AN
The suggested method demonstrated better performance as compared to conventional wavelet
approaches at the cost of slightly more computational load. Contrary to the suggested
framework, Cvejic et al. used ICA bases for region based multimodal image fusion [35, 36].
M
The authors used different training subsets to identify the most significant regions in the
source images. Subsequently, they combined the ICA coefficients utilizing the fusion metrics
ED
for enhanced results. In [35], a combined approach is used to obtain the most significant
regions from the source images. The fused image is obtained by merging the ICA coefficients
PT
from the obtained regions. The authors used the Piella metric to enhance the quality of
results. The performance improved with an increase in the computational complexity.
CE
These proposed methods exhibit the following drawbacks: (i) the approaches cannot
be easily extended to multiple sensor fusion application, (ii) there is no theoretical
AC
justification for the presence of the global optimum of the objective function derived from the
Piella index. To overcome this, in [37], the authors used optimization of the Piella index for
multiple input sensors. To get the optimal solution, the authors revisited the previously
proposed work in [34] and further suggested a method to find the optimum intensity range
through optimization of a fusion index. The suggested method improves the original ICA-
based framework and produces a better fused image. In [38], the authors extended their work
to a more advanced region based fusion approach. A group of fusion rules using textural info
is presented. The suggested method enhances the performance than the max-abs fusion rule in
11
ACCEPTED MANUSCRIPT
case of multifocus images. On the contrary, it was not so good for multimodal image fusion.
The reason may be the various modality images have different texture properties. In [39],
Omar et al. used the combination of Chebyshev polynomial (CP) and ICA depending on
regional information of source images. The proposed method used segmentation technique to
identify features such as texture, edge, etc. The fused image is found using distinct fusion
rules as per the selected regions. The advantage of this method is that it offers an autonomous
denoising property, combining the benefits of both CP and ICA. Nirmala et al. proposed a
region based multimodal (visible and Infrared ) image fusion scheme using ICA and SVM
[40]. The source images are jointly segmented in the spatial domain. The significant features
T
of every region are calculated. The ICA coefficients of the particular regions are combined to
IP
form the fused ICA representation. As the ICA bases are computed and SVM is trained, the
CR
suggested technique may seem to multiply the computational load.
In recent studies, many optimization methods have been utilized by researchers for
image fusion. Neural networks (NN) have been extensively employed for image fusion [43-
US
45]. Hsu et al. suggested a multi-sensor image fusion method using ANN, which merges the
features of the feature and pixel level fusion scheme [43]. The basic concept is to segment
AN
only the far infrared image. The information from each segmented region is added to the
visual image. The different fused parameters are determined according to the different
regions. In [46], PCNN has been utilized for the fusion of multi-sensor images. The
M
procedure begins with the segmentation of source images using PCNN and the output is used
to direct the fusion process. The suggested method outperforms the pixel based methods in
ED
terms of blurring effect, sensitivity to noise and misregistration. Saeedi and Faez [47]
proposed a fusion of visible and IR image using fuzzy logic and PSO. The high frequency
PT
wavelet coefficient of IR and visible images is fused using fuzzy based approach. The PSO is
suggested for the low frequency fusion rule. The low frequency and high frequency parts of
CE
the wavelet coefficient are fused. The authors have not considered noise in their work. It is
evident that for noisy images, the segmentation process will result in over segmentation with
AC
inaccurate regions. Aslantas et al. described a new method for thermal and visible images
[48]. Instead of using a single weighting factor, the authors have used multiple weighting
factors for distinct regions to get the fused image. The weighting factors are optimized using
DE. A new image fusion metric – sum of correlation difference is formed to assess the
performance of the fused images during the optimization procedure.
Some other methods have also been proposed by many researchers. Zheng and Qin
used BEMD technique for region based image fusion [49]. The source images are partitioned
into several intrinsic mode functions and a residual image. The process of fusion is carried
12
ACCEPTED MANUSCRIPT
out as per the segmentation of the source images, which produces a combined BEMD
representation. The fused image is found by using the inverse BEMD. Because of the finite
length of wavelet function, DWT induces energy leaking. This problem does not occur in
BEMD because it is considered as an adaptive highpass filter. Li and Yang [50] proposed a
novel method for multifocus images utilizing region segmentation and spatial frequency. The
normalized cut method is utilized to segment the intermediate fused image. The two input
images are segmented as per the segmenting results of the intermediary fused image. The
fused image is obtained by merging the segmented regions using spatial frequency. The
advantage of the suggested technique is that it does not use the multiresolution approach, as
T
few information of the input image may be missed while performing the inverse
IP
multiresolution operation. The limitation of the proposed method is that the computational
CR
time is more as compared to the wavelet based approaches, as the segmentation procedure
takes more time.
In most of the papers, a common segmentation technique for different input images
US
taken from distinct sensors have not been used for region based image fusion. These
approaches are restricted to the particular input images only. Luo et al. proposed the method
AN
of region partition strategy in which the segmentation is performed on the similar features of
input images (irrespective of the kind of input images) [51]. The complementary and
redundant correlations of the source images are distinguished by using fusion methods. On
M
the basis of the similar components, the small homogeneous regions are merged. The final
region map is obtained by comparing the resemblance between the images. The method has
ED
several advantages like the generality of application, superior visual perception and simple
realization without parameter setting.
PT
Chen and Qin [52] used CS theory in the region based fusion framework. The authors
considered both compression capabilities of sensors and the intelligent understanding of the
CE
image features for fusion. In dynamic scene, it is very hard to accurately determine whether a
pixel or region is blurry or not by utilizing only the focus information only. Besides, another
AC
limitation of the pixel based method is that the fusion results obtained is not accurate when
the image patterns become complex. In contrast to this, Li et al. proposed image matting
fusion technique for the fusion of multifocus images in dynamic scenes [53]. The algorithm
uses morphological filters to get rough segmentation results followed by image matting to
obtain the accurate focussed regions. The fused image is found by merging the focussed
regions. Chen et al. [54] used the SR method for the fusion of multifocus images. The
suggested technique merges the advantages of regional and sparse representation based
fusion to obtain the fused image. The fusion of high resolution images using mean shift
13
ACCEPTED MANUSCRIPT
segmentation method is described in [55]. The authors used SSIM for the measurement of
regional similarity.
These methods need a precise selection of segmentation technique. The choice of
segmentation technique is crucial to obtain a fused image. In these approaches, the regions
are chosen and fused based on some regional characteristics. The significance of the
statistical characteristics of the regions has not been considered, which is utilized to enhance
the precision of the decision process in image fusion applications.
Most of the region based image fusion procedures suggested by different researchers
have not applied the estimation theory approach rigorously. In the statistical and estimation
T
based algorithms, first the input images are partitioned into regions by using some
IP
sophisticated region segmentation algorithms. A joint region map is developed by analysing
CR
the region map of each source image to produce the fused image. A statistical image
formation model is developed for every region in the joint region map. The estimation
procedure is utilized in combination with the model to build an iterative fusion process to
US
determine the model constraints and to generate the fused image. A typical flow diagram of
statistical and estimation based algorithm explained in [58] is shown in Fig.7.
AN
M
Region
Image 1
Segmentation
Region EM Fused
ED
Region
Image 2
Segmentation
PT
CE
Sharma et al. [56] suggested a Bayesian fusion scheme inspired by estimation theory.
AC
A statistical signal processing method to image fusion is suggested by Yang and Blum in
[57]. Many researchers have proposed image fusion using the EM algorithms [58-60]. In
[58], the authors used the EM algorithm, which utilizes the features of regions for fusion in
an optimal way. Each region is built by a statistical image formation model. The region-level
EM algorithm is developed by using the EM fusion procedure in combination with the model
to produce the fused image. Zhang and Ge [61,62] proposed image fusion method employing
energy estimation approach. To partition the source images into regions on the basis of
14
ACCEPTED MANUSCRIPT
homogeneity, piecewise smooth Mumford and Shah Model is used. To speed up
convergence, a level set based optimization procedure is combined with it. The fusion quality
is determined using an energy model. The methods proposed in [27,28,58], fail to consider
the importance of the statistical features of regions, which is utilized to enhance the precision
of the decision process in image fusion applications. In some of the research articles,
statistical model for fusion and the segmentation of images are integrated. Wan et al.
integrated the multi-scale image segmentation technique and statistical feature extraction in
the suggested framework [63]. A region map of input images is obtained using DT-CWT and
statistical region combining algorithm. The source images are divided into significant regions
T
that contain salient information using symmetric alpha stable distribution. By employing the
IP
BαS technique, region features are modelled. The fused image is obtained by applying a
CR
segmented driven approach in the complex wavelet domain.
In most of the practical applications, the dimensions of the training data set are too
large which accounts for a large computation time in statistical image fusion. This constraint
US
requires data reduction. This is achieved by choosing a suitable subset of the prime training
data set without compromising the image fusion precision appreciably. Zribi [64] proposed a
AN
non-parametric and region based image fusion method following the principle of bootstrap
sampling. This brings down the dependence effect of pixels in true image and also reduces
the fusion time. The image sensors are expressed as a real scene degraded by distortion in the
M
estimation approaches. The quality of image segmentation is vital for determining accurate
segmented regions. The ROI and boundary detection of focused regions is not accurately
CE
obtained with the methods discussed above. Many researchers have also used saliency map
based algorithms for region based image fusion. In general, salient object detection is an
AC
15
ACCEPTED MANUSCRIPT
Infrared Visible light
Image Image
T
Fused low Fused high
frequency frequency
IP
CR
Primary fused
Object region
image
Final fused
image
US
AN
Fig. 8. Flow diagram of region based image fusion scheme of NSCT based method [76].
The problem of the spatial domain based methods is that they may generate artefacts
M
or imprecise results at the focused border areas, as the boundary of focused areas cannot be
estimated correctly. This problem is solved by the use of multi-scale transform method.
ED
However, the problem with this method is to choose the proper fusion rule. Additionally,
some information of the input image may be lost while implementing the inverse multi-scale
PT
transform. So as to avoid these problems, the authors proposed methods combining the
benefits of spatial region based and transform domain [70,71]. Chai et al. suggested a
CE
multifocus image fusion technique utilizing the focus region detection and multiresolution
approach [70]. The authors have used the focus region detection and LSWT to combine the
benefits of spatial and transform domain based fusion approaches. The idea of local visibility
AC
(LV) in LSWT domain is used for the fusion of the lowpass subband coefficients. The sum
modified Laplacian inspired local visual contrast rule is applied for the fusion of the highpass
subband coefficients. However, the fused image consists of several imprecise results at the
boundary of the focused region. To overcome this, the authors proposed a multifocus image
fusion method based on NSCT and focus region detection [71].
The image visibility rule in NSCT domain is used for the fusion of lowpass subband
coefficients. The local area spatial frequency rule is employed for the fusion of highpass
16
ACCEPTED MANUSCRIPT
subband coefficients. The limitation of the methods is that the post- processing phase
utilizing morphological procedure is not robust. In order to avoid this problem, Liu et al.
suggested a multifocus image fusion utilizing QWT [72]. Initially, the authors used the local
variance of the phases to identify the focus or defocus for each pixel. They segmented the
focus detection result using normalized cut to eliminate the detected errors. Finally, the fused
image is found utilizing the spatial frequency as fusion weight along the edge of the focus
region. However, the method may create false information and irregular phenomenon at the
boundaries of the focused areas, as the boundary cannot be estimated precisely. To perfectly
determine the boundaries of the focused region, Yang et al. suggested a novel hybrid
T
multifocus image fusion method based on NSCT and focus area detection [73]. The authors
IP
modified their work in the sense that they used the log Gabor filter for high frequency
CR
subband coefficients. The limitation of the method is that the NSCT procedure is consuming
more time. Nejati et al. proposed a new focus measure depending on the surface area of the
region surrounded by juncture points for multifocus image fusion [74]. The objective of this
US
metric is to differentiate focus regions from blurred regions. The juncture points of the source
images are computed and segmented utilizing these juncture points. The surface area of every
AN
region is used as a quantity to get the focused regions. An initial selection map for the fusion
is obtained using this measure which is refined using morphological operations.
The most challenges task in region based image fusion is the proper image
M
among them is the salient region detection approaches which may focus the object region as
well as some of the background region. Detection of the visual salient region has been an
PT
ongoing research process for a long time. The saliency map is an emerging technique to
identify the salient region, which can overcome the above problem. Zhang et al. suggested a
CE
multifocus image fusion technique based on focus region extraction. The authors used
saliency analysis [75]. The GBVS procedure is utilized to identify the focused region in the
AC
input image. Subsequently, watershed and morphological techniques are employed to find the
bounded area of saliency map and eliminate the pseudo focus area. In the final step, the
focused area is fused straight and the residual area are fused using shearlet transform.
In [76], Meng et al. suggested a new procedure based on object region detection and
the NSCT. The FT saliency detection map is used to acquire the saliency map for the IR
image. The significant object region is extracted from the IR image using free regions
removal method. The input images are divided via NSCT and distinctive fusion rules are
applied for the lowpass and highpass subband coefficients. The inverse NSCT is utilized to
17
ACCEPTED MANUSCRIPT
produce a primary fused image. At last, the final fused image is found by combining the
primary fused image with the object region. Again in [77], Meng et al. presented the region
based image fusion technique to merge IR and visible image by employing the saliency map
and interest point. A saliency map is built using a saliency detection process for IR image.
Further, it is explored to identify interest points. To get the salient region, a convex hull of the
salient interest point is computed. The first saliency map is developed by merging the convex
hull of the salient interest points. Finally, the various fusion rules are employed for object
region and background. In [78], Han et al. proposed a saliency aware fusion procedure for
multimodal image fusion (IR and visible image) to enrich the visualization of the visible
T
image. The process is used for saliency recognition followed by a bias fusion. The
IP
information of these two sources is combined using the Markov random fields. The following
CR
fusion phase is used to bias the end result favouring the visible image, excepting when a
region has distinct IR saliency. The fused image represents both salient foreground from the
IR image and background as provided by the visible image. The fused image obtained with
US
these methods has significant improvement compared to other methods, because it avoids the
traditional segmentation process. The transform used in these methods may consume more
AN
time. Still, the different methods have their own importance. The further improvement in the
fusion quality may require a suitable selection of the regional features.
The region partition based algorithms are simpler than the other two approaches –
M
statistical and estimation based, focus region detection and saliency map based. The region
partition based algorithms need a precise selection of segmentation technique. The choice of
ED
segmentation technique is crucial to obtain a fused image. In these approaches, the regions
are selected and fused based on certain regional characteristics. However, the importance of
PT
the statistical characteristics of the regions are not considered, which is utilized to enhance
the precision of the decision process in image fusion applications. The statistical and
CE
estimation approach uses the statistical characteristics to improve the precision of the
decision process. It is needed to choose an efficient image segmentation technique for
AC
determining accurate segmented regions. However, the ROI and boundary detection of
focused regions is not accurately obtained with these approaches. The focus region detection
and saliency map approaches uses the ROI and boundary detection to accurately obtain the
fused image. The selection of segmentation technique is not vital in these approaches.
However, these approaches are complex and application specific.
18
ACCEPTED MANUSCRIPT
3. Performance evaluation
There are two methods in which the quality of the fused image is computed i.e.,
qualitative or subjective and quantitative or objective. Qualitative analysis means visual
analysis in which the fused image is matched with the source images by a group of observers.
The analysis of the fused image considers different optical parameters like spatial details, the
size of the object, colour etc. The qualitative analysis is typically accurate if it is done
correctly. However, these methods are inconvenient, expensive and consumes more time. It is
a very difficult task in most of the image fusion applications due to lack of availability of a
T
ground truth image that is perfectly fused. So, another technique to calculate the fusion
performance is the quantitative or objective evaluation. But again, it remains an issue as how
IP
to measure the performances of the fused image objectively. In this survey, the quantitative
CR
fusion assessment metrics are classified in two groups based on the presence or absence of a
reference fused image.
US
3.1. Objective fusion evaluation metrics with reference image
The reference image is the ideal fused image that is taken as a ground truth image for
AN
validating the image fusion algorithm. The ground truth image may be available or manually
constructed. A list of commonly used objective fusion evaluation metrics with reference
image is illustrated in Table 2. The quality metrics employed to compute the fusion
M
performance are peak signal to noise ratio (PSNR), root mean square error (RMSE), mutual
information (MI), structural similarity index measure (SSIM), correlation coefficient (CC)
ED
and universal quality index (UQI). The symbols used in the expressions carry the same
meaning as depicted in the corresponding reference papers mentioned in table 2 and 3.
PT
Table 2
Objective fusion evaluation metrics with reference image.
CE
number of intensity levels in PSNR 20 log10
L2
the image to the related pixels 1
I i, j I i, j
M N
2
MN r f
in the ideal and the fused i 1 j 1
image. A higher PSNR value
indicates superior fusion.
2 RMSE It estimates the quality of the M 1 N 1 [80]
R i, j F i, j
1 2
fused image by relating the RMSE
MN i 0 j 1
ideal and the fused image.
The value of RMSE should be
lower i.e. close to zero for a
good fused image.
19
ACCEPTED MANUSCRIPT
3 MI The similarity of image MI FAB MI FA MI FB [81]
intensity between ideal and
fused image is measured
using mutual information. MI
should be high for a better
fusion performance.
4 SSIM The SSIM is employed to
SSIM ( x, y )
2 x y C1 2 C xy 2
[82]
calculate the quality of one
image compared with the
2
x y2 C C
1
2
x
2
y 2
T
5 CC It is used to compute the
spectral feature similarity
X i, j
x Xˆ i , j xˆ .
[83]
IP
i, j
CC
between the reference and the
X
x Xˆ i , j xˆ
2 2
CR
i, j
should be high i.e. close to 1.
6 UQI The UQI is used to calculate xy 2x y 2 x y [84]
how much amount of relevant UQI
x y x y
x2 y2
2 2
data is transformed from ideal
image into the fused image.
The value of this metric
US
AN
ranges between -1 to 1.
In very rare case, the ground truth image is available. Hence, it is highly desirable to
assess the quality of the fused image without taking the ideal image. The meaning of quality
ED
metric with the reference image is that either the ground truth image or the ideal image is
available. The meaning of quality metric without the reference image is that either the ground
PT
truth image or the ideal image is not available. In such cases the quality metric is computed
using the source or input images and the output or fused image. A commonly used list of the
CE
entropy (H), cross entropy (CE), spatial frequency (SF), fusion mutual information (FMI),
sum of the correlation of difference (SCD), Piella metric and Petrovic metric.
Table 3
Objective fusion evaluation metrics without reference image.
Sl. Quality Description Formula References
No. metric
20
ACCEPTED MANUSCRIPT
1 SD The change in intensity [85]
M N 2
variation in the fused image is f i, j
i 1 j 1
measured using standard SD
MN
deviation. The SD value should
be high.
2 H The information content of the L 1 [85]
fused image is computed using H Pi log Pi
i 0
entropy. The fused image
containing rich information has
high entropy value.
3 CE To know the resemblance in D hA hF D hB hF [86]
information content between the CE A, B, F
input and fused image, cross 2
T
entropy is employed. The cross
entropy should be low.
IP
4 SF Spatial frequency is measured [87]
RF CF
2 2
SF
only using the fused image. It is
CR
measured by calculating the
row and column frequency of
the fused image. The value of
5 FMI
SF should be high.
It is utilized to compute the
level of dependency between
the input and the fused image.
US
FMI I FA I FB [86]
AN
A larger value of FMI indicates
an improved quality of the
fused image.
6 SCD The sum of the correlation of SCD r D1 , S1 r D2 , S2 [88]
M
wW
significance of edge
information. The dynamic range
of Piella’s metric is [0,1] and QE a, b, f
the value should be near to 1 for
QW a, b, f QW a , b , f
improve fusion performance.
8 Petrovic It provides the matching Q AB / F [90]
metric between the edges transmitted N M
Q n, m wA n, m Q n, m wB n, m
AF BF
i 1 J 1
21
ACCEPTED MANUSCRIPT
4. Discussions
Image fusion is very important and beneficial for various image processing steps such
as object extraction, identification and computer vision. It is a tedious job due to
misregistration, distortion and other artefacts. Recently, in many applications like
photography [91-93], medical diagnosis [94-96], surveillance [97-99], remote sensing [100-
102] etc., the region based image fusion has been widely used. Different methods are
proposed in the literature for region based image fusion, as discussed in section 2. A
comparison among various methods is quite a difficult task, as they use different modalities,
databases and performance indices. The assessment of quality of fused image is carried out in
T
two ways i.e. qualitative and quantitative. In this paper, a comparison of various methods is
IP
carried out for different applications using quantitative and qualitative analysis. This
CR
comparison would benefit the researchers to apply different methods in the various
applications.
The quantitative assessment for the existing methods is provided in form of tables.
US
The different tables are prepared after considering different references using similar test
conditions, same application and using the same source images. The symbol ‘–’ in all the
AN
tables represent unavailability of the particular information. The particular performance
metric is not evaluated due to non-availability of the fused image. A comparison of different
M
Method MI QAB/F H SF Q0 Qw QE
RSSF [29] 6.9279 0.7119 8.7813 10.3350 0.7138 0.8312 0.6618
BEMD [49] 5.9016 0.6250 7.4562 8.8864 0.6170 0.7690 0.6500
PT
DWT and highboost 7.7344 0.7018 8.8066 10.0048 0.7608 0.8119 0.6521
filter [29]
SR [54] 5.6106 0.7533 7.0657 8.4682 0.6911 0.7814 0.5918
AC
22
ACCEPTED MANUSCRIPT
It is seen that most of the methods perform well for the multifocus fusion example in
terms of MI. However, the methods listed under the focus region detection group performs
better as compared to other methods. A similar trend is observed for the parameter QAB/F. For
instance, QWT and normalized cut, NSCT and focus area detection and Surface area based
methods give a value of 8.9971, 8.6580, and 8.8280 respectively for MI. The reason may be
(i) presence of superiorities such as multiresolution, multidirection and shift-invariance, (ii)
use of the detected focus area as a fusion decision map to drive the fusion process, (iii)
preventing artefact and imprecise results at the boundary of the focus region. Further, few
T
researchers have computed the remaining metrics. So, these metrics may be calculated for the
IP
given methods for a fair comparison. A comparison of various methods for IR and visible
CR
image fusion application is illustrated in Table 5.
Table 5
Method QAB/F MI H
US
Comparison of different methods for IR and visible images (Reference: UN Camp Image).
SD Piella Qw
AN
ICA-SVM [40] 0.6100 7.1700 7.1800 31.4100 0.9500 0.7118
DT-CWT [40] 0.4900 2.9800 6.8300 32.9900 0.9100 0.6512
Region based ICA 0.5700 4.1600 6.5300 30.4100 0.9200 0.6602
[40]
M
[36]
DT-CWT [36] 0.4600 3.0173 6.6727 34.1523 0.9100 0.6835
FPSO [47] 0.5520 2.7028 7.4100 32.3155 0.9128 0.6918
DT-CWT [47] 0.5030 3.1603 6.6600 36.9536 0.8910 0.6721
PT
It is observed that the ICA based methods outperform other methods in the IR and
CE
visible image fusion application. For instance, the ICA-SVM method gives a value of 0.61
for QAB/F, 7.1700 for MI, 7.1800 for H, 31.4100 for SD, 0.9500 for the Piella metric and
AC
0.7118 for Qw which is better as compared to other methods. The reason may be the ICA
based methods perform segmentation and uses numerous statistical properties of the regions
to make intelligent decisions. The researchers have not computed most of the metric values.
Therefore, these values may be computed for the above mentioned methods for a fair
comparison. A comparison of different methods for medical image fusion application is
depicted in Table 6.
Table 6
Comparison of different methods for Medical images (Reference: CT and MRI).
23
ACCEPTED MANUSCRIPT
Method MI QAB/F Q0 Qw QE H
RF-SSIM [51] 5.6274 0.6259 0.8928 0.8027 0.6283 6.8614
RSSF [51] 3.4326 0.5017 0.4372 0.4034 0.4128 5.4359
BEMD [51] 2.1028 0.4604 0.5293 0.7145 0.5612 6.8879
CS [52] 2.8958 0.5195 0.5634 0.7231 0.6275 6.3807
BαS [63] - 0.6808 - 0.7541 - -
DT-CWT [63] - 0.6339 - 0.6722 - -
The results indicated in the table shows that the region partition based algorithms
perform better than the other methods. For example, the RF-SSIM method gives a value of
5.6274 for MI, 0.8928 for Q0 , 0.8027 for Qw, 0.6283 for QE and 6.8614 for H which is better
T
than the other methods. The reason may be the methods merge the homogenous regions
IP
based on SSIM. The statistical and estimation based algorithms also show promising results
CR
but using other metrics. So researchers may compute these metrics for a fair comparison.
The subjective or qualitative evaluation reflects the visual perception of the image,
which varies from viewer to viewer. It is evaluated properly by the experts who have long
US
experience in the field. However, this type of evaluation is not always preferred, as it is
inconvenient and time consuming. The fused images for different image fusion applications
AN
are depicted below. The experimental results obtained for clock images carried out by
different researchers are shown in Fig. 9.
M
ED
24
ACCEPTED MANUSCRIPT
T
clock is out of focus and in (b) the right clock is out of focus. The fused images of different
IP
methods are illustrated in fig (c)-(p). It is observed that almost all the images look similar.
Therefore, the subjective evaluation, in general, is treated as an ineffective tool for
CR
comparison. However, it is seen that the visual quality of the images in (c), (e), (f) and (g) is
better as compared to other methods. The image in (c) and (f) contain sharp edges at the
US
upper portion of the right clock. The fused image obtained with these methods have retained
more relevant information of the source images. In (i) and (j), the edge at the top of the right
AN
clock is not sharp i.e. fractional edge information is lost. Further, the images in (d) and (k)
have blurry clock hands. The experimental results obtained for IR and visible image fusion
application carried out by different researchers are shown in Fig. 10.
M
ED
PT
25
ACCEPTED MANUSCRIPT
As shown in Fig. 10, (a) represents the visible image and (b) represents the IR image.
The fused images from different methods are shown in (c)-(j). It is observed that the person is
identifiable in (b) and trees and fence are visible in (a). The fused images in (d),(f),(g), (h)
and (k) are demonstrating more details as compared to other methods. However, the images
in rest of the methods do not show the details clearly. In some cases such as (i), the person’s
image is blurry. The experimental results obtained for medical image fusion application
carried out by different researchers are shown in Fig. 11.
T
IP
CR
(a) (b) (c) (d)
US
AN
(e) (f)
Fig. 11. Experimental results for medical image fusion application, (a) CT image, (b) MRI
image, (c) CS [52], (d) RF-SSIM [51], (e) RSSF [51], (f) BEMD [51].
M
As presented in Fig. 11, (a) represents the CT image, (b) represents the MRI image.
ED
The fused images from different methods are shown in (c)-(f). It is observed that the image in
(d) is more prominent as compared to other methods. The method preserved more salient
PT
source information. In the middle portion of (e), some part is lost. Further, some artefacts are
introduced at the top portion of the fused image (f). It is wise to reiterate the fact that the
CE
reason of choosing the special fusion methods for multifocus, IR, visible, and for medical
images are – i) the tables in this work are prepared application wise for showing a fair
comparison among different methods, ii) the respective fusion methods are chosen based on
AC
including more number of quality metrics, and iii) the fusion methods are grouped based on
the results available using the same input images.
5. CONCLUSION
In this paper, we have presented a survey on region based image fusion methods. The
region based image fusion algorithms are classified, for the first time, into three classes:
region partition based, statistical and estimation based, and focus region detection and
26
ACCEPTED MANUSCRIPT
saliency map based. A comparison of different methods in terms of various metrics for
different applications is done. Based on this comparison, an idea about the various image
fusion methods is developed for different applications. The focus region detection and
saliency map based algorithm is mostly suitable for the multifocus image fusion applications.
The ICA based methods perform better in case of the multimodality image fusion
applications. The region partition algorithms are used in medical image fusion applications
producing better fusion results. It is observed that the saliency map method is an emerging
technique that can be used in many applications. The quality assessment metrics QAB/F and
MI is mostly preferred in all the applications. However, the other metrics can also be
T
computed for a fair comparison. The problems existing in different methods is discussed. The
IP
survey carried out in this paper may help the researchers in further research in the domain of
CR
region based image fusion.
US
AN
M
ED
PT
CE
AC
27
ACCEPTED MANUSCRIPT
References
[1] A.Goshtasby, S. Nikolov, Image fusion: Advances in the state of the art,
Information Fusion. 8 (2) (2007) 114-118.
[2] H. Shen, X. Meng, L. Zhang, An integrated framework for the spatio-temporal-
spectral fusion of remote sensing images, IEEE Transactions on Geoscience and
Remote Sensing. 54(12) (2016) 7135-48.
[3] X. Meng , J. Li, H. Shen , l. Zhang , H. Zhang , Pansharpening with a guided filter
based on three-layer decomposition, Sensors. 16(7) (2016) 1068.
[4] C.Thomas, T. Ranchin, L. Wald, J. Chanussot, Synthesis of multispectral images to
high spatial resolution: A critical review of fusion methods based on remote sensing
T
physics, IEEE Transactions on Geoscience and Remote Sensing. 46(5) (2008)1301-
IP
1312.
[5] S. Li, X. Kang, L. Fang, J. Hu, H. Yin, Pixel-level image fusion: A survey of the
state of the art, Information Fusion. 33 (2017) 100-112.
CR
[6] B. Yang, Z.L. Jing, H.T. Zhao, Review of pixel-level image fusion, J. Shanghai
Jiaotong Univ. 15 (2010) 6-12.
[7] H. Li, L. Li, and J. Zhang, Multi-focus image fusion based on sparse feature matrix
[8] US
decomposition and morphological filtering, Opt. Commun.,342 (2015) 1-11.
F. Mirzapur, H. Ghassemian, Improving hyperspectral image classification by
combining spectral, texture, and shape features, International Journal of Remote
Sensing, 36 (4) (2015)1070–1096.
AN
[9] S. S. Malik, S.P. Kumar, G.B. Maruthi, DT-CWT: Feature level image fusion based
on dual-tree complex wavelet transform, in: Proceedings of IEEE International
Conference on Information Communication and Embedded Systems (ICICES),
M
[12] M. Fauvel , J. Chanussot , J. Atli Benediktsson , Decision Fusion for the Classifi-
cation of Urban Remote Sensing Images, IEEE Transactions on Geoscience and
Remote Sensing. 44 (10) (2006) 2828-2838.
CE
[13] Y. Liu, X. Chen, H. Peng, Z. Wang, Multi-focus image fusion with a deep
convolutional neural network. Information Fusion. 36(1) (2017) 191-207.
[14] A. Dogra, B. Goyal, S. Agrawal, From Multi-Scale Decomposition to Non-Multi-
AC
28
ACCEPTED MANUSCRIPT
review, Spat. Inf. Res. 24 (4) (2016) 475-483.
[19] H. Ghassemian, A review of remote sensing image fusion methods, Information
Fusion. 32 (2016) 75-89.
[20] G. Vivone, L. Alparone, J. Chanussot, M.D. Mura, A. Garzelli, S. Member, G.A.
Licciardi, R. Restaino, L. Wald, A Critical Comparison Among Pansharpening
Algorithms, IEEE Transaction on Geoscience and Remote Sensing. 53 (5) (2014)
2565-2586.
[21] Z. Wang, S. Wang, Y. Zhu, Y. Ma, Review of Image Fusion Based on Pulse-
Coupled Neural Network, Arch. Comput. Methods Eng. 23 (4) (2015) 659-671.
[22] J. Du, W. Li, K. Lu, B. Xiao, An overview of multi-modal medical image fusion,
Neurocomputing. 215 (2016) 3-20.
[23] A.P. James, B.V. Dasarathy, Medical image fusion: A survey of the state of the
art, Information Fusion. 19 (2014) 4-19.
T
[24] Z. Zhang, R. Blum, Region-based image fusion scheme for concealed weapon
IP
detection, in: Proceedings of the 31st Annual Conference on Information Sciences
and Systems, 1997, pp. 168-173.
CR
[25] B. Matuszewski, L.K. Shark, M. Varley, Region-based wavelet fusion of ultrasonic,
radiographic and shearographic non-destructive testing images, in: Proceedings of
the 15th World Conference on Non- Destructive Testing, 2000, pp. 15-21.
[26] G. Piella, A region-based multiresolution image fusion algorithm, in: Proceedings of
[27] US
IEEE International Conference on Information Fusion, Vol. 2, 2002, pp. 1557-1564.
G. Piella, A general framework for multiresolution image fusion: from pixels to
regions. Information fusion. 4(4) (2003) 259-280.
AN
[28] J.J. Lewis, R.J.O.Õ. Callaghan, S.G. Nikolov, D.R. Bull, Pixel- and region-based
image fusion with complex wavelets, information fusion. 8(2) (2007) 119-130.
[29] T. Zaveri, M. Zaveri, A novel region based multimodality image fusion method,
Journal of Pattern Recognition Research. 2 (2011)140-53.
M
[30] B.L. Guo,Q.Zhang,Y.Hou,Region based fusion of infrared and visible images using
nonsubsampled contourlet transform,Chin.Opt.Lett. 6 (5) (2008) 338-341
[31] Z. Yu, L. Yan, N. Han, A Region-based Image Fusion Algorithm for Detecting
ED
image fusion using ICA bases, In: Proceedings of IEEE International Conference
on Information Fusion. 2006, 10, pp. 1-6.
[36] N. Cvejic, D. Bull, N. Canagarajah, Region-based multimodal image fusion using
ICA bases, IEEE Sensors Journal. 7(5) (2007) 743-51.
[37] N. Mitianoudis, T. Stathaki, Optimal contrast correction for ICA-based fusion of
multimodal images, IEEE sensors journal. 8(12) (2008) 2016-26.
[38] N. Mitianoudis, S.A. Antonopoulos, T. Stathaki, Region-based ICA image fusion
using textural information, In: Proceedings of IEEE International Conference on
Digit. Signal Process. (DSP) 2013, pp. 1-6.
[39] Z. Omar, N. Mitianoudis, T. Stathaki, Region-based image fusion using a
combinatory Chebyshev-ICA method, in: Proceedings of IEEE International
29
ACCEPTED MANUSCRIPT
Conference on Acoustics, Speech and Signal Processing (ICASSP), 2011, pp. 1213-
1216.
[40] D.E. Nirmala, A.B.S. Paul, V. Vaidehi, Improving independent component analysis
using support vector machines for multimodal image fusion, Journal of computer
science. 9 (2013) 1117-1132.
[41] S. Agrawal, S. Swain, L. Dora, BFO-ICA based multi focus image fusion, In IEEE
Symposium on Swarm Intelligence (SIS), 2013, pp. 194-199.
[42] N. Cvejic, D. Bull, N. Canagarajah, A novel ICA domain multimodal image fusion
algorithm, In Proc. Of International Society for Optics and Photonics on Multi-
sensor, Multisource Information Fusion: Architectures, Algorithms, and
Applications Vol. 6242, 2006, p. 62420W.
[43] S.Li Hsu, P. Gau, Region-based Image Fusion with Artificial Neural Network, Int. J.
Inf.Math.Sci..5(2009)264-267.
T
[44] S. Li, J. T. Kwok, Y. Wang, Multifocus image fusion using artificial neural
IP
networks. Pattern Recognition Letters, 23(8) (2002) 985-997.
[45] M. Pagidimarry, K. A. Babu, An all approach for multi-focus image fusion using
neural network, Artificial Intelligent Systems and Machine Learning,3(12) (2011)
CR
732-739.
[46] M. Li, W. Cai, Z. Tan, A region-based image fusion scheme using pulse-coupled
neural network, Pattern Recognition Letters. 27(16) (2006)1948-56.
[47]
[48]
US
J. Saeedi, K. Faez, Infrared and visible image fusion using fuzzy logic and
population-based optimization, Applied Soft Computing. 12(3) (2012) 1041-1054.
V. Aslantas, E. Bendes, R. Kurban, A.N. Toprak, New optimised region-based
AN
multi-scale image fusion method for thermal and visible images, IET Image
Process.8(5) (2014) 289-299.
[49] Y. Zheng, Z. Qin, Region-based image fusion method using bidimensional
empirical mode decomposition, Journal of Electronic Imaging. 18 (1) (2009)
M
013008-013008.
[50] S. Li, B. Yang, Multifocus image fusion using region segmentation and spatial
frequency, Image Visual Computing. 26 (7) (2008) 971-979.
ED
[53] S. Li, X. Kang, J. Hu, B. Yang,Image matting for fusion of multi-focus images in
dynamic scenes,Information Fusion. 14 (2) (2013)147-162.
[54] L. Chen, J. Li, C.P. Chen, Regional multifocus image fusion using sparse
CE
30
ACCEPTED MANUSCRIPT
processing, In IEEE 2nd International Congress on Image and Signal Processing,
2009, pp. 1-4.
[60] J. Yang, R.S Blum, Image fusion using the expectation-maximization algorithm and
a hidden Markov model in: Proceedings of IEEE Conference on Vehicular
Technology, Vol. 6, 2004, pp. 4563-4567.
[61] Y. Zhang., L. Ge, Region-based image fusion using energy estimation, in:
Proceedings of IEEE International Conference on Software Engineering, Artificial
Intelligence, Networking, and Parallel/Distributed Computing, Vol. 1,2007, pp. 729-
734.
[62] Y. Zhang, Adaptive region-based image fusion using energy evaluation model for
fusion decision, Signal, Image and Video Processing. 1(3) (2007)215-23.
[63] T. Wan, N. Canagarajah, A. Achim, Segmentation-driven image fusion based on
alpha-stable modeling of wavelet coefficients, IEEE Transactions on
T
Multimedia.11(4) (2009)624-33.
IP
[64] M. Zribi, Non-parametric and region-based image fusion with Bootstrap sampling,
Information Fusion. 11(2) (2010) 85-94.
[65] L. Itti, C. Koch, E.Niebur,A model of saliency-based visual attention for rapid scene
CR
analysis, IEEE Transactions on pattern analysis and machine intelligence. 20(11)
(1998) 1254-1259.
[66] J. Harel, C. Koch, P. Perona, Graph-based visual saliency, In Advances in neural
[67] US
information processing systems, 2007, pp. 545-552.
R. Achanta, S. Hemami, F. Estrada, S. Susstrunk, Frequency-tuned salient region
detection, in: Proceedings of IEEE Conference on Comput. Vis. Pattern
AN
Recognition, 2009, pp. 1597-1604.
[68] H. Jiang, J. Wang, Z. Yuan, Y. Wu, N. Zheng, S. Li, Salient object detection: A
discriminative regional feature integration approach, in: Proceedings of the IEEE
conference on computer vision and pattern recognition, 2013, pp. 2083-2090.
M
[69] J. Zhang, S. Sclaroff, Saliency detection: A boolean map approach, in: Proceedings
of the IEEE international conference on computer vision ,2013, pp. 153-160.
[70] Y. Chai, H. Li, Z. Li, Multi-focus image fusion scheme using focused region
ED
fusion using quaternion wavelet and normalized cut, Signal Processing. 97 (2014) 9-
30.
[73] Y. Yang, S. Tong, S. Huang, P. Lin, Multi-focus image fusion based on NSCT and
CE
31
ACCEPTED MANUSCRIPT
(2010) 48-54.
[80] L. F. Zoran, Quality evaluation of multiresolution remote sensing images fusion,
UPB Sci Bull Series C,71 (2009) 38-52.
[81] X. L. Zhang, Z.F. Liu, Y. Kou, J.B. Dai, Z.M. Cheng, Quality assessment of image
fusion based on image content and structural similarity, in: Proceedings of IEEE 2nd
International Conference on Information Engineering and Computer Science
(ICIECS), 2010, pp. 1-4.
[82] Z. Wang, A. C. Bovik, H. R. Sheikh, E. P. Simoncelli, Image quality assessment:
From error visibility to structural similarity, IEEE Trans. Image Process. 13(4)
(2004) 600-612.
[83] X. X. Zhu, R. Bamler, A sparse image fusion algorithm with application to pan-
sharpening, IEEE transactions on geoscience and remote sensing.51(5) (2013) 2827-
36.
T
[84] Z. Wang, A. C. Bovik, A universal image quality index, IEEE signal processing
IP
letter,9 (3) (2002) 81-84.
[85] W. Wang, F. Chang, A multi-focus image fusion method based on laplacian
pyramid. JCP 6(12) (2011) 2559-66.
CR
[86] M.B.A. Haghighat, A. Aghagolzadeh, H. Seyedarabi, A non-reference image fusion
metric based on mutual information of image features, Computers & Electrical
Engineering.37 (5) (2011) 744-56.
[87]
[88]
US
S. Li, J. T. Kwok, Y. Wang, Combination of images with diverse focuses using the
spatial frequency, Information Fusion. 2(3) (2001)169-176.
V. Aslantas, E. Bendes, A new image quality metric for image fusion: The sum of
AN
the correlations of differences, AEU-International Journal of Electronics and
Communications. 69(12) (2015) 1890-1896.
[89] G. Piella, H. Heijmans, A new quality metric for image fusion, in: Proceedings of
IEEE International Conference on Image Processing, 2003, pp. 173-176.
M
116.
[93] Y. Liu, S. Liu, Z. Wang, Multi-focus image fusion with dense SIFT, Information
Fusion. 23 (1) (2015) 139-155.
CE
[94] Z. Xu, Medical image fusion using multi-level local extrema, Information Fusion.
19(2014)38-48.
[95] L. Wang, B. Li, L. Tian, Multi-modal medical image fusion using the inter-scale
andintra-scale dependencies between image shift-invariant shearlet coefficients,
AC
32
ACCEPTED MANUSCRIPT
[100] G. Simone, A. Farina, F.C. Morabito, S. B. Serpico, L. Bruzzone, Image fusion
techniques for remote sensing applications, Information fusion. 3 (1) (2002) 3-15.
[101] J. Cheng, H. Liu, T. Liu, F. Wang, H. Li, Remote sensing image fusion via wavelet
transform and sparse representation, ISPRS Journal of Photogrammetry and Remote
Sensing. 104 (2015) 158-73.
[102] W. Wenbo, Y. Jing, K. Tingjun, Study of Remote Sensing Image Fusion and Its
Application in Image Classification, The International Archives of the
Photogrammetry, Remote Sensing and Spatial Information Sciences, 2008, pp.
1141-1146.
T
IP
CR
US
AN
M
ED
PT
CE
AC
33