Professional Documents
Culture Documents
www.elsevier.com/locate/inus
Abstract
The task of enhancing the perception of a scene by combining information captured by dierent sensors is usually known as image
fusion. The pyramid decomposition and the Dual-Tree Wavelet Transform have been thoroughly applied in image fusion as analysis and
synthesis tools. Using a number of pixel-based and region-based fusion rules, one can combine the important features of the input images
in the transform domain to compose an enhanced image. In this paper, the authors test the eciency of a transform constructed using
Independent Component Analysis (ICA) and Topographic Independent Component Analysis bases in image fusion. The bases are obtained
by oine training with images of similar context to the observed scene. The images are fused in the transform domain using novel pixelbased or region-based rules. The proposed schemes feature improved performance compared to traditional wavelet approaches with
slightly increased computational complexity.
2005 Elsevier B.V. All rights reserved.
Keywords: Image fusion; Image segmentation; Independent component analysis; Topographic ICA
1. Introduction
Let I1(x, y), I2(x, y), . . . , IT(x, y) represent T images of
size M1 M2 capturing the same scene. Each image has
been acquired using dierent instrument modalities or capture techniques, allowing each image to have dierent
characteristics, such as degradation, thermal and visual
characteristics.
In this scenario, we usually employ multiple sensors that
are placed relatively close and are observing the same
scene. The images acquired by these sensors, although they
should be similar, are bound to have some translational
motion, i.e. miscorrespondence between several points
of the observed scene. Image registration is the process
of establishing point-by-point correspondence between a
number of images, describing the same scene. In this study,
we will assume that the input images Ii(x, y) have negligible
registration problems, which implies that the objects in all
images are geometrically aligned [5].
Corresponding author. Tel.: +44 207 594 6199; fax: +44 207 594 6234.
E-mail address: n.mitianoudis@imperial.ac.uk (N. Mitianoudis).
1566-2535/$ - see front matter 2005 Elsevier B.V. All rights reserved.
doi:10.1016/j.inus.2005.09.001
The fusion operator g() describes the merging of information from the dierent input images. Many fusion rules
132
T
1 X
TfI i x; yg
T i1
based approach that segments and fuses active and non-active areas of the image is introduced.
The paper is structured, as follows. In Section 2, we
introduce the basics of the Independent Component Analysis technique and how it can be used to generate analysis/
synthesis bases for image fusion. In Section 3, we describe
the general method for performing image fusion using ICA
bases. In Section 4, we present the proposed pixel-based
weighted combination scheme and a combinatory regionbased scheme. In Section 5, we benchmark the proposed
transform and fusion schemes, using common fusion testbed. Finally, in Section 6, we outline the advantages and
disadvantages of the proposed schemes together with some
suggestions about future work.
2. ICA and topographic ICA bases
Assume an image I(x, y) of size M1 M2 and a window
W of size N N, centered around the pixel (x0, y0). An
image patch is dened as the product between a N N
neighbourhood centered around pixel (x0, y0) and the window W
I w k; l W k; lIx0 bN =2c k; y 0 bN =2c l;
8k; l 2 0; N 1
8k; l 2 0; N 1
K
X
uj bj k; l
j1
1
ut B I w t AI w t
10
T
1
133
11
12
Image Patch
Iw(t)
I(x0,y0)
Iw(k,l)
Iw(t)
Lexicographic
ordering
Fig. 1. Selecting an image patch Iw around pixel (x0, y0) and the lexicographic ordering.
134
0:5
16i6K
15
16
where /(x) = oG(x)/ox. We randomly initialise the update rule in (15) for each projecting vector ai. The new updates are then orthogonalised, using the symmetric like
orthogonalisation scheme in (16). These two steps are iterated, until ai have converged.
2.1.3. Topographical independent component analysis
(TopoICA) bases
In practical applications, one can very often observe
clear violations of the independence assumption. It is possible to nd couples of estimated components such that
they are clearly dependent on each other. This dependence
structure, however, is very informative and it would be useful to somehow estimate it [9].
Hyvarinen et al. [9] used the residual dependency of the
independent components, i.e. dependencies that could
not be cancelled by ICA, to dene a topographic order between the components. Therefore, they modied the original ICA model to include a topographic order between the
components, so that components that are near to each
other in the topographic representation are relatively
strongly dependent in the sense of higher-order correlations or mutual information. The proposed model is usually known as the Topographic ICA model. The
topography is introduced using a neighbourhood function
h(i, k), which expresses the proximity between the ith and
the kth component. A simple neighbourhood model can
be the following:
1; if ji kj 6 L
hi; k
17
0; otherwise
where L denes the width of the neighbourhood. Consequently, the estimated coecients ui are no longer assumed
independent, but can be modelled by some generative random variables dk, fi that are controlled by the neighbourhood function and shaped by a non-linearity /() (similar
to the one in the FastICA algorithm). The topographic
source model, proposed by Hyvarinen et al. [9], is the
following:
!
K
X
ui /
hi; kd k fi
18
k1
ai gEfzaTi zri g;
T
AA A
0:5
16i6K
19
20
where g denes the learning rate of the gradient optimisation scheme and
ri
K
X
k1
hi; k/
K
X
135
!
2
hj; kaTi z
21
j1
136
Estimated
Transform
Transformed
Images
uk t TfI k tg AI k t
22
23
Optional
Denoising
T{}
Fused Image
T{}
Fusion
rule
T -1{}
u k(t)
T{}
Fig. 3. The proposed fusion system using ICA/topographical ICA bases.
137
edges. In contrast, small values for Ek(t) denote the existence of almost constant background in the frame. Using
this idea, we can segment the image in two regions: (i)
active regions containing details and (ii) non-active
regions containing background information. The threshold
that will be used to characterise a region as active or
non-active can be set heuristically to 2meant{Ek(t)}.
Since we are not interested in creating the most accurate
edge-detector, we can allow some tolerance around the real
edges of the image. As a result, we form the following segmentation map mk(t) from each input image:
1; if Ek t > 2meant fEk tg
mk t
27
0; otherwise
The segmentation map of each input image is combined to
form a single segmentation map, using the logical OR operator. As mentioned earlier, we are not interested in forming
a very accurate edge detection map, but instead we have to
ensure that our segmentation map contains all the edge
information
mt ORfm1 t; m2 t; . . . ; mT tg
28
k1
P
There might be some cases, where Tk1 Ek t is very small,
denoting small energy activity in the corresponding patch.
As this can cause numerical instability, we can use the
max-abs or mean fusion rule for those patches.
4.2. Region-based image fusion using ICA bases
In this section, we will use the analysis of the input
images in the estimated ICA domain to perform some regional segmentation and then we will fuse these regions
using dierent rules, i.e. perform region-based image fusion.
During, the proposed analysis methodology, we have
already divided the image in small N N patches (i.e. regions). Using the splitting/merging philosophy of regionbased segmentation [17], we can nd a criterion to merge
the pixels corresponding to each patch in order to form
contiguous areas of interest.
One could use the energy activity measurement, as introduced by (25), to infer the existence of edges in the corresponding frame. As the ICA bases tend to focus on the
edge information, it is clear that great values for Ek(t), correspond to great activity in the frame, i.e. the existence of
5. Experiments
In this section, we test the performance of the proposed
image fusion schemes based on ICA bases. It is not our
intention to provide an exhaustive comparison of the many
dierent transforms and fusion schemes that exist in literature. Instead, a comparison with fusion schemes using
wavelet packets analysis and the Dual-Tree (Complex)
Wavelet Transform are performed. In these examples we
will test the fusion by absolute maximum (max-abs),
the fusion by averaging (mean), the Weighted Combination (weighted) and the Region-based (Regional) fusion,
where applicable.
We present three experiments, using both articial and
real image data sets. In the rst experiment, we have the
Ground Truth image Igt(x, y), which enable us to perform
numerical evaluation of the fusion schemes. We assume
that the input images Ii(x, y) are processed by the fusion
schemes to create the fused image If(x, y). To evaluate
the schemes performance, we can use the following
Signal-to-Noise Ratio (SNR) expression to compare the
ground truth image with the fused image
138
PP
SNRdB 10 log10 P P
x
2
y I gt x; y
y I gt x; y
I f x; y
29
4rI gt I f mI gt mI f
m2I gt m2I f r2I gt r2I f
30
where
r2I
rIJ
M1 X
M2
X
1
2
Ix; y mI
M 1 M 2 1 x1 y1
M1 X
M2
X
1
Ix; y mI J x; y mJ
M 1 M 2 1 x1 y1
31
32
139
Table 1
Performance comparison of several combinations of transforms and fusion rules in terms of PSNR (dB)/Q0 using the airplane example
Max-abs
Mean
Weighted
Regional
WP (Sym7)
DT-WT
ICA
TopoICA
13.66/0.8247
22.79/0.9853
13.44/0.8178
22.79/0.9853
14.48/0.8609
17.41/0.9565
17.56/0.9531
17.56/0.9533
14.80/0.8739
17.70/0.9580
17.70/0.9547
17.69/0.9549
Fig. 4. Region mask created for the region-based image fusion scheme.
The white areas represent active segments and the black areas nonactive segments.
Fig. 5. Three articially-distorted input images and various fusion results using various transforms and fusion rules.
140
Fig. 6. The Toys data-set demonstrating several out-of-focus examples and various fusion results with various transforms and fusion rules.
141
Fig. 7. Multi-modal image fusion: three images acquired through dierent modality sensors and various fusion results with various transforms and fusion
rules.
cially in the max-abs and weighted schemes. However, it seems that the fused image created using the
ICA and the TopoICA bases looks sharper and less
blurry.
142
6. Conclusion
References
[1] A. Cichocki, S.I. Amari, Adaptive Blind Signal and Image Processing. Learning Algorithms and Applications, John Wiley &
Sons, 2002.
[2] R.R. Coifman, D.L. Donoho, Translation-invariant de-noising,
Technical report, Department of Statistics, Stanford University,
Stanford, California, 1995.
[3] The Image fusion server. Available from: <http://www.imagefusion.org/>.
[4] Lehigh fusion test examples. Available from: <http://www.ece.lehigh.edu/spcrl/>.
[5] A. Goshtasby, 2-D and 3-D Image Registration: For Medical,
Remote Sensing, and Industrial Applications, John Wiley & Sons,
2005.
[6] P. Hill, N. Canagarajah, D. Bull, Image fusion using complex
wavelets, in: Proceedings of the 13th British Machine Vision
Conference, Cardi, UK, 2002.
[7] A. Hyvarinen, Fast and robust xed-point algorithms for independent component analysis, IEEE Transactions on Neural Networks 10
(3) (1999) 626634.
[8] A. Hyvarinen, Survey on independent component analysis, Neural
Computing Surveys 2 (1999) 94128.
[9] A. Hyvarinen, P.O. Hoyer, M. Inki, Topographic independent
component analysis, Neural Computing 13 (2001).
[10] A. Hyvarinen, P.O. Hoyer, E. Oja, Image denoising by sparse code
shrinkage, in: S. Haykin, B. Kosko (Eds.), Intelligent Signal Processing, IEEE Press, 2001.
[11] A. Hyvarinen, J. Karhunen, E. Oja, Independent Component
Analysis, John Wiley & Sons, 2001.
[12] J.J. Lewis, R.J. OCallaghan, S.G. Nikolov, D.R. Bull, C.N.
Canagarajah, Region-based image fusion using complex wavelets,
in: Proceedings of the 7th International Conference on Information
Fusion, Stockholm, Sweden, 2004, pp. 555562.
[13] H. Li, S. Manjunath, S. Mitra, Multisensor image fusion using the
wavelet transform, Graphical Models and Image Processing 57 (3)
(1995) 235245.
[14] S.G. Nikolov, D.R. Bull, C.N. Canagarajah, M. Halliwell, P.N.T.
Wells, Image fusion using a 3-d wavelet transform, in: Proceedings of
the 7th International Conference on Image Processing and its
Applications, 1999, pp. 235239.
[15] G. Piella, A general framework for multiresolution image
fusion: from pixels to regions, Information Fusion 4 (2003) 259
280.
[16] O. Rockinger, T. Fechner, Pixel-level image fusion: the case of image
sequences, SPIE Proceedings 3374 (1998) 378388.
[17] M. Sonka, V. Hlavac, R. Boyle, Image Processing, Analysis
and Machine Vision, second ed., Brooks/Cole Publishing Company,
1999.
[18] A. Toet, Targets and Backgrounds: Characterization and Representation VIII, The International Society for Optical Engineering, 2002,
pp. 118129.
[19] Z. Wang, A.C. Bovik, A universal image quality index, IEEE Signal
Processing Letters 9 (3) (2002) 8184.
Acknowledgements
This work is supported by the Data Information Fusion
project of the Defence Technology Centre, UK. The
authors would like to thank the three anonymous reviewers
for their insightful comments and suggestions.