Tempered Image Detection
Tempered Image Detection
Visualisation
175
rotation, but also partial invariant to changes in the We propose a Reconstruction Similarity test to
illumination and the 3D camera viewpoint. avoid the need for the information of the data set. The
flow chart is shown in Figure 1. We define a non-
2.3. Matching overlapping feature to represent a kind of feature
which the distance to any other features is larger than a
Exact NN search, such as the Hash Table and the threshold (a threshold of 4 pixels is chosen in our
KD-Tree, may bring “curse of dimensionality” into our experiments), and the rest are overlapping features.
image matching system due to the high dimensionality After performing an image matching task between two
of the SIFT-like descriptors. In the standard SIFT, the images, we firstly remove the matched overlapping
Best-Bin-First (BBF) approach [10] is used, which can features in one reference image, and a Delaunay
reduce the search time through maintaining a priority triangulation mesh [11, 12] is built on the non-
queue of bins that are not visited. We adopted a new overlapping features. Then, a piece-wise affine
approximate NN approach called the hybrid SP-Tree deformation is performed on the mesh to reconstruct
[3] which has been shown to outperform other existing image 2 using the materials from image 1. The last step
approaches. The hybrid SP-Tree is considered as a is similarity measurement, in which, features are
combination of the “Defeatist search” in the non- extracted using 2D-PCA [13], and the similarity is
overlapping nodes and the SP-Tree in the overlapping computed using a cosine classifier.
nodes. The key idea of this approach is to properly set
an overlapping buffer that the child-nodes can share Image 1 Image 2
their points, with which the time cost of back-tracking Image Matching
can be greatly reduced.
Finally, an effective way of eliminating the error Non-overlapping
Feature
matches, such as RANSAC or epipolar geometry Selection
consistent, should not be omitted.
The comparisons of computational time between Delaunay
Triangulation
different image matching algorithms are conducted and
the results are shown in Table 2. The results show that Affine
Reconstruction
the hybrid SP-Tree works more efficiently than the
BBF while maintaining the same accuracy. Image 1 Image 2*
Trim Trim
Table 2. Comparisons of different
image matching algorithms Similarity Measurement
Reconstruction
Similarity
LoG+MSER Proposed
Algorithm SIFT GLOH MSER +BBF Algorithm
Time (sec) 2.312 2.507 1.475 3.144 2.389 Figure 1. Flow chart of
Reconstruction Reconstruction Similarity
Similarity 0.966 0.969 0.911 0.972 0.972
176
(1) (2)
(1) (2)
We tested our algorithm on the image library [8] when the camera is closer to A in image 1 than it in
with three highly recognized methods, namely, SIFT, image 2, accordingly, the camera will be closer to 1B
GLOH and MSER. The experimental result shown in than to 2B. It can be decomposed as follows:
Table 2 demonstrates that algorithm we proposed E[1 A / E[1B | E[ 2 A / E[ 2 B
performs better than others, benefiting from the ® .
efficient searching engine and the fusion of region- ¯ E[1 A E[ 2 A | E[1B E[ 2 B
based and point-based detectors. However, it is more complicated in a real system
due to changes in the camera settings and the
3. Order regularity of scale space viewpoint. More significantly, the durations of certain
features in a scale space are not the same. Some
According to the theory of scale space, the linear features can be detected with large scale variances
scale space mainly reflects the distance between the (stable), while other features only exist in a narrow
camera and the object. In a simplified model, the value range of the scale change (unstable). Thus, there is a
of (denote = log2 as an indicator of the scale space) common phenomenon that a feature from a higher
is proportional to the distance. Assuming object A and scale space has been matched with one from a lower
B simultaneously appeared in image 1 and 2 is scale space, which makes E usually smaller than its
adjacent, the order regularity can be expressed as actual value. Wang et al. investigated the relation
E[1 A E[ 2 A between the two types of extrema, ɮ1 and ɮ2, and the
!0, information content in the scale space [14]. They
E[1B E[ 2 B concluded that the higher the variation rate of the
where E is the average of the scale space indicators in information content is, the lower is the ratio of the
an image patch. Intuitively, as illustrated in Figure 2, stable to unstable features. Therefore, we can
177
compensate the change of E using the information radius should be adopted to avoid the problem (rX and
theory. Owing to the instability of the cardinality of ɮ1 rY are set to two times of rZ).
and ɮ2, we substitute the entropy for the cardinality. Figure 4 demonstrated the use of this rotation
( E[1 A E[ 2 B ) /( E[ 2 A E[1B ) R , expression. The horizontal view angles vary much of
>1, 1 D @ En'1 A ! En'2 A or En'1B ! En'2 B
the two given images, but it is abnormal that Y:x,y of
R® , region A and B remains the same. In other words, the
¯>1 D , 1@ En'1 A En'2 A or En'1B En'2 B change of Y:x,y does not reflect the rotation relative to
where is set to be 0.16 in our experiments. the camera, so the authenticity is questionable.
For instance, as shown in Figure 2, it is easy to find
that the change of EC from images 1 to 2 cannot meet 5. Conclusions and future work
the order regularity with EA and EB. Hence, if image
1 is authentic, we can say that region C in image 2 has In conclusion, a hybrid matching algorithm is
been tampered. proposed in this paper. It has been proven to be very
suitable for scientific image analysis due to its
4. Description for object rotation desirable properties. We also propose the order
regularity of the scale space and the description for
Y object rotation, which can be used as powerful tools
for tampered image detection. In future, the following
three aspects will be further studied.
z Finding reliable correspondences from images of
X a scene taken from arbitrary viewpoints in
different illumination conditions is still a difficult
Z and critical research issue. A potential solution is
to extend the current algorithm, usually designed
for gray level images, to the color space. Thus,
Figure 3. Rotation in a 3D space we are planning to replace gray-level based
detectors by color-based detectors in our
In Section 2.2, the orientation x,y and the gradient algorithm.
magnitude mx,y of a descriptor are computed using the z While most researchers have concentrated on
pixel differences in order to ensure that they are obtaining more reliable matches, the useful
invariant to rotation. information hidden in matched pairs has been
Z : mx , y ( Lx 1, y Lx 1, y ) 2 ( Lx , y 1 Lx , y 1 ) 2 w ignored. An attempt to explore this is made in this
paper, but it is a very small percentage of the total
Z : T x, y tan 1 (( Lx , y 1 Lx , y 1 ) /( Lx 1, y Lx 1, y ) matches. More embedded information can be
where mx,y is weighted by the Gaussian-weighted mined and utilized in image processing.
circular window. z Completing a semi-automatic system for
rZ
2 detecting tampered images is now under
w Exp( ) consideration.
2 (1.5 V ) 2
Z:x,y and Z:mx,y can be treated as an indicator of the Acknowledgements
rotation of an object with respect to the Z-axis as
illustrated in Figure 3. We extend another two This research is supported by the National High-
orthogonal orientations to fully describe the rotation in Tech Research and Development Plan of China (863
a 3D space: Program) under Grant No. 2007AA01Z423, by the
X : mx , y ( Lx , y 1 Lx , y ) 2 ( Lx , y 1 Lx , y ) 2 w' Defense Basic Research Project of the ‘Eleventh Five-
Year-Plan’ of China under Grant No. C10020060355,
X : T x, y tan 1 (( Lx , y 1 Lx , y ) /( L x , y 1 Lx , y )) by the Key Project of Chinese Ministry of Education
under Grant No. 02057 and by the Key Research
Y : mx , y ( Lx1, y Lx , y ) 2 ( Lx1, y Lx , y ) 2 w' Project of the Natural Science Foundation of
.
Y : T x, y tan 1 (( Lx1, y Lx , y ) /( Lx1, y Lx , y )) Chongqing Municipality of China under Grant No.
CSTC2005BA2002, CSTC2007AC2018. In addition,
The absence of the depth information usually the research is conducted at the CIPMAS AR
causes the instability of the two presented orientations. Laboratory at the National University of Singapore.
Therefore, a larger Gaussian-weighted circular window
178
References
[13] J. Yang, D. Zhang, A.F. Frangi, and J.Y. Yang, “Two-
[1] J. Matas, O. Chuma, M. Urbana, and T. Pajdla, “Wide- dimensional PCA: a new approach to appearance-
Baseline Stereo from Maximally Stable Extremal based face representation and recognition”, IEEE
Regions”, Image and Vision Computing, 2004, 22(10), Transactions on Pattern Analysis and Machine
pp.761-767. Intelligence, 2004, 26(1), pp.131-137.
[2] D.G. Lowe, “Distinctive Image Features from Scale- [14] Z.Y. Wang, Z.X. Cheng, and S.J. Tang, “Information
Invariant Keypoints”, International Journal of Measures of Scale-Space based on Visual Characters”,
Computer Vision, 2004, 60(2), pp.91-110. Journal of Image and Graphics, 2005, 10(7), pp.922-
928.
[3] T. Liu, A.W. Moore, A. Gray, and K. Yang, “An
Investigation of Practical Approximate Nearest
Neighbor Algorithms”, Advances in Neural
Information Processing Systems, 2005, pp.825-832.
179