Satellite_imagery_classification_with_LIDAR_data
Satellite_imagery_classification_with_LIDAR_data
net/publication/268378551
CITATIONS READS
17 1,965
2 authors, including:
María C. Alonso
University of Alcalá
69 PUBLICATIONS 1,348 CITATIONS
SEE PROFILE
All content following this page was uploaded by María C. Alonso on 26 May 2015.
M. C. Alonso *, J. A. Malpica
School of Geodesy and Cartography, University of Alcalá, Ap. Correos 20, 28871 Alcalá de Henares, Madrid, Spain –
(mconcepcion.alonso, josea.malpica)@uah.es
Commission VIII, WG 8
KEY WORDS: LIDAR, Satellite Imagery, Classification, Support Vector Machine, Feature Extraction, SPOT5
ABSTRACT:
This paper shows the potential of LIDAR for extracting buildings and other objects from medium resolution satellite imagery. To
that end, the study integrated multispectral and LIDAR elevation data in a single imagery file and then classified it using the Support
Vector Machine. To determine the method’s potential, the study used a SPOT5 satellite from an area situated southeast of Madrid,
Spain. First, with the four multispectral bands and the panchromatic band of the SPOT5 image, a multispectral four bands
pansharpening was performed with Principal Component Analysis. Once integrated, these four pansharpening images and LIDAR
data, were treated as independent multiple band imagery to perform the classification.
Using five classes, a sample of ground truth pixels was taken for training and testing. The study used 10% of the ground truth for
training and the entire ground truth for testing the robustness of the classification with and without LIDAR data. To assess and
compare the classification results numerically, confusion matrices and Receiver Operating Characteristic (ROC) were calculated for
the five classes, for both classifications, with and without LIDAR.
Generally, when using only multispectral imagery, some confusion among classes occurs; for instance, buildings with flat asphalt
roofs represent a separate problem in classification, since they are extremely difficult to discern from roads. This is mostly solved
when integrating LIDAR data to the multispectral imagery. In general, when adding LIDAR, the classification results show a more
realistic and homogeneous distribution of geographic features than those obtained when using multispectral SPOT5 alone.
* Corresponding author.
730
International Archives of the Photogrammetry, Remote Sensing and Spatial Information Science, Volume XXXVIII, Part 8, Kyoto Japan 2010
2. MATERIALS AND STUDY AREA and Mathur, 2004; Melgani and Bruzzone, 2004; Theodoridis
and Koutroumbas, 2003; Perkins et al., 2001). These works led
SPOT5 satellite imagery has been used to determine the us to apply SVM rather than other methods for classification.
method’s potential. Launched in 2002, this satellite captures
panchromatic images with a 2.5 m resolution and multispectral 3. CLASSIFICATION OF SPOT5 WITH LIDAR
images with a 10 m resolution (for the bands R, G, and NIR),
and 20 m (for the MIR band). This experiment captured the With the four multispectral bands and the panchromatic image,
scene at 11:20 a.m. on 12 August 2006; it represents an urban we carried out a multispectral four bands pansharpening
area southeast of Madrid, Spain, on mostly flat terrain. Figure 1 performed with PCA (Figure 2).
(a) shows the study area in a false color SPOT image generated
with bands 1, 2, and 3 for the RGB monitor channels. The SPOT5 multispectral pansharpening images and the
LIDAR data are integrated and treated as independent multiple
band imagery to carry out the classification.
Input
Panchromatic
2.5 m LIDAR 1 m
R, G, NIR, MIR
10, 10, 10, 20 m
Resampling and
Registration
Pansharpening
R, G, NIR, MIR,
R, G, NIR, MIR LIDAR, 2.5 m
(a) 2.5 m
Training sample
Test sample
731
International Archives of the Photogrammetry, Remote Sensing and Spatial Information Science, Volume XXXVIII, Part 8, Kyoto Japan 2010
(b)
Figure 4. Results of the SVM classifications with LIDAR (a) (a) (b) (c)
and without LIDAR (b)
Figure 6. Aerial image for reference (a), results of SVM
Figures 5 (a), 6 (a), and 7 (a) show three aerial images classifications with LIDAR (b), and without LIDAR (c)
representing three details from Figure 4 (a) and Figure 4 (b),
represented with rectangles in Figure 4. Figures 5 (b), 6 (b), and Figure 7 (a) reveals a bullfighting arena (the round object
7 (b) show the classifications of the three details using toward the top of the figure). Figure 7 (b) correctly classifies it
multispectral bands plus LIDAR. Figures 5 (c), 6 (c), and 7 (c) as a building, while it has been mistaken as artificial ground in
represent the classifications of multispectral bands without Figure 7 (c). The surroundings of the bullfighting arena
adding LIDAR. corresponding to parking lots have been detected correctly as
artificial ground in Figure 7 (b) and wrongly classified as
The left upper corner of Figure 5 (a) shows a motorway buildings in Figure 7 (c).
correctly detected in Figure 5 (b) but not in Figure 5 (c), where
732
International Archives of the Photogrammetry, Remote Sensing and Spatial Information Science, Volume XXXVIII, Part 8, Kyoto Japan 2010
(a) (b) (c) Table 4 shows the buildings class with an omission error of
19.59% and a commission error of 30.59% without LIDAR,
Figure 7. Aerial image for reference (a), results of SVM whereas the addition of LIDAR reduces these errors to 3.27%
classifications with LIDAR (b), and without LIDAR (c) and 3.70%, respectively. This is a consequence of introducing
elevation data, which allows building extraction to take place
To assess and compare the classification results numerically, we more accurately, as seen in the comparison between the
calculated confusion matrices for both classifications, with and classifications presented in Figure 4 (a and b). To a lesser
without LIDAR. Table 2 shows the confusion matrix for the degree than with the buildings, we can observe an improvement
classification of Figure 4 (a), while Table 3 shows the confusion in the classification of vegetated areas between high vegetation
matrix corresponding to Figure 4 (b). (trees) and low vegetation (lawns, grass, or shrubs).
Ground Truth Spectral signatures of some urban materials are similar in
Class (Pixels)
A B C D E Total
composition. For instance, the roofs of some buildings and
Trees (A) 2091 0 0 15 0 2106 roads and parking lots are made of asphalt, therefore the
Shrubs (B) 0 958 0 6 0 964 multispectral classifier (without LIDAR) assigns them to the
Natur. g(C) 0 3 2310 39 0 2352 same class; however, when we consider LIDAR elevation data,
Artif. g (D) 0 21 0 2115 109 2245 the classifier discriminates between buildings and artificial
Build. (E) 15 0 0 109 3229 3353 grounds, and these errors disappear.
Total 2106 982 2310 2284 3338 11020
As a measure of performance of supervised classification, we
Table 2. Confusion matrix for classification with the LIDAR use the receiver operating characteristic (ROC) curve (see
band. Figure 8). See also Metz (1978), for additional information.
Overall Accuracy = (10703/11020) ~ 97.1234% and Kappa Many applications use the area under the ROC curve (AUROC)
Coefficient = 0.9630 as a measure of accuracy (Hanley and McNeil, 1982). For the
five classes considered in this work, the AUROCs are bigger
Ground Truth using LIDAR than without it (Table 5).
Class (Pixels)
A B C D E Total
Trees (A) 2010 78 0 23 2 2113 LIDAR
Area under the ROC curve
Shrubs (B) 79 878 0 2 0 959 With Without
Natur. g(C) 17 1 2225 104 166 2513 Trees (A) 0.97784 0.94885
Artif. g (D) 0 25 0 1057 486 1568
Shrubs and lawns (B) 0.99475 0.99405
Build. (E) 0 0 85 1098 2684 3867
Total 2106 982 2310 2284 3338 11020
Natural grounds (C) 0.93395 0.90910
Artificial grounds (D) 0.82263 0.71577
Table 3. Confusion matrix for classification without LIDAR Buildings. (E) 0.90492 0.80629
Overall Accuracy = (8854/11020) ~ 80.3448% and Kappa
Table 5. Area under the ROC curve for each class with LIDAR
Coefficient = 0.7454
and without LIDAR classification
The overall accuracy and the Kappa coefficients are 97.12%
and 0.96 (see Table 2) for the SVM classification with LIDAR
data, shown in Figure 4 (a), while overall accuracy reaches
80.34% and 0.7454 (see Table 3) for the SVM classification
without LIDAR data, as shown in Figure 4 (b). Therefore, a
gain of 16.77% is obtained in merging the LIDAR data with the
multispectral bands for this data set. The confusion matrices
allow us to examine the differences of the classifications in
detail. For all the classes considered, the omission and
commission errors are improved to some degree by the
integration of LIDAR (see Table 4).
733
International Archives of the Photogrammetry, Remote Sensing and Spatial Information Science, Volume XXXVIII, Part 8, Kyoto Japan 2010
REFERENCES
734
International Archives of the Photogrammetry, Remote Sensing and Spatial Information Science, Volume XXXVIII, Part 8, Kyoto Japan 2010
Lee, D.S. and Shan, J., 2003. Combining Lidar Elevation Data
and IKONOS Multispectral Imagery for Coastal Classification
Mapping. Marine Geodesy, an International Journal of Ocean
Surveys, Mapping and Remote Sensing, Vol. 26(1-2), pp.117-
127.
Zeng, Y., Zhang, J., Wang, G. and Lin, Z., 2002. Urban landuse
classification using integrated airborn laser scanning data and
high resolution multispectral imagery, Pecora 15/Land Satellite
Information IV/ISPRS Commssion I/FIEOS.
ACKNOWLEDGEMENTS
735