Disocclusion of 3D LiDAR Point Clouds Using Range Images
Disocclusion of 3D LiDAR Point Clouds Using Range Images
images
Pierre Biasutti, Jean-François Aujol, Mathieu Brédif, Aurélie Bugeau
Figure 1: One result of our proposed method. (left) original point cloud, (center) segmentation, (right) disocclusion.
1
Our first contribution is to provide a fast segmentation tech- leads to a fast segmentation but it might fail when the scale
nique for dense and sparse point clouds to extract full objects of the objects in the scene is too different. Another simpli-
from the scene by leveraging the implicit range image topol- fied model of the point cloud is presented by [31]. The au-
ogy (Figure 1, center). A second contribution is to introduce thors take advantage of the implicit topology of the sensor
a fast and efficient variational method for the disocclusion to represent the point cloud as a 2-dimensional range im-
of a point cloud using range image representation while tak- age in order to segment it before performing classification.
ing advantage of an horizontal prior without any knowledge The segmentation is done through a graph-based method
of the color or texture of the represented objects (Figure 1, as the notion of neighborhood is easily computable on a
right). 2D image. Although the provided segmentation algorithm
The paper is organized as follows: after a review on related is fast, it suffers from the same issues as geometry-based
state-of-the-art, we detail how the point cloud is turned into algorithms such as over-segmentation or incoherent segmen-
a range image. In section 3, both segmentation and disoc- tation. Moreover, all those categories of segmentation tech-
clusion aspects of the framework are explained. We then niques are not able to treat efficiently both dense and sparse
validate these approaches on different urban LiDAR data. LiDAR point clouds e.g. point clouds aquired with high or
Finally a conclusion and an opening are drawn. low sampling rates compared to the real-world feature sizes.
In this paper, we present a novel simplified model for seg-
mentation based on histograms of depth in range images.
2 Related Works
The growing interest for MMS over the past decade has lead 2.2 Disocclusion
to many works and contributions for solving problems of Disocclusion of a scene has only been scarcely investigated
segmentation and disocclusion. In this part, we present a for 3D point clouds [28, 24, 2]. These methods mostly work
state-of-the-art on both segmentation and disocclusion. on complete point clouds rather than LiDAR point clouds.
This task, also referred to as inpainting, has been much
2.1 Point cloud segmentation more studied in the image processing community. Over the
past decades, various approaches have emerged to solve the
The problem of point cloud segmentation has been exten- problem in different manners. Patch-based methods such as
sively addressed in the past years. Three types of methods [10] (and more recently [8, 21]) have proven their strengths.
have emerged: geometry-based techniques, statistical tech- They have been extended for RGB-D images [7] and to Li-
niques and techniques based on simplified representations of DAR point clouds [13] by considering an implicit topology
the point cloud. in the point cloud. Variational approaches represent an-
other type of inpainting algorithms [9, 5, 29, 3]. They have
Geometry-based segmentation The first well-known been extended to RGB-D images by taking advantage of the
method in this category is region-growing where the point bi-modality of the data [15, 4]. Even if the results of the dis-
cloud is segmented into various geometric shapes based on occlusion are quite satisfying, those models require the point
the neighboring area of each point [20]. Later, techniques cloud to have color information as well as the 3D data. In
that aim at fitting primitives (cones, spheres, planes, cubes this work, we introduce an improvement to a variational
...) in the point cloud using RANSAC [26] have been pro- disocclusion technique by taking advantage of a horizontal
posed. Others look for smooth surfaces [25]. Although those prior.
methods do not need any prior about the number of objects,
they often suffer from over-segmenting the scene and as a re-
sult objects are segmented in several parts. 3 Methodology
The main steps of the proposed framework, from the raw
Statistical segmentation The methods in this category point cloud to the final result, are described in Figure 2.
analyze the point cloud characteristics [12, 30, 6]. They con- We detail each of these steps in this section.
sider different properties of the PCA of the neighborhood of
each point in order to perform a semantic segmentation. It
leads to a good separation of points that belongs to static 3.1 Range maps and sensor topology
and mobile objects, but not to the distinction between dif- The key point of the proposed approach is to work on a
ferent objects of the same class. simplified representation of the point cloud known as a 2D
range map. The acquired dataset simply consisted in a map-
Simplified model for segmentation MMS LiDAR ping of the scene, the range map is obtained using the im-
point clouds typically represent massive amounts of unor- plicit topology of the sensor. The fact that most raw Li-
ganized data that are difficult to handle, different segmen- DAR acquisitions offer an intrinsic 2D sensor topology is
tation approaches based on a simplified representation of the rarely considered. Namely, LiDAR points may obviously
point cloud have been proposed. [23] proposes a method in be ordered along scanlines, yielding the first dimension of
which the point cloud is first turned into a set of voxels the sensor topology, linking each LiDAR pulse to the im-
which are then merged using a variant of the SLIC algo- mediately preceding and succeeding pulses within the same
rithm for super-pixels in 2D images [1]. This representation scanline. For most LiDAR devices, one can also order the
2
a. b.
3
define the direction in which the diffusion should be done to
respect this prior. Note that most of MLS systems provide
georeferenced coordinates of each point that can be used to
define ~η .
We aim at extending the level lines of u along ~η . This can
be expressed as h∇u, ~η i = 0. Therefore, we define the energy
F (u) = 21 (h∇u, ~η i)2 . The disocclusion is then computed
as a solution of the minimization problem inf u F (u). The
2
Figure 5: Example of point cloud segmentation using our gradient of this energy is given by ∇F (u) = −h(∇ u)~η , ~η i =
model on various scenes. −uη~η~ , where uη~η~ stands for the second order derivative of
u with respect to ~η and ∇2 u for the Hessian matrix. The
minimization of F can be done by gradient descent. If we
of the sub-image uR
s as follows: cast it into a continuous framework, we end up with the
P following equation to solve our disocclusion problem:
b × hs (b) ∂u
b∈Csi
Csi = P (1) ∂t − uη
~η~ = 0 in Ω × (0, T )
(4)
hs (b) u(0, x) = u0 (x) in Ω
b∈Csi
using previously mentioned notations. We recall that ∆u =
where b are all bins belonging to class Csi .
The distance uη~η~ +uη~T η~T , where ~η T stands for a unitary vector orthogonal
between two classes Csi and Crj , of two consecutive windows to ~η . Thus, Equation (4) can be seen as an adaptation the
can be defined as follows: Gaussian diffusion equation (3) to respect the diffusion prior
in the direction ~η . Figure 7 shows a comparison between the
d(Csi , Crj ) = |Csi − Crj | (2) original Gaussian diffusion algorithm and our modification.
The Gaussian diffusion leads to an over-smoothing of the
Finally, we can set a threshold such that if d(Csi , Crj ) ≤ scene, creating an aberrant surface whereas our modification
τ , classes Csi and Crj should be merged. Results of this provides a result that is more plausible.
segmentation procedure can be found in Figure 5. We argue The equation proposed in (4) can be solved iteratively.
that the choice of Ws , B and τ mostly depends on the type The number of iterations simply depends on the size of the
of data that is being treated (sparse or dense). For sparse area that needs to be filled in.
point clouds, B has to remain small (e.g. 50) whereas for
dense point clouds, this value can be increased (e.g. 200).
3.4 Range image to 3D point cloud
In practice, we found out that good segmentations may be
obtained on various kind of data by setting Ws = 0.5×B and After the segmentation and the disocclusion, we need to turn
τ = 0.2 × B. Note that the windows are not required to be the range image back to the 3D space. For every point pi
overlapping in most cases, but for very sparse point clouds, of the original point cloud, we define poi and pei respectively
an overlap of 10% is enough to reach good segmentation. the point of emission and the point of echo of pi . We denote
dorig (pi ) the original range of pi and drec (pi ) its range after
3.3 Disocclusion disocclusion. The new coordinates pfi inal of each point can
be obtained using the following formula:
The segmentation technique introduced above provides
drec (pi )
masks for the objects that require disocclusion. As men- pfi inal = poi + (pei − poi ) × (5)
tioned in the beginning, we propose a variational approach d orig (pi )
to the problem of disocclusion of the point cloud. The Gaus- The range image can then be easily turned back to a 3D
sian diffusion algorithm provides a very simple algorithm for point cloud while including the disocclusion.
the disocclusion of objects in 2D images by solving partial
differential equations. This technique is defined as follows:
∂u
∂t − ∆u = 0 in Ω × (0, T ) (3)
u(0, x) = u0 (x) in Ω
4
4 Results
In this part, the results of the segmentation of various ob-
jects and the disocclusion of their background are detailed.
c.
5
a.
a.
b.
b.
6
sical imaging technique that takes the nature of the point
cloud into account (horizontality prior on the 3D embed-
ding), leading to better results. The segmentation step can
be done in streaming any time a new window is acquired,
leading to great speed improvement, constant memory pro-
cessing and the possibility of online processing during the
acquisition. Moreover, our model is designed to work semi-
automatically using very few parameters in reasonable com-
putational time. Finally, we have shown that this work per-
a. b. forms well in various cases, both on dense and sparse point
clouds.
Considering the range image derived from the sensor
topology enabled a simplified formulation of the problem
from having to determine an unknown number of 3D points
to estimating only the 1D depth in the ray directions of a
fixed set of range image pixels. Beyond simplifying drasti-
cally the search space, it also provides directly a reasonable
sampling pattern for the reconstructed point set.
Although the average results of the method are more than
c. d. acceptable, it can underperform in some specific cases. In-
deed, the segmentation step first relies on the good extrac-
tion of non-ground points, which can be tedious when the
data quality is low. Moreover, when the object that needs
to be removed from the scene is hiding complex shapes, the
disocclusion step can fail recovering all the details of the
background and the result ends up being too smooth. This
is likely to happen when disoccluding very large objects.
In the future, we will focus on improving the current
model to perform better reconstruction by taking into ac-
e. f. count the neighborhood of the background of the object to
remove either by using a variational method or by extending
Figure 11: Example of results obtained for the quantitative patch-based method.
experiment. (a) is the original point cloud (ground truth),
(b) the artificial occlusion in dark, (c) the disocclusion result
with the Gaussian diffusion, (d) the disocclusion using our 6 Acknowledgement
method, (e) the Absolute Difference of the ground truth
against the Gaussian diffusion, (f) the Absolute Difference J-F. Aujol is a member of Institut Universitaire de France.
of the ground truth against our method. Scales are given in This work was funded by the ANR GOTMI (ANR-16-CE33-
meters. 0010-01) grant.
7
[6] M. Brédif, B. Vallet, and B. Ferrand. Distributed [22] N. Paparoditis, J-P. Papelard, B. Cannelle, A. Devaux,
dimensionality-based rendering of LIDAR point clouds. In- B. Soheilian, N. David, and E. Houzay. Stereopolis II: A
ternational Archives of the Photogrammetry, Remote Sens- multi-purpose and multi-sensor 3D mobile mapping system
ing and Spatial Information Sciences, 40, 2015. for street visualisation and 3D metrology. Revue française
de photogrammétrie et de télédétection, 200, 2012.
[7] P. Buyssens, M. Daisy, D. Tschumperlé, and O. Lézoray.
Depth-aware patch-based image disocclusion for virtual view [23] J. Papon, A. Abramov, M. Schoeler, and F. Worgotter.
synthesis. In ACM SIGGRAPH, 2015. Voxel cloud connectivity segmentation-supervoxels for point
clouds. In IEEE Conf. on Computer Vision and Pattern
[8] P. Buyssens, M. Daisy, D. Tschumperlé, and O. Lézo-
Recognition, 2013.
ray. Exemplar-based inpainting: Technical review and new
heuristics for better geometric reconstructions. IEEE Trans. [24] S. Park, X. Guo, H. Shin, and H. Qin. Shape and appearance
on Image Processing, 24, 2015. repair for incomplete point surfaces. In IEEE Int. Conf. on
Computer Vision, volume 2, 2005.
[9] A. Chambolle and T. Pock. A first-order primal-dual al-
gorithm for convex problems with applications to imaging. [25] T. Rabbani, F. Van Den Heuvel, and G. Vosselmann. Seg-
Jour. of Math. Imag. and Vis., 40(1), 2011. mentation of point clouds using smoothness constraint. In-
ternational Archives of the Photogrammetry, Remote Sens-
[10] A. Criminisi, P. Pérez, and K. Toyama. Region filling and ing and Spatial Information Sciences, 36, 2006.
object removal by exemplar-based image inpainting. IEEE
Trans. on Image Processing, 13, 2004. [26] R. Schnabel, R. Wahl, and R. Klein. RANSAC based out-
of-core point-cloud shape detection for city-modeling. Pro-
[11] J. Delon, A. Desolneux, J-L. Lisani, and A. B. Petro. A ceedings of “Terrestrisches Laserscanning", 2007.
nonparametric approach for histogram segmentation. IEEE
Trans. on Image Processing, 16, 2007. [27] A. Serna and B. Marcotegui. Urban accessibility diagnosis
from mobile laser scanning data. ISPRS Journal of Pho-
[12] J. Demantke, C. Mallet, N. David, and B. Vallet. Dimension- togrammetry and Remote Sensing, 84, 2013.
ality based scale selection in 3D LIDAR point clouds. Inter-
[28] A. Sharf, M. Alexa, and D. Cohen-Or. Context-based sur-
national Archives of the Photogrammetry, Remote Sensing
face completion. ACM Trans. on Graphics, 23, 2004.
and Spatial Information Sciences, 38, 2011.
[29] J Weickert. Anisotropic diffusion in image processing, vol-
[13] D. Doria and R. Radke. Filling large holes in lidar data
ume 1. Teubner Stuttgart, 1998.
by inpainting depth gradients. In Int. Conf. on Pattern
Recognition, 2012. [30] M. Weinmann, B. Jutzi, S. Hinz, and C. Mallet. Semantic
point cloud interpretation based on optimal neighborhoods,
[14] S. El-Halawany, A. Moussa, D. Lichti, and N. El-Sheimy. relevant features and efficient classifiers. ISPRS Journal of
Detection of road curb from mobile terrestrial laser scan- Photogrammetry and Remote Sensing, 105, 2015.
ner point cloud. In Proceedings of the ISPRS Workshop on
Laserscanning, Calgary, Canada., volume 2931, 2011. [31] X. Zhu, H. Zhao, Y. Liu, Y. Zhao, and H. Zha. Segmentation
and classification of range image from an intelligent vehicle
[15] D. Ferstl, C. Reinbacher, R. Ranftl, M. Rüther, and in urban environment. In IEEE Trans. on Intelligent Robots
H. Bischof. Image guided depth upsampling using and Systems, 2010.
anisotropic total generalized variation. In IEEE Int. Conf.
on Computer Vision, 2013.
[16] A. Geiger, P. Lenz, C. Stiller, and R. Urtasun. Vision
meets robotics: The KITTI dataset. International Journal
of Robotics Research, 2013.
[17] F. Goulette, F. Nashashibi, I. Abuhadrous, S. Ammoun, and
C. Laurgeau. An integrated on-board laser range sensing
system for on-the-way city and road modelling. Interna-
tional Archives of the Photogrammetry, Remote Sensing and
Spatial Information Sciences, 34, 2006.
[18] A. Hervieu and B. Soheilian. Semi-automatic
road/pavement modeling using mobile laser scanning.
ISPRS Annals of the Photogrammetry, Remote Sensing and
Spatial Information Sciences, 2, 2013.
[19] A. Hervieu, B. Soheilian, and M. Brédif. Road marking ex-
traction using a MODEL&DATA-DRIVEN Rj-Mcmc. IS-
PRS Annals of the Photogrammetry, Remote Sensing and
Spatial Information Sciences, 2, 2015.
[20] J. Huang and C-H. Menq. Automatic data segmentation for
geometric feature extraction from unorganized 3D coordi-
nate points. IEEE Trans. on Robotics and Automation, 17,
2001.
[21] L. Lorenzi, F. Melgani, and G. Mercier. Inpainting strategies
for reconstruction of missing data in VHR images. IEEE
Trans. on Geoscience and Remote Sensing Letters, 8, 2011.