Robust Autonomous Navigation of Unmanned Aerial Vehicles (Uavs) For Warehouses' Inventory Application
Robust Autonomous Navigation of Unmanned Aerial Vehicles (Uavs) For Warehouses' Inventory Application
Authorized licensed use limited to: Zakir Husain College of Engineering & Technology. Downloaded on July 10,2021 at 11:32:00 UTC from IEEE Xplore. Restrictions apply.
244 IEEE ROBOTICS AND AUTOMATION LETTERS, VOL. 5, NO. 1, JANUARY 2020
Authorized licensed use limited to: Zakir Husain College of Engineering & Technology. Downloaded on July 10,2021 at 11:32:00 UTC from IEEE Xplore. Restrictions apply.
KWON et al.: ROBUST AUTONOMOUS NAVIGATION OF UAVs FOR WAREHOUSES’ INVENTORY APPLICATION 245
Fig. 3. The sensing system of our autonomous UAV. Fig. 4. April Tags (in yellow circles) attached to the warehouse rack. Their po-
sition and orientation information is saved in an xml-based tag map beforehand
for the autonomous navigation of the UAV.
B. Hardware Setup
For the purpose of robust indoor autonomous navigation, our
Since this tag map is an xml file, its size is very small
UAV’s sensing system is composed of a 2D laser scanner, a 1D
(under 800 KB per warehouse) when compared with 3D feature
range sensor, three cameras(forward, upward and downward),
map. Also there is no need for piloted flights to build a map
and an IMU(Inertial Measurement Unit) as shown in Fig. 3. A
periodically in contrast to [1].
barcode reader is equipped for the inventory barcode scanning.
The 2D lidar(laser scanner) is Hokuyo UST-10LX with
D. Multi-Sensor Fusion Framework
270 degrees field of view and 10 meters range, which is much
cheaper than 3D lidars. The 1D range sensor(lidar), Terabee We suggest a multi-sensor fusion framework to optimally
TeraRanger One, measures the distance up to 14 meters between estimate the 6D pose of the UAV in the indoor(GPS-denied)
the UAV to the ground, to be used as an altitude measurement environments. The framework is based on the EKF where the
through UAV posture compensation. The forward and upward IMU pose is used as state variables of the time update phase -
cameras, mvBluFox-MLC, are 752 × 480 monochrome cameras prediction - and the other sensors’ data are processed in the
with global shuttering and equipped with flat and wide angle measurement update – correction - phase in a modular way [9]
lenses respectively. The downward camera, oCam-1CNG-U, is as depicted in Fig. 2.
1280 × 960 global shuttering color camera. We use the IMU We assume that the pose of the IMU sensor is equivalent to
sensor embedded within Pixhawk3 FCU. All the cameras and the that of the UAV. The state of the EKF consists of the position
sensors are also low-cost under 400 USD. The barcode reader, vector of the IMU piw in the world frame {W}, its velocity vector
i i
Cognex DM363, is an image-based barcode scanner with the vw , its orientation quaternion qw describing the rotation of IMU
maximum range of 500 mm. with respect to W, gyro bias bω and accelerometer bias ba . The
All the sensor drivers, software modules and algorithms run entire state yields a 16-dimensional state vector x:
on a single board computer, Intel NUC7i7BNH with 16 GB
x = [piw T vw q w bw T ba T ] T .
i T i T
(1)
RAM. We use the custom-made quadrotor frames as the flying
mechanism. The horizontal size of the UAV is 74 cm × 71 cm We derive the following differential equations:
and its weight including battery is 3.2 kg.
ṗiw = vw
i
(2)
i i
C. Concept of Tag Map v̇w = Rw (am − ba − na ) − g (3)
The April Tag [8] (abbreviated to “tag” later), which is shown i 1 i
q̇w = Ω(ωm − bω − nω )qw (4)
in yellow circles of Fig. 4, was devised to precisely calculate its 2
relative 3D positions and 3D orientations with respect to the cam- ḃω = nbω (5)
era frame by means of a computer vision algorithm. Conversely,
the forward camera and the tag recognition algorithm in Fig. 2 ḃa = nba (6)
output relative 6D poses (3D positions and 3D orientations) i
where Rw is a rotation matrix corresponding to quaternion
of the camera with respect to the tag frame. Through simple i
qw , am is a measured acceleration from IMU, Ω(ω) is the
coordinate transformation, we have the absolute 6D pose of the
quaternion-multiplication matrix of ω [10], ωm is a measured
UAV with respect to the world coordinate frame if the poses of
angular velocity from IMU, g is the gravity vector in the world
the tags in the world frame are provided beforehand.
frame, and na , nω , nbω , and nba are Gaussian noises. Equations
Prior to the operation, we build an xml-based map storing
(5) and (6) mean that the bias terms are modeled as random
each pose of tags in the world frame, which are attached to the
walk. Taking the expectations of differential equations (2)∼(6),
predetermined locations of the every rack as shown in Fig. 4.
and defining the error state vector Δx of (1), we have a linearized
Thus, the absolute 6D pose data of the UAV calculated from tag
error state equation
recognitions are used as one of the measurement update of the
EKF, once tags are observed. Δẋ = AΔx + Bn (7)
Authorized licensed use limited to: Zakir Husain College of Engineering & Technology. Downloaded on July 10,2021 at 11:32:00 UTC from IEEE Xplore. Restrictions apply.
246 IEEE ROBOTICS AND AUTOMATION LETTERS, VOL. 5, NO. 1, JANUARY 2020
Fig. 5. Images from the forward camera showing the recognition of an April
Tag while the UAV is flying horizontally. As the tag is seen far from the image
center, the possibility of wrong detection (i.e., outlier) increases.
Authorized licensed use limited to: Zakir Husain College of Engineering & Technology. Downloaded on July 10,2021 at 11:32:00 UTC from IEEE Xplore. Restrictions apply.
KWON et al.: ROBUST AUTONOMOUS NAVIGATION OF UAVs FOR WAREHOUSES’ INVENTORY APPLICATION 247
Fig. 7. Images from the upward camera showing the ORB SLAM2 results.
The numbers of matching points in motion model are (a) 288 and (b) 47. As Fig. 9. Images from the downward camera showing the results of lane
the number decreases, the uncertainty of pose estimation tends to increase. recognition.
Therefore, we use this number to calculate the pseudo-covariance.
ratio of the physical gap of both lanes to their gap on the image
plane. The latter is used as yaw measurements of the UAV.
These two absolute measurements are very helpful to remove
the influences of drift of the estimation. This lane recognition is
especially effective in flights at the 2nd and 3rd floor altitude and
in the middle of corridors.
Authorized licensed use limited to: Zakir Husain College of Engineering & Technology. Downloaded on July 10,2021 at 11:32:00 UTC from IEEE Xplore. Restrictions apply.
248 IEEE ROBOTICS AND AUTOMATION LETTERS, VOL. 5, NO. 1, JANUARY 2020
Fig. 10. An exemplar motion demonstrating the improvement of using the robust data fusion methods. While the previous UAV had shown a large disturbance
motion approaching to the rack dangerously (the 2nd red circle), the current UAV maintains the center path safely. The yellow dotted lines are drawn along the
vertical image center line for the visual understanding.
Fig. 11. A variety of tasks for inventory inspection conducted by our UAV system at the warehouse. For the easy visual understanding, we reconstruct the voxel
map [14], [15] and the path of the UAV from the data gathered in flights. The colors of the voxels reflect their altitude, and the black trajectory is the UAV’s
estimated pose. Note that, since a 2D laser scanner is used for environmental mapping, the 3D voxel map is created only within the laser scanner’s field of view.
and 6 floors high, and has 180 bins for pallets. Full inspection task motions done by our autonomous UAV targeting to bins at
means that all the bins of the racks are scanned at once. Common the 3rd floor. The UAV may start cycle counting tasks by taking
warehouses do full inspection periodically for complete inven- off at a place within the corridor in front of the target rack as in
tory management. Fig. 11(a) depicts a full inspection task of a Fig. 11(b), and at a home position outside of the aisles to reach
single rack performed by our autonomous UAV. In Fig. 11, for multiple racks in turn as shown in Fig. 11(c).
the intuitive understanding, we reconstruct the 3D voxel map With our autonomous UAV, we have been doing both full
[14], [15] and the path of the UAV from the data gathered in inspection and cycle counting tasks successfully in the materials
real autonomous flights. The colors of the voxels reflect indicate warehouse without any problems for one year, replacing the con-
their altitude, and the black trajectory is the estimated pose of ventional manual inventory inspection done by human workers.
the UAV. Since the environmental mapping is done by the 2D Fig. 12 shows several snapshots captured during a variety of the
laser scanner, the 3D voxel map is created only within the laser inventory inspection motions of our autonomous UAVs. Prior
scanner’s field of view. manual inspection tasks typically required high-reach forklifts,
The other category is cycle counting which is to count items its licensed driver, and one or two workers on board holding a
in specific bins within the warehouse without having to count hand-held barcode scanner. The application of the autonomous
the entire inventory. Fig. 11(b) and (c) describe cycle counting UAV enabled the warehouse to save such large equipment costs
Authorized licensed use limited to: Zakir Husain College of Engineering & Technology. Downloaded on July 10,2021 at 11:32:00 UTC from IEEE Xplore. Restrictions apply.
KWON et al.: ROBUST AUTONOMOUS NAVIGATION OF UAVs FOR WAREHOUSES’ INVENTORY APPLICATION 249
Fig. 12. Photo captures of our UAV system while doing various motions for the inventory inspection tasks at the warehouse.
and human labors efficiently. The inspection time is also reduced [3] DroneScan, 2019. [Online]. Available: www.dronescan.co
since it takes only 9 minutes and 14 seconds to complete a [4] L. Campos-Macias et al., “Autonomous navigation of MAVs in unknown
cluttered environments.” [Online]. Available: https://arxiv.org/abs/1906.
full inspection of a single rack containing 180 bins. One of 08839
the most important consequences of our autonomous UAV is [5] E. Welburn et al., “A navigational system for quadcopter remote inspection
to get rid of the risk of human workers’ unpredicted fatal injury of offshore substations,” in Proc. 15th Int. Conf. Autonomic Auton. Syst.,
2019.
while doing barcode scanning on a vertically moving high-reach [6] I. Kalinov, E. Safronov, R. Agishev, M. Kurenkov, and D. Tset-
forklift. serukou, “High-precision UAV localization system for landing on a mo-
bile collaborative robot based on an IR marker pattern recognition,” in
Proc. IEEE 89th Veh. Technol. Conf., Kuala Lumpur, Malaysia, 2019,
VI. CONCLUSION pp. 1–6.
[7] S. Kohlbrecher et al., “Hector open source modules for autonomous
We proposed an autonomous UAV with a low-cost sensing mapping and navigation with rescue robots,” in Robot Soccer World Cup.
system and multi-sensor fusion framework to be used effectively Berlin, Germany: Springer, 2013, pp. 623–631.
for narrow and dark warehouse environments. To address the [8] J. Wang and E. Olson, “AprilTag 2: Efficient and robust fiducial detection,”
in Proc. IEEE/RSJ Int.Conf. Intell. Robots Syst., 2016, pp. 4193–4198.
problems of naïve UAV localization methods, we suggested [9] S. Lynen, M. W. Achtelik, S. Weiss, M. Chli, and R. Siegwart, “A robust
robust data fusion methods: outlier rejection using component- and modular multi-sensor fusion approach applied to MAV navigation,”
based Mahalanobis norm test, incorporation of visual SLAM by in Proc. IEEE/RSJ Int.Conf. Intell. Robots Syst., Tokyo, Japan, 2013,
pp. 3923–3929.
introducing pseudo-covariance, and recognition of floor lanes [10] N. Trawny and S. Roumeliotis, “Indirect Kalman filter for 3D attitude
for absolute lateral position and yaw measurements. These allow estimation,” Dept. Comput. Sci. Eng., Univ. Minnesota, Minneapolis, MN,
us to perform safe and fully autonomous flights for the cyclic USA, Tech. Rep. 2005-002, 2005.
[11] S. Weiss, “Vision based navigation for micro helicopters,” Ph.D. disser-
inventory inspections at our materials warehouse. As a fully tation, Autonomous Syst. Lab, Dept. Mech. Process Eng., ETH Zürich,
autonomous system, our method does not need piloted flight, Zürich, Switzerland, 2012.
and also reduces the cost, time and the risk of human’s high-floor [12] R. Mur-Artal and J. D. Tardós, “ORB-SLAM2: An open-source SLAM
system for monocular, stereo, and RGB-D cameras,” IEEE Trans. Robot.,
tasks. For further study, we plan to improve the system reducing vol. 33, no. 5, pp. 1255–1262, Oct. 2017.
the cost more, and extend our method to various indoor and [13] J. Engel, J. Sturm, and D. Cremers, “Scale-aware navigation of a low-
outdoor applications. cost quadrocopter with a monocular camera,” Robot. Auton. Syst. vol. 62,
pp. 1646–1656, 2014.
[14] H. Oleynikova, Z. Taylor, M. Fehr, R. Siegwart, and J. Nieto, “Voxblox:
REFERENCES Incremental 3D euclidean signed distance fields for on-board MAV plan-
ning,” in Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst., Vancouver, BC,
[1] M. Beul, D. Droeschel, M. Nieuwenhuisen, J. Quenzel, S. Houben, and Canada, 2017, pp. 1366–1373.
S. Behnke, “Fast autonomous flight in warehouses for inventory applica- [15] A Hornung et al., “OctoMap: An efficient probabilistic 3D mapping
tions,” IEEE Robot. Autom. Lett., vol 3, no. 4, pp. 3121–3128, Oct. 2018. framework based on octrees,” Auton. Robots, vol. 34, pp. 189–206, 2013.
[2] EyeSee, 2019. [Online]. Available: eyesee-drone.com [16] Pixhawk, 2019. [Online]. Available: pixhawk.org
Authorized licensed use limited to: Zakir Husain College of Engineering & Technology. Downloaded on July 10,2021 at 11:32:00 UTC from IEEE Xplore. Restrictions apply.