Convolutional Neural Network-Based Real-Time ROV Detection Using Forward-Looking Sonar Image
Convolutional Neural Network-Based Real-Time ROV Detection Using Forward-Looking Sonar Image
Abstract—Agent system is strategy to enhance the underwater make underwater maps or investigate the particular underwater
manipulation. The conventional manipulation is generally robot resources. The underwater robots’ control theory, navigation
arm-based configuration which has singular points. On the other methods and sensing techniques are continuously studied by
hand, the agent system is an armless manipulation that the agent many researchers. The underwater manipulation issue is also
vehicle works as the end-effector. If the location of the agent can rising field for some physical operations.
be measured, the end effector is able to be place to any position.
To implement this system, the method of an agent vehicle Underwater robots have some ways of underwater
localization is proposed. The method uses the sonar images of manipulation. The robot arm method is the most widely used
moving agent obtained by forward-looking sonar. To detect the tools [1]. The AUVs or ROVs are equipped with a robotic arm
location of the agent in the sonar images, the convolutional neural and conduct some underwater physical works. It can precisely
network is applied. We applied the state-of-art object-detection control the joint angle. However, it has limitation of moving,
algorithm to the agent vehicle system. The fast object-detection because joint structure which is called singular point takes heavy
algorithm based on neural network can fulfil the real-time volume and weight [2]. For this reason, the armless
detection and show the remarkable validity. It means the manipulation method is developed.
underwater robot can begin navigation under its feed-back.
Through field experiment, we confirm the proposed method can The small ROV is a deployable and maneuverable as end-
detect and track the agent in the successive sonar images. effector robot which is strap-on main AUV [3]. The ROV with
long tether is free from heavy battery and conduct many
Keywords—armless manipulation; agent vehicle; convolutional missions without particular embedded-intelligences. The AUVs
neural network; object detection; forward-looking sonar; sonar have extensive action radius related to freedom of tether, and
image processing. they can use variety of instruments and sensors. The small ROV
is beneficial to detail operations. It can perform armless
I. INTRODUCTION
underwater manipulation or execution of agent docking. It can
Exploring the deep sea is a fascinating field in that it has play a part of manipulation hand to grip something or do
unknown environments. The autonomous underwater vehicles precisely controlling works. We call its small ROV as “Agent
(AUVs) and the remotely operated vehicles (ROVs) were vehicle” and its system as ‘Agent vehicle system’ (Fig. 1). For
created for exploring deep seas where humans cannot explore this system, we should detect agent’s localization by its position
directly, and this technology developed rapidly over the past few sensor and main AUV’s forward looking sonar. The precise
decades. These robots collect a variety of sensor data and it can location data can lead to accurate manipulation.
In this study, we proposed neural network based real-time
object-detection for localization of the agent vehicle. The state-
of-art and fast object-detection algorithm You Only Look Once
(YOLO) shows the high-speed and exact detection [4]. We
conducted this algorithm to our forward-looking sonar data. As
a result, we found the possibility of using it as input of feed-back
control.
II. BACKGROUND
A. Forward-looking Sonar
The forward-looking sonar obtains the acoustic video images
in real-time [5]. It has longer visual range than that of optical
imaging, so it is a prospective solution for underwater object
detection. However, the image quality of the forward-looking
Fig. 1. The agent vehicle manipulation system. sonar is lower than that of optical images. Because of the
Fig. 3. The YOLO algorithm structure conducting our custom data-set [4].