Agricultural Robotics for Field Operations
Agricultural Robotics for Field Operations
3 Wageningen University and Research, Wageningen Plant Research, PO Box 644, 6700AP Wageningen,
Abstract: Modern agriculture is related to a revolution that occurred in a large group of technologies
(e.g., informatics, sensors, navigation) within the last decades. In crop production systems, there are
field operations that are quite labour-intensive either due to their complexity or because of the fact
that they are connected to sensitive plants/edible product interaction, or because of the
repetitiveness they require throughout a crop production cycle. These are the key factors for the
development of agricultural robots. In this paper, a systematic review of the literature has been
conducted on research and commercial agricultural robotics used in crop field operations. This
study underlined that the most explored robotic systems were related to harvesting and weeding,
while the less studied were the disease detection and seeding robots. The optimization and further
development of agricultural robotics are vital, and should be evolved by producing faster
processing algorithms, better communication between the robotic platforms and the implements,
and advanced sensing systems.
1. Introduction
Major technological advancements in agriculture have drastically transformed several
processes, both in crop and livestock production systems, during the last decades. These
advancements are mainly related to the minimization of operational and production costs, reduction
of environmental impact and optimization of the overall production cycle. Focusing on crop
production, a series of optimization models and software tools have been developed so far on field-
operation level. This progress, in parallel with the technological advancements and equipment in
field machinery, has provided radical solutions to several challenges that modern farmers face. In
crop production systems, one of the most significant issues is connected to human labour-intensive
operations. These are, mainly, field tasks (such as sensitive fruits harvesting and intra-row weed
control) that are more difficult to be executed by traditional field machinery, and human workers are
employed. This has brought the increased need for autonomous tractors and robotic platforms to be
used in the crop field operations, currently developed at research stage [1].
Field operations in agriculture are quite complex, and various issues should be addressed to
allow an effective transition towards the robotics era. To build a robotic solution, an overall system
analysis of the field operation should be conducted, together with a cost–benefit analysis [2,3]. Such
a system should comply with very specific requirements, such as lightweight, small size, autonomy,
intelligence, communication, safety, and adaptability, to execute the potential task effectively [4]. The
relatively smaller size of autonomous machines, compared to conventional tractors and implements,
contributes to the reduction of the soil-related issues and in particular soil erosion and soil
compaction, that was provoked by large and heavy modern farm machinery [5]. To automate these
field operations, a decomposition of these tasks should be performed to transform them into discrete
robotic functionalities. This can be achieved by classifying the agricultural tasks into deterministic
(tasks that can be designed and optimized, in advance) and reactive (tasks that are associated with
behaviors that should deal with unexpected conditions) [6].
The main challenges that agricultural robots face are associated with both universal and task-
specific issues. Universal issues are related to terrain assessment [7,8], route planning [9,10], safety
issues, especially focusing on human detection [11], and fleet of robots [11–13]. Task-specific issues
are related to specifications related to crop architecture, crop or pests detection and classification, and
precise application of inputs. Most of these issues are related to the vision system, the robotic
actuation system, the navigation system in semi-structured agricultural environments, and the
intelligence to control both the robotic platform and the implement.
There have been several reviews focusing on agricultural robots for field operations based on
specific tasks (such as agricultural navigation, transplanting/seeding, pruning, weed control,
harvesting, traceability and multi-robot interactions) and systems assessment [14]. One study
reviewed a group of field operations applied only in open arable farming systems [15], while the
digital perspective of mainly commercial agricultural robots has been reviewed as well [16]. Finally,
a short review based on the methodological approach, the field operation and the farming
environment has been presented, under the demonstration of an emulation tool for agri-robotics [17].
As weed control robots are among the most developed agricultural robots so far, a review has been
conducted regarding only autonomous weed control systems [18], while there has also been a crop-
specific literature review about robotic systems applied to strawberries [19]. Even though the
scientific contribution of these reviews is highly important, the contribution of this study is to provide
a holistic review of agricultural robotic systems that execute all the main field operations, for all
cropping systems and in both open-field and greenhouse environments.
The aim of this paper was to study the specific issues related to main field operations in crop
production. More specifically, the main challenges leading to the concept of the current review are:
(i) to identify and emphasize the currently existing robotic systems developed for each field
operation, regarding different crop production systems; (ii) to provoke further development and
optimization of existing agricultural robotic systems in general and, more specifically, the
development of smarter perception and actuation systems; and (iii) to propose the key challenges
and directions for future agricultural robotic developments.
This review paper covers both scientific and commercial agricultural robotic systems applied in
various field operations (such as weeding, seeding, disease detection, crop scouting, spraying and
harvesting) for different farming environments (arable, greenhouses, vineyards and orchards). The
structure of the present work is as follows: the followed framework in this review is showcased in
Section 2. In Section 3, the operational classification and analysis are both presented and allocated in
the main field operations sub-sections. Finally, the overall discussion of the review is shown in
Section 4, followed by Section 5 where general conclusions are underlined.
2. Review Framework
In the pre-processing stage of this review, a series of theoretical considerations were taken into
account. Throughout the analysis, a series of steps were established, as presented in Figure 1.: (i)
development of the review protocol regarding the eligibility criteria; (ii) search for research studies
and commercial resources, and selection of the ones satisfying the eligibility criteria; (iii) definition
of the classification framework to be applied in the literature review, in order to classify the material
and build the structure; (iv) selection of studies to be included in each category of the classification
framework; (v) analysis of the selected studies; and (vi) representation of results by comparing
studies.
Sensors 2020, 20, 2672 3 of 27
Regarding the eligibility criteria, the studies that are related to robotic systems applied in crop
production systems have been included as a wide boundary of this study. More specifically, the
included references should be related to robotic systems that conduct specific field operation(s) in
crop production systems in agriculture. The referenced robotic systems may have been developed in
various environments (i.e., open-field, greenhouses, laboratory, etc.). Hence, they should be
applicable to any in-field operation, excluding out-of-field transportation operations.
In addition, a number of literature-related eligibility criteria are set to be followed. Those are: (1)
the work should be published in the English language; (2) the included studies should be research
articles published in scientific journals, conference proceedings, or other sources (commercial
websites, research projects, etc.); (3) they should have been published within the last 20 years (without
including the current year), as any technology used prior to this is probably obsolete today.
Publications in journals that are not research articles (such as reviews) are out of the scope of this
review and they are excluded.
The included studies have been withdrawn from open online sources (such as open-access
journals, websites, conference proceedings, etc.) and by a systematic literature review in various
electronic depositories, namely Web of Science, ScienceDirect, Scopus and SpringerLink. The primary
literature research was conducted at the end of 2019. Thus, any newly emerging publication will have
been possibly missed. Due to the nature of the reviewed field, grey literature (technical/research
reports, doctoral dissertations, etc.) have been knowingly permitted by the authors in order to cover
as much as possible in this review.
accounts for more than 40% of the labor effort [20]. Besides being one of the most difficult tasks, it is
one of the most costly, including among other factors significant labor cost, especially in crops that
are highly labor-demanding. Moreover, it is a field operation that has a substantial negative impact
on the environment, as large quantities of herbicides are used to keep the field weed-free, to promote
plant health and yield. As an example, global estimated loss potential in rice, wheat and maize
indicates that weeds account for 46.2–61.5% of potential yield losses, and 27.3–33.7% of actual yield
losses caused by all pests together [21]. For all the above-mentioned reasons, significant attention has
been given to weeding robots both by private companies and academia. A number of such weed
robots have been presented in a short review [22], while in the current study a wider review has been
conducted regarding the major field operations, including weeding robots.
Traditionally, there have been two approaches for weeding; namely, mechanical weeding, and
chemical weeding. Mechanical weeding describes the task of destroying weed plants, mainly by
plucking them out, burning them or cutting them. For each of the three above-mentioned methods of
mechanical weeding, a variety of tools are available. In order to remove the weeds by mechanical
means, the system should first detect the row that it will work on, and then cut or pluck out the
weeds. The precision in row detection recorded with mechanical weeding robots was less than 25
mm [23,24], followed by other systems with precision less than 6 mm, [25,26] or higher, up to 3 cm
[27,28]. In addition, the performance of the weed removal (expressed also as weeding efficiency) has
been evaluated in systems that presented rates higher than 90% [24,29], and 65% and 82%—for two
different presented methods [30]. Apart from those, there are a series of mainly commercial
mechanical weeding robots that have no performance metrics provided [31–34].
On the other hand, chemical weeding refers to spraying the weeds with herbicides that have a
toxic effect, thus eliminating the weeds, and it has been popular because it requires considerably less
labor. Herbicides manufacturers have also developed active ingredients that are only toxic to weeds
and not crops, allowing for extensive spraying over the entire field. However, a number of chemical
substances, especially in the EU, have already been restricted in their use, or are going to be restricted
in the near future, due to the negative impact on the environment and human health. In addition,
weeds have already developed chemical resistances, making them harder to eradicate. The main
solution proposed to reduce the pollution caused by herbicides and delay the appearance of resistant
weeds is selective/spot spraying, spraying only the weed and not the entire field. This would require
even more working hours to be spent on weeding. On the bright side, recent developments in
computer vision technologies have allowed manufacturers to automate that process, and integrate
high-level software into spraying/weeding robots.
Similar to mechanical weeding robots, the effectiveness of the system for spotting weeds is
among the most critical performance metrics for the chemical weeding robots. By using a Drop on
Demand (DoD) system—a system that detects the weeds within the plant row and selectively shoots
droplets of herbicide onto those weed leaves—100% effectiveness was achieved, which is the highest
among chemical robots [22,35]. In parallel, 98% and 89% detection accuracy were presented in a
double experiment of a different vision-based weeding robot [36]. Finally, effectiveness (weed
detection and destruction) of more than 85% under ideal conditions is underlined by the EcoRobotix
weeding robot [37]. While a series of commercial chemical weeding robotic systems are available,
almost no performance metrics exist [38–41]. The perception sensors used for weeding offer similar
levels of performance. All robots make use of a camera, where RGB (Red–Green–Blue), infrared (IR)
and web cameras are the most commonly used. There also exist sensors that allow the robot to
identify its surroundings and its relative position to the weeds, using acoustic distance sensors,
gyroscopes, laser range finders, Inertial Measurement Unit (IMU), etc. Table 1 outlines the most
significant chemical and mechanical weeding robotic systems with their main characteristics in terms
of perception sensors, level of weed detection, control method and performance rates (where they
exist).
Sensors 2020, 20, 2672 5 of 27
To conclude, weeding robots have already been under development over the last years, and
there are also a number of commercial robots on the market. Most of the major problems related to
the existing weeding robots are being dealt with—mainly, the use of recent advancements in
computer science (e.g., deep learning) and the integration of more accurate navigation systems (e.g.,
RTK GPS). What remains to be done is to improve their performance, increase working area and
working speed, and increase the accurate detection of weeds for selective spot spraying. There are a
few key issues that should be taken into account; i.e., how the existing robotic systems could perform
on crops with similar operational requirements, to expand their potential use and the overall
development of weeding robots in crops with low competitiveness towards weeds, such as specific
vegetables. In most weeding robots, the vision sensors are currently used in combination with
distance identification sensors. This, though, could be potentially enhanced by even higher resolution
weed detection sensors for selective spot spraying. Generally, existing robots focus on above-ground
weed detection. Possible solutions for future weeding robots may include the detection of weeds in
their primary growth, when the weed sprouts are still below the ground. Such systems could be
equipped with sensors measuring, for instance, the soil electric conductivity.
3.2. Seeding
Seeding is one of the fundamental tasks in a crop production cycle and contributes significantly
to the labor cost. Seeding machinery has evolved to offer precision seeding solutions. Since recent
publications have shown the importance of plant densities to a high yield [43], precision seeding has
gained more interest in the scientific community, which as a result has led to the development of
several seeding robots in recent years (Table 2).
The main performance metrics presented in the existing literature are connected either to
accuracy rates (%), regarding the number of seeds that have been accurately placed, or to position
errors. The main targeted crops for the existing reported seeding robots are mainly cereals, with high
accuracy rates reported regarding wheat and rice crops [44,45]. On the other hand, an autonomous
Sensors 2020, 20, 2672 6 of 27
seeding robot more suitable for wet-fields has been proposed, but with no performance metrics [46].
Finally, there has been progress in commercial seeding robots made by Fendt [47,48], which can be
used as an intelligent fleet of seeding robots or individual units. The most important information for
seeding is the depth and the position that the seed is going to be planted. As a result, force, pressure,
angle and displacement sensors are used in combination with wheel encoders, compasses and
displacement sensors.
Cited
Crop Perception Results
Work
Force sensor, displacement sensor, The path tracking errors are +/− 5 cm and the
Wheat [49]
angle sensor angle errors are about zero.
Signal sensor, angle, pressure &
Wheat Qualified rate: 93.3% [44]
infrared sensors
92% accuracy & 5 cm error in the dropping
Rice Compass, wheel encoder [45]
position.
In conclusion, the last years’ seeding robots have been developed for seeding both at
com0mercial and research level, but to a limited number due to the complexity of the operation. Even
though the current advancements in sensors play a crucial role in facing the issues related to these
robots, they should be optimized further by minimizing the errors and maximizing the accuracy
rates. There are a few key challenges identified for seeding robots, such as the working speed of the
robot, as it is one of the main factors for precision seeding, making its calibration crucial for each
robot. Moreover, the need to improve speed and accuracy, with better vision systems. Better image
detection systems that use more advanced algorithms are already under development [50]. Before
seeding, such vision systems should be able to identify the soil surface quality and structure, the
existence of any weeds, and possible obstacles that may occur. After seeding, the seeding
performance should be evaluated, especially in fields with an uneven surface.
detection on tomato and rice plants with accuracy of about 94% [55]. Moreover, eAGROBOT was
used to detect a series of diseases, and presented about 90% accuracy for both cotton plants and
groundnut plants [56].
Table 3. Disease and insect detection agrirobotic systems features for various crops.
Highest Cited
Crop Perception Detected Disease
Accuracy Work
RGB camera, multispectral camera, Powdery mildew & 95% &
Bell pepper [51,52]
laser sensor tomato spotted wilt virus 90%
Cotton (Bacterial blight,
Cotton, magnesium deficiency), ~90%, 83–
RGB camera [56]
groundnut groundnut (leaf spot & 96%
anthracnose)
Two DSLR cameras (one in BNDVI
mode), a multispectral camera, a
Xylella fastidiosa
Olive tree hyperspectral system in visible and NIR N/A [53]
bacterium
range, a thermal camera, LiDAR, an
IMU sensor(*)
Tomato, rice RGB camera Pyralidae insect 94.3% [55]
Strawberry RGB camera Powdery mildew 72%-95% [54]
* DSLR: Digital Single-lens Reflex. BNDVI: Blue Normalised Difference Vegetation Index. NIR: near-infrared.
( )
To sum up, disease detection robot technology is in its infant stages. From what is available at
this point, there are three major problems related to disease detection in robots: (i) the lack of image
databases for each disease in order to train the detection/classification models; (ii) the slow image
processing, especially when using large volumes of images such as hyperspectral; and (iii) the non-
uniform lighting conditions in the field. The lack of available image datasets is gradually being
addressed, as open-access agricultural databases are becoming available [57] and new data synthesis
methods are presented [58]. At the same time, novel lighting systems are also being proposed for the
detection of various diseases such as powdery mildew [54]. As for insect detection, it has not gained
attention due to the constraints related to insect features (such as size, color, shape, etc.). Potential
future vision systems could detect insects by using multiple ultra high definition cameras, that could
identify insects covered among the leaves, underneath or elsewhere on the plant. Even in this case,
such an insect detection robot should include high-performance algorithms, given the high volume
and size of images, in order to perform accurately.
been on high-value crops and how robotic systems can effectively replace humans in monitoring
tasks. In Table 4, various plant monitoring robots are presented.
The majority of crop monitoring robots have been developed specifically for orchard/tree crops
or vineyards. A crop monitoring robot has been experimentally evaluated in laboratory conditions
[62] and semi-structured environments [63], and proved to be highly reliable for the estimation of
orchard canopy volume and vigor, by using a series of sensors. Crop monitoring robots for vineyards
have been widely developed, even though there are no clear quantified performance metrics. Most
of them are commercial crop monitoring robots (such as [64]), or robots that have been developed
under EU-funded projects ([65–67]). In all cases, a series of vision-related sensors are included, such
as various types of cameras (RGB, hyperspectral, multispectral, etc.), and LiDAR sensors are
incorporated on the robots, in order to monitor volume, vegetation, and health of the plants. Given
that vineyards are usually cultivated in hilly or mountainous terrains, a vineyard monitoring robot
deals with a series of challenges related to terrain traversing, mapping, and localization [68].
Similar perception systems have been implemented in crop monitoring robots for crops such as
Canola, which presented maximum measurement error up to 2.5% [59]. In this specific case, the
platform is mounted on a swather. In addition, Earthsense has commercially developed the
Terrasentia crop monitoring robot, suitable for various crops such as corn, soybean, wheat, etc.[69].
More generic crop monitoring robots handle agricultural operations, such as soil sampling [70].
CO2 concentration levels may have a significant effect on crop growth. Under this scope, a hexapod
robot has been developed for the collection of agronomical information, such as soil nutrients and
plant growth. The monitored CO2 concentration levels were up to 2500 ppm [71].
Cited
Crop Perception Results
Work
No specific High-resolution stereo-
Soil sampling. No performance metrics provided. [70]
crop cameras, 3D LiDAR
CO2 gas sensor,
Gas source tracking. CO2 concentration levels up
No specific anemoscope, IR
to 2500 ppm were recorded while the robot was [71]
crop distance measuring
moving at a speed of 2 m/min.
sensor
Canopy volume estimation. The system is
Orchards LiDAR, luxmeter independent of the light conditions, it is highly [62]
reliable and data processing is very fast.
RGB & IR camera, laser
Crop monitoring tasks. No performance metrics
Grapes ranger finder, IMU, [68]
provided.
pressure sensor, etc.
Monitor health status and canopy thickness.
Orchards and
LiDAR, OptRX sensor Terrestrial Laser Scanning (TLS): 2 mm distance [63]
vineyards
accuracy
Ultrasonic sensors,
NDVI sensors, IR Gather phenotypic data. Maximum measurement
Canola [59]
thermometers, RGB error: 2.5%
camera
To conclude, plant vigor monitoring robots face challenges such as illumination, background
separation, and data manipulation. Various lighting conditions due to weather conditions have been
faced so far, by artificial means and by following strict timetables on image collection processing.
Background separation is due to neighbor plants that create occlusions in the visual sensors. This
problem can be faced either by better algorithms, that could capture these visual interpolations, or
by using multiple cameras. The huge amount of data collected is an issue that could be manipulated
by faster processing algorithms and more performant hardware. In the future, plant vigor monitoring
robots could include smarter vision systems, i.e., high definition cameras that will automatically
Sensors 2020, 20, 2672 9 of 27
correct low-light images and apply artificial light where needed, in order to produce uniform image
data.
3.4.2. Phenotyping
Parallel to the development of plant monitoring robots, phenotyping robots have been under
development. Plant monitoring and plant phenotyping have a lot in common, as in both cases, vision
systems are used to gather as much information as possible about the plant/fruit under inspection.
The difference is that phenotyping robots focus more on morphology, plant growth and traits that
contribute to crop performance, while plant monitoring robots focus on the state, health and chemical
composition of the plant. Numerous sensors used in phenotyping robots have been reviewed in [72].
The developed phenotyping prototypes are presented in Table 5, for various levels of autonomy.
Fully autonomous phenotyping robots have been developed in sorghum plant field trials,
presenting stalk detection up to 96% [73]. These robots also measure light penetration, leaf erectness,
leaf necrosis and GRVI (Green-Red Vegetation Index). Similar autonomous phenotyping robots in
sugar beet crop (Mobile and Bettybot) analyze in detail the geometric and colorimetric evolution of
the plants [74]. Bonirob, an autonomous phenotyping robot, uses multisensor data fusion to measure
population density, plant distribution, plant height, stem thickness, etc. [75].
In addition to the autonomous robots, a fixed-site, fully automated phenotyping platform has
been applied in rice, maize and wheat plants [76]. This platform may provide highly accurate
measurements on canopy growth throughout crop lifecycles, including plant height root mean
square (RMS) error up to about 1.9 cm. In sorghum crop, a tractor-mounted autonomous robot
resulted in highly repeatable measurements (such as plant height, width, convex hull volume and
plant surface area) that are quite close to the in-field manual measurements [77].
Not all of the phenotyping robots are fully automated. Two phenotyping robots are able to work
collaboratively in cotton crop, as an autonomous ground vehicle for individual plant data collection,
and as a mobile observation tower that inspects the entire field [78], including RMS (Root mean
square) errors for plant height less than 0.5 cm, RGB to IR image calibrations of 2.5 pixels, and less
than 1 °C for temperature. Finally, another semi-autonomous phenotyping robot has been tested in
energy sorghum crop [79]. The average absolute error achieved by the algorithms used in this robot
was 13% for stem width estimation, and 15% for plant height estimation.
Autonomy Cited
Crop Perception Results
Level Work
Cameras, spectral imaging
Maize &
systems, laser sensors, 3D Autonomous No performance metrics provided [75]
wheat
time-of-flight cameras
Stereo RGB & IR thermal RMS error:
camera, temperature, Plant height: < 0.5 cm (Vinobot)
Semi-
Cotton humidity, light intensity RGB to IR calibration: 2.5 px [78]
autonomous
sensors, pyranometer, (Vinoculer)
quantum, LiDAR Temp: < 1 °C (Vinoculer)
Stereo camera, RGB camera
Sorghum with fish eye lenses, Autonomous Stalk detection: 96% [73]
penetrometer
RGB camera, chlorophyll
Rice, fluorescence camera, NDVI
Fixed site fully
maize & camera, thermal infrared Plant height RMS error: 1.88 cm [76]
automated
wheat camera, hyperspectral
camera, 3D laser scanner
Mobile robot: Webcam
Sugar camera, gigaethernet camera
Autonomous No performance metrics provided [74]
Beet Bettybot: Color camera,
hyperspectral camera
Sensors 2020, 20, 2672 10 of 27
In conclusion, plant vigor monitoring and plant phenotyping are two similar tasks, and therefore
face identical challenges. Illumination is a major problem, as lighting conditions are constantly
changing. So far, the way researchers have dealt with this problem is by using artificial lighting to
ensure uniform illumination, and by collecting their data at the same time of the day [77]. Another
challenge is the separation of the background, mainly due to occlusions caused by neighbor plants.
Various solutions for the above-mentioned problem have been suggested. Some focus on the crop,
for example, cultivating and using wider rows, but the major focus is on the technology used, and
mainly on improving the algorithms [80] used, as this does not require changes in cultivation practice.
The final challenge is to deal with the amount of data generated. Datasets have to be efficiently
processed, the workflow has to be automated, especially when data from different types of sensors
are gathered, and pipelines that translate data to the consumer must be built. Computer scientists are
already introducing better and faster image processing algorithms, and pipelines that could find use
in agricultural applications [81,82].
3.5. Spraying
Besides weed management, where herbicide sprayers are used, spraying robots can be used for
pests (diseases and insects) and also for liquid fertilizers (foliar). As a result, the farmer is exposed to
large quantities of those toxic active ingredients no matter the protective measures that are taken. It
is, therefore, crucial to introduce spraying robots to prevent possible health hazards. Due to the rapid
development of computer vision and artificial intelligence, robotic sprayers feature novel intelligence
systems that enable selective spraying, compared to conventional uniform spraying across the crop.
Robots could make use of such technologies, reducing agriculture’s environmental impact as well as
consumer exposure to pesticides, and preventing the development of resistance to those substances
by the targeted organisms. An overview of the spraying robots can be found in Table 6.
Various approaches have been proposed when designing a spraying robot, mainly due to
restrictions and possibilities of the operating environment. In greenhouses, two approaches have
been identified: robots that move using the greenhouse piping system [83,84], and robots that
navigate between rows without being mounted on those pipes [85–87]. Moreover, the ability to spray
selectively requires an accurate detection system, and therefore advanced sensors need to be
mounted on the robot [88,89]. Finally, the spraying system (nozzles) could be mounted directly on
the spraying platform [83–87,90], or it could be mounted on a robotic manipulator with various
degrees of freedom (DOF). The manipulator used by [91] had 3 DOF, the one used by [89] 6, while
the one from [88] had 9 DOF.
Based on the provided results, the execution time of the spraying operation is presented [86,89],
while in most cases the effectiveness of the system is reported. The effectiveness of the systems is
demonstrated by providing the successful spraying rates, the error of the system [85,87], or both
performance metrics [83,88,90]. Finally, only a few of the systems consider real-time detection of the
potential spots to be sprayed [88–90], while the majority of them do not include real-time detection
[83,85–87,91].
Sensors 2020, 20, 2672 11 of 27
Real-time Cited
Crop Perception Results
Detection Work
NSGA-II execution time was better
Cantaloupe Robot controller No 1.5–7% than NSGA-III for the same [86]
test cases. (*)
27% off-target shots, 99.8% of the
N/A Web camera Yes targets were sprayed by at least one [90]
shot.
Run success:90% & 95%, topside leaf
Bump sensors, infra-red coverage:95% & 90%, underside leaf
Cucumber No [83]
sensors, induction sensors coverage:95 & 80%, over-spray:20% &
10% for Tests 1 & 2
Maximum error varied from 1.2–4.5
cm for self-contained mode on
N/A Ultrasonic sensors No [87]
concrete, while for trailer varied from
2.2–4.9 cm.
Ultrasonic sensor, color TV
Grapevine No No performance metrics. [91]
camera
Middle-range sonar, short- Longitudinal, lateral and orientation
N/A range sonar, radar, No errors are close to zero, by including [85]
compass, radar slip varied 10–30%.
The sensitivity of the robotic selective
RGB camera, R-G-NIR spraying was 85%, the selectivity was
Grapevine Yes [88]
multispectral camera 92% (8% of the total healthy area was
sprayed unnecessarily).
Hyperspectral camera, The greedy sort and raster methods
Vegetable stereo vision, thermal IR are substantially faster than back-to-
Yes [89]
crops camera, monocular color front scheduling, taking only 68% to
camera 77% of the time.
* NSGA: Non-dominated Sorting Genetic Algorithm
( )
It can be concluded from the metrics provided in the reviewed papers that spraying robots
perform reasonably well, with non-selective sprayers achieving coverage of about 92% and selective
sprayers being able to identify and spray 85% of the diseased areas [83]. The speed, however, of
selective sprayers is low, with the total spraying time for every seedling in a hectare cultivated with
vegetable crops being higher than 30.5 hours [89]. Although, it has to be mentioned that the time
needed can be considerably less since not all plants need to be sprayed. Some of the solutions
suggested to decrease the time needed for spraying involve using more spray nozzles or more than
one platform, and increasing the speed of the detection algorithms. All these are directly connected
to the further development of detection sensors, in order to reach even better accuracy rates in the
future.
The majority of the harvesting robots focus on strawberries, a high-value crop, that suffers from
a high production cost mainly due to labor cost, particularly during harvesting [93,94]. For
strawberries, the fastest picking robots were the strawberry harvesting robots in [95] and in [92], with
picking speeds of 7.5 and 8.6 seconds per strawberry, respectively, and the Berry 5 robot from Harvest
Croo, with a claimed picking speed of 8 seconds per fruit [96]. Other strawberry harvesting robots
achieved picking speed up to 10 s/fruit [97] and 11.5 s/fruit [98]. However, having a high picking
speed is not enough, as a high picking rate is also important. Having unharvested fruit means that
the grower will have to manually harvest these fruits, which is not additional labor compared to
manual harvesting, but it is still labor which adds to the total labor cost. The highest picking rate
presented was 86% [64]. A range in picking success rates, from 0% up to 64%, has been shown for
various categories to evaluate robots’ harvesting performance [99]. It is worth mentioning that for
strawberry harvesting, it appears to be a trade-off between picking speed and picking rate, as fast
picking speeds are accompanied by poor picking rates and vice versa. Apart from the
abovementioned robots, there is, also, a wide variety of commercial strawberry harvesting robots
with no performance metrics provided. Dogtooth technologies developed a selective picking robot
for table-top growing systems [100], Agrobot developed the E-series, an autonomous selective
strawberry harvester with 24 robotic arms [101], while Octinion presented Rubion harvesting robot
[102].
Other crops that have gained attention are apples and tomatoes. The apple fruit has attributes
that make it easy to harvest robotically. Red varieties are easy to distinguish among the canopy, while
the hard nature of the fruit makes it easier for grippers to harvest it without damaging it. The fastest
apple harvesting robots have reported picking speeds of 7.5 s per apple [103] and 9 s per apple [104],
while the highest picking rates showcased were 89% [105] and about 90% in high-density apple
orchards [106]. Tomatoes are cultivated worldwide, and in terms of vegetable crop production, the
tomato only falls behind the potato. The fastest tomato harvester had a picking speed of 23 s per
tomato [107], while the most reliable one had a picking rate of 87% [108]. In cherry tomatoes, the
picking speed was 8 s per tomato bunch without including the moving time [109]. Metomotion [110]
and Root-AI [111] have their own selective commercial tomato harvesters, as well.
Similar to apple robotic harvesters, other references regarding various fruits’ robotic harvesters
are included in this study. Citrus harvesters, such as the commercial one by Energid, manage to
collect each orange in 2–3 s [112]. Another fruit that has gained attention is cherries, which are also a
fruit that can be easily detected by robotic harvesters, with picking speeds up to 14 s per fruit having
been shown in cherry orchards [113]. Moreover, Agribot, a manually driven fruit robotic harvester,
needs two seconds to grasp and detach each fruit [114].
Vegetables, in general, represent a promising group of crops in which robotic systems could be
implemented. Considering this, a series of different robotic harvesters have been developed. A
cucumber robot claimed picking rate of 80%, while its average picking speed was 45 s per cucumber
[115]. For eggplants, the picking rate was about 62%, while it needed 64 seconds to harvest an
eggplant [116]. The Sweeper robot, that was an EU-funded project, is a selective sweet peppers
harvesting robot that, achieved an average picking speed of 24 s/pepper, based on the project’s
preliminary results [117]. In asparagus crop, the achieved picking speed was 11.9 s per asparagus,
when the robot harvested two asparagus at the same position, while in the case that there was only
one asparagus to be harvested in each position, the picking speed was 13.7 s/asparagus [118]. Sparter,
an asparagus harvesting robot, is also commercially available by Cerescon [119].
Apart from the aforementioned vegetables and fruits, there are certain vegetables that require
different harvesting principles due to the challenging features they have, e.g., size and weight. As a
result, kinematic analysis and simulation have attracted interest, in order to allow the robotic
harvesting of heavy agricultural products, such as pumpkin and cabbage [120]. Regarding fruits,
melon and watermelon are representative with similar characteristics. For a melon robotic harvester,
the picking rate was about 85% [121], while Stork, that was developed for watermelon harvesting,
achieved a picking rate up to 67% [122].
Sensors 2020, 20, 2672 13 of 27
A global illustration of the overall performance of each reviewed robot, where the performance
metrics are presented, can be found in Figure 2 (multiple robots for each crop). The figure is not
provided for a direct comparison but to get a better understanding of how variation in shape, size
and fruit characteristics, such as firmness, affect the picking rate and picking speed.
To conclude, harvesting robots cannot directly compete with human labor. Harvesting robotic
systems do not perform as intended due to two major problems: localization of the object to be
harvested because leaves, fruits and other parts of the plant obstruct the view of the robot; and
damage to the fruit due to the sensitive nature of agricultural products. In the future, the performance
of harvesting robots should be comparable to those of human workers. In order to do so, they should
be able to deal with clusters of fruit and occlusions (better software), and harvest the product reliably
without damaging it (hardware). Faster and more accurate vision algorithms should be developed,
that work together with soft, conflict-free end effectors, as well as manipulators suitable for trees.
Already, various papers are trying to address the above-mentioned problems, for example, Cornell
University recently published a paper presenting a deep learning method that improves the
Sensors 2020, 20, 2672 15 of 27
performance of a robotic arm end-effector in unstructured and occluded environments [127], and at
the same time there is research ongoing for novel end-effector designs that fit the needs of specific
crops [128]. At this point it is worth mentioning that it is essential for developers to work closely with
horticulturists to come up with suitable tree training methods, that facilitate robotic harvesting and
at the same time provide high yield and high-quality products. Finally, the picking speed which
appears low could be overcome if the robots are able to work for an extensive period of time. Their
overall performance can also be solved if human workers harvests the ends of the field and robots
the inner parts, to increase their speed.
In conclusion, due to the limited number of publications for each task included in this section,
the challenges that have to be dealt with in the future are not clear and solid. However, pruning
robots should be able to deal with specific key challenges related to crop architecture, such as height
and branches diameter. In this light, suitable sensors should be available on the platform; i.e., force
sensors, a vision system, a multi-degrees of freedom end-effector, and, in the background, algorithms
that will decide the optimal pruning per plant. Finally, the key challenges that string twining robots
face are connected mostly to their performance (faster operational speed and success rate), which is
directly connected with the vision system, the end-effector, and the overall platform performance.
Overall, there should be more experimental and research studies in the future regarding plant
management and architecture robots, in order to optimize their sensing parts and software. In this
light, in the current systems, faster execution time cycles could be achieved by adapting higher
precision vision systems and faster algorithms.
Sensors 2020, 20, 2672 16 of 27
Working Cited
Perception Operation(s)
Environment Work
Arable crops N/A Ploughing, seeding [145]
Monitoring, spraying, fertilization,
Arable crops Sonar sensor, temperature [138]
disease detection
Τhree cameras with IR filter, humidity Ploughing, seeding, harvesting,
Arable crops [146]
& ultrasonic sensors spraying
Arable crops, Harrowing, soil sampling,
2D LiDAR, ultrasonic sensors, an RGB
polytunnels, phenotyping, additional tasks by [135]
camera & a monochromatic IR camera
greenhouse combining modules
Arable crops Soil sensors Ploughing, irrigation, seeding [147]
Two color cameras, machine vision Spraying, weeding, additional tasks
Greenhouse [134]
sensor by adding or removing components
Humidity and temperature air sensor, Ploughing, seeding, irrigation,
Arable crops [136]
IR sensors spraying, monitoring
Color camera, temperature, humidity Sowing, irrigation, fumigation,
Urban crops [139]
and luminosity sensors pruning
Seeding, spraying, ploughing,
Arable crops Voltage sensors [148]
mowing
Arable crops 2D LiDAR, ultrasonic sensor, IR camera Seeding, weeding, ploughing [141]
Sowing, sprinkling, weeding,
Arable crops N/A [149]
harvesting
Arable crops CloverCam, RoboWeedCamRT Seeding, weeding [137]
Vineyards N/A Pruning, weeding, mowing [150]
Precision seeding, ridging discs &
Arable crops N/A [151]
mechanical row crop cleaning
To conclude, multipurpose robots perform more than one field operation of the production
cycle, and at this point in time, no solutions have achieved the level of modularity and good
Sensors 2020, 20, 2672 17 of 27
performance needed, as being able to perform multiple tasks adds complexity to hardware and
software. On the bright side, the same solutions that can be used for the individual task robots can
also be applied to the multipurpose ones. Besides, for those problems and solutions that have already
been mentioned in previous chapters, multipurpose robots have to perform adequately/accurately
for each and every task they are developed for. In addition, the ability to fulfill more than one task
increases the cost of the robot, due to complex configurations and the variety of sensors needed. A
way to justify higher costs is by extending their functionalities using relatively cheap sensors to
measure additional parameters.
4. Discussion
The reviewed literature of robotic systems in agricultural field operations was classified
according to major field operations; weeding, seeding, disease and insect detection, crop scouting,
spraying, harvesting, plant management, and multi-purpose robots. An indicative representation of
each category robot is depicted in Figure 3.
Figure 3. Representation of various types of agricultural robots ((1) weeding [23], (2) seeding [47], (3)
disease and insect detection [55], (4) plant monitoring [69], (5) phenotyping [78], (6) spraying [86], (7)
harvesting [100], (8) plant management [152], (9) multi-purpose [137]).
The number of references included in this review was 153 in total. They have been allocated,
according to the referenced sources, to four main categories, namely research journal sources,
conference proceedings, commercial and other sources (such as EU-projects, book chapters, etc). The
research journal studies were retrieved from various journals, such as “Computers and Electronics in
Agriculture”, “Biosystems Engineering”, “Precision Agriculture”, “Journal of Robotics”,
“International Journal of Agricultural and Biological Engineering”, etc. In addition to this, part of
these references have been published in conference proceedings, while others are related to
commercial robots and sources such as international research projects.
Agricultural robots have been recognized as one of the most promising solutions for solving the
challenges agriculture is currently facing. Agricultural robots have been developed for every
operation in the crop production cycle, starting with the establishment of a crop by seeding, and
ending with the harvesting of the final agricultural product. However, the tasks that have attracted
most of the attention are the most labor-intensive ones, namely harvesting and weeding. As most of
the agricultural tasks are, in principal, seasonal, within a specific period of time, the only way to allow
Sensors 2020, 20, 2672 18 of 27
for a quick return of investment (ROI)—which is the most convincing argument for a grower to
invest—is by adding value to the robot by allowing it to perform more than one task, which explains
why multi-purpose robots have been recently brought into the spotlight. The extent to which the ROI
and labor can affect the development of robots is apparent, especially when comparing the number
of commercial robots for each task and the number of robots developed by academia. Companies
focus on the two operations mentioned above, weeding and harvesting, while academia focuses
mostly on crop scouting, weeding, harvesting and multipurpose. Harvesting and multipurpose tasks
usually require a complicated solution that attracts a lot of attention, and therefore provides exposure
of the university or research center to a wider part of the population. All of the above can be clearly
seen in Figure 4.
35
30
25
20
15
10
5
0
In all these operations there are specific issues that have to be tackled. Regarding the
communication issue, which applies in all agricultural robots, the focus is on the way data is
transmitted between the various components or the way this information reaches the user. The most
popular technologies that address this challenge are WiFi [22,37,56,70,78], Ethernet
[63,73,75,83,107,123], CAN-bus [13,22,27,38,75,108,153], and RS-232 [59,63,71,85,134]. For seeding
robots, the optimal working speed according to crop and field features, and the image detection
accuracy, are the key factors that should be further improved. A variety of sensors have been used to
increase precision and speed in the seeding operation, such as RGB cameras, force, displacement and
pressure, infrared sensors, and wheel encoders. However, there is still research to be carried out for
better fusion of sensors, and potentially novel sensors and techniques should be tested, e.g.,
spectroscopy to estimate soil texture for the optimum seed depth and space placement. Weeding
robots (chemical or mechanical) have been widely developed in the last years, and they offer the most
commercially available systems, as weeding is tedious and labor-intensive work, especially for
weeding within the crop row. However, there are optimization issues that should be faced, i.e., robots
provide a small working width and slow working speed, restricting the deployment in large-scale
farms and higher accuracy weed detection for selective spot spraying. So far, weeding robots have
been using, in the vast majority, RGB cameras in combination with a sensor to identify the distance
from the object, such as optical and acoustic distance sensors, laser range finders or stereo vision
systems. Weeding robots should also be assessed and customized to be able to operate in crops with
similar operational requirements.
For disease and insect detection robots, three core issues have been identified: the lack of image
databases; the slow image processing (for large volumes of images); and the non-uniformity of field
lighting conditions. However, these issues are gradually addressed by the use of available open-
Sensors 2020, 20, 2672 19 of 27
access image databases, by applying new methods on data synthesis processing, and by the use of
innovative lighting systems. Similar illumination-related issues have been presented in monitoring
and phenotyping robots, and they are faced by artificial lighting systems. Another challenge for the
monitoring and phenotyping robots is the separation of the background. This issue has been
addressed by using wider rows in the field, or by implementing other technological approaches, such
as imaging of the same plant from multiple angles, and the development of better and faster
algorithms. It is obvious that the amount of generated data should be efficiently handled by suitable
dataset processing, automated workflows and pipelines. Spraying robots have presented similar
issues to disease detection robots. Additional problems are the low spraying tank capacity and the
autonomy of the robot. Even though the working speed may be considerably low, as not all plants
need to be sprayed, the total spraying time of selective sprayers can be lower than conventional non-
selective ones. In addition, selective sprayers can reduce drift and runoff using micro sprayers. In this
light, suggested solutions include more spray nozzles, multiple operating platforms and, of course,
increased speed of detection algorithms. All the above-mentioned task-specific robots use optical
sensors to be able to perform their tasks, mainly RGB cameras, multi-/hyper-spectral cameras, and
LiDAR sensors, therefore as optics technology advances, so will the performance of the robots.
Harvesting robots have gained wide interest globally, even though they should be further
optimized. Two major problems have been underlined: the localization of the object to be harvested
because leaves, fruits and other parts of the plant obstruct the view of the robot; and the potential
damage to the fruit due to the sensitive nature of agricultural products. In order to develop harvesting
robots with performance comparable to human workers, they should be equipped with better
software and hardware; better software to deal with clusters of fruit and occlusions, and better
hardware to operate harvesting reliably, without causing damages to the product and the plant. In
order to harvest the ripe fruit, a vision system is needed to evaluate the maturity, and locate and
grasp the fruit, and as a result, all harvesting robots include a camera (RGB, Time-of-Flight, 3D or
Stereo camera) and an additional sensor to measure distance if the camera cannot provide that, such
as a LiDAR sensor. Moreover, faster and more accurate vision algorithms should be developed, that
work together with soft, conflict-free end effectors, as well as manipulators suitable for trees.
For the plant management robots, due to the limited number of references for each task, the
challenges that have to be dealt with in the future cannot be outlined and summarized. However,
given the challenges that other task-specific robots face, it could be stated that both software (e.g.,
better machine vision algorithms) and hardware (e.g., lightweight platforms and tools) development
are needed in order to improve current robots’ performance. Moreover, further scientific research
and experimental trials should be held, targeting the development of new robotic solutions in other
agricultural tasks where no robotic systems exist.
Multipurpose robots perform more than one field operation, and up to now, there have been no
solutions for achieving the required level of modularity and good performance. Besides those
problems and solutions that have already been mentioned in the single-task robots, multipurpose
robots have to perform adequately/accurately for each task they have been developed for. More tasks
means more complexity hardware-wise and software-wise, and this should be a tech improvement
for future robots. Because of the need to perform more than one task, the sensors used by these
multipurpose robots should have a universal value; this leads to the wide use of optical sensors, such
as IR and RGB cameras, in combination with more task-specific ones, like humidity, temperature and
luminosity sensors. Generally, the ability to fulfill more than one task increases the cost of the robot,
due to complexity of configurations and the variety of components needed. A way to justify the higher
costs is by extending their functionalities and by making them compatible with various groups of crops.
Agricultural robots are associated with high investment costs from the farmer’s point of view.
Consequently, there is a need for new business models in agriculture, such as robotic services
contractors. As has been mentioned above, a way to return the investment fast is by investing in
multi-purpose agricultural robots. In addition to this, another way to enable a quick ROI is to target
high-value crops that pay back the amount invested. Furthermore, the crops that the robot will
operate must require extensive crop care in order to justify replacing humans. As a result, most of the
Sensors 2020, 20, 2672 20 of 27
robots have been developed under this scope, with strawberries and grapes being the most popular
amongst the agro-robotics community (Figure 5).
15
10
Crops
Figure 5. The main crops in correlation with the number of robotic systems.
Agriculture is one of the more unstructured environments, with constant changes, and most
systems are not designed in an adaptive way; therefore it is hard to robotize/automate tasks. Moreover,
only recently have learning systems been able to adapt to new situations, including agriculture. Plant
growth cannot be predicted accurately, obstacles may appear suddenly (animals, rocks, etc.), and
occlusions are a common problem due to the growth of the plant and movement of branches. Finally,
lighting conditions are constantly changing, due to clouds and the day/night cycle. As a result, one
would assume that robots operating in a semi-structured environment, such as greenhouses or orchards
and vineyards, would be developed as front-runners. Nevertheless, that is not the case, as almost half
of the developed robots focus on open field crops (Figure 6). This could be attributed to the fact that the
majority of the crops worldwide are cultivated as open field crops, or to the fact that weeding, which is
one of the most popular robot categories, is mainly an issue in open fields and not greenhouses.
Greenhouse
Open field
1% Orchard
Vineyard
11% Polytunnels
27%
12%
49%
5. Conclusions
Overall, throughout all the reviewed robots, the most-studied robot categories were the
harvesting and weeding robots, while the least-studied were disease detection and seeding robots.
From the studied robots, specific and universal conclusions can be pointed out in order to address
the current key issues and potential future challenges. A general overview of an agricultural robot
includes part or all of the following components: a navigation system; a vision system; a control
system; communication components; the robotic arm with its components; and, of course, a
computer, a safety system, a remote assistance/tele-robotics system (to assist the robot when in failure
mode), an edge/cloud-AI adaptive learning system and a farmer-friendly brilliant simple interface
system. The majority of the robots make use of RTK-GPS for navigation, and LiDAR in order to
reconstruct the environment digitally. Moreover, RGB, stereo and multi- and hyper-spectral cameras
have been widely used, to allow the localization of important environmental objects and features, so
that the robot can perform its task consistently and with high accuracy. Better vision systems are
needed for all operations in parallel with faster image processing methods/algorithms. In particular,
image processing is directly connected with processing methods, communication means, vision
systems and the volume of the images. A key challenge related to the robotic vision systems is to
provide uniform lighting conditions by artificial illumination means. Furthermore, new deep
learning methods could be developed to improve the performance of the robotic arm end-effectors,
and, more specifically, improve the eye-hand coordination of the robot by using learning methods
(such as Grasp Quality Convolutional Neural Network). As for the computer hardware, it should be
able to deal with the huge amount of data quickly, especially where high-resolution images are
needed.
Author Contributions: S.F. contributed in the conceptualization, methodology, writing—review and editing,
supervision, funding acquisition. N.M. contributed in the methodology, investigation, writing—original draft
preparation, I.M. contributed in the methodology, investigation, writing—original draft preparation. E.R.
contributed in the investigation, writing—original draft preparation, writing—review and editing. C.H.S.
contributed to writing—review and editing, funding acquisition. E.P. contributed to writing—review and
editing, funding acquisition. All authors have read and agreed to the published version of the manuscript.
Funding: This research was funded by the European Commission H2020 “AgROBOfarm” project “Business-
Oriented Support to the European Robotics and Agri-food Sector, towards a network of Digital Innovation Hubs
in Robotics”, Grant Contract number 825395 https://agrobofood.eu/
References
1. Marinoudi, V.; Sørensen, C.G.; Pearson, S.; Bochtis, D. Robotics and labour in agriculture. A context
consideration. Biosyst. Eng. 2019, 184, 111–121.
2. Pedersen, S.M.; Fountas, S.; Have, H.; Blackmore, B.S. Agricultural robots—System analysis and economic
feasibility. Precis. Agric. 2006, 7, 295–308.
3. Pedersen, S.M.; Fountas, S.; Sørensen, C.G.; Van Evert, F.K.; Blackmore, B.S. Robotic seeding: Economic
perspectives. In Precision Agriculture: Technology and Economic Perspectives; Pedersen, S.M., Lind, K.M., Eds.;
Springer International Publishing: Cham, Switzerland, 2017; pp. 167–179. ISBN 978-3-319-68715-5.
4. Blackmore, B.S.; Fountas, S.; Gemtos, T.A.; Griepentrog, H.W. A specification for an autonomous crop
production mechanization system. In Proceedings of the International Symposium on Application of
Precision Agriculture for Fruits and Vegetables, Orlando, FL, USA, 1 April 2009; Volume 824, pp. 201–216.
5. Fountas, S.; Gemtos, T.A.; Blackmore, S. Robotics and Sustainability in Soil Engineering. In Soil Engineering;
Springer: Berlin, Germany, 2010; pp. 69–80.
6. Fountas, S.; Blackmore, B.; Vougioukas, S.; Tang, L.; Sørensen, C.; Jørgensen, R. Decomposition of
agricultural tasks into robotic behaviours. Available online:
https://cigrjournal.org/index.php/Ejounral/article/viewFile/901/895 (accessed on 6 May 2020).
7. Reina, G.; Milella, A.; Galati, R. Terrain assessment for precision agriculture using vehicle dynamic
modelling. Biosyst. Eng. 2017, 162, 124–139.
Sensors 2020, 20, 2672 22 of 27
8. Fernandes, H.R.; Garcia, A.P. Design and control of an active suspension system for unmanned agricultural
vehicles for field operations. Biosyst. Eng. 2018, 174, 107–114.
9. Bochtis, D.D.; Vougioukas, S.G.; Griepentrog, H.W. A mission planner for an autonomous tractor. Trans.
ASABE 2009, 52, 1429–1440.
10. Bochtis, D.; Griepentrog, H.W.; Vougioukas, S.; Busato, P.; Berruto, R.; Zhou, K. Route planning for orchard
operations. Comput. Electron. Agric. 2015, 113, 51–60.
11. Yang, L.; Noguchi, N. Human detection for a robot tractor using omni-directional stereo vision. Comput.
Electron. Agric. 2012, 89, 116–125.
12. Noguchi, N.; Will, J.; Reid, J.; Zhang, Q. Development of a master-slave robot system for farm operations.
Comput. Electron. Agric. 2004, 44, 1–19.
13. Zhang, C.; Noguchi, N.; Yang, L. Leader-follower system using two robot tractors to improve work
efficiency. Comput. Electron. Agric. 2016, 121, 269–281.
14. Bechar, A.; Vigneault, C. Agricultural robots for field operations. Part 2: Operations and systems. Biosyst.
Eng. 2017, 153, 110–128.
15. Aravind, K.R.; Raja, P.; Pérez-Ruiz, M. Task-based agricultural mobile robots in arable farming: A review.
Span. J. Agric. Res. 2017, 15, 1–16.
16. Shamshiri, R.R.; Weltzien, C.; Hameed, I.A.; Yule, I.J.; Grift, T.E.; Balasundram, S.K.; Pitonakova, L.;
Ahmad, D.; Chowdhary, G. Research and development in agricultural robotics: A perspective of digital
farming. Int. J. Agric. Biol. Eng. 2018, 11, 1–14.
17. Tsolakis, N.; Bechtsis, D.; Bochtis, D. Agros: A robot operating system based emulation tool for agricultural
robotics. Agronomy 2019, 9, 403.
18. Slaughter, D.C.; Giles, D.K.; Downey, D. Autonomous robotic weed control systems: A review. Comput.
Electron. Agric. 2008, 61, 63–78.
19. Defterli, S.G.; Shi, Y.; Xu, Y.; Ehsani, R. Review of robotic technology for strawberry production. Appl. Eng.
Agric. 2016, 32, 301–318.
20. FAO Recommendations for improved weed management. Plant Prod. Prot. Div. Rome 2006, 1–56.
21. Oerke, E.C. Crop losses to pests. J. Agric. Sci. 2006, 144, 31–43.
22. Utstumo, T.; Urdal, F.; Brevik, A.; Dørum, J.; Netland, J.; Overskeid, Ø.; Berge, T.W.; Gravdahl, J.T. Robotic
in-row weed control in vegetables. Comput. Electron. Agric. 2018, 154, 36–45.
23. Dino. Available online: https://www.naio-technologies.com/en/agricultural-equipment/large-scale-
vegetable-weeding-robot/ (accessed on 6 May 2020).
24. Bakker, T.; Van Asselt, K.; Bontsema, J.; Müller, J.; Van Straten, G. An autonomous weeding robot for
organic farming. In Field and Service Robotics; Springer: Berlin, Germany, 2006; Volume 25, pp. 579–590.
25. Kim, G.H.; Kim, S.C.; Hong, Y.K.; Han, K.S.; Lee, S.G. A robot platform for unmanned weeding in a paddy
field using sensor fusion. In Proceedings of the IEEE International Conference on Automation Science and
Engineering, Seoul, Korea, 20–24 August 2012; pp. 904–907.
26. Vitirover. Available online: https://www.vitirover.fr/en-home (accessed on 6 May 2020).
27. Jørgensen, R.N.; Sørensen, C.G.; Pedersen, J.M.; Havn, I.; Jensen, K.; Søgaard, H.T.; Sorensen, L.B. Hortibot:
A system design of a robotic tool carrier for high-tech plant nursing. Available online:
https://ecommons.cornell.edu/bitstream/handle/1813/10601/ATOE%2007%20006%20Jorgensen%2011July
2007.pdf?sequence=1&isAllowed=y (accessed on 6 May 2020).
28. Bochtis, D.D.; Sørensen, C.G.; Jørgensen, R.N.; Nørremark, M.; Hameed, I.A.; Swain, K.C. Robotic weed
monitoring. Acta Agric. Scand. Sect. B Soil Plant Sci. 2011, 61, 201–208.
29. OZ. Available online: https://www.naio-technologies.com/en/agricultural-equipment/weeding-robot-oz/
(accessed on 6 May 2020).
30. Reiser, D.; Sehsah, E.S.; Bumann, O.; Morhard, J.; Griepentrog, H.W. Development of an autonomous
electric robot implement for intra-row weeding in vineyards. Agriculture 2019, 9, 18.
31. Gokul, S.; Dhiksith, R.; Sundaresh, S.A.; Gopinath, M. Gesture controlled wireless agricultural weeding
robot. In Proceedings of the 2019 5th International Conference on Advanced Computing and
Communication Systems, San Mateo, CA, USA, 4–8 November 2019; pp. 926–929.
32. TED. Available online: https://www.naio-technologies.com/en/agricultural-equipment/vineyard-weeding-
robot/ (accessed on 6 May 2020).
33. Anatis. Available online: https://www.carre.fr/entretien-des-cultures-et-prairies/anatis/?lang=en
(accessed on 6 May 2020).
Sensors 2020, 20, 2672 23 of 27
60. Xue, J.; Su, B. Significant remote sensing vegetation indices: A review of developments and
applications..Sens. 2017. doi.org/10.1155/2017/1353691.
61. Wang, Y.; Hu, X.; Hou, Z.; Ning, J.; Zhang, Z. Discrimination of nitrogen fertilizer levels of tea plant
(Camellia sinensis) based on hyperspectral imaging. J. Sci. Food Agric. 2018, 98, 4659–4664.
62. Bietresato, M.; Carabin, G.; Vidoni, R.; Gasparetto, A.; Mazzetto, F. Evaluation of a LiDAR-based 3D-
stereoscopic vision system for crop-monitoring applications. Comput. Electron. Agric. 2016, 124, 1–13.
63. Vidoni, R.; Gallo, R.; Ristorto, G.; Carabin, G.; Mazzetto, F.; Scalera, L.; Gasparetto, A. Byelab : An
agricultural mobile robot prototype for proximal sensing and precision farming. In Proceedings of the
ASME International Mechanical Engineering Congress and Exposition, Tampa, FL, USA, 3–9 November
2017; Volume 4A-2017.
64. OUVA. Available online: https://www.polariks.com/text?rq=ouva (accessed on 6 May 2020).
65. Vinbot. Available online: http://vinbot.eu/ (accessed on 6 May 2020).
66. Vinescout. Available online: http://vinescout.eu/web/ (accessed on 6 May 2020).
67. Grape. Available online: http://www.grape-project.eu/home/ (accessed on 6 May 2020).
68. Dos Santos, F.N.; Sobreira, H.; Campos, D.; Morais, R.; Moreira, A.P.; Contente, O. Towards a reliable monitoring
robot for mountain vineyards. In Proceedings of the 2015 IEEE International Conference on Autonomous Robot
Systems and Competitions, ICARSC 2015, Vila Real, Portugal, 8–10 April 2015; pp. 37–43.
69. Terrasentia. Available online: https://www.earthsense.co (accessed on 6 May 2020).
70. Vaeljaots, E.; Lehiste, H.; Kiik, M.; Leemet, T. Soil sampling automation case-study using unmanned
ground vehicle. Eng. Rural Dev. 2018, 17, 982–987.
71. Iida, M.; Kang, D.; Taniwaki, M.; Tanaka, M.; Umeda, M. Localization of CO2 source by a hexapod robot
equipped with an anemoscope and a gas sensor. Comput. Electron. Agric. 2008, 63, 73–80.
72. Deery, D.; Jimenez-Berni, J.; Jones, H.; Sirault, X.; Furbank, R. Proximal remote sensing buggies and
potential applications for field-based phenotyping. Agronomy 2014, 4, 349–379.
73. Mueller-Sim, T.; Jenkins, M.; Abel, J.; Kantor, G. The Robotanist: A ground-based agricultural robot for
high-throughput crop phenotyping. In Proceedings of the IEEE International Conference on Robotics and
Automation, Singapore, 29 May–3 July 2017; pp. 3634–3639.
74. Benet, B.; Dubos, C.; Maupas, F.; Malatesta, G.; Lenain, R. Development of autonomous robotic platforms
for sugar beet crop phenotyping using artificial vision. Available online: https://hal.archives-
ouvertes.fr/hal-02155159/document (accessed on 6 May 2020).
75. Ruckelshausen, A.; Biber, P.; Dorna, M.; Gremmes, H.; Klose, R.; Linz, A.; Rahe, R.; Resch, R.; Thiel, M.;
Trautz, D.; et al. BoniRob: An autonomous field robot platform for individual plant phenotyping. In
Proceedings of the Precision Agriculture 2009—Papers Presented at the 7th European Conference on
Precision Agriculture, Wageningen, The Netherlands, 6–8 July 2009; pp. 841–847.
76. Virlet, N.; Sabermanesh, K.; Sadeghi-Tehran, P.; Hawkesford, M.J. Field Scanalyzer: An automated robotic
field phenotyping platform for detailed crop monitoring. Funct. Plant Biol. 2017, 44, 143–153.
77. Bao, Y.; Tang, L.; Breitzman, M.W.; Salas Fernandez, M.G.; Schnable, P.S. Field-based robotic phenotyping
of sorghum plant architecture using stereo vision. J. Field Robot. 2019, 36, 397–415.
78. Shafiekhani, A.; Kadam, S.; Fritschi, F.B.; Desouza, G.N. Vinobot and vinoculer: Two robotic platforms for
high-throughput field phenotyping. Sensors 2017, 17, 214.
79. Young, S.N.; Kayacan, E.; Peschel, J.M. Design and field evaluation of a ground robot for high-throughput
phenotyping of energy sorghum. Precis. Agric. 2019, 20, 697–722.
80. Zhang, S.; Wang, H.; Huang, W.; You, Z. Plant diseased leaf segmentation and recognition by fusion of
superpixel, K-means and PHOG. Optik 2018, 157, 866–872.
81. Vacavant, A.; Lebre, M.A.; Rositi, H.; Grand-Brochier, M.; Strand, R. New definition of quality-scale
robustness for image processing algorithms, with generalized uncertainty modeling, applied to denoising
and segmentation. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial
Intelligence and Lecture Notes in Bioinformatics); Beijing, China, 2019; Volume 11455 LNCS, pp. 138–149.
82. Rajbhoj, A.; Deshpande, S.; Gubbi, J.; Kulkarni, V.; Balamuralidhar, P. A System for semi-automatic
construction of image processing pipeline for complex problems. In Lecture Notes in Business Information
Processing; Rome, Italy,2019; Volume 352, pp. 295–310.
83. Sammons, P.J.; Furukawa, T.; Bulgin, A. Autonomous pesticide spraying robot for use in a greenhouse. In
Proceedings of the 2005 Australasian Conference on Robotics and Automation, Sydney, Australia, 5–7
December 2005.
Sensors 2020, 20, 2672 25 of 27
108. Wang, L.L.; Zhao, B.; Fan, J.W.; Hu, X.A.; Wei, S.; Li, Y.S.; Zhou, Q.B.; Wei, C.F. Development of a tomato
harvesting robot used in greenhouse. Int. J. Agric. Biol. Eng. 2017, 10, 140–149.
109. Feng, Q.; Zou, W.; Fan, P.; Zhang, C.; Wang, X. Design and test of robotic harvesting system for cherry
tomato. Int. J. Agric. Biol. Eng. 2018, 11, 96–100.
110. Metomotion. Available online: https://metomotion.com (accessed on 6 May 2020).
111. Root-AI. Available online: https://root-ai.com (accessed on 6 May 2020).
112. Energid. Available online: https://www.energid.com/industries/agricultural-robotics (accessed on 6 May
2020).
113. Tanigaki, K.; Fujiura, T.; Akase, A.; Imagawa, J. Cherry-harvesting robot. Comput. Electron. Agric. 2008, 63,
65–72.
114. Ceres, R.; Pons, J.L.; Jiménez, A.R.; Martín, J.M.; Calderón, L. Design and implementation of an aided fruit-
harvesting robot (Agribot). Ind. Rob. 1998, 25, 337–346.
115. Van Henten, E.J.; Hemming, J.; Van Tuijl, B.A.J.; Kornet, J.G.; Meuleman, J.; Bontsema, J.; Van Os, E.A. An
autonomous robot for harvesting cucumbers in greenhouses. Auton. Robot. 2002, 13, 241–258.
116. Hayashi, S.; Ganno, K.; Ishii, Y.; Tanaka, I. Robotic harvesting system for eggplants. Jpn. Agric. Res. Q. 2002,
36, 163–168.
117. Sweeper. Available online: http://www.sweeper-robot.eu (accessed on 6 May 2020).
118. Irie, N.; Taguchi, N.; Horie, T.; Ishimatsu, T. Asparagus harvesting robot coordinated with 3-D vision
sensor. In Proceedings of the IEEE International Conference on Industrial Technology, Gippsland,
Australia, 10–13 February 2009.
119. Cerescon. Available online: https://www.cerescon.com/EN/sparter (accessed on 6 May 2020).
120. Roshanianfard, A.R.; Noguchi, N. Kinematics analysis and simulation of A 5DOF articulated robotic arm
applied to heavy products harvesting. Tarım Bilim. Derg. 2018, 24, 91–104.
121. Edan, Y.; Rogozin, D.; Flash, T.; Miles, G.E. Robotic melon harvesting. IEEE Trans. Robot. Autom. 2000, 16,
831–835.
122. Umeda, M.; Kubota, S.; Iida, M. Development of “STORK”, a watermelon-harvesting robot. Artif. Life Robot.
1999, 3, 143–147.
123. Pilarski, T.; Happold, M.; Pangels, H.; Ollis, M.; Fitzpatrick, K.; Stentz, A. The Demeter system for
automated harvesting. Auton. Robots 2002, 13, 9–20.
124. Rowley, J.H. Developing Flexible Automation for Mushroom Harvesting (Agaricus Bisporus): Innovation Report;
University of Warwick: Coventry, UK, 2009.
125. Siciliano, B.; Khatib, O. Springer Handbook of Robotics; Springer: Berlin, Germany, 2016. ISBN 3319325523.
126. Abundant Robotics. Available online: https://www.abundantrobotics.com/ (accessed on 6 May 2020).
127. Zapotezny-Anderson, P.; Lehnert, C. Towards active robotic vision in agriculture: A deep learning
approach to visual servoing in occluded and unstructured protected cropping environments. arXiv 2019,
arXiv:Prepr/1908.01885.
128. Mu, L.; Cui, G.; Liu, Y.; Cui, Y.; Fu, L.; Gejima, Y. Design and simulation of an integrated end-effector for
picking kiwifruit by robot. Inf. Process. Agric. 2019, 7, 58–71.
129. Harvest Automation. Available online: https://www.public.harvestai.com/ (accessed on 6 May 2020).
130. Vision Robotics. Available online: https://www.visionrobotics.com/vr-grapevine-pruner (accessed on 6
May 2020).
131. Botterill, T.; Paulin, S.; Green, R.; Williams, S.; Lin, J.; Saxton, V.; Mills, S.; Chen, X.; Corbett-Davies, S. A
robot system for pruning grape vines. J. F. Robot. 2017, 34, 1100–1122.
132. Zahid, A.; He, L.; Zeng, L. Development of a robotic end effector for apple tree pruning. In Proceedings of
the 2019 ASABE Annual International Meeting, Boston, MA, USA, 7–10 July 2019.
133. He, L.; Zhou, J.; Zhang, Q.; Charvet, H.J. A string twining robot for high trellis hop production. Comput.
Electron. Agric. 2016, 121, 207–214.
134. Grimstad, L.; From, P.J. The Thorvald II agricultural robotic system. Robotics 2017, 6, 24.
135. Xue, J.; Fan, B.; Zhang, X.; Feng, Y. An agricultural robot for multipurpose operations in a greenhouse.
DEStech Trans. Eng. Technol. Res. 2017. doi: 10.12783/dtetr/icmme2017/9073.
136. Narendran, V. Autonomous robot for E-farming based on fuzzy logic reasoning. Int. J. Pure Appl. Math.
2018, 118, 3811–3821.
137. Robotti. Available online: http://agrointelli.com/robotti-diesel.html#rob.diesel (accessed on 6 May 2020).
Sensors 2020, 20, 2672 27 of 27
138. Durmus, H.; Gunes, E.O.; Kirci, M.; Ustundag, B.B. The design of general purpose autonomous agricultural
mobile-robot: “AGROBOT.” In Proceedings of the 2015 4th International Conference on Agro-
Geoinformatics, Agro-Geoinformatics 2015, Istanbul, Turkey, 20–24 July 2015; pp. 49–53.
139. Peña, C.; Riaño, C.; Moreno, G. RobotGreen: A teleoperated agricultural robot for structured environments.
J. Eng. Sci. Technol. Rev. 2018, 11, 87–98.
140. Murakami, N.; Ito, A.; Will, J.D.; Steffen, M.; Inoue, K.; Kita, K.; Miyaura, S. Development of a teleoperation
system for agricultural vehicles. Comput. Electron. Agric. 2008, 63, 81–88.
141. Jasiński, M.; Mączak, J.; Radkowski, S.; Korczak, S.; Rogacki, R.; Mac, J.; Szczepaniak, J. Autonomous
agricultural robot—Conception of inertial navigation system. In International Conference on Automation;
Springer: Cham, Switzerland, 2016; Volume 440, pp. 669–679.
142. Chang, C.L.; Lin, K.M. Smart agricultural machine with a computer vision-based weeding and variable-
rate irrigation scheme. Robotics 2018, 7, 38.
143. Moshou, D.; Bravo, C.; Oberti, R.; West, J.S.; Ramon, H.; Vougioukas, S.; Bochtis, D. Intelligent multi-sensor
system for the detection and treatment of fungal diseases in arable crops. Biosyst. Eng. 2011, 108, 311–321.
144. Quan, L.; Feng, H.; Lv, Y.; Wang, Q.; Zhang, C.; Liu, J.; Yuan, Z. Maize seedling detection under different
growth stages and complex field environments based on an improved Faster R–CNN. Biosyst. Eng. 2019,
184, 1–23.
145. Gollakota, A.; Srinivas, M.B. Agribot—A multipurpose agricultural robot. In Proceedings of the 2011
Annual IEEE India Conference, Hyderabad, India, 16–18 December 2011; pp. 1–4.
146. Praveena, R.; Srimeena, R. Agricultural robot for automatic ploughing and seeding. In Proceedings of the
2015 IEEE Technological Innovation in ICT for Agriculture and Rural Development (TIAR), Chennai, India,
10–12 July 2015; pp. 17–23.
147. Sowjanya, K.D.; Sindhu, R.; Parijatham, M.; Srikanth, K.; Bhargav, P. Multipurpose autonomous
agricultural robot. In Proceedings of the International Conference on Electronics, Communication and
Aerospace Technology, Coimbatore, India, 20–22 April 2017; pp. 696–699.
148. Ranjitha, B.; Nikhitha, M.N.; Aruna, K.; Murthy, B.T.V. Solar powered autonomous multipurpose
agricultural robot using bluetooth/android app. In Proceedings of the 2019 3rd International conference on
Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 12–14 June 2019; pp.
872–877.
149. Wall-ye (MYCE_Agriculture). Available online: http://wall-ye.com/index-6.html (accessed on 6 May 2020).
150. Wall-ye (Myce_Vigne). Available online: http://wall-ye.com/index.html (accessed on 6 May 2020).
151. Green, O.; Schmidt, T.; Pietrzkowski, R.P.; Jensen, K.; Larsen, M.; Jørgensen, R.N. Commercial autonomous
agricultural platform. In Proceedings of the Second International Conference on Robotics and associated
High-technologies and Equipment for Agriculture and Forestry, Bergamo, Italy, 20–23 October 2014; pp.
351–356.
152. TrimBot 2020. Available online: http://trimbot2020.webhosting.rug.nl/ (accessed on 6 May 2020).
153. Bak, T.; Jakobsen, H. Agricultural robotic platform with four wheel steering for weed detection. Biosyst.
Eng. 2004, 87, 125–136.
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).