University of Dortmund, Computer Science Dept. LS VIII, D-44221 Dortmund, Germany Volker, Morik @ls8.informatik - Uni-Dortmund - de
University of Dortmund, Computer Science Dept. LS VIII, D-44221 Dortmund, Germany Volker, Morik @ls8.informatik - Uni-Dortmund - de
Abstract. The recognition of objects and, hence, their descriptions must be grounded in the
environment in terms of sensor data. We argue, why the concepts, used to classify perceived objects
and used to perform actions on these objects, should integrate action-oriented perceptual features
and perception-oriented action features. We present a grounded symbolic representation for these
concepts. Moreover, the concepts should be learned. We show a logic-oriented approach to learning
grounded concepts.
Key Words. Operational concepts, inductive logic programming, combining sensing and action,
symbol grounding
very similar to the one before, the control program sensors effectors
roles and the roles of other and be able to recip- pose that features at all levels should integrate ac-
rocate actions of the other with actions of their tion and perception. Perceptual features require
own, " (Nelson, 1983, p. 135). Objects should the integration of the action that is performed
:::
be represented as relations between actions and while perceiving an object. Suppose that you are
reactions, where an object implies a specic ac- looking at a cup from above. This position does
tion, like a \ball implies throwing". Nelson re- not allow to determine whether the bottom is
at
gards three kinds of relations between objects and or not. The perception is restricted by the action
actors, where the two higher levels are the most during which an object is perceived. Action fea-
appropriate ones in our scenario. In the medium tures, in turn, require the integration of percep-
level, \dierent objects may occur in the same po- tion. Particular sensor patterns express the appli-
sition of an event", like a specic tennis ball or a cability conditions, the conditions for successful
red rubber ball in the event \ball implies throw- performance of the action, and the conditions for
ing". In our representation, the dierent objects ending the action. In this way, an action is ex-
that can occur in an event are represented by their pressed by perceptions.
perceptual features, and the action these objects
imply are integrated in the denition of opera- 3. DATA FROM THE NAVIGATION
tional concepts as action features. Additionally, SCENARIO
\an object or category of objects may recur in a
number of dierent events, in similar or dierent Before describing the representation hierarchy and
relationships". The category of doors may occur the learning algorithms, we present the scenario
in dierent events like moving along them or mov- and the robot we used for getting data. PRI-
ing through them, so we have to relate dierent AMOS, developed by the University of Karl-
actions with the perceptions of the concept door. sruhe (Dillmann et al., 1993), is a mobile robot
with three degrees of freedom for motion, i.e., it
This integration of perceptual features and action can move into every direction and rotate simul-
features also supports the classication of an ob- taneously. It has 24 sonar sensors measuring the
ject as an instance of a concept. In a conventional distance to the nearest object within the emission
representation, a cup, for example, is dened by cone. Errors can occur, if the sonar beam is mul-
having a
at bottom and a handle and being con- tiply re
ected or the angle between the beam and
cave. But it is easy to nd arbitrarily many ob- the object is inconvenient. At every side of the
jects with these properties that are not cups be- robot three sensor are installed, emitting in par-
cause it is impossible to drink from them, e.g., if allel. Three sensors are mounted at every corner
the handle bridges over the opening (De Jong and emitting with a dierence of 15 .
Mooney, 1986). Finding a complete description
of a cup excluding all exceptions is impossible be- The aim of our rst learning trials was to learn
cause of the innite number of these exceptions. descriptions for the concepts move along door and
This is the qualication problem (McCarthy and move through door. We used data from 28 traces
Hayes, 1969). So, how to dene a cup suitably? in a simple room, most of them being paths
and their relations. The classical learning algo-
rithms for attribute-value representations cannot
solve our learning tasks.
10 2
9 3
Measured Distance
decreasing(trace1, sensor1, t5, t7,−0.3).
no_measurement(trace1, sensor1, t7, t10, _).
stable (trace1, sensor1, t10, t13, 0.01). perceptual features
3m
sensor features action features
basic features
2m
Fig. 3 Measurements and basic features. maximal sensing distance. This situation, in the
example the interval from t7 to t10, is labeled
no measurement. Other basic features are stable
Basic features are calculated for each sensor using (describing more or less equal measurements), de-
the following algorithm: creasing, decreasing peak, no movement (we can-
not compute the gradient, if the robot does not
1. Start with the rst measurement. move), and straight to and straight away (if the
2. Get the next measurement and calculate the gradient is about 1 or ,1, respectively).
gradient between the previous and the cur-
rent one, i.e., the quotient of the dierence of Note, that the basic features compress a sequence
the measurements and the moved distance. of measurements to a statement about a time in-
1 and 2 becomes an interval.
terval. The nely grained basic features summa-
rize only a short sequence of measurements where
T T
3. If the gradient is close enough to the average the rough features summarize a long sequence of
gradient of the interval: measurements. In that way, the basic features de-
then enlarge the interval by the new time termine the time intervals used by concept learn-
point and adapt the average gradient de- ing. The time intervals, in turn, determine what
pending on the way dened by a param- can be achieved by concept learning. It is, for in-
eter. stance, impossible to learn good concepts descrip-
else close the previous interval at the previ-
tions, if the time intervals determined by the basic
ous time point and start a new one at features do not match the time intervals used in
that time point with the new gradient the examples to be classied.
as average gradient.
\Close enough" is dened by a parameter, 4.2 The Representation Hierarchy
too.
4. A closed interval will be named by a symbol, Several levels of abstraction link the level of raw
depending on the average gradient attached sensor and motion data with the level of opera-
to that interval. The set of symbols and the tional concepts. The hierarchy of more and more
gradients they belong to are determined by a abstract representations comprises perceptual fea-
further parameter. tures at several levels which are all related with
5. Go to step 2. actions as well as action feature which become in-
tegrated with perceptual features at a higher level.
Figure 4 shows the representation hierarchy.
In Figure 3, an example for a short sequence of
measurements is displayed together with the cor- In the rst step, basic features are constructed
responding basic features. First, the gradient be- based on the measurements and the robot's po-
tween two measurements is more or less 0 7. The : sition. At the next higher level, the perceptions
measurements are increasing linearly. That is while moving are represented by single sensor fea-
represented by the basic feature increasing(trace1, tures.
sensor1, t1, t4, 0 7). Then, from time point t4
:
to t5, the gradient is greater than 1. Since the s line: a single edge.
dierence between two measurements cannot be s jump: two parallel edges, sensed in sequence.
greater than the moved distance, if the sensor
sensed the same edge, two dierent edges must s convex: two orthogonal edges, sensed from the
have been sensed. This situation is expressed by convex side of this corner.
the basic feature incr peak. Sometimes a sensor s concave: two orthogonal edges, sensed from the
gets no echo, e.g., if there is no object within the concave side.
Each sensor feature is characterized by sequences verication conditions, and end-of-action condi-
of basic features. The sequences are represented tions for an abstract action and its corresponding
by rules. An example rule for s jump is: object.
stable(Trace, Or, Sensor, Time1 , Time2 , Grad1 ) & standing(Trace, T1, T2, in front of door,
incr peak(Trace, Or, Sensor, Time2 , Time3 , Grad2 ) & PDirect, small side, PrevP) &
stable(Trace, Or, Sensor, Time3 , Time4 , Grad3 ) parallel moving(Trace, T2, T3, MSpeed,
! s jump(Trace, Sensor, Time1, Time4, parallel). PDirect,through door,right and left) &
standing(Trace, T3, T4, in front of door, rear,
At the next level, situations are represented where small side, through door)
dierent sensors of a class sensed the same con- ! move through door(Trace, T1, T4).
stellation of edges while moving. Single sensor fea-
tures are grouped together in order to increase the The rule describes the operational concept of mov-
evidence for an edge constellation. This helps to ing through a doorway. It combines the recogni-
cope with the noise of single sensor measurements. tion of the doorway with the action of moving
It can be sucient for a sensor group feature to be through it on a parallel track. Moving diagonally
derived, that two of three sensors of a sensor class through a doorway is a dierent operational con-
sensed the same constellation. Additionally, we cept. Action and perception are linked by the
abstract from the actual sensor numbers in this perceptual features occurring as arguments in the
step. Instead, the orientation of the perception action predicates. In our example rule, the per-
relative to the orientation of the motion is used. ceptual features are written in italics.
s jump(Trace, Sensor1 , Start1 , End1 , parallel) & The rst premise of the rule states that the robot
sclass(Trace, Sensor1 , right side) & is standing in a time interval from T1 to T2 of a
s jump(Trace, Sensor2 , Start2 , End2 , parallel) & particular (Trace) and senses the perceptual fea-
sclass(Trace, Sensor2 , right side) & ture in front of door from the sensors of a small
succ(Start1 , Start2 ) & succ(End1, End2 )
! sg jump(Trace, right side, Start1 , End2, parallel). side after having perceived the perceptual feature
PrevP prior to T1. This premise denotes the pre-
At the next level of abstraction, the level of action- condition for moving through a doorway on a par-
oriented perceptual features, perceptions at dier- allel track. The precondition can also be viewed
ent sides of the robot will be combined. as a trigger for planning: whenever recognizing
that it stands in front of a door, the robot may
execute the action of parallel moving through the
sg jump(Trace, right side, T1 , T2 , parallel) & doorway.
sg jump(Trace, left side, T1 , T2 , parallel)
! through door(Trace, T1, T2, both sides, parallel). The second premise states that in a following time
interval the action parallel move is executed, mea-
At the action side of the representation hierarchy, suring by the sensors at the right and the left side
only basic actions occur. They represent intervals of the robot the perceptual feature through door.
of uniform motions. From the absolute orientation Note, that the particular values for the time in-
and speed of the motion features are calculated tervals do not matter for the operational concept
that relate the orientation to the previous direc- but are instantiated by the specics of a particular
tion of motion. The basic action features used are path. Only the time relation is important. This
move, rotate, and stand, with the arguments trace, makes the operational concept independent of the
time interval, speed, and direction. particular depth of a doorway and the robot's par-
ticular rate of advance.
Action features and perceptual features are then
combined to perception integrating features repre- The third premise describes the end of the move-
senting single actions and the corresponding per- ment through the doorway: the robot is stand-
ceptions. An example of such an action is a steady ing in front of the door, now sensing with its
movement while measuring through door in a spe- rear sensors mounted at its small side the feature
cic direction PDirect. in front of door after having perceived the feature
through door. PDirect, the orientation of percep-
through door(Trace, PDirect, T1 , T2 , parallel) & tion (e.g., left, right, front, rear) is not xed by
move(Trace, T1 , T2 , Speed, Dir) the rule. It is only relevant that the orientation
! move parallel(Trace, T1 , T2 , Speed, Dir, should not change during the parallel move and
through door, PDirect). that with respect to this orientation the doorway
is sensed in the rear after the movement. That
Finally, operational concepts use perception inte- means, whatever the orientation was at the be-
grating features for characterizing preconditions, ginning of the action, it is dened as being front
so that after completing the action the doorway is creases as the representation hierarchy is climbed.
sensed by the opposite, the rear sensors. This shows that the hierarchy makes learning ro-
bust against the noise which is produced by the
weak ultrasonic sensors. The levels can be re-
4.3 Learning Perceptual Features garded as a lter { they only pass the more reliable
information to the next higher level.
Learning of higher-level features and, nally,
of operational concepts is performed in the
paradigm of inductive logic programming (ILP, 5. CONCLUSION
see, e.g., (Muggleton, 1992)). The learning task is
the following: In this paper, we introduced operational concepts
that integrate action-oriented perceptual features
and perceptual-oriented action features. Both
Learning higher-level features:
the target feature for various sit- psychological and practical arguments indicate
Given
uations and background knowledge { the necessity of this integration. We sketched
learn rules that characterize the tar-
a representation for operational concepts that is
get feature in terms of the background grounded in the perception and action of a mo-
knowledge. bile robot. We proposed two learning tasks: one
that nds appropriate basic features in order to
summarize sequences of sensor data, and one that
The background knowledge is a set of facts from characterizes higher-level features (or operational
the level below the target feature and general in- concepts) in terms of features from the level below.
formation about the robot (e.g., sensor classes). Experiments have shown that learning becomes
This learning task is performed several times: more accurate and the coverage of learned rules
learning sensor features from basic features, learn- increases when progressing to higher levels of the
ing sensor group features from sensor features, representation hierarchy. This means that even
learning action-oriented perceptual features from noisy input to learning leads to results that sum-
sensor group features, and learning operational marize the data such that at the next higher level
concepts from perception-integrating action fea- the noise is reduced. These results are promising,
tures. but further learning experiments in other environ-
ments need be conducted.
We applied learning algorithms that search a re-
stricted hypothesis space completely. The hy- In order to achieve rst results, we restricted sens-
pothesis space is restricted syntactically by rule ing and action of the mobile robot. For rep-
schemata that express the form of learnable rules. resenting the sensory data, we reduced the de-
Rules that are not instantiations of a rule schema grees of freedom of the robot allowing only one-
cannot be learned. This prohibits learning rules dimensional motions at once. That will lead to
that cannot be put to use by the robot. The rule problems, if we incorporate a reactive module ex-
schemata are ordered according to their general- ecuting the elementary operations, if this module
ity. Starting with the most general rule schema, generates traces where the robot deviates from
all its instantiations (i.e. rules) are tested whether straight courses. Then, the calculation of basic
they cover positive but not negative examples of features is probably not able to counterbalance
the target feature. Until no negative (or only few) the divergence of the sensor measurements, be-
negative examples are covered, the next special cause in the moment, we only use linear functions
rule schema is instantiated and tested. The most to calculate the features.
applicable learning algorithms were RDT (Ki-
etz and Wrobel, 1992) and a modication of it, Another restriction concerns the sensors used.
GRDT (Klingspor, 1994). Sonar sensors are not able to get precise mea-
surements, they cannot be used to discriminate
In a nutshell, the results are as follows. Given ne structured objects. E.g., it is not possible
1004 examples for the four sensor features, we to detect closed doors, so we used only simple
learned 129 rules, covering 87% of the given exam- structured objects in the moment. An interest-
ples. For learning sensor group features, we had ing future investigation might be the integration
given 956 examples, from which GRDT learned of another sensor system like a vision system to
136 rules. Using the learned rules for sensor fea- get further information in specic situations.
tures and sensor group features, we got a coverage
of 64%. For the operational concept through door, The next challenge will be to validate the con-
we had given ten examples. We learned three tribution of learning to easing robot applications.
rules, two of them covering exactly the ten ex- Until now we have a small planning component
amples. The third rule was a poor one that could that computes basic actions of a robot from an
be dropped. The quality of the learned rules in- operational concept. This component uses learned
rules in a backward-chaining manner. The basic Giordana, A. and Saitta, L. (1990). Abstraction {
perceptual features are incrementally calculated a general framework for learning. In Procs. AAAI
from sensor measurements as shown in this pa- { Workshop on Automatic Generation of Approxima-
per (Section 4.1). Using the learned rules in a tions and Abstractions.
forward-chaining manner, higher-level features are Harnad, S. (1990). The symbol grounding problem.
derived. Using these techniques, learned rules are Physica D, 42:335{346.
made available for the robot's processing. First, Jong, G. D. and Bennett, S. (1993). Permissive plan-
experiments with the robot are to be made where ning { a machine learning approach to linking internal
the robot uses learned rules in the environment in and external worlds. In AAAI-93.
which learning took place. Then, the same rules Kedar-Cabelli, S. (1988). Toward a computational
should be used by the robot in similar but dier- model of purpose-directed analogy. In Prieditis, A.,
ent environments from the one in which learning editor, Analogica. Morgan Kaufmann.
took place. Finally, the eorts of preparing these Kietz, J.-U. and Wrobel, S. (1992). Controlling the
robot applications need be compared. complexity of learning in logic through syntactic and
task-oriented models. In Muggleton, editor, Inductive
ACKNOWLEDGEMENTS Logic Programming. Academic Press.
Klingspor, V. (1994). GRDT: Enhancing model-based
This work is partially funded by the European learning for its application in robot navigation. In
Community under the project B-Learn II (P7274) Wrobel, S., editor, Proc. of the Fourth International
and the Ministry for Sciences and Research of the Workshop on Inductive Logic Programming, GMD-
German federal state Nordrhein-Westfalen. Studien Nr. 237, St. Augustin, Germany.
We would like to thank Anke Rieger for her con- McCarthy, J. and Hayes, P. J. (1969). Some philo-
sophical problems from the standpoint of articial in-
tributions to representing and learning perceptual telligence. Machine Intelligence, 5:463{502.
features, Stefan Sklorz for developing the repre-
sentation for action features and programming a Millan, J. and Torras, C. (1992). A reinforcement
meta-interpreter for planning, and Stephanie Wes- connectionist approach to robot path nding in non-
maze-like environments. Machine Learning, 8.
sel for her work on learning basic features. Thanks
to Joachim Hertzberg and Stephanie Wessel for Mitchell, T. M. (1982). Generalization as search. Ar-
reading drafts of this paper. ticial Intelligence, 18(2):203{226.
Muggleton, S. (1992). Inductive Logic Programming.
Academic Press, London.
REFERENCES Nelson, K. (1983). The derivation of concepts and cat-
Abelson, R. P. (1963). Hot cognition. Computer Sim- egories from event representations. In Scholnick, edi-
ulation of Personality. tor, New Trends in Conceptual Representation: Chal-
Bennett, S. W. (1989). Learning uncertainty toler- lenges to Piaget's Theory? Lawrence Erlbaum.
ant plans through approximation in complex domains. Segre, A. (1988). Machine Learning of Robot Assembly
Tech. Rep. UILU-ENG-89-2204, University of Illinois Plans. Kluwer, Boston.
at Urbana-Champaign. Steels, L. (1993). Building agents out of autonomous
Brooks, R. A. (1986). A robust layered control system behavior systems. In Steels and Brooks, editors, The
for a mobile robot. IEEE Journal of Robotics and 'articial life' route to 'articial intelligence' { Build-
Automation, 2. ing situated embodied agents. Lawrence Erlbaum.
Brooks, R. A. (1991). The role of learning in au- Wallner, F., Kaiser, M., Friedrich, H., and Dillmann,
tonomous robots. In Valiant and Warmuth, editors, R. (1994). Integration of topological and geometrical
COLT'91, Proc. of the fourth Annual Workshop. Mor- planning in a learning mobile robot. In Proc. of IROS-
gan Kaufmann. 94.
Cottrell, G. W., Bartell, B., and Haupt, C. (1990). Wrobel, S. (1991). Towards a model of grounded con-
Grounding meaning in perception. In Marburger, cept formation. In Proc. 12th Int. Joint Conf. on AI.
editor, GWAI{90, 14th German Workshop on AI. Morgan Kaufman.
Springer.
Zercher, K. (1992). Wissensintensives Lernen fur
De Jong, G. and Mooney, R. (1986). Explanation- zeitkritische technische Diagnoseaufgaben. inx, Sankt
based-learning: An alternative view. Machine Learn- Augustin.
ing, 2(1):145{176.
Dillmann, R., Kreuziger, J., and Wallner, F. (1993).
PRIAMOS { an experimental platform for re
exive
navigation. In Groen, Hirose, and Thorpe, editors,
IAS-3: Intelligent Autonomous Systems. IOS Press.
Gil, Y. (1994). Learning by experimentation { incre-
mental renement of incomplete planning. In Cohen,
W. and Hirsh, H., editors, Procs. 11th Int. Machine
Learning Conference.