Artificial Intelligence For Everyone
Artificial Intelligence For Everyone
Artificial
Intelligence
for Everyone
Artificial Intelligence for Everyone
Christian Posthoff
© The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Switzerland
AG 2024
This work is subject to copyright. All rights are solely and exclusively licensed by the Publisher, whether
the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse
of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and
transmission or information storage and retrieval, electronic adaptation, computer software, or by similar
or dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication
does not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
The publisher, the authors and the editors are safe to assume that the advice and information in this book
are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or
the editors give a warranty, expressed or implied, with respect to the material contained herein or for any
errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
v
vi Foreword
Joachim Reiss
Preface
The development of programs that are now called intelligent has accompanied
me since my student days. In the 1960s, computer science did not yet exist as a
subject of study. For mathematics students, there was an introductory course called
“Programming of Calculating Automata.” Programming was done in machine code,
and punched cards served as the input medium.
My PhD thesis had the title “The Application of Mathematical Methods in
Communicative Psychotherapy” (1976). It attempted to develop mathematical
methods to visualize and objectify the problems in groups of psychologically
vulnerable patients. Some computer programs supported the evaluation of the data.
Overall, the need for interdisciplinary cooperation could already be seen at this
stage.
In 1982, the Computer Science Department was founded at the TU Karl-Marx-
Stadt (today Chemnitz) in East Germany with a chair “Theoretical Computer
Fig. 1 The
Zeiß—Calculating
Automaton ZRA 1
vii
viii Preface
Science and Artificial Intelligence,” which I held from 1983 to 1993. At first, of
course, it was important to secure the teaching for the 5 years of study up to the
MSc level. The research work led to good results from 1986 on. In 1988, the setup
was essentially complete; everything that belongs to a normal university routine,
i.e., lectures, MSc as well as PhD theses, was in place.
In 1993, I switched to The University of The West Indies in Trinidad &
Tobago. Here the same sequence was followed. The teaching was developed into
a full-fledged course of studies in Computer Science (B.Sc., M.Sc., Ph.D.), and
after that I started again research on problems of artificial intelligence, without a
clear understanding or definition of the area. I simply followed the international
development and concentrated on some areas which I found interesting and useful.
The current discussions about future development are strongly driven by emo-
tions, wishes, fears, and premonitions that are not always based on knowledge and
not useful. But this is not a new phenomenon. There was such a wave in the 1980s
with computer chess. The development was followed with interest until the then
world champion G. Kasparov finally lost one game against the IBM system Deep
Blue in 1996. At first there were lively discussions, but subsequently the scene
calmed down, and a normal scientific-technical development began.
Autonomous driving of vehicles of various kinds is another focus of discussion.
Here, the focus is already on all possible real or fictitious dangers and possibilities,
without taking into account that robots have already been used for several years
without any problems.
There are currently very heated discussions about GPT-4 and similar programs.
This is a model that accepts text input and produces text output. Here one talks
almost exclusively about negative consequences—infringements of copyright, use
of plagiarism, distortion and falsification of facts, restrictions on creativity. From
this, one very often already deduces the necessity of legal steps, the justification of
bans, etc. But all these problems have existed before. The fact that people can also
use any technical progress in a negative way is not a new phenomenon, but it is
people themselves who are to blame.
To excuse the general public, it must of course be said that the media publish a
veritable deluge of opinions ranging from the destruction of humanity by artificial
intelligence to the colonialization of space. Since an ordinary mortal has only
taken a superficial note of the developments in computer science and is not
very knowledgeable, it naturally scares him very quickly, which is completely
inappropriate.
This book should enable many people to discuss and follow these developments
in an informed way and to draw conclusions for their own workplace and to acquire
the necessary new knowledge. This book is intended to be understandable for a wide
range of readers. If one wants to acquire special in-depth knowledge, then one must
resort to corresponding textbooks and courses. Many programs in the most diverse
fields are available online; one can then experiment with them at will.
Contents
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
2 How to Define Intelligence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2.1 Different Concepts for Defining Intelligence . . . . . . . . . . . . . . . . . . . . . . . 5
2.2 Intelligence Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
Reference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3 The Historical Development . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.1 The Dual System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
3.2 The Dartmouth Conference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
3.3 Algorithms and Programming Languages . . . . . . . . . . . . . . . . . . . . . . . . . . 20
3.3.1 The 1930s and 1940s: Logical Calculi . . . . . . . . . . . . . . . . . . . . . . 21
3.3.2 The 1950s: First Modern Programming Languages . . . . . . . . 22
3.3.3 The 1960s and 1970s: Development of New Paradigms . . . 23
3.3.4 Object-Oriented Programming . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
3.3.5 The 1980s . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
3.3.6 Further Development . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
3.4 The Turing Machine. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
3.5 The First Computers: The Turing Test, the Single-Tape
Turing Machine. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
3.6 LaTeX . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
3.7 The Evolution of Computers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
4 Mathematics on the Computer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
4.1 Microsoft Mathematics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
4.2 Mathematica . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
4.3 Maple . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
5 Polynomial and Exponential Complexity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5.1 Introductory Considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5.2 Mathematical Descriptions of Complexity . . . . . . . . . . . . . . . . . . . . . . . . . . 44
ix
x Contents
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 217
List of Figures
xiii
xiv List of Figures
M82589933 = 282589933 − 1
.
is (probably) a prime number, the largest prime number found so far. Written out, it
would have over 24 million decimal places. The number was found by a participant
in the GIMPS (Great Internet Mersenne Prime Search) project, a network in which a
large number of participants on distributed computers search for so-called Mersenne
primes. These Mersenne numbers are of the form .2n − 1 and are named after the
French mathematician Marin Mersenne (1588–1648). Until 1947, the numbers
p1 = 22 − 1 = 3, p2 = 23 − 1 = 7, p3 = 25 − 1 = 31, p4 = 27 − 1 = 127
.
Then computers began to be used, and the record given above shows the gap that
could be closed by computers. It is impossible for humans to verify such a value, and
the acceptance of such a solution becomes a matter of faith or trust. One must trust
both the programmer and the electronic engineer, and there must be no misbehavior
of any circuit during the calculations. We will encounter this situation of not being
able to check results very often.
The fact that technology is outperforming humans is not new. These problems are
now being discussed so fiercely mainly because the achievements of computers are
now playing a role in areas that were previously the exclusive preserve of humans,
their creativity and their intelligence. The intelligence of humans is more and more
supported by computers or even only effective with their help, which finally gave
rise to the name “artificial intelligence.”
It is probably quite natural that no completeness can be achieved in this
book. This area is already too large, and the number of qualified publications
grows exponentially. For some time now, it has been disintegrating into relatively
autonomous sub-areas:
• Application of data science in combination with AI and machine learning
methods,
• AI in medicine,
• Robotics,
• Speech and language technology,
• Educational technology,
• Innovative factory systems,
• Intelligent networks,
• Agents and simulated reality,
• Interactive machine learning,
• Augmented reality,
• Language technology and multilingualism.
The representation is further complicated by the fact that many areas overlap
and cooperate. It is precisely this cooperation that is an essential feature of artificial
intelligence. The book strives to provide basic knowledge that will objectify the
discussions and relieve some of the creepiness of utopian films. It must also be
understood that research results are a necessary condition for progress; they are
not sufficient until they can be translated into practice embedded in programs. This
difficult relationship between theory and practice has been known for a long time.
Chapter 2
How to Define Intelligence
The basic prerequisite for discussing intelligence is, of course, a healthy human
brain and associated nervous system. It has an enormous complexity: a human
being has about 100 billion brain cells that build up the central nervous system, our
brain, and they are interconnected. The number of these connections is estimated at
100 trillion. One speaks of artificial intelligence when the complexity of a solution
procedure on a computer is in similar dimensions. Microelectronic structures can
also reach this order of magnitude, and this is the fundamental prerequisite for
computer solutions to reach or increasingly surpass humanity.
herd in their buying and selling decisions and to invest or disinvest in a trading
object by a majority. Herd behavior is a manifestation of mass psychological
contagion effects and can thus be a cause of financial market crises or economic
crises. Hoarding purchases also exhibit herd behavior, as before natural disasters
or during the COVID-19 pandemic starting in March 2020, when there were shelf
gaps for certain goods (e.g., flour, pasta, toilet paper) in German stores.
Various mass-psychological or market-psychological causes may underlie
herd behavior. Consumers may be driven by the fear of not being able to meet
their needs in the face of shelf gaps if they do not buy immediately. A consumer’s
expectation that other consumers will also hoard after him also pushes him to
hoard purchases. Likewise, his fear that supply shortages may occur in the future
forces him to make purchasing decisions that are not in line with his needs.
Sometimes, consumers’ feelings of powerlessness are also seen as the cause.
This behavior is irrational, especially since food and beverages or toilet paper are
mass products that can be reproduced at any time. In France and Italy, one of the
products affected by hoarding is red wine, a product that cannot be reproduced at
any time.
The consequence of herd behavior is strong price fluctuations of the trading
object concerned. In addition, hoarding purchases accelerate the rotation of
goods and reduce the logistical range. As market behavior, herd behavior is
particularly known among noise traders, who are often guided by herd behavior
and motivated by sentiment or groups to buy or sell into falling ones. This is
what is known as “mood noise.” Rising or falling prices are an indication that
other market participants have previously made the same decision. This noise
can underlie both buy and sell decisions and also hold decisions. Herd behavior
is thus a sign of a lack of market efficiency.
Speculation only becomes problematic for a market when it is no longer
speculated on the basis of fundamental data but when herd behavior sets in. Then
speculative bubbles can arise, which are usually due to herd behavior. Speculative
bubbles can be justified by the expectation of the majority of market participants
of future profit opportunities.
Profit taking can also be based on herd behavior, when a large number of
investors take advantage of a high price level to sell and other investors join
in. The bank run is also a typical herd behavior, as investors observe a perhaps
random mass withdrawal of cash and blindly join it, trusting that it must have
a specific reason; the mass withdrawals eventually culminate in the domino
effect. Investors withdraw their deposit because they fear that, as a result of the
sequential payout principle (“first come, first served”), they will otherwise not
be able to withdraw this deposit because their cash reserves have been depleted.
Consequently, it is rational for any depositor to follow the herd. A bank run is
more likely the less informed bank customers are and the more they overreact.
Hoarding is likely to contribute to the scarcity of certain goods or services and
thus to market tightness by sharply increasing demand. Herd behavior can lead to
self-fulfilling prophecies: If market participants behave in a certain way, this can
cause the fundamentals underlying an investment to change as a result of the herd
2.1 Different Concepts for Defining Intelligence 9
→ = ρ→
div D
.
div B→ = 0
∂ B→
rot E→ + =0
∂t
∂D →
rot B→ − = j→.
∂t
The question immediately arises, of course, as to how to compare the various forms
of intelligence defined in the previous section. After all, they are not completely
congruent and are naturally expressed differently in each person. For this reason, a
variety of intelligence tests has been defined.
An intelligence test is a psychological diagnostic tool used to measure a person’s
intelligence. There are many different types of intelligence tests. Differences in
performance on intelligence tests are thought to reflect differences in cognitive
performance in daily life. Psychological tests are used as a tool to find an optimal
decision in the context of various diagnostic issues. For example, the assessment of
intelligence can be helpful in predicting career success or career aptitude (personnel
selection or career counseling) or in recommending a school education or choice of
study. Various clinical issues (presence of diseases such as dementia or reduced
intelligence or the application of vocational rehabilitation measures) may also
require assessment of intelligence.
The best known result of some of these tests is the intelligence quotient (IQ).
Because of the danger of making IQ an absolute designation for a person (like height
or weight—i.e., independent of a particular question), this measure is no longer used
in professional language, and other standard scales are used to describe intelligence
performance.
Intelligence tests are often controversial. This is due, among other things, to the
multitude of factors that influence intelligence, a possible link to heredity issues,
and doubts about objectivity. It is very easy to pass the time with intelligence tests
offered on the Internet, but it is not easy to draw a reasonable conclusion from
them. Very often, math comprehension is asked; also the ability to quickly remember
a larger set of figures and to classify them according to not immediately visible
features is helpful. Often, the time available also plays a role. If one performs the
same test with a lower or increased thinking time, the result can be quite different.
Even this brief overview shows that it is very difficult to find a generally valid
and accepted definition of intelligence that meets all requirements. Of course, this
also shows when one wants to speak of “artificial intelligence.”
There is still a fundamental quality to be considered that falls under the umbrella
of many discussions, but which is also part of the concept of competence. Very often,
people say a person can do this or that. But what is actually meant is there are some
people who can do this or that. These are the specialists for a certain activity, and
very often, they are extremely rare. If you want to introduce (intelligent) hardware
and software in a certain area, you cannot make sure early enough that specialists
are available (Fig. 2.2).
Reference 11
Reference
1. Gunter Dueck. Schwarmdumm—So blöd sind wir nur gemeinsam, Goldmann 2. Auflage, 2018,
978-3-442-15950-5
Chapter 3
The Historical Development
This chapter will show which mathematical and technical means were necessary so
that intelligent applications could be achieved in a wide variety of fields. It is also, it
should be emphasized here, not at all meaningful to speak of “artificial intelligence.”
There are many applications that are only related to each other in that they are based
on the use of computers.
Many modern scientific theories related to mathematics find their beginning with
Gottfried Wilhelm Leibniz (1646–1716, Fig. 3.1).
Some of his research findings and initiatives were:
• description of the dual system,
• development of decimal classification,
• plans for a submarine,
• improvement of technology of door locks,
• device for determining wind speed,
• advice to physicians to take fever regularly,
• establishment of a widows and orphans’ fund,
• proof of the unconsciousness of man,
• infinitesimal calculus (integral calculus and differential calculus),
• matrices and determinants,
• invention of the staggered roller for a mechanical calculating machine,
• development of the endless chain for mining ore extraction.
As a friend, professional colleague, and encourager of the European scribes of the
time and author of linguistically influential writings, he was one of the key initiators
in the founding of modern linguistics, particularly Indo-European studies.
In the letter to Duke Rudolph August of January 12, 1697, Leibniz describes
and outlines for the first time his idea of the “dual system.” At the center of his
explanations is the Christian belief that one can create everything (1) from nothing
(0), which leads to the famous sentence omnibus ex nihilo ducendis sufficit unum.
Today we are familiar with the fact that any number greater than 1 can serve as
a base for the representation of any number. Currently, the number 10 is used in
human everyday life and the base 2 for all digital devices.
The decimal representation assigns powers of 10 to the positions from right to
left: .. . . 104 103 102 101 100 .
If you use base 10, you need the ten digits .0, 1, 2, 3, 4, 5, 6, 7, 8, 9.
The binary (dual, two-valued) representation of the same number:
728 = 1011011000 = 1·29 +0·28 +1·27 +1·26 +0·25 +1·24 +1·23 +0·22 +0·21 +0·20 .
.
We can see that the dual representation requires only the digits 0 and 1. But since
each power of 2 contributes much less to the value of the number than the powers
of 10, you need a larger number of powers of two, and the representation becomes
much longer.
The addition of dual numbers is quite simple:
0+0 = 0
.
0+1 = 1
1+0 = 1
1+1 = 0 carry 1 to the next position on the left
0·0 = 0
.
0·1 = 0
1·0 = 0
1·1 = 1
The dual system can also be approached from a completely different angle.
The English mathematician, logician, and philosopher George Boole (1815–1864,
Fig. 3.2) founded in his work The Mathematical Analysis of Logic (1847), the first
algebraic logic calculus, which has fundamental importance for the development of
computer technology.
The premises for propositional logic are simple and self-evident:
• Repeating a statement does not change its truth value. A false statement remains
false, and a true statement remains true.
His further conclusion is brilliant: he transfers this fact to arithmetic, which
has worked without error for millennia. He introduces a logical variable x and
formulates:
x 2 = x.
.
This equation has the two solutions .x = 0 and .x = 1. Thus, if arithmetic is free of
contradictions, then a calculus of logics can have only two values.
This also has a (funny) side effect. Repeating a message adds nothing in terms of
truth or falsehood. A true message remains true, and a false message remains false.
16 3 The Historical Development
On these two values, he built the Boolean algebra. Today it is the basis
for the design of microelectronic circuits, but it also plays a very important role
in programming and in artificial intelligence. The following logical functions are
available:
• conjunction .∧,
• disjunction .∨,
• exclusive-or .⊕,
• equivalence .∼
• negation .¬.
The symbol .¬ is rarely used and is usually replaced by a horizontal line .x.
The symbol .∧ is usually omitted (as is often the case with multiplication). So
we understand ab as .a ∧ b. The use of these operations in logic gave .∧ the name
and: the conjunction of two statements by and is true if either statement is true;
otherwise, it is false.
.∨ is denoted by or: to obtain a true statement, the first statement or the second
statement must be true; it is allowed that both statements are true. We can see that
the equivalence function is equal to 1 when x and y have the same value. The value
of the exclusive-or is equal to 1 if the values of x and y are different. It can be seen
that antivalence and equivalence merge by negation.
Negation
x .x
0 1
1 0
The comparison of the table for the addition of binary numbers with the table
for the exclusive-or shows that they are exactly the same; therefore, the formula
3.1 The Dual System 17
sign .⊕ is used for them. This is also true for the tables of conjunction and
multiplication of binary numbers. The disjunction and the negation can also be
expressed arithmetically:
.a ∨ b = a + b − a · b, x = 1 − x.
The available circuits became smaller and faster, and larger and larger amounts of
data could be stored.
The fastest supercomputer in the world at the moment of writing is probably
Fugaku at the RIKEN Center for Computational Science in Kobe, Japan. It
performs 415.5 petaflops with its 7,299,072 cores. One petaflop corresponds to a
speed of .250 floating-point operations (operations with real numbers) per second.
3.2 The Dartmouth Conference 19
The other side of the coin: there has been a huge development in presenting
information digitally, as was the case with numbers. Nowadays, one can assume
that any physical phenomena can be represented digitally and thus be displayed
and processed by computers: sound waves, electromagnetic waves, X-ray images,
ultrasound images, images from a tomographer, and images from space are stored
in digital form immediately after they are obtained and can be used accordingly.
The first time the term artificial intelligence has been used explicitly was in the
proposal
• “A Proposal for the Dartmouth Summer Research Project on Artificial Intelli-
gence”
by J. McCarthy, Dartmouth College, M. L. Minsky, Harvard University, N.
Rochester, I.B.M. Corporation, C.E. Shannon, Bell Telephone Laboratories,
dated August 31, 1955 (Fig. 3.6).
The following topics were discussed:
• Automatic computers,
• How a computer can be programmed to use a language,
• Neural networks,
• Theory of the size of a computation,
• Self-improvement,
• Abstractions,
• Randomness and creativity.
However, there were already some publications before that, which used terms
like thinking computer, thinking machines, or similar. It can therefore be assumed
that this idea of intelligent machines had already been around for a while but that
the conference immediately put it on the agenda.
carried out by machines, especially since these processes could also be carried out
even faster.
Exemplary for this necessity is the introduction of the programmable loom by
Joseph Marie Jacquard, (1752–1834), at the beginning of the nineteenth
century. The loom worked with punched cards, which contained the information
about the pattern to be woven. The craft of weaving could now be continued as a
“programming art” on a medium analogous to the fabric pattern (Fig. 3.7).
The lambda calculus was developed by Alonzo Church (1903–1995) and Stephen
Kleene (1909–1984) in the 1930s. It was proved early on that the lambda calculus
is, in principle, a universal programming language.
With the completion of the first electronic calculators and the use of Boolean
algebra, the development of programming languages continued noticeably. Notable
in this regard are the patents of Konrad Zuse (Fig. 3.8) in about 1937, describing a
computer architecture that would later become known as the von Neumann machine.
In the 1940s, Zuse completed his programming language plan calculus, in
German Plankalkül for it, incorporating ideas from the lambda calculus. The
von Neumann architecture (VNA) is a reference model for computers, according to
which a shared memory holds both computer program instructions and data. von
Neumann systems belong to the class of SISD (Single Instruction, Single Data)
architectures according to Flynn’s classification, as distinct from parallel processing.
The von Neumann architecture forms the basis for the way most computers
known today work. It is named after the Austro-Hungarian mathematician John von
Neumann, who later worked in the USA and whose major work on the subject was
22 3 The Historical Development
However, new concepts also emerged during this time. Great importance was
given to object-oriented programming, which unites data, procedure,
and reference aspects in the single concept of the object. Thinking and concepts
of object orientation first appeared in SIMULA 67, a language for simulation
purposes, which was the first language to introduce object-oriented procedures. The
story goes that its developers, Ole-Johan Dahl (1931–2002) and Kristen
Nygaard (1926–2002), had been working on ship simulations. In the process,
the unmanageable number of parameter relationships resulted in a bewildering
variety of ways in which the various attributes of the different ships could influence
each other. So they came up with the idea of treating the different ship types
as independent objects, with each class of objects being responsible for its own.
However, in SIMULA, the new concepts were not yet clearly distinguished from
their implementation. Their concepts and procedures had already been refined and
implemented more consistently than in SIMULA since the early 1970s at the Xerox
Palo Alto Research Center with the language SMALLTALK. Smalltalk was finally
released to the public at large in the 1980s. SMALLTALK was designed as a fully
dynamic system, where objects could be created and changed interactively—in
contrast to the system of static programs used before. Remarkable also compared
to its successors is the integration of the language in an innovative graphical user
interface, which for the first time allowed real interaction.
After Niklaus Wirth (Fig. 3.9) was disappointed while working on ALGOL,
he developed PASCAL together with Kathleen Jensen and consistently used
PASCAL for teaching language concepts from 1971 on. After realizing how
difficult it was to implement software development of larger projects with multiple
developers, he published MODULA-2, a further development with a strict module
and interface concept, in 1978.
Alain Colmerauer (1941–2017), Phillipe Roussel, and Robert Kowalski (b. 1941)
founded logical programming starting in 1972, which is expressed in several
realizations of the PROLOG language.
In the 1970s, the US Department of Defense expressed concern about the number of
programming languages used in its projects, which numbered over 450. Many of the
programming languages were also not standardized but rather vendor dependent. A
working group was tasked with clearing this jungle and finding a language that met
the department’s military requirements. Many existing languages were reviewed,
but in 1977, the working group concluded that none of the existing languages were
suitable. After inviting bids for four draft languages, it was decided in 1980 to
use the language ADA composed from the best ideas of the drafts. This draft was
standardized under the designation MIL-STD 1815, because in 1815, the eponym
Ada Lovelace (1815–1852) was born. The US Department of Defense temporarily
mandated that any software project with more than 30% of new code be written
in ADA. To support the spread of the standard, the US Air Force funded the
development of the free GNAT compiler. The number of different programming
languages in the Department eventually reduced to 36.
Object-oriented programming began to become more popular in the mid-1980s,
largely due to the influence of C++, which was designed as a syntactic extension
of the C language. Bjarne Stroustrup (b. 1950) had introduced C++ in
1983. Many existing programming languages received object-oriented extensions
since that time, such as PASCAL or LISP.
Especially for the development of artificial intelligence, the languages PROLOG
and LISP were intended. PROLOG realizes a logical programming based on facts
and rules, while LISP realizes a functional programming. It is used for problems
where a prototype of the solution must be created quickly.
From 1945 to 1948, Turing was employed at the National Physical Laboratory
in Teddington, where he worked on the development of the ACE (Automatic
Computing Engine). The name of the machine is derived from the analytical engine
of the mathematician Charles Babbage, whose work Turing admired throughout his
life (Fig. 3.11).
In 1948, Turing taught at the University of Manchester and became deputy head
of the computer department in 1949. Here he worked on the software for one of
the first real computers, the Manchester Mark I, while continuing to write various
theoretical papers. In Computing Machinery and Intelligence (Mind, October 1950),
Turing took up the problem of artificial intelligence and proposed
the Turing test as a criterion for whether a machine is capable of thinking like
a human. Since the thinking process cannot be formalized, the test looks only at
the responses of a machine in dialogue with a human, i.e., the communicative
behavior of the machine. If this appears indistinguishable from human behavior,
one should speak of machine intelligence. With his publication, he significantly
influenced the development of artificial intelligence. In 1952, he wrote the chess
program Turochamp. Since there were no computers with sufficient power to run
it, Turing took over its function and calculated each move himself. This took up
to 30 minutes per move. He lost the only game documented in writing against a
colleague.
From 1952 until his death in 1954, Turing worked on mathematical problems
in theoretical biology. In 1952, he published a paper on the chemical basis of
morphogenesis. In this work, he described for the first time a mechanism by
which reaction-diffusion systems can spontaneously evolve structures. This process,
work tape
control
M = (Z, E, A, d, q, F )
.
with
• Z a finite nonempty set of states of the control system,
• E the input alphabet,
30 3 The Historical Development
The working alphabet A contains the special characters $ and the space; they do
not occur in the input alphabet. The characters L and R do not occur as alphabet
characters.
• .E ⊆ A means that every element of the set E is an element of the set A; possibly,
as a limiting case, .E = A.
• The sign .∪ means the union of two sets; one obtains the set .A' if one adds to the
set A the movements L (to the left) and R (to the right).
In addition to the computation of functions, the Turing machine is also used for
decision problems, i.e., for questions that have to be answered with “yes” or “no.”
Certain final states are defined as “accepting” while others as “not accepting.” The
input is accepted exactly when the Turing machine terminates in an accepting final
state.
With his model, Turing defined the terms “algorithm” and “computability”
as mathematical concepts. It is generally assumed that Turing’s computability
corresponds to the intuitive understanding of computability; this statement is known
as the Church-Turing thesis . The distinctive feature of a Turing machine
is its structural simplicity. It requires only three operations (reading, writing, and
moving the read-write head) to simulate all the operations of common computer
programs.
3.6 LaTeX
Scientific papers are very often difficult to write. Here we find two languages that are
extremely significant. LaTeX describes documents quite precisely and in all details.
The basic program of LaTeX is TeX and was developed by Donald E. Knuth during
his time as a computer science professor at Stanford University. Building on TeX,
Leslie Lamport developed LaTeX in the early 1980s, a collection of TeX macros
that simplified and enhanced its use for the average user compared to TeX. The
name LaTeX is an abbreviation for La—import—TeX. The details of the document
are described and converted by Tex into the document, into a form in which it is then
to be presented. To support the creation of the description, an editor that supports
the input of the LaTeX text is still necessary. For this book, MiKTex was used
as a program for translation, and TeXnicCenter is the used editor. However, there
are other equivalent systems available. They are all provided free of charge on the
Internet. In contrast to other word processing programs, which work according to
the what-you-see-is-what-you-get principle, LaTeX works with text files in which
3.6 LaTeX 31
O
H
H
. The formula for water
The introduction of the euro sign was quickly taken into account, since it was
not present on the keyboards at the beginning: you simply write \euro{ } and get e.
There is plenty of teaching material on the Internet, which you have to learn step by
step.
Moore’s law shows that computing power is constantly increasing. Where have
we gotten to now? According to Wikipedia [2], the Japanese computer Fugaku
RIKEN from the Center for Computational Science, Kobe, Japan, is in first place,
with 415,530.00 teraflops. It is followed by two American and two Chinese systems.
The German system, Hawk, is located in Stuttgart, Germany, and has a performance
of 26,000 teraflops.
The SC21 supercomputing show comes with confirmation that China is running
the world’s first two exascale supercomputers: OceanLight and Tianhe-3. According
to this, OceanLight achieves a peak computing performance of 1.3 ExaFlops with
double precision (FP64) in the Linpack benchmark—over a longer period of time,
1.05 Exa-Flops are possible. If the National Research Center of Parallel Computer
Engineering and Technology (NRCPC) were to include the supercomputer in its
Top500 list, it would be around twice as fast as the previous leader, Japan’s Fugaku.
Development in the United States is proceeding in much the same way. A kind of
race and prestige thinking has emerged (see Figs. 3.14 and 3.15) [2, 3].
The costs are quite high: the cost of the Summit supercomputer was about 44
million euros, while the annual electricity costs are up to 4 million euros per year.
The next step might be quantum computers. Unlike the classical computer, it
operates on the basis of quantum mechanical states rather than electrical ones. Here,
firstly the superposition principle and secondly the quantum entanglement are of
importance.
Theoretical studies show that by exploiting these effects, certain problems in
computer science, such as searching extremely large databases and factoring large
numbers, can be solved more efficiently than with classical computers. This would
make it possible to significantly reduce the computation time for many mathematical
and physical problems.
3.7 The Evolution of Computers 33
For a long time, the quantum computer has been a theoretical concept. There
have been various proposals on how a quantum computer could be realized; on a
small scale, some of these concepts have been tested in the laboratory, and quantum
computers with a few qubits have been realized. The record was 127 qubits for the
processor in November 2021. However, in addition to the number of qubits, other
34 3 The Historical Development
important factors exist, for example, a low error rate during computation and readout
and how long the states in the qubits can be maintained without errors.
qubits or quantum bits are the computational units of a quantum com-
puter. Qubits can store more information than just 1 or 0 compared to conventional
bits in computers. Qubits also exist in a superposition of these values—in what is
known in technical jargon as superposition. Unlike conventional bits, quantum bits
can assume any number of states simultaneously. Quantum computers thus achieve
great computing power with just a few qubits. Theoretically, the computing power
of a quantum computer increases exponentially with the number of qubits. Even
with just a few qubits, complex problems can be solved in a very short time.
In addition to the technology giant IBM, the industry itself has long recognized
the great potential of quantum computers and has joined forces in a consortium. The
aim of the consortium, called Quantum Technology and Application
Consortium (QUTAC), is to test industrial applications with high potential for
quantum computers.
Overall, quantum computers are expected to be used in the following areas:
• Logistics
Quantum computers could help optimally allocate limited resources in logistics,
traffic flow and transportation routes, network structures of telecommunication
companies, or the distribution of medical resources for hospitals.
• Artificial intelligence and machine learning
According to the Fraunhofer Alliance “Big Data and Artificial Intelligence,”
quantum computers could revolutionize artificial intelligence and machine learn-
ing: “Procedures of artificial intelligence and machine learning can be adapted
for quantum computers in such a way that they can follow several solution paths
at the same time. This allows quantum computers to process large data sets in
a single step, detect patterns in the data that classical computers cannot, and
provide reliable results even on incomplete or uncertain data.”
• Robotics
BMW’s example is just a taste of the potential of quantum computing for
robotics. Quantum computing could determine the most efficient paths for
employees or robots (AGVs, Autonomous Guides Vehicles) to navigate through
a warehouse.
• Chemistry, pharmacy, and materials science
The Fraunhofer “Cluster of Excellence Cognitive Internet Technologies” is
certain: “With the help of the simulation of molecules, catalysts could be
developed in a targeted manner in the future, for example, to make chemical
production processes more efficient. Opportunities of a similar magnitude arise
for the pharmaceutical industry.” And battery research also falls within the
scope of quantum computer-based simulations—the development of novel, more
powerful batteries for electric cars.
• Engineering
References 35
Are materials with certain properties desired and is a trade-off between stability
and weight important?—a case for quantum computing, because the properties of
any material depend fundamentally on its quantum mechanical building blocks.
• Finance
Faster financial simulations and better portfolio optimization solutions are
potential applications that make quantum computing attractive to the financial
industry.
• The German government has released a total of two billion euros for the devel-
opment of quantum computers for the international race for quantum computing.
The goal is to build a competitive quantum computer in Germany within the
next five years. So far, there is no quantum computer in Germany that has
been built entirely without technology from abroad. Currently, the German mid-
sized companies Trumpf and Sick are considered leading competence carriers in
quantum optical sensors [4].
So the path to ever more powerful computers is also far from having reached
its end. However, the design of circuits based on this principle is much more
complicated.
References
The solution of mathematical problems with the help of a computer is a very original
concern of the development of calculating machines (Fig. 4.1) and computers. To the
extent that mathematics is applied in many fields, computers and their programs can
be applied as well.
At the beginning, it was essentially a matter of calculating with numbers, but
very quickly the programs were extended to symbolic arithmetic, solving equations,
differential and integral calculus, and all other areas of mathematics. There are now
a large number of such systems, usually grouped together under the name “systems
for computer algebra.”
It is difficult to choose. It is best to start with a smaller system and work your
way through it. This already costs a lot of effort. Later one can decide whether the
system is already sufficient or whether one needs a larger, more elaborated system.
Small systems have above all the advantage that one can use them without costs. The
existence of such systems will have a significant impact on mathematics education
in the coming years.
Before presenting individual software packages with which one can work,
it should be mentioned that one can obtain valuable information by searching
the Internet. Formulas and explanations are provided there for all subfields of
mathematics. A prudent high school graduate can complete his homework in a short
time.
arises: on the one hand, it is sensible not to spend a lot of time on calculations that
a PC has at its fingertips. But on the other hand, it could be that the independent
calculations also add to the amount of knowledge.
However, one can also consider that in this way, it is possible to no longer regard
mathematics as an “‘enemy” that one must avoid as soon as possible. One can
make the lessons much more interesting by including the relations of mathematics
to architecture, painting, and others. There are also two textbooks [1, 2] for this
system, which can be found on the Internet at bookboon.com.
4.3 Maple 39
4.2 Mathematica
4.3 Maple
The first version of Maple was programmed in 1980 by Keith O. Geddes, Gaston
H. Gonnet, and their collaborators from the Symbolic Computation Group at the
University of Waterloo in the Canadian city of Waterloo (Ontario). At the end of
1987, Maple already existed in the version 4.2. (Fig. 4.4).
In 2005, Maple 10 introduced a new document mode within the standard version
of Maple. Since then, it is possible to edit Maple input in normal mathematical
notation. Here, texts and mathematical symbols can be combined in the same input
line.
40 4 Mathematics on the Computer
The main component of the graphical user interface of Maple is the respective
worksheet, in which interactive work is done. It appears as a window in which
calculation instructions are entered. The Maple engine interprets these instructions
and returns corresponding outputs. Typical Maple outputs are numerical values,
terms, functions, tables, two- and three-dimensional graphics, animation objects,
and diagrams.
Maple includes a core of commonly used standard computational instructions
and additional packages loadable at runtime with the with command. It is exhilarat-
ing to have almost the “entire currently known mathematics” on your computer. A
technician, engineer, geologist, biologist, physician, etc. is of course faced with the
problem of selecting the subfields that are important to him. Therefore, in practice,
one already encounters some sub-domains such as business mathematics, financial
mathematics, statistics, etc., but further such specializations and narrowing down
will be necessary. Education can focus on teaching the basic ideas; the calculations
themselves and their presentation can be left to computer technology.
References
1. Christian Posthoff, Bernd Steinbach. Mathematik—Ohne Sorgen an der Uni I—Nutze Microsoft
Mathematics, bookboon.com, 2017, 978-87-403-1596-7
2. Christian Posthoff, Bernd Steinbach. Mathematik—Ohne Sorgen an der Uni II—Nutze Microsoft
Mathematics, bookboon.com, 2017, 978-87-403-1949-1
Chapter 5
Polynomial and Exponential Complexity
The interrelationships presented in the next sections and chapters make it clear that
artificial intelligence is always up for discussion when one is faced with complex
systems whose investigation far exceeds the capabilities of humans. Worldwide
globalization, the fact that everyone is networked with everyone else, offers great
opportunities on the one hand but also harbors dangers. The possibilities are often
pushed into the distance, while disaster scenarios are immediately painted on the
wall when it comes to the risks.
The high level of interconnectedness means that a change at one point can
propagate throughout the system almost unhindered. Very often, a butterfly
effect occurs. It states that in complex, nonlinear dynamic systems, even the
smallest changes in the initial conditions, can lead to the fact that predictability
with respect to the further development of a system can basically be ruled out. The
butterfly effect has become known especially in connection with the topic weather
forecast. This is to express figuratively that already the wing beat of a butterfly at a
point X on earth can influence the weather event at another place Y. As astonishing
as this may seem at first glance, this effect is understandable. The simple scheme
cause .→ effect is not applicable to complex systems; even small causes can have a
large effect (Fig. 5.1).
The catchy term butterfly effect originated with US meteorologist Edward N.
Lorenz, who in 1972 gave a talk to the American Association for the Advancement
of Science titled “Predictability: Does the Flap of a Butterfly’s Wings in Brazil Set
Off a Tornado in Texas?” In its original form, however, it used the flap of a seagull’s
wings instead of the butterfly.
Preliminary work on the theory was done by Edward N. Lorenz (1917–
2008) with a 1963 paper in which he undertook a computer weather prediction
calculation. In connection with long-term weather forecasts, he used a simplified
convection model to investigate the behavior of liquids or gases when they are
heated: here, rolls initially form (hot gas rises on one side, loses heat, and sinks
again on the other side), which become unstable when further heat is added.
He characterized this behavior using three connected differential equations. He
projected the numerical result into phase space and obtained that strange attractor
that later became known as the Lorenz attractor: an infinitely long trajectory in three-
dimensional space that does not intersect itself and has the shape of two butterfly
wings.
Lorenz came across the chaotic behavior of his model rather by accident. In
order to save computing time, he had used intermediate results of already performed
calculations for the numerical solution of the equations but only took three decimal
places into account, although the computer calculated with an accuracy of six
decimal places. The differences between the old and the new calculations became
greater and greater, although they could only be based on deviations in the fourth
digit.
This led Lorenz to his statements about the sensitivity to the initial conditions.
From nearly the same starting point, the weather curves diverged until they finally
showed no commonality. In his first calculation, he gave an initial value for one
iteration to six decimal places (0.506127) and in the second calculation to three
(0.506), and although these values differed only by about .1/10,000, as time went
on, this calculation diverged sharply from the first.
If you consider that the circumference of the earth at the equator is 40,000 km,
then one ten-thousandth of that is equal to 4 km. A change in the landscape along
this distance is enough to trigger a chaotic event somewhere (Fig. 5.2).
The varying trajectories .x(t), y(t), z(t) are solution of the following differential
equation system:
dx
. = σ · [y(t) − x(t)]. (5.1)
dt
5.1 Introductory Considerations 43
dy
= [ρ · x(t)] − [x(t) · z(t)] − y(t). (5.2)
dt
dz
= [x(t) · y(t)] − β · z(t). (5.3)
dt
Here .σ, ρ and .β are parameters to be defined beforehand. The nonlinearity can be
seen in the terms .[x(t) · z(t)] and .[x(t) · y(t)].
Because of the finite mantissa in floating point representation, numbers cannot
be represented with arbitrary precision on a computer. They must be rounded.
Instead of .x, the computer uses the number .rd(x) for another calculation. Today’s
computers usually work according to IEEE 754. The machine precision for the
data types used in this process is . = 2−24 ≈ 6 · 10−8 for single precision and
. = 2
−53 ≈ 1.1 · 10−16 .
We begin with a simple example. We imagine a wall board with a row of books.
If one is now looking for a particular book, one starts from the left and checks one
book after another to see if it is the wanted book. If you are lucky, it is the first one,
but if you are unlucky, it is the last book in the row, or it is not there at all. So the
number of necessary steps is between 1 and n, if n is the number of books on the wall
board. To compare different algorithms in a meaningful way, one always uses the
worst case (worst-case complexity) and says that the linear search
has complexity .O(n).
In the next step, we imagine a shelf that contains n boards for books. So there are
.n · n books there, and in the worst case, it takes .n search steps. The complexity of
2
2
the search is of order .O(n ). If we now go to a library, there are many such shelves
in a row. We again assume that there are n pieces, and the complexity of the search is
now equal to .O(n3 ). However, one can also imagine a very long wall shelf where all
the books are in a row, and one uses the same linear search algorithm. Therefore, we
group all these problems into the set for which the running time .t (n) of a solution
algorithm is bounded by a polynomial of n-th degree:
t (n) ≤ nk .
. (5.4)
M = {a, b, c}.
.
For some reason, we are interested in determining the power set of M, that is, the
set of all subsets. This is not difficult.
5.2 Mathematical Descriptions of Complexity 45
⎛ ⎞
a b c Menge
⎜0 0 0 ∅ ⎟
⎜ ⎟
⎜1 0 {a} ⎟
⎜ 0 ⎟
⎜ ⎟
⎜0 1 0 {b} ⎟
⎜ ⎟
. ⎜0 0 1 {c} ⎟
⎜ ⎟
⎜1 1 0 {a, b} ⎟
⎜ ⎟
⎜1 0 1 {a, c} ⎟
⎜ ⎟
⎝0 1 1 {b, c} ⎠
1 1 1 {a, b, c}
You form a matrix where each row has four positions. The first three columns are
assigned to the three elements, a, b and c; the last column shows the set represented
by the row. The value 0 indicates that the element does not belong to the subset; the
number 1 indicates that it belongs to the subset. Since both possibilities are possible
for each element, there are .2 · 2 · 2 = 23 = 8 possibilities. In real life, you can see
this in a family tree: if you disregard all misfortunes and other abnormalities, you
have
• one child,
• two parents,
• four grandparents,
• eight great-grandparents.. . .
• sixteen great-great grandparents, etc.
So here one sees the same powers of the number 2 as before. More generally, the
power set of a set with n elements contains .2n elements.
We consider the function
f (a, b, c, d) = ab ⊕ (c ∨ d).
.
Here four variables occur with possible values of 0 or 1. A table of the function
values now has .24 = 16 rows. So the size of the tables for the values of binary
(logical) functions also grows exponentially.
46 5 Polynomial and Exponential Complexity
⎛ ⎞
a b c d f (a, b, c, d
⎜0 ⎟
⎜ 0 0 0 0 ⎟
⎜ ⎟
⎜0 0 0 1 1 ⎟
⎜ ⎟
⎜0 0 1 0 1 ⎟
⎜ ⎟
⎜0 0 1 1 1 ⎟
⎜ ⎟
⎜0 1 0 0 0 ⎟
⎜ ⎟
⎜0 ⎟
⎜ 1 0 1 1 ⎟
⎜0 ⎟
⎜ 1 1 0 1 ⎟
⎜ ⎟
. ⎜0 1 1 1 1 ⎟
⎜ ⎟
⎜1 0 0 0 1 ⎟
⎜ ⎟
⎜1 0 0 1 0 ⎟
⎜ ⎟
⎜1 0 1 0 0 ⎟
⎜ ⎟
⎜1 ⎟
⎜ 0 1 1 0 ⎟
⎜ ⎟
⎜1 1 0 0 0 ⎟
⎜ ⎟
⎜1 1 0 1 1 ⎟
⎜ ⎟
⎝1 1 1 0 1 ⎠
1 1 1 1 1
Since the field of circuit design makes exclusive use of binary functions, one
constantly encounters exponential complexity there (Fig. 5.3).
Even in everyday life, one is confronted with exponential complexity, as can
easily be seen in the family tree of a human being (Fig. 5.4).
It is an important characteristic of problems that are handled by means and
methods of artificial intelligence that they have a very high complexity. It is not
always known which function describes the complexity; in any case, it is far above
the complexity of problems that can be solved by humans.
The result is a huge web of relationships in which only individual parts can be
seen through separately. Not for nothing do court cases go on for years. Inheritance
disputes can destroy entire families and much more. And there is a long and
interesting history of how such systems of rules are created, introduced, or enforced.
Unfortunately, they do not always follow humane and correct principles either. In
any case, one can immediately see that this is the next extremely complex issue.
In mathematics and the natural sciences, implication is defined as a function to
formalize the application of rules. Linguistically, the implication is expressed by a
sentence with the structure “if assumption then conclusion.”
Implication
x y .x ⇒ y .x ∨ y
0 0 1 1
0 1 1 1
1 0 0 0
1 1 1 1
If one accepts a rule as valid and the assumption is true, then one receives with
certainty a true conclusion, about which one does not have to think further! If the
premise is not fulfilled .(x = 0), then the rule must not be applied.
Now, a scientific way of working is to have a set of rules to apply and a set of true
propositions. The application of the rules results in new true propositions, which one
adds to the set of true propositions one already has. You continue this process until
you have reached a certain goal.
And of course, it is questionable whether a theorem can really be proved with the
help of the premises and whether the proof is also correct. One famous example is
the proof of Fermat’s last theorem by Andrew Wiles.
Theorem No three positive integers a, b, and c can satisfy the equation
a n + bn = cn
. (5.5)
(x → y) = (x ∨ y).
. (5.6)
If .x = 0, i.e., the assumption is false, but the implication is true and does
not depend on x. From a premise and its converse, two parties reach the same
conclusion. What is the effect of such a situation? The use of this rule is in the
given context without any meaning, the rule can be skipped.
When correct assumptions are used, the number of true conclusions should be
growing.
5.4 Axioms 49
5.4 Axioms
the axioms were common to many sciences, the postulates of each particular
science were different. Their validity had to be established by means of real-world
experience. Aristotle warns that the content of a science cannot be successfully
communicated if the learner is in doubt about the truth of the postulates.
The classical approach is well illustrated by Euclid’s Elements, where a list of
postulates is given (common-sensical geometric facts drawn from our experience),
followed by a list of common notions (very basic, self-evident assertions).
• It is possible to draw a straight line from any point to any other point.
• It is possible to extend a line segment continuously in both directions.
• It is possible to describe a circle with any center and any radius.
• It is true that all right angles are equal to one another.
• (Parallel postulate) It is true that if a straight line falling on two straight lines
make the interior angles on the same side less than two right angles, the two
straight lines, if produced indefinitely, intersect on that side on which are the
angles less than the two right angles.
By making small changes in the axiom system, one can define a new subfield of
mathematics. One impressive example is the definition of non-Euclidean geome-
tries.
Non-Euclidean geometry arises by replacing the parallel postulate with an
alternative. In this case, one obtains hyperbolic geometry and elliptic geometry.
The beginning of the nineteenth century would finally witness decisive steps
in the creation of non-Euclidean geometry. Circa 1813, Carl Friedrich Gauss
and, independently around 1818, the German professor of law Ferdinand Karl
Schweikart had the germinal ideas of non-Euclidean geometry worked out, but
neither published any results. Schweikart’s nephew, Franz Taurinus, did publish
important results of hyperbolic trigonometry in two papers in 1825 and 1826, yet
while admitting the internal consistency of hyperbolic geometry, he still believed in
the special role of Euclidean geometry.
Then, in 1829–1830, the Russian mathematician Nikolai Ivanovich Lobachevsky
and, in 1832, the Hungarian mathematician János Bolyai separately and inde-
pendently published treatises on hyperbolic geometry. Consequently, hyperbolic
geometry is called Lobachevskian or Bolyai-Lobachevskian geometry, as both
mathematicians, independent of each other, are the basic authors of non-Euclidean
geometry. Gauss mentioned to Bolyai’s father, when shown the younger Bolyai’s
work, that he had developed such a geometry several years before, though he did
not publish. While Lobachevsky created a non-Euclidean geometry by negating the
parallel postulate, Bolyai worked out a geometry where both the Euclidean and
the hyperbolic geometry are possible depending on a parameter k. Bolyai ended his
work by mentioning that it is not possible to decide through mathematical reasoning
alone if the geometry of the physical universe is Euclidean or non-Euclidean; this is
a task for the physical sciences.
Bernhard Riemann, in a famous lecture in 1854, founded the field of Riemannian
geometry, discussing in particular the ideas now called manifolds, Riemannian met-
ric, and curvature. He constructed an infinite family of non-Euclidean geometries by
5.4 Axioms 51
giving a formula for a family of Riemannian metrics on the unit ball in Euclidean
space. The simplest of these is called elliptic geometry, and it is considered a non-
Euclidean geometry due to its lack of parallel lines.
By formulating the geometry in terms of a curvature tensor, Riemann allowed
non-Euclidean geometry to apply to higher dimensions. Beltrami (1868) was the
first to apply Riemann’s geometry to spaces of negative curvature.
It was Gauss who coined the term non-Euclidean geometry. He was referring
to his own work, which today we call hyperbolic geometry or Lobachevskian
geometry. Several modern authors still use the generic term non-Euclidean geometry
to mean hyperbolic geometry.
Arthur Cayley noted that distance between points inside a conic could be defined
in terms of logarithm and the projective cross-ratio function. The method has
become called the Cayley-Klein metric because Felix Klein exploited it to describe
the non-Euclidean geometries in articles in 1871 and 1873 and later in book form.
The Cayley-Klein metrics provided working models of hyperbolic and elliptic
metric geometries, as well as Euclidean geometry.
Klein is responsible for the terms hyperbolic and elliptic (in his system, he called
Euclidean geometry parabolic, a term that generally fell out of use). His influence
has led to the current usage of the term non-Euclidean geometry to mean either
hyperbolic or elliptic geometry.
Euclidean geometry can be axiomatically described in several ways. Unfortu-
nately, Euclid’s original system of five postulates (axioms) is not one of these, as his
proofs relied on several unstated assumptions that should also have been taken as
axioms. Hilbert’s system consisting of 20 axioms most closely follows the approach
of Euclid and provides the justification for all of Euclid’s proofs. Other systems,
using different sets of undefined terms, obtain the same geometry by different paths.
All approaches, however, have an axiom that is logically equivalent to Euclid’s
fifth postulate, the parallel postulate. Hilbert uses the Playfair axiom form, while
Birkhoff, for instance, uses the axiom that says that there exists a pair of similar
but not congruent triangles. In any of these systems, removal of the one axiom
equivalent to the parallel postulate, in whatever form it takes, and leaving all the
other axioms intact, produces absolute geometry. As the first 28 propositions of
Euclid (in The Elements) do not require the use of the parallel postulate or anything
equivalent to it, they are all true statements in absolute geometry.
To obtain a non-Euclidean geometry, the parallel postulate (or its equivalent)
must be replaced by its negation. Negating the Playfair’s axiom form, since it is a
compound statement (.. . . there exists one and only one .. . .), can be done in two
ways:
Either there will exist more than one line through the point parallel to the given
line or there will exist no lines through the point parallel to the given line. In the
first case, replacing the parallel postulate (or its equivalent) with the statement In
a plane, given a point P and a line l not passing through P, there exist two lines
through P, which do not meet l and keeping all the other axioms yields hyperbolic
geometry. The second case is not dealt with as easily. Simply replacing the parallel
postulate with the statement In a plane, given a point P and a line l not passing
52 5 Polynomial and Exponential Complexity
through P, all the lines through P meet l does not give a consistent set of axioms.
This follows since parallel lines exist in absolute geometry, but this statement says
that there are no parallel lines. This problem was known (in a different guise) to
Khayyam, Saccheri, and Lambert and was the basis for their rejecting what was
known as the obtuse angle case. To obtain a consistent set of axioms that includes
this axiom about having no parallel lines, some other axioms must be tweaked.
These adjustments depend upon the axiom system used. Among others, these tweaks
have the effect of modifying Euclid’s second postulate from the statement that line
segments can be extended indefinitely to the statement that lines are unbounded.
Riemann’s elliptic geometry emerges as the most natural geometry satisfying this
axiom.
On a sphere, the sum of the angles of a triangle is not equal to 180.◦ . The surface
of a sphere is not a Euclidean space, but locally, the laws of the Euclidean geometry
are good approximations. In a small triangle on the face of the earth, the sum of the
angles is very nearly 180.◦ . Models of non-Euclidean geometry are mathematical
models of geometries, which are non-Euclidean in the sense that it is not the case
that exactly one line can be drawn parallel to a given line l through a point that is
not on l. In hyperbolic geometric models, by contrast, there are infinitely many lines
through A parallel to l, and in elliptic geometric models, parallel lines do not exist.
Euclidean geometry is modeled by our notion of a flat plane. The simplest model
for elliptic geometry is a sphere, where lines are great circles (such as the equator or
the meridians on a globe), and points opposite each other are identified (considered
to be the same). The pseudosphere has the appropriate curvature to model hyperbolic
geometry (Fig. 5.5).
We note here that these axiom systems for geometry branch out.
Until the 1970s, it was completely self-evident that proofs should consist of
logical steps that approached the goal to be proven. But suddenly computers came
into play. Their high speed made it impossible to retrace the steps individually;
only the result counted. This was initially controversial among mathematicians, but
since a correct result was kept in mind, such proofs were later accepted without any
problems.
In mathematics, the four-color theorem, or the four-color map theorem, states
that no more than four colors are required to color the regions of any map so that no
two adjacent regions have the same color. Adjacent means that two regions share a
common boundary curve segment, not merely a corner where three or more regions
meet. It was the first major theorem to be proved using a computer. Initially, this
proof was not accepted by all mathematicians because the computer-assisted proof
was infeasible for a human to check by hand. The proof has gained wide acceptance
since then, although some doubters remain.
The four-color theorem was proved in 1976 by Kenneth Appel and Wolfgang
Haken after many false proofs and counterexamples (unlike the five-color theorem,
proved in the 1800s, which states that five colors are enough to color a map). To
dispel any remaining doubts about the Appel-Haken proof, a simpler proof using
the same ideas and still relying on computers was published in 1997 by Robertson,
Sanders, Seymour, and Thomas. In 2005, the theorem was also proved by Georges
Gonthier with general-purpose theorem-proving software.
The intuitive statement of the four-color theorem—given any separation of a
plane into contiguous regions, the regions can be colored using at most four colors
so that no two adjacent regions have the same color—needs to be interpreted
appropriately to be correct.
First, regions are adjacent if they share a boundary segment; two regions that
share only isolated boundary points are not considered adjacent. (Otherwise, a
map in a shape of a pie chart would make an arbitrarily large number of regions
“adjacent” to each other at a common corner and require arbitrarily large number
of colors as a result.) Second, bizarre regions, such as those with finite area but
infinitely long perimeter, are not allowed; maps with such regions can require more
than four colors. (To be safe, we can restrict to regions whose boundaries consist of
finitely many straight line segments. It is allowed that a region has enclaves and that
it entirely surrounds one or more other regions.) Note that the notion of contiguous
region (technically, connected open subset of the plane) is not the same as that
of a country on regular maps, since countries need not be contiguous (they may
have exclaves, e.g., the Cabinda Province as part of Angola, Nakhchivan as part of
Azerbaijan, Kaliningrad as part of Russia, France with its overseas territories, and
Alaska as part of the United States are not contiguous). If we required the entire
territory of a country to receive the same color, then four colors are not always
sufficient (Fig. 5.6) [6].
However, the verifiability of a proof by humans also has its rough edges. This is
shown by the results of A. Wiles in the proof of Fermat’s Great Theorem. In number
theory, Fermat’s Last Theorem (sometimes called Fermat’s conjecture, especially
in older texts) states that no three positive integers, a, b, and c, satisfy the equation
54 5 Polynomial and Exponential Complexity
the title that Fermat’s last theorem would be discussed,” Dr. Ribet said. .. . . Finally,
at the end of his third lecture, Dr. Wiles concluded that he had proved a general
case of the Taniyama conjecture. Then, seemingly as an afterthought, he noted that
meant that Fermat’s last theorem was true. q.e.d.
In August 1993, it was discovered that the proof contained a flaw in one area.
Wiles tried and failed for over a year to repair his proof. According to Wiles,
the crucial idea for circumventing—rather than closing—this area came to him on
September 19, 1994, when he was on the verge of giving up. Together with his
former student Richard Taylor, he published a second paper that circumvented the
problem and thus completed the proof. Both papers were published in May 1995 in
a dedicated issue of the Annals of Mathematics.
The problems with the proof of the abc assumption were even more difficult. The
abc conjecture refers to numerical expressions of the type .a + b = c. The statement,
which comes in several slightly different versions, concerns the prime numbers that
divide each of the quantities a, b, and c. Every whole number, or integer, can be
expressed in an essentially unique way as a product of prime numbers—those that
cannot be further factored out into smaller whole numbers: for example, 15 = 3 .× 5
or 84 = 2 .× 2 .× 3 .× 7. In principle, the prime factors of a and b have no connection
to those of their sum, c. But the abc conjecture links them together. It presumes,
roughly, that if a lot of small primes divide a and b, then only a few, large ones
divide c.
The publication of the proof comprised 520 pages and is not yet fully confirmed.
In the daily business life, it is always necessary to define all the objects included
and the respective laws and rules as precisely as possible.
Many problems that have a combinatorial character are defined by rules; sometimes,
they look like a game. When the implications are transformed into disjunctions,
then the classical search algorithms disappear, and they can be solved by logical
equations. This is very typical for the results of artificial intelligence; some fields
are completely replaced by other methods.
Example We take a chessboard of size .4 × 4 (Fig. 5.8). We want to position four
queens on this board so that none threatens another. We know that a queen can
move horizontally, vertically, and diagonally, so there must be only one queen in
each horizontal, each vertical, and each diagonal. This can be expressed by rules:
.a1 = 1 expresses the fact that there is a queen on this field and .a1 = 0 the fact that
a1 ∨ a2 ∨ a3 ∨ a4.
.
In the first line a, the queen can be on one of the four squares, a1, a2, a3, or a4.
2. One formulates the constraints that eliminate impermissible solutions:
a1 a2 a3 a4 ∨ a2 a1 a3 a4 ∨ a3 a1 a2 a4 ∨ a4 a1 a2 a3
.
∨a1 b1 c1 d1 ∨ a2 b2 c2 d2 ∨ a3 b3 c3 d3 ∨ a4 b4 c4 d4
∨a1 b2 c3 d4 ∨ a2 b1 b3 c4 ∨ a3 b2 c1 b4 ∨ a4 b3 c2 d1.
Thus, a queen on a1 prevents a queen on one of the squares a2, a3, a4 b1, c1,
d1, b2, c3, and d4.
Analogously, one proceeds with the other verticals b, c, and d. The conjunctions
for a vertical are connected by .∨, and the verticals themselves are connected by .∧.
So we get an equation
that are no longer available due to the position of the queen. The positions of the
queen on b2, b3, c2, and c3 leave only four squares open; all other positions still
allow the occupation or non-occupation of six squares.
If one uses an appropriate solution program [2], then this is a simple matter. A
search for a solution can even be done by hand.
• Queen on a1: the next possibility is a queen on b3; with this, you can’t place
anything on the vertical c.
• Queen on a2: the next possibility is a queen on b4; this commits one to c1 and a3
and gives the solution shown in Fig. 5.6.
• In a mirror image, one gets the second solution Qa3 – Qb1 – Qc4 – Qd1.
• There are no further possibilities.
Following this methodology, one can extend this to chessboards of any size
(Fig. 5.9). Modifications of the problem are also possible (Fig. 5.10).
One suspects that there is no solution with a queen on a1 or a2. For reasons of
symmetry, the squares a7 and a8 are also omitted. A modification of the problem
Fig. 5.10 Nine queens and a pawn—ten queens and two pawns
5.7 Coloring Problems 59
is, for example, to place additional pawns on the board, in the first example on d7
and in the second on c3 and e5. These pawns interrupt the lines of action of the
queens, so one suspects that the number of queens on the board might increase. You
can see that one pawn allows nine queens; ten queens are possible with two pawns.
The methodology remains completely the same; only the corresponding equation
changes.
The following problem shows how far the possibilities of computers reach today.
The background is a problem from electronics. On two conducting rails, there is a
row of contacts. Each contact of one rail is conductively connected to each contact
of the other rail. Each connection has one of four possible properties, which is
described by the four colors red, green, blue, and amethyst. The circuit
does not work if four edges (1 – 2 – 3 – 4 – 1) have the same properties, i.e., the
same color. It is important that the last edge returns to the starting point, so the edge
train must be closed (Fig. 5.11).
In 2010, it was known that for 16 contacts on both tracks, it is possible to avoid
such conflicts. It was also known that for 19 contacts, it is not possible to work
without conflicts. The interesting question was how it looks like for 17 and 18
contacts. A logical model for this problem is constructed in the following way with
a matrix .18 × 18:
• the one track is assigned to the columns of a matrix;
• the second track is assigned to the rows of a matrix.
The difficulty of the solution is due to the exponential complexity of the problem:
there are .18 × 18 = 324 positions in the matrix. Each position can be colored with
one of four colors, resulting in .4324 = 2648 possibilities for possible colorings. This
number is about twice the number of atoms in the universe.
One proceeds exactly as in the checker problem. One begins with a matrix .18 ×
18, and first write down what possibilities you have.
• For each position (m,n) from (1,1) to (18,18) holds:
track 2
1 3
60 5 Polynomial and Exponential Complexity
is the constraint that must be satisfied based on the rule: not one rectangle with
the same color. The same procedure is used for the other colors and all possible
rectangles. The formulas of all rectangles are again linked by .∧.
It is absolutely impossible for a person to find a solution with paper and pencil,
but he can still check the solution. With other problems, even that is no longer
possible.
Here, another interesting feature of such problems becomes apparent. Every
solution for the matrix .18 × 18 is also a solution for matrices that are smaller. You
just have to delete the corresponding number of rows and columns. The also open
question for .17 × 17 did not have to be examined at all, because one only has to
delete one row and one column of the solution for .18 × 18.
The solutions make up only a very small part of all possibilities; nevertheless,
there are very many solutions, because each interchange of rows and columns results
in a solution again. In addition, you can also swap the colors with each other. This
results in a total of
.18! × 18! × 4!
solutions. The solution of the problem took the years 2010–2012. The basic logical
algorithm still had to be supplemented by many subtleties, but in the end, a solution
with tolerable effort was found: 20 graphics processors working in parallel and
1 week of computing time were sufficient for the solution (Fig. 5.12).
What comes out of this solution: the original search algorithms for this problem
belong to AI. The development of this area made it a mathematical problem, which
is solved by deterministic algorithms. And still, a lot of computer power is required.
There are many other combinatorial problems of this kind.
• The problem of the traveling salesman
Figure 5.13 shows the optimal travel route of a traveling salesman through the
15 largest cities in Germany. The given route is the shortest of 43,589,145,600
possible routes. The problem of the traveling salesman is a com-
5.7 Coloring Problems 61
References
Two other outstanding scientists, Norbert Wiener and John von Neumann (Fig. 6.1),
add to the list of famous mathematicians and computer scientists who are of
outstanding importance in the field of artificial intelligence.
Work on quantum mechanics established John von Neumann’s reputation in
America. From 1933, he worked at the newly founded, sophisticated Institute
for Advanced Study in Princeton as a Professor of Mathematics. His
colleagues there included Albert Einstein and Hermann Weyl. Like them, von
Neumann finally emigrated to the USA after Hitler came to power.
John von Neumann made outstanding contributions in many areas of mathemat-
ics. As early as 1928, an essay by the mathematician Émile Borel on minimax
properties had given him ideas that later led to one of his most original designs,
game theory. von Neumann proved the minimax theorem in 1928 for the
existence of an optimal strategy in “zero-sum games”. Zero-sum games
describe situations in which the sum of all players’ combined wins and losses is
zero. Together with the economist Oskar Morgenstern, he wrote the classic
Theory of Games and Economic Behavior in 1944, which also deals with the
generalization to n-person games, which is important for economics. Decisive
for representation and solution is the level of information of the players. Three
notions are distinguished here, complete, perfect (or perfect) information, and
perfect memory, depending on whether the player is aware of the rules of the
game, the moves of the other players, and his own past information. Standard is the
game with complete information as well as perfect memory. Perfect information is
not one of the standard assumptions, as it would be obstructive in explaining many
simple conflicts.
Complete information, the knowledge of all players about the rules of the game,
is an assumption that is considered a prerequisite for joint play in normal games.
Perfect memory is each player’s knowledge of all information that was already
available to him in the past. Although this assumption always seems to be fulfilled,
at least in principle, there are counterexamples: If a game involves a team of
cooperating players, as in Skat, the individual player at the time of his own decision
no longer knows the information context of past moves made by a partner on the
basis of his cards.
6.2 Cybernetics
If one wants to single out one chapter in particular, it is perhaps the chapter
Computing Machines and the Nervous System. There he justifies the
necessity of working with the dual system, but from a completely different angle. He
looks at the cost of storing and processing numbers and reasons that the dual system
is the least costly in both technical realization and work, at a time when there were
no computers!
He writes: “The ideal computing machine must then have all its data inserted at
the beginning, and must be as free as possible from human interference to the very
end. This means that not only must the numerical data be inserted at the beginning,
but also the rules for combining them, in the form of instructions covering every
situation which may arise in the course of computation.”
He justifies the dual system by basing it on the way neurons (nerve cells) work:
They are either in a resting state or, when they fire, they undergo a series of changes
that are almost independent of the nature and intensity of the stimulus.
The chapter “Cybernetics and Psychopathology” is extremely interesting and
reaches far into the future. He poses the question, “How the brain avoids gross
errors, gross malfunctions resulting from the malfunction of individual compo-
nents.” Similar questions, relating to computing machines, are of great practical
importance. “It is quite possible for a chain of arithmetic operations to consist of .109
individual steps. Under these circumstances, the chance that at least one operation
will go wrong is anything but negligible.”
And he is not at all afraid to consider social problems; remember that these
sentences were written over 70 years ago:
66 6 Game Theory, Cybernetics, and Information Theory
The psychology of the fool has become a subject worthy of the attention of scoundrels.
Instead of keeping his own interest in mind, as von Neumann’s player does, the fool acts
in a way that is, on the whole, as predictable as the struggles of a rat in a maze. This
policy of lies - or rather, of truth-telling claims - will get him to buy a particular brand
of cigarettes; this policy will get him, the party hopes, to run for a particular office -
or to participate in a political witch hunt. A certain mix of religion, pornography, and
pseudoscience will sell an illustrated newspaper. A certain mix of ingratiation, bribery,
and intimidation will persuade a young scientist to work on guided missiles or the atomic
bomb. To find this out we have our machinery of radio fan ratings, straw-man polls, opinion
surveys, and other psychological investigations which have the common man as their object;
and there are always statisticians, sociologists, and economists offering their services to
these undertakings.
One can only admire his enormous foresight. In fact, everything he described
with foresight has come to pass. In the penultimate chapter, written in 1961, he
writes:
There is nothing more dangerous than World War III. It’s worth thinking about whether part
of the danger might not lie in the ill-considered use of learning machines. Time and again
I’ve heard the statement that learning machines can’t expose us to new dangers because
we can turn them off when we feel like it. But can we? In order to effectively shut down
a machine, we must be in possession of information about whether the danger point has
been reached. The mere fact that we built the machine does not guarantee that we have the
right information to do so. This is already contained in the statement that the chess playing
machine can defeat the man who programmed it. Moreover, the very speed of operation of
modern digital machines is opposed to our ability to perceive and think through the signs of
danger.
He also discusses the question of whether the machines can create successors
that possess an even higher level of intelligence than their parent generation, and his
answer is yes! The intelligent systems can continue to improve on their own.
All of these problems still exist today and have been exacerbated many times
over by the technical and social developments from 1948 to the present day.
We return once again to C. Shannon, who also founded information theory, a math-
ematical theory from the field of probability theory and statistics. It deals with con-
cepts such as information and entropy, information transmission,
data compression, encoding, and related topics. In addition to mathemat-
ics, computer science, and communications engineering, the theoretical consider-
ation of communication by information theory is also used to describe commu-
nication systems in other fields; as examples, we see the media in journalism,
nervous systems in neurology, DNA and protein sequences in molecular biology,
knowledge in information science, and documentation. Shannon’s theory uses the
term entropy to characterize the information content (also called information
density) of messages. The more non-uniform the structure of a message is, the
higher is its entropy.
6.3 Information Theory 67
log is the logarithm function that gives 0 surprise when the probability of the event
.
is 1. In fact, the logarithm function is the only function that satisfies this specific set
of characterization. Hence, we can define the information, or surprisal, of an event
E by the entropy (Fig. 6.3)
Amazing results have been obtained in recent years by using neural networks.
They are currently the key to great successes in many fields, combined with the
hope of obtaining further important results. Behind this is the effort to develop
mathematical models that will allow the activity of neurons to be described and
then recreated on computers.
A neural network is a mathematical model that mimics the way the human
brain works. Of course, the replication of its functioning on computers is of interest
to artificial intelligence.
In 1943, mathematicians Warren McCulloch (Fig. 7.1) and Walter
Pitts (Fig. 7.2) introduced the neuron as a logical threshold element with multiple
inputs and a single output. It could take on the states true and false as Boolean
variables and fired when the sum of the input signals exceeded a threshold
(Fig. 7.3).
The neuron can be divided into different sections [1]:
• the neuron,
• dendrites,
• the nucleus,
• the soma,
• the axon,
• the axon hillock,
• the myelin sheath,
• the Ranvier’s lacing ring,
• the Schwann cell,
• synaptic terminal knob.
The dendrites are cell extensions of the cell body and form the contact with other
cells or neurons. They receive the initial excitation signals and transmit them to the
cell body.
The soma is the cell body of the neuron. It forms an area filled with cytoplasm
that contains the important cell organelles. These include, for example, the cell
nucleus, the mitochondria, the endoplasmic reticulum, and the
Golgi apparatus.
The axon hill forms the transition from the soma to the axon. Electrical
signals are collected and summed here until a certain threshold or threshold potential
is exceeded. Only then is a signal transmitted to the axon. These signals are called
7.1 Neural Networks 71
action potentials. This prevents our body from passing on even the smallest
signal. Without this filtering mechanism, processing the relevant stimuli would be
impossible.
The long extension of the nerve cell that emerges from the axon hillock is
called axon or neurite. The task of the axon is to transmit action potentials
to nerve or muscle cells. To ensure that the transmission of electrical signals works
as quickly as possible and without losses, the axon is insulated like an electrical
cable. For this purpose, the process is sheathed by supporting or sheath cells. In the
peripheral nervous system (outside the brain and spinal cord), they are also called
Schwann cells. In the central nervous system (brain and spinal cord), they are called
oligodendrocytes.
As an electrical insulation layer, the sheath cells form a so-called myelin sheath
around the axons. The sheath is repeatedly interrupted by exposed axon areas. The
non-enveloped area of an axon is called Ranvier’s lacing ring. They allow
the speed of excitation conduction to be increased. This is because excitation can
also be “jumped” from one lacing ring to the next. Thus, the long sheathed areas are
simply bridged.
The synaptic terminal knobs form the end of a neuron. The electrical signal is
transmitted here to the next neuron or, for example, to a sensory or muscle cell. For
this purpose, the electrical signal is usually converted into a chemical signal. The
connection at the end of a nerve cell with another cell is called synapse. In most
cases, these are chemical synapses. The terminal button brings chemical molecules
into the synaptic cleft, the gap between the two cells. There they meet receptors and
pass on the excitation. Electrical signals are thus triggered again in the next nerve
cell.
According to their function, neurons can be divided into three different types:
72 7 Machine Learning and Neural Networks
• sensory neurons conduct information from sensory organs or other organs in the
body to the brain;
• motor neurons conduct information from the brain or spinal cord to muscles or
glands;
• interneurons, as part of the central nervous system, interconnect neurons with
each other; they have a mediating function.
These concepts are the neurobiological analogy of an action potential, that a
neuron emits upon a critical change in its membrane potential. McCulloch and Pitts
showed that by suitably combining several such neurons, any propositional function
(.∧, ∨, ¬) is realizable.
In 1957, Frank Rosenblatt (Fig. 7.4) published the perceptron model,
which is still the basis of artificial neural networks (Fig. 7.5). It consists in the
basic version (simple perceptron) of a single artificial neuron with adjustable
weights and a threshold.
A human is assumed to have 100 billion dendrites with a total length of many
hundreds of kilometers. We consider a single neuron with input values .x1 . . . xn , a
threshold .θ , and weights .w1 . . . wn for the connections of the input values with the
output.
⎧ n
⎪
⎪
⎨ xi · wi + b > θ
o=
. i=1 (7.1)
⎪
⎪
⎩
0 otherwise
The number of synapses in the brain of an adult is about 100 trillion (.1014 )—
related to a single neuron, it varies between 1 and 200,000. Despite this large
number, it is amazing at first glance that these simple linear operations form the
basis of intelligent behavior.
7.1 Neural Networks 73
structure allows the network to learn nonlinear and linear relationships between
input and output vectors. The linear output layer allows the network to obtain values
outside the range of .−1 to .+1 (Fig. 7.6).
For the learning process, the data must be split into two sets: the training
dataset, which is used to calculate the deviations and update the weights, and the
validation dataset , which is used to select the optimal number of iterations.
As the number of iterations increases, the training error decreases, while the error of
the validation dataset first decreases, then reaches a minimum, and finally increases.
If the learning process continues after the validation error reaches a minimum,
this will result in overlearning. Once the learning process is complete, another
dataset (test dataset) is used to validate and confirm the prediction accuracy
(Fig. 7.7).
The difficulty in understanding exactly how changing weights and biases affect
the overall behavior of an artificial neural network was one of the factors that
prevented a wider application of neural networks until the early 2000s. Today,
backpropagation algorithms are applied in many areas of artificial intelligence,
for example, optical character recognition, natural language processing, and image
processing.
This is how you learn as a child; after touching a hot stove top once, you either
keep your distance or use an oven mitt, which you most likely learned from your
parents. After a few tries, animals learn to avoid an electric fence. Pain is considered
unpleasant or harmful in many biological systems, and the neural network of
humans or animals seeks a way out.
There are different types of neural networks adapted to different tasks. One such
type is convolutional neural networks (folding neural networks, CNN).
It is a concept inspired by biological processes in the field of machine learning. It is
primarily applied to machine processing of image or audio data.
7.1 Neural Networks 75
The fully trained neural network then plays the game on a single machine with
only four TPUs, which is a reasonably fair comparison with the respective strongest
chess and shogi programs, which had 64 CPU cores available after all. An important
point of view is shown here: the training requires a very large effort, but the
application of the finished net is possible with much less effort.
After the comparison with human players has become irrelevant, only one
interesting question remains: Is there an optimal strategy for these games?
• Can White necessarily win from the initial position, for any of Black’s moves?
One might assume so, since White makes the first move and is always one move
ahead in mobilizing his forces;
• Does Black own a move for any move by White, so that the game ends in a draw?
• Can Black necessarily win from the initial position, for every move by White?
It could be possible, because White may irreversibly weaken himself on the first
move.
The question of an optimal strategy has already been answered by the German
mathematician Georg Cantor (1845–1918). He referred to the following steps:
• Build up all the positions in which White has mated, one after the other.
• One move back. White can mate in one move.
• Black takes back a move, and white again chooses a move in which he can
forcibly mate. This results in an optimal strategy. But it cannot be traced back
to the initial position, because that is of course too many possibilities.
The first ending tablebases—for all four-piece endgames—were built by the end
of the eighties. In the beginning of the nineties, the same task was done for five
pieces. In 2005, six-piece endings were solved in Nalimov Tablebases, which are
now used by many professional chess programs and services.
Experts didn’t expect seven-piece endings to be cracked and catalogued until
after 2015, but Convekta Ltd, namely, programmers Zakharov and Makhnichev,
managed to solve this task in just 6 months using a new algorithm designed
specifically for this purpose and run on the Lomonosov supercomputer based in
the Moscow State University, Lomonosov Tablebases.
As a result, we now have 525 tablebases of the 4 vs. 3 type and 350 tablebases of
the 5 vs. 2 type. The calculations for six pieces playing against a lone king weren’t
done because the outcome is rather obvious.
The total volume of all tablebases is 140 000 gigabytes, which is obviously
too much for personal computers. Lomonosov Tablebases will be accessible online
from the Aquarium interface and on the ChessOK website. All users of ChessOK
Aquarium, Houdini Aquarium, and Chess Assistant products get free access to the
service for the period specified in their product description.
Very interesting theoretical and mathematical questions appear here again. In
all previous chess programs, the moves were selected with evaluation functions
whose components had a chess background. For example, the value of the pieces,
the safety of the king, the occupation of the center, etc. were evaluated. The move
that corresponded to the best value of the evaluation function was selected.
78 7 Machine Learning and Neural Networks
The training of neuronal networks generates the weights .wi , which obviously
guarantee an excellent playing style but where one does not know what the
individual weights mean for chess. The requirement that the results of AI programs
must be explainable cannot be realized. Either you believe the computer and act
according to its suggestion or you leave it alone. The result in many areas will
be a computer faith of which we do not yet know where it will lead. In any case,
one must keep in mind that all the existing knowledge of a field has been used.
New knowledge can easily be added, and the finished systems can in principle be
effortlessly duplicated and made available to a wide range of users.
The algorithm for creating the elements of a database and for proving the
existence of an optimal strategy was published by the mathematician Ernst Zermelo
at a mathematics congress in 1912. Later, it found itself as a special case in
mathematical game theory.
Poker has always been considered a game between people. It differs from other
games in that psychological thinking plays a role here, at least when there are only
humans at the table. In a casino in the American city of Pittsburgh, a computer
program beat four of the best poker players in the world. For nearly three weeks,
four poker professionals each spent many hours a day competing with the software
called “Libratus,” betting chips, analyzing their strategies, and still lost. “It felt like
I was playing against someone who was cheating, like they could see my cards,”
said 28-year-old American-Korean poker player Dong Kim.
Thus, neural networks form a central role in the development of other intelligent
applications. The strategic games are just to try out the techniques. The results are
easily grasped by humans. One can now trust that similar results can be achieved
for other problems and other fields.
nachzudenken) oder etwas zu kritisch gesehen (weil sie angeblich alle Übersetzer
um ihre Arbeit bringen wird). Der weltweite Übersetzungsmarkt verzeichnet jedes
Jahr einen Umsatz von etwa 40 Milliarden Dollar und wird bis 2025 voraussichtlich
auf über 60 Milliarden Dollar anwachsen. Durch die fortschreitende Globalisierung
rücken Regierungen, Unternehmen und sogar Privatpersonen in verschiedenen
Ländern immer enger zusammen, und die ständig steigende Nachfrage nach Über-
setzungen kann nicht mehr von Menschen allein gedeckt werden.
English
The existence of machine translation is often either overlooked (because people
have actually long been used to it and use it without thinking about it) or viewed
somewhat too critically (because it will supposedly deprive all translators of their
jobs). The global translation market records about 40 billion Dollars in revenue
every year and is expected to grow to over 60 billion Dollars by 2021. Governments,
companies, and even private individuals in different countries are growing ever
closer together as a result of ongoing globalization, and the ever-increasing demand
for translations can no longer be met by people alone.
Latvian
Mašintulkošanas pastavešana bieži vien tiek vai nu ignoreta (jo cilveki jau sen pie tas
ir pieraduši un izmanto to, par to nedomajot), vai ari uz to raugas parak kritiski (jo
ta it ka atnems darbu visiem tulkotajiem). Pasaules tulkošanas tirgu katru gadu tiek
registreti aptuveni 40 miljardu dolaru ienemumi, un paredzams, ka lidz 2025. gadam
tie pieaugs lidz vairak neka 80 miljardiem dolaru. Globalizacijas rezultata dažadu
valstu valdibas, uznemumi un pat privatpersonas klust arvien ciešak saistitas, un
arvien pieaugošo pieprasijumu pec tulkojumiem vairs nevar apmierinat tikai cilveki.
Hungarian
A gépi fordítás létezésérol gyakran vagy nem vesznek tudomást (mert az emberek
már régóta hozzászoktak, és gondolkodás nélkül használják), vagy túlságosan
kritikusan szemlélik (mert állítólag minden fordítót megfoszt a munkájától). A
globális fordítási piac évente mintegy 40 milliárd dollár bevételt könyvel el, és
2025-re várhatóan 80 milliárd dollár fölé no. A különbözo országok kormányai,
vállalatai, sot magánszemélyei is egyre közelebb kerülnek egymáshoz a globalizáció
következtében, és a fordítások iránti egyre növekvo igényt már nem lehet egyedül
kielégíteni.
Translations into Japanese and Chinese are also possible (Figs. 7.10 and 7.11).
Of course, there are still problems where improvements may be necessary:
• problems with technical terminology and with terminology consistency,
• problems with homonyms and context recognition.
And if you are still not convinced, you can look for a Chinese or Japanese journal
on the Internet and transfer the text to DeepL’s window with Ctrl + C and Ctrl
+ V. There, it will be translated, and you can easily see if you understand the text.
After doing this several times, one is convinced that such systems already have a
80 7 Machine Learning and Neural Networks
very good level that far exceeds the level of a normal mortal. You can now (as a
scientist, as a reporter, as someone interested in everyday life in foreign countries)
get information that previously would only have been possible with enormous effort.
Google has also invested a lot in translation programs. There are more than
100 languages offered, including many languages that are very rare, especially
from Africa and from Asia, up to Xhosa, Yoruba, and Zulu. The Yoruba are a
West African people living mainly in southwestern Nigeria. The Xhosa are a South
African people who linguistically belong to the Bantu. The Zulu are a subgroup of
the Bantu with over eleven million people today and are the largest ethnic group in
South Africa.
When travelling to Africa, you can be pretty sure to encounter knowledge of
English, French, Spanish, or Portuguese, depending on the former colonial powers.
You get a very friendly welcome if you make an effort to learn the language of the
local people. You write down on your computer what you want to know and then
present it in the local language. And there are also occasions when you have to rely
on communicating with a local who doesn’t know another language.
• Where is the next opportunity to buy something to eat or drink here?
• Dab tsi yog qhov tuag nyob rau hauv Gelegenheit, thiab yog dab tsi.
Hmong is an ethnic group living in southern China.
Reference 81
Reference
1. https://studyflix.de/biologie/nervenzelle-2737
Chapter 8
Other Applications
8.1 Databases
Databases represent a frontier area with regard to artificial intelligence. Here, one
uses the large storage possibilities available and the search algorithms available.
A database consists of two parts:
1. the management software, the database management system (DBMS),
2. the set of data to be managed, the database (DB) in the narrower sense,
often also called “database.” The management software internally organizes the
structured storage of the data and controls all read and write access to the
database.
3. A database system offers a database language for querying and managing the
data. SQL (System Query Language) is a database language for defining
data structures in relational databases and for editing (inserting, modifying,
deleting) datasets.
SQL is based on a relational algebra, its syntax is relatively simple and
semantically based on English colloquial language. A joint committee of ISO and
IEC standardizes the language with the participation of national standardization
bodies such as ANSI or DIN. The use of SQL aims to make applications independent
of the database management system used.
A very natural form of databases are, for example, dictionaries. The dictionary
dict.cc [2] offers for German as source language 26 languages on the opposite
side. The corresponding page allows you to select a language and provides the
answer when you enter a German word. The advantage of such databases is obvious.
You save a lot of space, a lot of time, and a lot of money. The databases available at
linguee work closely with DeepL.
The existence of dictionaries on the Internet has a major impact on the business
world. Fewer dictionaries are printed, as the Internet provides much more compre-
hensive dictionaries for many more languages. The Duden and its special editions
are still considered the norm for spelling. However, sales have also declined, as it is
of course also available on the Internet at https://www.duden.de.
For generations, the Brockhaus Encyclopedia has been synonymous with the
world’s collected knowledge, with lexical competence at the highest level. It is
the basis for work, education, and cultural experience and for discussions in
everyday life. But it has also adapted to technical developments: The “Brockhaus
Enzyklopädie Audiothek” is a novelty: over 70 hours of audio examples on two
audio DVDs—one for PC, one for the DVD player—make audible for the first time
what can only be described in words to a limited extent: sounds of nature, speeches,
pieces of music, animal voices, literary lectures, etc.
The online portal https://brockhaus.de ensured that the entire keyword inventory
was accessible to encyclopedia owners anytime and anywhere free of charge until
the end of 2010. In addition, the content is constantly updated here, so that the
Brockhaus Enzyklopädie is and remains not only the most comprehensive
but also the most up-to-date German-language reference work. The same principle
is followed by Wikipedia on the Internet.
There are many more databases on the Internet that have made many branches of
business more or less superfluous. At www.stampworld.com, you can find a stamp
catalogue for all countries and all times. It was a huge job to produce this relational
database, and it is no easy task to keep it constantly updated [3].
In the simplest case, one proceeds in the following way:
• Given a single object selected from a set of similar objects, we take as an example
a single stamp, which was issued somewhere at some time. All stamps of all
countries and times form the set M.
For stamps, one now defines sets of properties that can be attributed to a stamp.
In Fig. 8.1, there are specified the following items:
The offer and use of these computing resources are defined and usually take place
via an application programming interface (API) or, for users, via a Web site or app.
With the emergence and development of the Internet, many new interlinked
industries have also emerged, which are heavily dependent on databases but
now combine many other characteristics, advantages and disadvantages. They are
summarized under the name “E-commerce”. The Internet is the basis for extensive
sales operations. Potentially, “online banking” also belongs here, but this has already
become a separate industry. Almost every bank today has an online presence.
E-commerce has developed very strongly over the last 30 years and now domi-
nates many business areas where traditional retailing has been severely impacted. A
well-known example is the “Kindle ebook.” Books offered there are cheaper than in
paper format, and they do not require space in the home. However, large pictures and
maps are usually not very clear, beautiful bindings are also lost, and the profession
of bookbinder is more or less dying out. The best known are probably “Amazon”
and “eBay”, but every industry is included. Many manufacturers of certain products
sell via the Internet but also run retail stores.
Amazon.com, Inc. (Amazon for short) is a listed US online mail order company
with a wide range of products. According to its own statements, Amazon, as the
market leader of commerce on the Internet, has the world’s largest selection of
books, CDs, and videos [4]. Via the integrated sales platform “Marketplace”, private
individuals or other companies can also offer new and used products as part of
online trading. Under its own brand, the company sells the “Amazon Kindle” as
a reading device for electronic books, the “tablet computer” Amazon Fire HD, the
“set-top box Fire TV” as well as the “HDMI stick Fire TV Stick,” and the voice
recognition system “Echo.” Through Amazon Web Services, the company is also
one of the leading service providers for cloud computing. The company generated
sales of $386 billion in 2020. In Germany, Amazon was the top-selling US company
in 2020 with sales of 29.5 billion euros. It is almost inconceivable that this company
was only founded in 1994 and has achieved this level of performance within just
under 30 years.
The special characteristics are market transparency and the absence of pref-
erences. In the creation and use of such systems, many previously unimagined
problems arise. For example, it is of great importance for the sales of a product when
an offer appears on the Internet. Research has shown that a customer searching for
something will only look at three to four pages and then make a decision or abandon
the search and postpone the decision. That’s why advertising on the Internet plays a
crucial role. It enables the relatively precise measurement of a specific advertising
success. However, it has probably often reached a stage where this form becomes
extraordinarily exaggerated. However, it is gradually replacing advertising with
catalogs, advertisements, and brochures.
8.2 Trade over the Internet 87
Amazon obviously proceeds in two steps: first, products are searched for that
semantically basically match the search query. This excludes the possibility that, for
example, a slipper is suddenly offered, although the customer was actually looking
for motorcycle accessories. In the second step, the products pre-filtered in this way
are then sorted by an algorithm according to various criteria. In this step, the most
important factors are “early reviews,” “product images,” or “availability.”
These factors are measured in the units “click rate” and “conversion frequency.”
The click rate shows how many people visited the product page. The second factor
shows how many people actually bought the product. Before you start searching for
a product, you should do some thinking yourself: what would you ask a saleswoman
in the store when searching for a certain product? Using this method, you can come
up with a few matching keywords to the product relatively quickly and build on
that list. This list can then be entered into the Amazon search box to find out more
matching terms through Amazon’s recommendations.
One can secure a place at the top of the search results through Amazon’s
“sponsored products,” which are based on keywords and auctions. The merchant
with the highest bid gets the coveted spot at the top of Amazon’s product search.
Anyone who has a professional Amazon Seller account has the opportunity to place
such ads.
The goal of Google’s search engine is to provide its users with the best possible
information in a fast and uncomplicated way. For this purpose, billions of Web
pages are searched. Numerous factors are used in this search: the words used in
a search query, the relevance and usefulness of pages, the expertise of sources, and
the location and settings.
It is becoming increasingly clear that what matters most is large amounts of data.
The existing algorithms for classifying data and finding factors relevant to decision-
making are already working very successfully. It’s almost a given that learning
programs will also be used to constantly improve advertising.
And, of course, one must not forget the entertainment industry. Here, YouTube
occupies a top position. YouTube is a video portal founded in 2005 by the US
company YouTube, LLC, a subsidiary of Google LLC since 2006. Users can watch,
rate, comment, and upload video clips themselves on the portal free of charge.
YouTube’s main source of income is advertising. It also earns through monthly
subscription services such as YouTube Premium. In 2019, YouTube generated $15
billion in annual revenue. Most of its revenue is generated by playing commercials.
YouTube offers three forms of accessibility: public, unlisted, or private. Public
videos are the default. These can be viewed by anyone and found in search results.
There is an infinite amount of material available, books, plays, classical music, or
anything your heart desires, and search time is very short.
Netflix, Inc. is a US media company engaged in paid streaming and production
of movies and series. It follows the same principles as Amazon and eBay but has a
more limited business scope [19]. Founded by Reed Hastings and Marc Randolph
in Los Gatos, California, in 1997, the company initially operated as an online video
store, shipping movies on DVD and Blu-ray to its subscribers. In 2007, Netflix
entered the video-on-demand business, making its content available to subscribers
8.2 Trade over the Internet 89
via streaming. By March 2020, the media service had offered nearly 3,000 movie
and TV shows for streaming. By the end of June 2021, Netflix had signed up more
than 200 million paid subscriptions since the company’s founding, including more
than 70 million in the United States and Canada. In 2020, the company’s stock
market value exceeded that of the Walt Disney Company for the first time, at just
under $195 billion.
Netflix categorizes its customers as starters who have watched only two
minutes of a movie or series, watchers who abandon a movie or series episode
after 70%, and completers who have watched movies or series seasons 90% of
the time. The numbers in the three categories are reported back to producers within
the first 7 days and 28 days after the release of a new movie or series season. Metrics
for production success beyond 1 month after release are not collected.
Through content-based and collaborative filtering, Netflix is able to make
recommendations to users about series that are tailored to them. To this end,
multilayered data about user behavior is collected each time Netflix is used. In
addition to the plot, actors, film location, etc., this also includes the exact length
of time spent consuming films.
Every driver appreciates the database hidden in his navigation device. It knows
every city, every there, every street, and every house number in every country. It’s
hard to remember that you used to have to carry a mountain of maps or a car atlas
with you. However, the art of finding your way around maps is slowly being lost.
The company eBay follows the same principles. eBay Inc. is a US company that
operates one of the largest online marketplaces. It was founded on September 3,
1995, by Pierre Omidyar in San José (California) under the name AuctionWeb.
The company also offers its services in Europe, for example, in Germany, Switzer-
land, and Austria. Until July 17, 2015, the payment service PayPal also belonged
to the company. Here, then, the character of an auction is added to Amazon’s
business approach. Anyone can offer a product on eBay for a certain commission,
along with a minimum amount at which you want to sell the object. Customers can
then raise the price, as in an auction. A closing date then ends this auction, and the
highest bid wins.
Over the years, the offer expanded from a marketplace with a flea market-like
character to a platform that also offers new goods from commercial dealers, as
opposed to secondhand goods from private sellers. The dependence on databases
and the Internet is quite obvious, and the intelligence here lies, among other things,
in the business idea.
The tourism industry is also firmly in the hands of the Internet and has very
much marginalized local travel agencies. The American company Tripadvisor
is an example of this. It is a US tourism Web site that provides users with
individual testimonials to help plan vacations [5]. There are reviews and testimonials
on 1.8 million stores, 4.2 million restaurants, 1.1 million hotels, and 830,000
vacation rentals, as well as 730,000 attractions in 135,000 destinations (as of mid-
2016). Offered are testimonials and reviews from travelers; links to reports from
newspapers, magazines, and travel guides; as well as travel forums (electronic
“bulletin boards”).
90 8 Other Applications
China is one of the most attractive e-commerce markets in the world. In 2018,
online retail in China generated sales of around 530 billion euros. The number of
online shoppers has increased more than tenfold in the past ten years. Today, nearly
85% of China’s Internet users already shop online. The company Alibaba is quite
comparable to Amazon.
Big data refers to large volumes of data originating from sectors such as the
Internet and mobile communications, the financial industry, the energy industry,
healthcare, and transport and from sources such as intelligent agents, social media,
credit and customer cards, smart metering systems, assistance devices, surveillance
cameras, and aircraft and vehicles, which are stored, processed, and analyzed
using special solutions. It is about “grid search,” “(inter-)dependency analysis,”
“environmental” and “trend research” as well as “system” and “production control.”
The main concern is the discovery of new knowledge. The worldwide volume of
data has swelled to such an extent that unprecedented opportunities are being opened
up. The networking of data sources is also leading to new kinds of applications, as
well as risks for users and organizations. Here we are only at the beginning of the
development. The far-reaching linkages are illustrated in Fig. 8.2.
The economy promises itself new insights into prospective customers and
customers, their risk potential, and their purchasing behavior and generates personal
profiles. It seeks to optimize production and make it more flexible and to better
bring innovations to markets through forecasting. Science studies climate change
and the occurrence of earthquakes and epidemics, as well as mass phenomena such
as population migrations and traffic congestion. With the help of supercomputers,
it simulates both atomic bomb drops and meteorite flights and their impacts.
Authorities and intelligence agencies use enormous amounts of data to detect
deviations and anomalies that could give criminals and terrorists away, as well as
similarities that allow groups to be identified and narrowed down.
Big data is a challenge for data protection and privacy. Often, there is no consent
from the data subject for the use of the data, and often, they can be identified and
controlled. The linking of information that is in itself unproblematic can lead to
problematic findings, so that one suddenly belongs to the circle of suspects, and
statistics can make someone appear uncreditworthy and risky because one lives in
the wrong neighborhood, uses certain means of transportation, and reads certain
books. Information ethics asks about the moral implications of big data, in terms
of digital paternalism, informational autonomy, and informational justice. Business
and legal ethics are also required. Data protection laws and institutions can go some
way toward preventing excesses and ensuring consumer protection.
The volume of data has assumed immense proportions: The amount of data that
mankind accumulated from the beginning until the year 2002 was already created
within just ten minutes in 2014. According to forecasts, this mountain of data
will continue to grow and double worldwide every two years. The flood of data
results from the increasing digitization of all areas of life. According to the big
data definition, the volumes of data are so extensive that normal software is simply
overwhelmed and the usual technology cannot process the volume efficiently. So
when dealing with mass data, special technical requirements arise for the software
used. Only special frameworks can analyze the data. To do this, the software must
be able to process as many datasets as possible at once and import large amounts of
data as quickly as possible. In addition, the software should make the data volumes
available to the user in real time if possible and, if necessary, also be able to answer
several database queries simultaneously.
Here it is particularly important to be able to process very large amounts of data.
To calculate the correlation between two data sequences, one uses the correlation
coefficient, for example. This is not very difficult mathematically [6].
We assume two datasets .(x1 . . . xn ) and .(y1 . . . yn ). First, we calculate the two
mean values:
1 1
n n
x=
. xi y= yi (8.1)
n n
i=1 i=1
• If deviations from the mean are predominantly of the same sign, the covariance
becomes positive (if positive deviations in one variable are usually accompanied
by positive deviations in the other variable).
• If deviations from the mean are mostly opposite in sign, the covariance becomes
negative (if positive deviations in one variable are mostly accompanied by
negative deviations in the other variable).
• When deviations vary non-systematically, the covariance is close to zero (when
positive deviations in one variable are sometimes accompanied by negative
deviations and sometimes by positive deviations in the other variable).
Then we calculate the standard deviations
n
i=1 (xi − x)
2
.sx = . (8.3)
n−1
n
i=1 (yi − y)
2
sy = . (8.4)
n−1
cov(x, y)
r(x,y) =
. (8.5)
sx · sy
It can be seen that the order determined by age differs from the order given by
the running times; three people are not in the place corresponding to their age in
the running times. If we calculate the correlation coefficient, we get r = 0.73, which
means that age determines the place in this run only in about 75% of the cases. This
value can be further refined by using additional data.
Regression analysis provides very extensive possibilities. With all these methods,
one can determine extensive correlations between very different pieces of infor-
mation. For example, Google has filtered out 45 factors for the occurrence of an
epidemic from several million queries.
Overall, statistics and probability theory provide many sophisticated, well-
studied means and methods for evaluating mass data that can provide very accurate
information about relationships and expected events. However, one must be very
careful in interpreting results. There are very many (intentional or unintentional)
misinterpretations.
Big data is used in the most diverse areas of life. Even everyday things that almost
every Internet user knows are based on it. An example of big data evaluation from
the area of online shopping: Anyone who has ever shopped on the Internet will
be familiar with the section “Customers who bought product XY also bought.. . ..”
These recommendations result from the evaluation of millions of purchase data from
other customers.
Other areas to benefit from big data:
1. Medical research: By analyzing mass data, physicians can find the best therapy
solutions for patients.
2. Industry: Thanks to the use of their own machine data, companies can increase
the efficiency of their production.
3. Business: Big bata enables companies to get to know their customers well and
tailor offers to them.
4. Energy: To tailor energy consumption to individual needs, you have to know
those needs. Collected consumption data ensures an appropriate energy supply
in the long term.
5. Marketing: Big data is often used in the marketing sector to determine targets.
Mostly, it is about improving relationships with customers.
6. Crime prevention: Government and state security are also turning to big data; the
mass use of surveillance cameras has reduced petty crime quite substantially.
94 8 Other Applications
7. Analysis method for finding patterns: It identifies clusters of similarities and then
forms groups of objects that are very similar in terms of certain aspects; unlike
classification, the groups (or clusters) are not predefined and can take different
forms depending on the data analyzed.
8. Association analysis: It uncovers correlations between two or more independent
factors, which, although not directly related, are increasingly found together.
9. Regression analysis: It uncovers relationships between a dependent variable (e.g.,
product sales) and one or more independent variables (e.g., product price or
customer income); it is used, among other things, to make predictions about the
dependent variable (e.g., a sales forecast).
10. Predictive analysis: This is actually a higher-level task aimed at making pre-
dictions about future trends; it works with a variable (predictor) measured for
individuals or larger entities.
The applications related to big data are very broad and sometimes show quite
surprising results [1]. Very often, there is no causal relationship between the
phenomena, but the results are quite surprisingly accurate.
In 2009, the H1N1 virus (responsible for bird and swine flu) was detected in the
United States. A few weeks before this virus hit the headlines, engineers working
at Google published an article in the journal Nature. There, the authors showed
how Google could predict the spread of the virus, not only at the country level but
also at the state level. They looked at correlations between the frequency of certain
searches and the temporal and spatial extent of infections. The software found 45
search terms that were used together in a mathematical model. The predictions made
on this basis and the actual national numbers were in high agreement.
This form of data analysis does not look for causal relationships cause .→
effect; it focuses its attention on the fact that many factors, each in itself, may
contribute a tiny bit to the final facts.
A company called 23andMe has been working in Silicon Valley since 2007. The
company’s name refers to the 23 pairs of chromosomes in a human being. Starting
in December 2007, it offered private individuals an examination of their genetic
information. The offer was initially aimed only at customers from the United States
but was expanded over time to numerous other countries. It is available in Germany,
Austria, and Switzerland, among other countries.
The saliva sample sent in is tested for around 200 genetic diseases and 99
other predispositions. Information on geographical origin is also provided. More
than 960,000 sections of the human genome are examined, which identify single
nucleotide polymorphisms and make up the personal characteristics [15, 19].
With the help of artificial intelligence, doctors could soon be able to more
accurately detect malignant tumors during a colonoscopy and better distinguish
them from harmless growths. Currently, all polyps (growths) are removed and then
examined histopathologically. With the help of artificial intelligence, an evaluation
should be available virtually in real time. Another advantage of artificial intelligence
is that it is capable of learning. This will enable the detection rate of polyps to
continue to improve. Small adenomas in particular are flat and often hidden in the
8.3 Big Data 95
folds of the intestinal wall. They may well be missed even with high-resolution
endoscopy equipment. So the use of AI should further improve the overall quality
of screening. It is also possible that the number of histopathological examinations
required may decrease, leading to a decrease in the overall cost of screening.
With the help of artificial intelligence, doctors could soon be able to more
accurately detect malignant tumors during a colonoscopy and better distinguish
them from harmless growths. Currently, all polyps (growths) are removed and then
examined histopathologically. With the help of artificial intelligence, an evaluation
should be available virtually in real time. Another advantage of artificial intelligence
is that it is capable of learning. This will enable the detection rate of polyps to
continue to improve. Small adenomas in particular are flat and often hidden in the
folds of the intestinal wall. They may well be missed even with high-resolution
endoscopy equipment. So the use of AI should further improve the overall quality
of screening. It is also possible that the number of histopathological examinations
required may decrease, leading to a decrease in the overall cost of screening.
Several tools for data analysis are already available and offered over the
Internet. Hadoop, MapReduce, and Spark are three offerings from the Apache
Software Projects. These, as well as other software solutions, can be used to
deploy big data in an enterprise.
• Hadoop is an open-source software solution for big data. Hadoop’s tools
help distribute the process load. This allows massive datasets to be run on
many separate computing nodes. Instead of transferring a petabyte of data to
a small processing site, Hadoop does the opposite. It thus provides much faster
processing of information.
• MapReduce supports the execution of two functions: compiling and organizing
(mapping) datasets and then refining them into smaller, organized datasets to
respond to tasks or queries.
• Spark is also an open-source project from the Apache Foundation. It is an
ultrafast distributed framework for big data processing and machine learning.
Spark’s processing—engine can be used as follows: as a stand-alone installation
and as a cloud service, in all popular distributed IT systems like Kubernetes or,
Sparks’ predecessor, Apache Hadoop.
Cloud technologies are constantly evolving and leading to ever greater floods
of data. In order to build forward-looking digital solutions, this information must
be processed. For virtual transactions, inventories, and IT infrastructures, a well-
thought-out big data approach with data from numerous sources is therefore
required. Only in this way can big data provide a holistic overview. Sources can
include:
• virtual network protocols,
• security-related events and patterns,
• global network traffic patterns,
• anomaly detection and resolution,
• compliance information,
96 8 Other Applications
nearby. WeChat is one of the most important communication tools in China. Even
for business matters, people prefer WeChat to email.
There is also a social feature called Moments. Users can upload a series of
pictures or videos, and their friends can view and comment on the post. From large
supermarkets to the smallest street vendors and cabs, you can pay almost anywhere
in China with WeChat if you have a Chinese bank account. There are two ways
to pay for something through the app. First, the store can scan the unique WeChat
barcode, or second, the buyer scans the merchant’s barcode.
When buying something online in China, one has the option to pay with
WeChat Pay. One has to enter a passcode or use a biometric authentication tool
to authorize the transaction. Instant money transfers to WeChat contacts can also be
made using the messaging feature, making it easier to share bills or move money
around China. It is possible to live almost cashless in China and go out all day
without a wallet.
The main competitor of WeChat Pay is Alipay, a company of Alibaba
subsidiary Ant Financial.
WeChat and Alipay are often referred to as Super Apps because everything is
integrated into a single service. Businesses can choose to launch mini-programs—or
apps within WeChat—instead of a stand-alone app. The program allows companies
to send promotional messages directly to users through WeChat and reach the app’s
more than one billion users.
The Chinese Big Data market is in a highly competitive phase. In the last ten
years, several start-ups have emerged and made their mark on the market. For
example, the big data company Terark, founded in 2015, has developed the
Terark Database, which reportedly outperforms similar products from Facebook
and Google. Several such start-ups have challenged the previous dominance of
a few major players and fragmented the market. This has also led to significant
consolidation in this industry as major players seek to gain market share and smaller
companies try to scale up. Increasing competition from the West has also forced
Chinese companies to look for opportunities around the world rather than limiting
themselves to China. Some of the recent developments are as follows:
• April 2019—ByteDance, known as the parent company of apps TikTok and
Toutiao, announced the acquisition of Beijing-based big data company Terark
for an undisclosed sum, allowing the company to expand its customer base to
market leaders such as Alibaba Group. Terark has solutions to compress data in
a much smaller space and with higher search speed.
• January 2019—Alibaba acquires Berlin-based start-up Data Artisans for
$103 million. Data Artisans provides distributed systems and large-scale data
streaming services for enterprises. Data Artisans was founded in 2014 by the
team leading the development of Apache Flink, a widely used open-source
technology for big data processing.
Perhaps the most big data evaluation software offerings are in Japan [8], with
239 systems on the market in the business intelligence category alone.
98 8 Other Applications
In Singapore, there are six companies producing software for big data applica-
tions.
Africa and Latin America are also making great strides in this area. Africa will
establish its first big data hub, which will promote knowledge sharing and the
extraction of information from complex datasets. The hub will enable the continent
to access and analyze up-to-date data related to the Sustainable Development Goals
(SDGs) to make evidence-based decisions, says Oliver Chinganya, director of the
Africa Statistics Centre at the United Nations Economic Commission for Africa
(UNECA).
According to one study, big data has a positive impact on almost all areas of
life, including health, aviation, banking, military intelligence, and space exploration.
The study says big data hosted in Africa could be a focal point for healthcare,
supporting diagnoses, treatment prescriptions, health systems monitoring, and
digital surveillance to track epidemics.
During the Ebola crisis in West Africa, data scientists tapped into phone
companies’ extensive data to visualize the population movements of Ebola patients
and predict the spread of the disease. Detailed maps showing the movements of
Ebola patients were created and used by authorities to find the best places to set up
treatment centers and develop means to limit travel to contain the disease.
In a recently signed memorandum of understanding between Rwanda and
Ethiopia-based UNECA, the data center will be established at the National Institute
of Statistics of Rwanda (NISR). “The establishment of this data hub will enable
Africa not to be left behind in the data revolution,” says Yusuf Murangwa, director
general of NISR. “At the moment, we have problems in Africa with access to big
data volumes.” Chinganya tells SciDev.Net that the center will promote knowledge
sharing on newly developed statistical methods and tools while providing training on
the use of big data and data science for development. Users will include members of
the African regional statistical community, the private sector, academic institutions,
and civil society.
The project is still in the early stages. The center is expected to be operational once the
implementation strategy is finalized by the end of the second quarter of 2020, but that
depends on many factors, including the COVID-19 global pandemic, which is restricting
movement and work patterns.
The Kigali center joins others planned in Brazil, China, and the United Arab
Emirates, according to a report by the Global Working Group on Big Data for
Official Statistics released in December by the United Nations Economic and Social
Council’s Statistical Commission.
The regional center in Rwanda “will support capacity building in Big Data
for official statistics in Africa, in collaboration with international and regional
agencies.” The biggest challenge in the field of data science is the lack of capacity
and appropriate infrastructure to support Big Data initiatives, as well as insufficient
investment. Under the global platform, many data sources will be openly accessible.
Mechanisms will also be developed to enable selective data sharing between the
various nodes.
8.3 Big Data 99
The repository will allow computer systems to share and use information; legal
regulations for access will be developed by the Economic Commission for Africa.
Data will be collected based on the UN Basic Principles for Official Statistics, which
protect civil liberties.
Dubai plans to use new technologies and mass data collection to optimize the
mobility of its citizens. The idea is to identify scenarios that will enable better traffic
flow while making all traffic in the city cleaner. The Roads and Transport Authority
(RTA) is an independent government agency responsible for Dubai’s roads and
traffic. It has recently made several announcements aimed at optimizing traffic flow
and making traffic cleaner throughout the emirate [9].
This extensive project is based on the collection of all kinds of data to transform
public and other transport. By collecting data on public transport, cab traffic, or hotel
occupancy, the “Dubai in Motion” program will be able to create optimized mobility
scenarios and thus make adjustments to improve people’s mobility and develop even
cleaner and more sustainable transport. In the long term, this program will also help
plan future green spaces and infrastructure for soft mobility.
But the RTA isn’t stopping there. The agency also plans to use artificial
intelligence for its bike lanes to count the number of users and monitor compliance
with traffic rules and basic safety regulations. Another upcoming project is a new
generation of smart displays that will provide information to public transportation
users and help them better organize their journeys. Dubai also aims to convert 25%
of its entire transportation system to autonomous modes by 2030, whether on the
ground or in the air! [22]
The corona crisis accelerates the use of artificial intelligence in Russia. First
applications to combat the pandemic are coming to market (Gerit School, Moscow,
4/29/2020).
• Pandemic increases the need for digital technologies. Financial resources for this
could be significantly increased in the wake of the corona crisis. The dramatic
drop in oil prices and the pandemic-induced need to work on the move are
leading to an increased demand for digital products in Russia. The Russian
Digital Ministry is proposing to spend up to an additional 67 billion rubles (804.4
million euros) over and above the planned budget.
• Russia’s companies have already presented the first applications for virus control.
The startup company Promobot from Perm offers service robots that measure
body temperature or sterilize surfaces without contact. Sberbank has developed
similarly adaptive robots that autonomously disinfect rooms with UV irradiation.
• The country’s largest financial institution is one of the pioneers. It initiates
important innovations and at the same time provides money for new projects.
Its partner is the Russian Direct Investment Fund (RDIF), which says it has
commitments of more than US $2 billion from foreign donors looking to invest
in artificial intelligence projects. Together with Yandex, Gazprom Neft, Mail.ru
Group, and mobile operator MTS, Sberbank and RDIF formed an “Alliance for
the Development of Artificial Intelligence” in November 2019.
100 8 Other Applications
• RDIF plans to invest, among other things, in analytical methods that will
make it possible to distinguish between COVID-19 disease and pneumonia.
The parastatal Rostelecom Group has unveiled an image recognition method
that evaluates X-ray images of lungs for symptoms of COVID-19 disease. The
technology is being developed with state-owned holding Rostec. RDIF also
plans to work with the city of Moscow to deploy cameras for contactless fever
measurement in urban areas.
• Russian Internet company Yandex and Nvidia, the US manufacturer of pro-
cessors and chipsets, have been providing Russian companies with software
libraries to work on artificial intelligence since mid-March 2020. Among them
is the Parabricks application for genomic analysis, which is used in COVID-
19 research. Yandex plans to invest at least US $40 million in building cloud
capacity for AI applications.
• That the corona pandemic could accelerate the market penetration of AI solutions
in Russia’s industry is shown by the example of Sibur. The country’s largest
chemical company is cooperating with Italian software provider Brochesia
to introduce augmented reality (AR, computer-assisted augmentation of reality
perception) in its factories. According to Sibur, AI-enhanced data glasses can
be used to monitor production processes more efficiently and perform many
maintenance tasks remotely. This has reduced the business travel activity of
technicians during times of pandemic.
• Moscow will become a pilot region from the summer of 2020. There, applications
will be used to control exit restrictions during the corona quarantine. For this, the
metropolis is relying on its area-wide camera network. On July 1, 2020, a 5-year
experiment started with a flexible legal framework designed to facilitate the use
of AI applications while ensuring the protection of personal data.
• According to the “Roadmap for the development of AI technologies in Russia,”
the share of products related to artificial intelligence in the gross domestic
product will increase noticeably, to 0.8 percent by 2024 and to 3.6 percent by
2030, which would be above the global average. According to research by market
research firm IDC, Russian companies invested around US $139 million in AI in
2019, half of which was in computing technology. The market for such products
is forecast to grow strongly.
• Russia’s government supports the new technology with the “National Strategy
for the Development of Artificial Intelligence by 2030.” It lists as its main goals
prosperity gain, higher quality of life, national security, competitiveness of the
Russian economy, and an international leadership role in artificial intelligence.
According to this strategy, the following main application areas are defined:
• production
• more efficient planning processes, supply chain optimization,
• predictive maintenance of machinery and equipment,
• automation of routine production,
• increased use of robots,
• improvement of logistics management,
8.3 Big Data 101
is getting older. As part of a pilot project, South Korea plans to use thousands of
surveillance cameras and facial recognition to track contacts of corona-infected
people, according to Reuters. The test will be conducted in the South Korean city of
Bucheon and is scheduled to begin in January.
This is an excellent way to obtain large amounts of data [18]. One models a problem
on the computer, where the process has a number of parameters that one can change.
A large set of changes produces quite different positive or negative results. The best
results then yield values for the parameters one can work with [11]. It is remarkable
that the results are available in a short time. There can be several reasons for using
simulations: an investigation on the real system would be too complex, too
expensive, ethically unacceptable, or too dangerous:
• driving simulator (too dangerous in reality),
• flight simulator for pilot training, reenactment of critical scenarios (engine
failure, emergency landing—too dangerous in reality),
• medical simulation for training and education of clinical staff, especially emer-
gency situations or complications,
• simulators in medical training (training on real patients is not ethically justifiable
in some areas),
• a power plant simulator in which mainly the operating crews of nuclear power
plants train the control of incidents up to GAU (too dangerous in reality),
• crash test (too dangerous or too costly in reality),
• simulation of manufacturing plants before a rebuild (multiple rebuilds of the plant
in reality would be too complex and too expensive),
• Simulation models can be modified much easier than the real system.
• The real system does not (yet) exist. Example: Wind tunnel experiments with
aircraft models before the aircraft is manufactured.
• The real system cannot be observed directly.
• The real system works too fast: simulation of circuits.
• The real system works too slowly: simulation of geological processes.
• The real system is not understood or is very complex.
• The real system is understood in its elementary dynamics, but the time evolution
is too complex; respectively, an exact solution of the equation of motion is not
possible. Examples are the three-body problem, the double pendulum, molecular
dynamics, and nonlinear systems.
• Simulations are reproducible.
• Technical simulations, for example, for strength calculation (finite element
method), flow simulation, of factory processes and complex logistic systems, for
virtual commissioning or circuit simulation.
8.4 The Role of Simulation 105
The current Earth system models used for the predictions of anthropogenic
climate change are designed to perform simulations of typically several centuries in
length. However, simulating a transition from a warm period to an ice age requires
about 10,000 years of simulation time, and simulating a complete glacial cycle
requires more than 100,000 years. For these much longer integrations, current Earth
system models are far too computationally intensive to perform the corresponding
experiments on today’s computers. In addition, key components that are essential to
this problem are missing, such as an explicit representation of the ice sheets or the
interaction between carbon dissolved in the deep ocean and marine sediment.
Previous attempts to solve the problem have been carried out with highly
simplified models. For example, the dependence on the geographical longitude is
neglected, i.e., only the so-called zonal mean is represented, or the atmosphere is
extremely simplified. Since probably not all relevant feedback mechanisms have
been identified yet, the simplifications could suppress important processes. We
therefore pursue the approach of developing a coupled Earth system model in which
the individual components should be as realistic as possible. This should capture all
important interactions between the different subcomponents as far as possible.
Our model will consist of the following components: the atmospheric circulation
model ECHAM, the oceanic circulation model LSG, the marine carbon cycle model
HAMOCC, the ice sheet model SICOPOLIS, and the terrestrial dynamic vegetation
model LPJ. The model will incorporate a closed carbon cycle. In particular, this
has the effect that the .CO2 concentration of the atmosphere will be simulated
(predicted) by the model, unlike most complex models where it must be prescribed.
Currently, the model is still in the testing phase. The last interglacial period
(about 125,000 years ago) and the subsequent transition to an ice age are used as
the test simulation. The baseline simulation was performed using only the coupled
atmosphere-ocean model. Earth orbit parameters were varied during integration,
starting with the year 132,000 BCE to 112,000 years BCE.
About 125,000 years ago, a warmer climate in the northern hemisphere compared
to today is documented in many geological climate archives. The eccentricity and
the earth’s slope were slightly increased, and the greatest approach to the sun
took place in the northern summer. Overall, this results in greater solar radiation
during the northern summer. On the northern hemisphere it becomes predominantly
warmer—compared with the present condition—particularly in the Arctic and on
the continents. The warming is strongest in summer and autumn, which also
corresponds to the change in insolation. A major positive feedback mechanism here
is the reduction in snow and sea ice cover. The darker surface absorbs a greater
fraction of the solar radiation striking the surface, which increases warming and
further reduces snow and sea ice cover. This interaction between ice cover and
albedo is an important positive feedback effect in the climate system. Over the
Sahara, the warming of the land surface leads to an intensification of monsoon
precipitation, especially in the Sahel region. A similar effect occurs in southern Asia.
The accompanying denser cloud cover leads to cooling at about 10 .◦ N to 20 .◦ N.
Albedo is the measure of diffuse reflection of solar radiation and is measured on a
8.4 The Role of Simulation 107
4
290
2
270
0
Temperature °C
250
CO2 ppm
–2
230
–4
–6 210
–8 190
–10 170
–420000 –370000 –320000 –270000 –220000 –170000 –120000 –70000 –20000
Time before present
concentration is also reduced, further enhancing cooling. The reasons for this
reduction are still unclear. We hope to find the answer in the Earth system model.
There is no better way to show the usefulness of simulation models.
The curves and surfaces shown in Fig. 8.3 clearly indicate a fractal structure. The
same shapes are always repeated; sometimes, they are larger, sometimes smaller,
but they do not change.
Global warming or climate change is at present an extraordinarily hotly discussed
subject. Therefore, a second model is to be presented here, which goes back even
3 million years. This work [21] was published by Matteo Willeit, paleoclimatologist
at the Potsdam Institute for Climate Impact Research in Spectrum.de SciLogs under
the title “3 million years of climate change in computer simulation”.
The Quaternary is the youngest geological period; it covers the last 2.6 million
years. It is characterized by massive climate fluctuations, the ice age cycles, the
dramatic alternation of growth, and decay of giant continental ice sheets in the
northern hemisphere. Quaternary climate fluctuations are best identified by oxygen
isotopes measured in deep-sea sediment cores, which reflect variations in global
ice volume and ocean temperature. These data clearly show that there has been a
general trend toward larger ice sheets and cooler temperatures over the past 3 million
years, accompanied by an increase in the amplitude of fluctuations between ice ages
(glacials) and interglacials (interglacials), as well as a transition about a million
years ago from mostly symmetric cycles with a periodicity of 40,000 years to
strongly asymmetric 100,000-year cycles. However, the causes of these transitions
in the dynamics of the glacial cycles remain controversial among experts.
8.4 The Role of Simulation 109
Among other things, the role of .CO2 changes in Quaternary climate dynamics
is not fully understood, largely because of imprecise data of atmospheric .CO2
concentrations for the period before 800,000 years ago, the time span for which
we have high-quality ice core data.
In a recent study, we were able to reproduce natural climate variability through-
out the Quaternary using an Earth system model of intermediate complexity. In
addition to ocean and atmosphere, the model includes interactive ice sheets for
the Northern Hemisphere and a fully coupled global carbon cycle. The model was
driven only by changes in Earth’s orbital configuration (i.e., by the well-known
Milanković cycles), as well as by several scenarios for slowly changing boundary
conditions, namely, .CO2 outgassing from volcanoes as a geological .CO2 source
and changes in sediment distribution on continents.
These model simulations provide a consistent reconstruction of .CO2 , climate,
and ice sheets that is matched to available observational data. The fact that the model
can reproduce the main features of climate history, including data on Earth’s ice
mass and surface temperatures, gives us confidence in our general understanding of
how the climate system works and provides estimates for the contribution of external
forcing and internal feedbacks to climate variability.
Our results show a strong sensitivity of the Earth system to relatively small
variations in atmospheric .CO2 . A gradual decrease in .CO2 to values below
.≈350 ppm led to the onset of glaciation of Greenland and, more generally, the
northern hemisphere at the beginning of the Quaternary. The growth and melting
of the ice sheets subsequently led to the gradual erosion of the thick layer of loose
sediments that had previously formed on the continents through the undisturbed
action of weathering over millions of years. The erosion of this sedimentary layer—
it was essentially pushed away by the flowing glaciers—influenced the development
of the ice age cycles in several ways. First, ice sheets lying on soft sediments
are generally more mobile than ice sheets on hard bedrock because the ice slides
over the sediments more easily than over the bedrock. In addition, the transport
of sediments to the edges of the ice sheet generates significant amounts of dust,
which, once deposited on the ice sheet surface, enhances ice sheet melting as the
ice becomes dirtier and thus darker. Our results show that the gradual increase in
the area of exposed bedrock over time led to more stable ice sheets that were less
responsive to Earth’s orbital cycles, eventually paving the way for the transition to
100,000-year cycles about a million years ago.
The evolution of ice sheets is very sensitive to temperature, and the onset of
northern hemisphere glaciation about 3 million years ago would not have been
possible in the model if global temperature in the early Quaternary had been higher
than 2 .◦ C relative to pre-industrial times. Because the model has been shown to
correctly predict sea-level variations over the past 400,000 years and also the spatial
distribution of ice sheets at the peak of the last Ice Age, we are confident that the
sensitivity of ice sheets to climate is realistic in the model.
One can see in Fig. 8.3 very clearly that warming and cooling are periodic
functions. The present struggle to reduce .CO2 emissions makes cities cleaner, which
is very desirable, but a significant impact on global warming is not visible. It is
110 8 Other Applications
impossible to prove this, but simulation of these processes does allow fairly reliable
conclusions to be drawn.
Of course, the problem of pandemic simulation has also been turned to very
quickly [12]. Pandemic simulations became very popular in the 2000s. Biosafety
specialists and public health experts wanted to stress test the public health system
and look at where problems were occurring. In running the simulation, scientists,
business people, and government officials sit together and must make decisions in
real time to deal with a widening crisis presented to them in the style of television
news. Some striking examples:
• 2001: As part of Dark Winter, researchers simulate a smallpox attack on the
United States—a few months before a series of anthrax attacks in the United
States.
• 2003: Outbreak of severe acute respiratory syndrome (Sars) in Asia. Triggered
by a coronavirus, SARS spreads to a dozen countries.
• 2005: WHO revises its International Health Regulations (IHR). Countries pledge
to better monitor and promptly report disease outbreaks.
• 2009: H1N1 swine flu emerges in the United States.
• 2014: An Ebola outbreak is reported in West Africa.
• 2015: Zika virus spreads in Brazil.
• 2017: A pandemic simulation is conducted as part of the World Economic
Summit in Davos.
• 2018: Ebola breaks out twice independently in the Democratic Republic of
Congo.
• 2019: Event 201, a simulation of a novel coronavirus pandemic, takes place in
New York City.
• 2020: The SARS-CoV-2 pandemic kills 670,000 people in the first half of 2020
alone.
Efforts are also underway in Germany to create simulation programs for pan-
demics. The “Chair of Innovation Economics” at the University of Hohenheim
presents such a program [13] entitled “Agent-based Model of Health Policy
Measures in the Corona Epidemic,” with the following text:
Exit barriers and social distancing or contamination? How is the course of the
Corona pandemic affected by policy actions? Test for yourself in an online policy
lab what impact the various tools have on the course of the epidemic. What mortality
rates can be expected, and what will be the time course of the crisis? Please note:
Our policy lab does not provide forecasts, but serves to strengthen the understanding
of complex interrelationships.
The corona pandemic is unsettling us all. For many days, we have been following
the publications of Johns Hopkins University and the Robert Koch Institute with
the latest numbers of people infected with the coronavirus. Worldwide, politicians
are reacting with different instruments: While in most countries slowing down the
infection rates is declared the top goal, elsewhere, people are also discussing a
8.4 The Role of Simulation 111
References
1. Viktor Mayer-Schönberger, Kenneth Cukier. Big Data: The Essential Guide to Work, Life and
Learning in the Age of Insight (English Edition), John Murray (Publishers), 2017, 978-1-848-
54793-3
2. https://browse.dict.cc
3. https://www.stampworld.com/de
4. https://de.wikipedia.org/wiki/Amazon/Geschichte
5. https://de.wikipedia.org/wiki/Tripadvisor
6. https://statistikgrundlagen.de/ebook/chapter/korrelation
7. https://www.cnbc.com/2019/02/04/what-is-wechat-china-biggest-messaging-app.html
8. https://sourceforge.net/software/big-data/japan
9. https://www.thestar.com.my/tech/tech-news/2021/10/29/how-big-data-is-helping-make-
transport-cleaner-in-Dubai
10. https://www.gtai.de/gtai-de/trade/specials/special/chile/in-chile-bringen-fast-nur-start-ups-
die-kuenstliche-intelligenz-22598
11. https://www.softguide.de/software/simulation
12. https://www.spektrum.de/news/covid-19-pandemiesimulationen-konnten-uns-nicht-
vorbereiten/1761094
13. https://inno.uni-hohenheim.de/corona
14. https://inno.uni-hohenheim.de/corona-modell
15. https://www.gtai.de/gtai-de/trade/specials/special/chile/in-chile
16. http://www.agendadigital.gob.cl/
17. https://www.conicyt.cl/fondequip/sobre-fondequip/que-es-fondequip
18. https://www.softguide.de/software/simulation
19. https://de.wikipedia.org/wiki/Netflix
20. https://www.mpg.de/847908/forschungsSchwerpunkt
21. https://scilogs.spektrum.de/klimalounge/3-millionen-jahre-klimawandel-in-der-
computersimulation
22. https://teletimesinternational.com/2021/rta-digital-dubai-dubai-in-motion-gitex-2021
Chapter 9
Digitization
It is currently possible to capture all information that exists in digital form. A digital
camera is a camera that uses a digital storage medium as the recording medium
instead of a film; the image is previously digitized by means of an electronic image
converter. Its price ranges from 100 to 2000 euros. The first camera was designed by
Steven J. Sasson, a US engineer, at Kodak in 1975. In August 1981, Sony CEO Akio
Morita introduced the first filmless magnetic camera under the name Mavica and
expanded the product range in the following years. Figure 9.1 shows an example.
If the information is in analog form, it can also be digitized using suitable tools.
Perhaps the best known is the possibility of scanning written documents. It can be
assumed that any information is available in digital form. The possibility of further
processing on computer systems is a principle common to all manifestations of the
Digital Revolution and the digital transformation in economic,
social, working, and private life. In fact, technically networked digital commu-
nication significantly increases the diversity of technical-organizational solution
options. It therefore does not create structures that are stable in the long term, but
increases their flexibility and complexity and reduces their predictability through
the processes of disruptive change it triggers.
For mass storage and processing, increasingly powerful storage media have existed
since the 1960s and microprocessors since the 1970s. It is estimated that by 2007,
94 percent of the world’s technological information capacity was digital (up from
just 3 percent in 1993).
The quantity to be digitized can be anything that can be measured by sensors.
Typical examples are as follows:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024 113
C. Posthoff, Artificial Intelligence for Everyone,
https://doi.org/10.1007/978-3-031-57208-1_9
114 9 Digitization
All word processors provide the ability to color the background as well as the
text.
The digitization of audio data is often referred to as “sampling.” Sound waves
from a microphone, previously transformed into analog electronic oscillations, are
randomly measured and stored in rapid succession as digital values. Conversely,
these values can also be played back in quick succession and assembled into an
analog sound wave that can then be made audible again. The measured values
would actually result in an angular waveform during the reconversion: the lower the
sampling frequency, the more angular the waveform or signal. This can be reduced
by mathematical methods. In sampling, the bitdepth denotes the “space” for
values in bits, which are necessary, among other things, for the resolution of the
dynamic range. From a sampling frequency of 44.1 kilohertz and a resolution of 16
bits, one speaks of CD quality.
Records can be read and digitized contact-free with software support by scanning
a high-resolution optical image of the sound carrier by a program. This process
is used in the reconstruction of historical sound recordings. So you can take your
records, which you have collected over many years, to the flea market or give them
away without any problems if you have digitized them beforehand.
Digitization of archaeological objects is the digital recording of archaeological
objects in writing and images. All available information (classification, dating,
dimensions, properties, etc.) about an archaeological object (vessel, stone tool,
sword) is digitally recorded, supplemented by electronic images and drawings and
stored in a database. Subsequently, the objects can be integrated in the form of a
data import into an object portal such as museum-digital, where the objects can
be freely researched by anyone (Fig. 9.2). The reason for digitizing archaeological
objects is usually the recording of larger holdings such as archaeological collections
at museums or the offices responsible for the preservation of archaeological
monuments in order to present them to the public. Since in everyday museum life
it is never possible to show all the objects in a collection in the form of exhibitions
or publications, digitization is a way of nevertheless presenting the objects to the
general public and also to the scientific world. In addition, an electronic preservation
9.2 Some Applications 117
of the collection is carried out, an aspect that is not insignificant in view of the
collapse of the historical archive of the city of Cologne.
Already in 1916, Albert Einstein postulated “gravitational waves.” These dis-
tortions of spacetime arise as a direct consequence of his “General Theory of
Relativity”. On September 14, 2015, the first direct detection of gravitational waves
was achieved in the USA with the Advanced LIGO detectors. The measured signal
came from two black holes orbiting each other, getting closer and closer, and
eventually merging.
On August 17, 2017, astronomers received both electromagnetic radiation and
gravitational waves from an event for the first time: in the galaxy NGC 4993, 130
million light-years away, two neutron stars had collided and merged. This marked
the beginning of a new era of gravitational wave astronomy. The result of their
digitization is shown in Fig. 9.3.
118 9 Digitization
forms a field, and the sum of all predictions, regulations, costs, and yields is the
economics of the field. By viewing the field as a sum of productive assets, digital
agriculture can apply advanced data and analytics algorithms in real time to the
management of each asset and thus to the field as a whole.
Digital agriculture is not about genetics, weather forecasting, etc. While these
factors are extremely important for predictions, they are generated externally and
used as inputs. For example, seed selection would be tested based on seed profiles
developed and produced by agribusinesses. Soil conditions, hydration, and other
known information would be obtained from sensors or from historical records for
the region and field. The region and field data are historical and measured, and these
data are used as the basis for forecasting. Digital agriculture is about using known
field, crop, nutrient, protection, and hydration data to forecast
Digitization in education is very extensive. The use of calculators, cell phones,
tablets, and computers is part of everyday life, and the use of computers and
mathematics software has already been presented in an earlier chapter. Distance
learning became important since the COVID-19 pandemic and was part of everyday
life in most educational institutions for quite some time (Fig. 9.5).
The fundamental advantages of digitization lie in the speed and universality of
information dissemination. Due to low-cost hardware and software for digitization
and the ever-increasing networking via the Internet, new application possibilities
in business, administration, and everyday life are emerging at a rapid pace. When
the logic of production and business models, value chains, economic sectors,
administrative routines, consumption patterns, or even everyday interaction and
the culture of a society are profoundly changed as a result, we speak of digital
transformation. This entails opportunities, but also risks.
Digitization places new demands on the legal system, although legal science only
began to address this problem a few years ago. The theory of fuzzy law assumes that
law as a whole is fundamentally changing in a digitized environment. According
to it, the importance of law as a means of controlling society is significantly
Reference 121
relativized, since society’s claims are additionally oriented toward intangible goods
that transcend national boundaries.
Advances in digital techniques mean that better and better copies, reproductions,
and even counterfeits can be produced. Copyright infringement and counterfeiting
can lead to significant economic damage. In the early 1990s, for example, an
internationally active counterfeiting ring was discovered in New York that had
produced fake works of art with a retail value of $ 1.8 billion. The crisis in the
music industry that began in 1997 due to unauthorized music downloads from the
Internet, which can be burned onto blank CDs and thus also distributed illegally,
coincided with the same period. The number of blank CDs in Germany rose from
58 million to 303 million between 1999 and 2004, while the number of recorded
original CDs fell from 210 million to 146 million during the same period.
New types of copyright problems are raised by meta-media such as search
engines, which select content (texts, images, etc.) from primary media, grab it,
recombine it, and distribute it. A class action lawsuit has been filed against Google
Books by US publishers and authors, but criticism of the Google corporation’s
behavior has also come from European countries. Media report a “creeping decline
of the book industry.”
Another copyright problem can be found in 3D printing: with the creation of a
3D-printed template (usually in the form of a digital template or CAD design), a
work exists in the sense of copyright law and is protected. The production of such a
template by a 3D print constitutes a reproduction. Unlawful sale by third parties (or
without the consent of the author) may constitute a violation of copyright law.
The Digital Revolution has enabled great advances in science in a wide variety
of fields:
• Successes in genome decoding
• Predictions of climate research
• Complex models in physics and chemistry
• Nanotechnology
• Neurophysiological bases of language development and cognitive functions
• Economic simulations
• Comparative studies in linguistics and literature
Actually, every scientific discipline has benefited from developments in computer
technology.
Reference
1. https://4919758967.naturavitalis.de/Mediathek/Magazine/Wie-viel-Digitalisierung-vertraegt-
das-Gesundheitswesen.html
Chapter 10
Image Processing
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024 123
C. Posthoff, Artificial Intelligence for Everyone,
https://doi.org/10.1007/978-3-031-57208-1_10
124 10 Image Processing
Every year in Germany, approximately 4,500 people develop glioma, the most
common and malignant brain tumor in adults. Because this type of tumor is
extremely resistant, chemotherapy or radiation therapy has limited effectiveness,
and the tumors often cannot be completely removed even by surgery. For this reason,
scientists have been researching new treatment approaches at full speed for a long
time.
One of the most important factors in assessing the efficacy of a new therapy
for brain tumors is growth dynamics, which are determined via standard magnetic
resonance imaging (MRI) scans. However, these scans are prone to error and easily
lead to discrepant results when tumor extent is measured manually. “This can nega-
tively affect the assessment of therapy response and subsequently the reproducibility
and precision of scientific statements based on imaging,” explains Martin Bendszus,
medical director of the Department of Neuroradiology at Heidelberg University
Hospital.
A team from Heidelberg University Hospital (UKHD) and the German Can-
cer Research Center (DKFZ) has developed a new, machine-based method for
automated image analysis of brain tumors. For this purpose, the scientists have
designed algorithms and neuronal networks that can reproduce the therapy response
in brain tumors based on MRI images more reliably and precisely than traditional
radiological methods (Fig. 10.2).
Using a reference database of MRI scans from nearly 500 brain tumor patients
at Heidelberg University Hospital, the algorithms learned to automatically detect
and localize the brain tumors. In addition, the algorithms were able to volumetri-
cally measure the individual areas (contrast-enhancing tumor fraction, peritumoral
edema) and precisely assess the response to therapy.
The results of the research project were extensively validated in cooperation
with the European Organisation for Research and Treatment
of Cancer (EORTC). “The evaluation of more than 2,000 MRI examinations
126 10 Image Processing
of 534 glioblastoma patients from all over Europe shows that our computer-based
approach enables a more reliable assessment of therapy response than would be
possible with the conventional method of manual measurement,” explains Philipp
Kickingereder from the Department of Neuroradiology at Heidelberg University
Hospital. “We were able to improve the reliability of the assessment by 36 percent.
This may be critical for imaging-based assessment of therapy efficacy in clinical
trials. Prediction of overall survival was also more accurate with our new method.”
Here another problem becomes apparent. It is not so easy to obtain larger amounts
of relevant images. Here, many organizations have to work together!
The researchers hope that this technique can soon be used in clinical trials and,
in the future, also in routine clinical practice for standardized and fully automated
assessment of the therapy response of brain tumors. To this end, they designed
and evaluated a software infrastructure so that the technique can be integrated into
existing radiological infrastructure. “With this, we are creating the conditions for
widespread use and fully automated processing and analysis of MRI examinations
of brain tumors within a few minutes,” explains Klaus Maier-Hein from the Medical
Image Processing Division at the German Cancer Research Center.
Currently, the new technology is being re-evaluated at the NCT Heidelberg
as part of a clinical trial to improve the treatment of glioblastoma patients. “For
precision therapies, a standardized and reliable assessment of the effectiveness of
new treatment approaches is of paramount importance. This is where the technology
we have developed can potentially make a decisive contribution,” explains Wolfgang
Wick, head of the Neurological Clinic at Heidelberg University Hospital.
Philipp Kickingereder says that with this work, he and his colleagues have
been able to demonstrate the great potential of artificial neural networks in
radiological diagnostics. Klaus Maier-Hein hopes to further advance the technology
for automated high-throughput analysis of medical image data and to “send it to
10.2 Tumors, Alzheimer’s Disease, and Heart Disease 127
other diseases, such as brain metastases or multiple sclerosis, in addition to its use
in brain tumors.”
Usually, mammography is used to diagnose breast cancer. Radiologists detect
about 80 percent of the tumors present—the remaining 20 percent are overlooked
for various reasons. Tests with DeepMind have shown that even the previously
overlooked tumors can be detected accurately. Comparative tests have shown that
thanks to DeepMind, around 9.4 percent more positive findings could be made in
the USA and as much as 2.7 percent in the UK. The number of false positives has
also been reduced. The number fell by 5.7 percent in the USA and by 1.2 percent in
the UK [6].
Researchers have trained an adaptive artificial intelligence to recognize symp-
toms of Alzheimer’s disease . These signs are visible in special brain scans, but can
hardly be spotted by medical professionals. In the pilot project, the AI proved highly
successful, with a hit rate of 100 percent [5].
Scientists at Boston Children’s Hospital have developed a catheter that
autonomously moves through blood vessels and the heart. Once in the heart,
the robotic catheter autonomously searches for the target previously defined by
physicians. To do this, the tip of the catheter has tactile sensors and a miniature
camera that provide orientation similar to that of rodents at night. The sensor
technology, which moves mainly by touch through the beating heart, is controlled
by self-learning algorithms.
During the movement to the target, the tip of the catheter regularly touches the
vessel walls with its tactile sensors. The data obtained in this way is evaluated by an
artificial intelligence system, which recognizes where the robot is currently located
on the basis of the tissue being scanned. This information is then used to calculate
the next movement until the target in the heart is reached. To prevent injuries, the
catheter tip also has pressure sensors that regulate the contact pressure and thus
avoid puncturing the vessel walls. The camera, which is also present, also enables
monitoring by physicians, who do not normally have to become active during the
surgical procedure [6].
According to the study published in the journal Science Robotics, the first animal
experiments with the surgical robot were very successful. Pigs, whose hearts are
very similar to the human organ, were used. During the tests, the catheter had
to navigate autonomously from the tip of the heart through the left ventricle to a
defective aortic valve and implant an artificial heart valve there. This application
option was chosen because defective aortic valves are also a common problem in
human medicine.
In order to be able to compare the results of the robot, the same operation was
performed in parallel by surgeons completely manually and via a catheter controlled
externally via joystick.
The results of the robotic catheter show great potential. Eighty-nine out of
90 surgeries were completed successfully and even faster than the surgeons who
navigated their catheter through the heart using a joystick. Only the surgeons who
performed the operation completely manually were even faster.
128 10 Image Processing
nothing changes for the patient. As before, the flexible colonoscope is inserted into
the intestine after bowel cleansing. The images from inside the intestine are output to
a monitor and examined by the physician for changes. However, the software, which
is simply switched on at the touch of a button, additionally detects problematic areas.
In this way, up to 15 percent of changes not noticed by the doctor can be further
examined and, in case of doubt, removed. However, there are already systems that
do not require a colonoscope and move a small robot with camera and surgical tools
in the clean intestine [13].
Figure 10.3 shows the application of this approach—the picture is taken from the
“Neue Züricher Zeitung.” An important feature of image processing by computer
programs is that the images can be greatly enlarged; they can reach the level of a
microscope.
The combination of 3D printing and image processing has now been extended to
the production of artificial hearts. Scientists from Tel Aviv University have presented
a prototype of an artificial heart in the journal Advanced Science. The heart from
the 3D printer is made of human tissue, but at present, the cells cannot contract
synchronously. The organ is the size of a rabbit’s heart and otherwise has all the
characteristics of a human heart and already has individual chambers. It is the first
complete heart that scientists have been able to 3D print from human tissue. In the
future, organs made in this way could help people who currently have to wait for
donor organs that are hard to come by.
According to Tal Dvir, biotechnologist and head of the study, the heart is
currently comparable to the heart of a human fetus. In addition to greater availability,
the artificial heart also has the advantage of being completely compatible with the
recipient. In many cases with donor organs, the organ is rejected by the body and
the immune system function in the recipient must be suppressed. Since the artificial
heart is made of fat cells from the future recipient, complete biocompatibility is
ensured.
In the course of the study, the scientists biopsied fat tissue from a patient’s
abdomen and then separated the cellular and acellular components. The fat cells
130 10 Image Processing
were then reprogrammed into pluripotent stem cells containing various types of
somatic cells, including heart cells. From the acellular material, such as glycopro-
teins and collagen, the scientists created a hydrogel that was used as a support tissue
for the artificial heart.
The scientists then created a bio-ink from the endothelial and cardiac muscle
cells and the hydrogel, from which the 3D printer created tissue patches (cardiac
patches). In the past, these tissue patches have been used after heart attacks to
supply dead areas of human hearts with new cells. Now, for the first time, the patches
have been successfully used to create a completely new organ.
According to Dvir, it is “principally possible to use the technology to make a
larger heart for humans as well.” However, several years of research will be needed
before hearts from 3D printers can help sick people. The currently manufactured
heart already looks like a natural heart, but its function is not yet complete.
In principle, the artificial heart muscle is already contracting, but the coordination
between the individual heart chambers required for pumping does not work. Should
the scientists solve this problem, hearts from the 3D printer could also be used in
practice. The Israeli scientists expect that animal trials with rabbits and rats can take
place within a year. Clinical trials with humans are not expected for several years.
Currently, heart disease is one of the leading causes of death worldwide and often
cannot be treated due to a lack of donor organs, even when detected.
Researchers at the Massachusetts Institute of Technology (MIT) have used artifi-
cial intelligence to discover a very special antibiotic that, unlike most antibiotics, is
not a conventional variation of an already known drug. The researchers named the
highly effective antibiotic halicin. The name is derived from the computer HAL
9000 from the cult film 2001 [8].
The antibiotic kills bacteria by limiting their ability to maintain an electrochem-
ical potential that is used to produce molecules for energy storage. The results of
tests with the developed agent are astonishing. For example, E. coli bacteria did not
develop resistance to the new antibiotic even after 30 days. By comparison, with the
conventional antibiotic ciprofloxacin, the bacteria adapted to the drug after
just 3 days.
The algorithms are able to find molecular structures with desired properties, such
as killing bacteria in this case. The system’s neural networks automatically learn the
molecules’ representations and classify them into specific domains, which makes it
possible to predict behavior very accurately.
The researchers then trained the system with 2,500 molecules, 1,700 of which
corresponded to those of already known drugs and the remaining 800 to natural
products. With this basis, the computer developed a library of around 6,000 possible
combinations, including the newly discovered antibiotic halicin, which actually
proved to be highly effective and efficient in the test.
Experience shows that it will be some time before the new antibiotic can be used
in human medicine. In mice, however, halicin has already proven that it can easily
combat a problematic strain of bacteria with which US soldiers frequently become
infected in the Middle East.
10.4 Medical Applications 131
The researchers are very confident that the new system will have a very positive
impact on future drug development. Just to see the potential of the system, the
researchers fed the neural network with a database of more than 100 million
molecules. After only a short time, the program was already suggesting 23
potentially highly effective new drugs. In the future, however, the system should
not only develop completely new drugs but also modify existing drugs so that they
have fewer side effects and increased efficacy.
A new calculation model precisely predicts the loss of performance of athletes
into old age. A research group led by Bergita Ganse, endowed professor of inno-
vative implant development at Saarland University, has developed a computational
model to determine the physical decline of people as they age. “We asked ourselves
whether we could succeed in predicting an athlete’s performance into old age, using
a single measurement,” Ganse explains.
According to the publication in the journal Geroscience, to do this, the scientists
trained a program with data from nearly 5,000 Swedish track and field athletes
whose performances were documented in the Swedish Veteran Athletics
database between 1901 and 2021. In total, the scientists were able to use 21,000
data points for their study. On average, there were four measurement data points
from each athlete. Only running disciplines were taken into account, because in
other disciplines such as javelin or discus throwing, the weight of the throwing
equipment changes depending on the age of the athletes. This makes comparability
difficult and would make a prediction of performance decline with age less accurate.
“Runners, on the other hand, run 100, 200, 800 meters whether they are 23, 40, or
70 years old” [9].
As Ganse explains, as a central result of the study, the scientists succeeded in
developing a model that predicts the decline in performance of athletes into senior
age with high accuracy. The precision achieved is significantly higher than that of
older models, which in most cases are based on a linear decline in performance.
“We were surprised to find that athletes who were very powerful and young had the
greatest relative decline in performance, which was also true for older athletes with
lower baseline performance. We found the lowest rate of decline in high-performing
athletes with high baseline age.” Moreover, the study shows that people should still
start exercising at an older age. Individuals who ran a good time in their discipline
at an advanced age were very likely to perform significantly better than their age
average, even at an even older age.
Image processing currently has perhaps the greatest significance for daily life,
for medicine and the use of robots, and much more. There are excellent textbooks,
software packages, equipment systems, and opportunities for comprehensive and
solid continuing education.
132 10 Image Processing
In many cities and countries, important streets and squares are secured by cameras.
With the aid of a learning program, the control of traffic lights can be adjusted to
reduce the waiting time of vehicles; in some cases, this significantly reduces the
amount of exhaust gas. Traffic jams can also be avoided or reduced in this way
(Fig. 10.4).
Singapore lives up to its reputation as an extremely clean city. Even skeptical
natures will not encounter litter or cigarette butts on the streets or in the subways,
not even in public restrooms. The reason for this is the rigorous policy that the
city-state has consistently pursued for years in matters of cleanliness: violations of
regulations that make carelessly throwing away cigarette butts or trash in public
a punishable offense are followed by draconian fines. The principle is effective:
anyone who throws trash on the street pays 1,000 Singapore dollars. Foreigners
have to leave the country the next day.
• Due to the level achieved in image processing, blind people can be greatly
assisted. They can be given a camera that captures the whole environment. This
recognizes when roads can be crossed or when obstacles appear on the way.
It guides its wearer safely to the entrance of a bus or a streetcar. In addition,
the camera is equipped with a navigation system otherwise used in cars. In this
way, the disabled person can reach their destination relatively easily. Their life is
arduous, yet relatively safe. Some such systems even respond to acoustic input.
• A modern car today has a greater number of cameras and microprocessors.
In autonomous driving, this equipment is fully responsible for safe driving. In
particular, it recognizes all traffic signs and follows them. The same equipment
can be installed in a normal vehicle. The driver thinks that he determines his
own driving; however, the observance of the mandatory and prohibitive signs is
enforced by the existing computer system.
One can always start from a raster graphic; this is an image that is already available
in digital form. The individual color values of the raster graphics are referred to as
pixels as well as the area elements necessary for the acquisition or representation
of a color value in the case of an image sensor or a screen with raster control
(Fig. 10.5) [1].
Each pixel is characterized by its shape, its coordinates, and its color value. The
encoding of the color used in the pixel is defined, among other things, by the color
space and the color depth. The simplest case is a binary image, where a pixel
stores a black and white value. The screen resolution specifies how many pixels
make up the image representation. The larger the resolution, the more detail
can be seen on the screen or the more space is available to display objects.
The resolution is specified in different notations, values, and units. For a digital
camera, for example, 12 megapixels is used. This is the product of the side lengths
(width .× height), for example, 4,256 .× 2,848 pixels. For screen resolution,
however, it is not the product that is interesting, but the number of pixels per page
length. Since we prefer widescreen display, our screens are wider than they are taller.
Known screen resolutions are VGA (640 .× 480), Super VGA (800 .× 600),
and Full HD (1,920 .× 1,080).
Both the pixel density (and thus the local sampling rate) and the size of the
information stored in the pixel (color depth, color channels, position, shape ) are
limited in practice, which is why a pixel can only represent an approximation of
reality.
The pixels of an image sensor or display usually consist of areas of one basic
color each (red, green, and blue). In the case of liquid crystal displays (LCD),
each visible picture element is controlled by a color value. The areas responsible
for the basic colors of the pixel (subpixel) are often arranged adjacent to
each other. The finer subpixel structure compared to the pixel can be used to
increase the horizontal resolution during rasterization (subpixel rendering).
Furthermore, triangular, irregularly sized, alternately arranged, or additional white
subpixels are also possible.
The physical size of a pixel depends on the device. The pixel density of a
screen or scanner is given in pixels per inch (ppi) or dots per inch
(dpi). Commercially available computer screens achieve a pixel density of about
100 ppi, corresponding to 0.3 millimeters per pixel. For televisions, the pixel
density is usually lower and several times higher for newer smartphones, while
the sensors of scanners and digital cameras can reach several thousand ppi. The
maximum number of pixels that can be used in image sensors is often specified in
megapixels, but this usually refers only to the color dots of a Bayer sensor and
not the pixels. The aspect ratio of a pixel on the screen (pixel aspect ratio)
does not necessarily have to be 1:1; most video standards prescribe irregular pixel
aspect ratios. The pixel size as well as the pixel pitch in relation to the image
resolution have a decisive influence on the readability and recognizability of texts
and graphics on computer monitors and televisions.
After scanning an image, a digital form is available for the image, which can
be further processed with the help of suitable software. Scanners are usually
integrated with a printer. Their quality is so good that many authorities accept
scanned documents. We recall the representation (red, green, blue) by three bytes.
Their content is between (0,0,0) and (256,256,256). (0,0,0) returns black, and
(256,256,256) corresponds to white. If the byte contents are the same, then different
gray levels are obtained. (127,127,127) is in the middle between all black and all
white.
In Fig. 10.7, red (upper circle), green (left circle), and blue (right circle) were
each used with the value 30. The superimposition of the different colors already
results in quite different color levels. Figure 10.6 focuses on white and black. These
images require less memory (Fig. 10.7).
In Fig. 10.8, red, green, and blue were each used with the value 60.
10.7 The Processing of Images 135
LaTeX
Design Coding
LaTeX
Design Coding
LaTeX
Design Coding
Figure 10.9 uses the three colors with the value 90. With this value, no shading
appears, and the respective colors are distinct. You can see that the contrast of the
individual colors increases.
136 10 Image Processing
The PixelMath program used in [1] and any similar program allows different
operations. One can enlarge or reduce a pixel with the screen coordinates .(x, y).
For example, one replaces the values .(x, y) with .(2 · x, 2 · y) to enlarge the image,
or one uses .(x/2,.y/2) to reduce the image. You can also do this for the values
individually, so replace .(x, y) with .(x/2, y) and so on. For example, you can make
an image gallery where all images have the same size. But you have to keep in
mind that the number of pixels does not increase, only the distance between them
becomes smaller or larger. Of course, this can lead to effects that change the image
quality.
Since a Cartesian coordinate system is present, one can carry out all transforma-
tions, which can be expressed by formulas of the analytic geometry . One can also
use polar coordinates , there one can express very nicely effects that start circularly
from a point. As an example, we look at the fish-eye projection .
If one uses spherical coordinates, then each point in space is determined by
.(r, ϕ, θ ). The spherical radius r has theoretically no influence on the mapping
function. Thus, all objects can be scaled to exactly one distance, and an ambient
sphere is created. For the radius of the surrounding sphere, one chooses the focal
length of the lens (.r = f ).
With Cartesian coordinates, the image plane is located at .z = f and touches
the surrounding sphere at the pole (.θ = 0). Plane polar coordinates are applied
to the image. The position of an image detail is described by .(r, φ). In this
geometric arrangement, the imaging functions can be well illustrated and usually
constructed. The object coordinates .ϕ (azimuth) and .θ (polar angle) map into the
image coordinates .φ (azimuth) and r (radius).
The principle is the same as in the azimuthal map mesh of the earth. There .φ = ϕ
is valid. However, in contrast to the earth sphere, the surrounding sphere is viewed
from the inside. The front side of the image plane then shows the environment
as a mirror image. For the backward view required here, the image coordinates
(polar and 2D Cartesian) are oriented differently, so the azimuth .−0.◦ direction and
azimuth-orbit sense may differ between the environment and the image. Azimuth
angular distances are imaged unaltered, and imaging function and azimuth do not
affect each other. Thus, an azimuthal imaging is present.
The imaging function .g describes how an object in the polar angle .θ appears on
the image displaced from the center by radius r. The lens has the (central) focal
length f . With .f = 1, .g becomes the normalized imaging function .h:
r = g(θ, f ) = f · h(θ ).
. (10.1)
dh
meridional scaling:
. SM = . (10.2)
dθ
10.7 The Processing of Images 137
h(θ )
sagittal scaling: Sθ = . (10.3)
sin(θ )
From this, solid angle scaling .SΩ , linear (effective) scaling .S, and deformation
D can be derived:
.
SΩ = Sm · Sθ .
. (10.4)
S = SΩ. (10.5)
Sm
D= . (10.6)
Sθ
θ
.angularly true: r = 2 · f · tan . (10.7)
2
equidistant : r = f · θ. (10.8)
θ
true to area: r = 2 · f · sin . (10.9)
2
orthographically: r = f · sin(θ ). (10.10)
There are a variety of ways to change the parameters and angle functions used,
creating a variety of interesting projections, for many different applications.
• Scientists and resource managers (e.g., biologists, foresters, geographers, and
meteorologists) use fish-eye lenses to image a hemispherical region of plant
vegetation in ecophysiology or to predict potential shortwave irradiance from
horizon elevation (sky view factor) and the longwave irradiance to be
derived from it. Relevant structural parameters of tree canopies such as leaf area
index (LAI), leaf angle distribution, and ground-level light availability can be
derived from the analysis of these images.
• Fish-eye lenses also help in evaluating forest health, detecting overwintering sites
of monarch butterflies, and managing vineyards.
• In topoclimatology, horizon elevation of ground depressions can be used to infer
the origin of deep frosts during inversion weather, as well as to make statements
about the causes of cold air lake phenomena.
• From data of the sky view factors parameters obtained from georefer-
enced fish-eye images, meteorological correlations of the radiation balance of
urban heat islands are studied in urban climatology.
• Meteorologists determine cloud cover (degree of coverage of the sky).
• Astronomers image a large portion of the sky to record constellations, the Milky
Way, meteors, auroras, and light pollution.
• Many planetariums use fish-eye projection lenses to project the night sky or other
digital content onto the inside of a dome.
138 10 Image Processing
• Surveillance cameras with fish-eye lenses can capture an entire room at once.
Unlike swiveling cameras, there is no intermittent dead zone and no vulnerable
drive.
• Video door phones with extra wide angle of view (door viewer function).
• In the IMAX dome system (previously “OMNIMAX”), film capture is through
a circular fish-eye lens, and projection of the motion picture is through similar
optics onto a hemispherical screen.
• Photographers and videographers use fish-eye lenses to get the camera as close
as possible to the decisive point for action shots while also capturing the overall
context. For example, in skateboarding, the focus is on the board, and the skater
is still in the frame.
• The first music video shot entirely with a fish-eye lens was the song “Shake Your
Rump” by the Beastie Boys in 1989.
• Flight simulators and visual combat simulators use fish-eye—projection lenses
to create a seamless environment for pilots, air traffic controllers, or military
personnel to train in.
• In computer graphics, circular fish-eyes can be used to create an environment
mapping of the physical world. A complete 180-degree fish-eye image can fill
half of a cubic environment mapping with an appropriate algorithm. Environment
maps can be used to render 3D objects embedded in virtual panoramas.
Many image transformations are used in painting, for example, the cylinder
anamorphosis. An image transformed in this way looks normal when reflected
in a cylindrical mirror. A famous example is Holbein’s painting “The Envoys”
(Fig. 10.10).
• The seventeenth century saw a revival of fantastic anamorphic imagery. Mag-
ical and religious connotations were largely abandoned, and the images were
understood as scientific curiosities. Two important works on perspective were
published: Perspective (1612) by Salomon de Caus and Curious
Perspective (1638) by Jean-Francois Niceron. Both contained extensive
scientific and practical information about anamorphic images. Niceron’s work
explains three types of large-scale anamorphism: optical (horizontal view),
anoptical (upward view), and catoptical (downward view, e.g., from a
mezzanine). A conic perspective is also described. Toward the end of the
century, the techniques for creating anamorphic images became widely known
through Charles Ozanam’s Mathematical Recreations.
Between 1669 and 1685, both perspective and mirror anamorphosis were
introduced by the Jesuits to Emperor K’ang-hi and the monks of the Beijing
Mission in China. However, Chinese production of anamorphic images was
already taking place on a large scale during the late Ming Dynasty. Unlike the
halftone system used in the West, the images were usually created freehand.
Since Chinese anamorphoses focused primarily on erotic subjects, a Jesuit
influence is unlikely. It is considered likely that Chinese catoptic techniques,
which are technically unrelated to geometric anamorphosis, influenced European
mirror anamorphosis and not vice versa.
10.7 The Processing of Images 139
from a tower at their base, the assembled people dissolved into recognizable
images.
Surrealist artist Salvador Dali used extreme foreshortening and anamor-
phosis in his paintings and works. A glass floor in a room next to his studio
allowed for radical perspective studies from above and below. In the Dali Theater
and Museum, there is a three-dimensional anamorphic living room installation,
Mae West Lips Sofa (Fig. 10.12), which looks like the movie star’s face
from a certain angle.
• In the twentieth century, artists began to play with perspective by drawing
impossible objects. These objects included stairs that always lead upward or
cubes where the back meets the front. Such works were popularized by the artist
M. C. Escher and the mathematician Roger Penrose. Although they are called
impossible objects, such objects as the Necker cube and the Penrose triangle can
be modeled in 3D by anamorphic illusion. When viewed from a certain angle,
such sculptures appear as the so-called impossible objects (Fig. 10.13).
• The Ames room was invented in 1946 by the American scientist Adelbert
Ames Jr.. When viewed through a peephole, the room appears to have a
normal perspective. From all other angles, however, one can see that the room
is made up of irregular trapezoids. Similar effects were achieved as early as
the Renaissance through the use of accelerated perspective in stage
design. These included productions by Scamozzi (1588–1589), Furtenbach
(1625), Sabbattini (1637), and Troili (1672).
10.7 The Processing of Images 141
One of the most interesting effects of an Ames space is that the distorted
perspective can make people and objects appear much larger or smaller than they
actually are. For this reason, Ames rooms are often used in cinema for practical
special effects. A well-known example is the houses in the Shire in the films
The Lord of the Rings and The Hobbit. By using forced perspective,
the character of Gandalf appeared much larger than the characters of Frodo and
Bilbo, without the use of digital effects (Fig. 10.14).
• Cinemascope, Panavision, Technirama and other widescreen formats
use anamorphosis to project a wider image from a narrower film frame. The
IMAX company uses even more extreme anamorphic transformations to project
moving images from a flat film image onto the inside of a hemispherical dome,
in its Omnimax or IMAX Dome process.
142 10 Image Processing
behind the “Meal of the Blind” in 1903. In the background of Picasso’s famous
painting “The Life,” the crouching woman can still be seen, but it was to take more
than a hundred years for the original to reappear (Fig. 10.16). The starting point
was a fluoroscopy of the original by X-ray fluorescence spectroscopy. But since
only the outlines of the crouching woman could be revealed in this way, a further
step was necessary. George Cann and Anthony Bourached, researchers at University
College London and founders of Oxia Palus, trained a convolutional neural network
with all the images from Picasso’s “Blue Period.” From this, the AI generated
Picasso’s typical style, brushstrokes, and color choices during this creative phase
and transferred them to the outlines. The result was then printed out in original size
on a 3D printer—and the new old Picasso was ready. Now, London-based start-up
Oxia Palus has achieved another step in reconstruction: it has reconstructed the
painted-over image and restored it using 3D printing.
References 145
References
Robots are a core issue of artificial intelligence for many; especially the topic self-
driving cars is the basis of discussions about autonomous systems independent of
humans. Opinions vary widely. The subject area of robotics deals with the attempt to
reduce the concept of interaction of a system with the physical world to principles of
information technology as well as to technically feasible kinetics. The term “robot”
describes a system that combines both concepts by implementing interaction with
the physical world on the basis of sensors, actuators, and information processing.
The core area of robotics is the development and control of such robots. It
encompasses subfields of computer science, electrical engineering, and mechanical
engineering. The goal of robotics is to establish a controlled cooperation of robot
electronics and robot mechanics through programming.
The term was invented by science fiction author Isaac Asimov, first men-
tioned in his short story Runaround in March 1942 in Astounding magazine.
According to Asimov’s definition, robotics refers to the study of robots or even
machines.
• Already in the antiquity, first experiments with automata were carried out
(Fig. 11.1). Known are automatic theaters and music machines of Heron of
Alexandria. With the decline of ancient civilizations, the scientific knowledge
of that time disappeared. Around 1205, Al-Jazari, a twelfth-century Arab
engineer and author, wrote his work on mechanical devices, the “Book of
Knowledge of Sensible Mechanical Devices” which also became known as
“Automata” in Western culture. Leonardo da Vinci is said to have been
influenced by the classical automata of Al-Jazari. Thus, records and sketches
from the fifteenth century are known from him, which can be interpreted as plans
for androids. However, the level of technical knowledge was not sufficient to
realize such plans. Around 1740, Jacques de Vaucanson designed and
built a flute-playing automaton, an automatic duck and the first programmable
fully automatic loom.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024 147
C. Posthoff, Artificial Intelligence for Everyone,
https://doi.org/10.1007/978-3-031-57208-1_11
148 11 Robotics
• At the end of the nineteenth century, increased efforts were made in the military
(remotely operated boats, torpedo controls). The writer Jules Verne (1828–
1905) wrote a story about a human machine. In 1920, the writer Karel Čapek
introduced the term robot for an android. After the end of World War II, the
field of robotics underwent rapid progress. Key factors were the invention of the
transistor in 1947 at Bell Laboratories and integrated circuits and, subsequently,
the development of powerful and space-saving computers.
• From around 1955, the first NC machines (devices for controlling machines)
appeared on the market, and in 1954 George Devol (1912–2011) applied for
a patent for a programmable manipulator in the USA. This date is considered
the birth of the development of industrial robots. Devol was also a co-founder of
Unimation, which introduced the first hydraulically operated industrial robot in
1960. In 1968, the first mobile robot was developed at MIT. In Germany, robot
technology was not used productively until the early 1970s. Around 1970, the
first autonomous mobile robot Shakey (the shaky one) was also developed
at Stanford Research Institute.
• In 1973, development of the WABOT-1 humanoid robot was launched at Waseda
University in Tokyo. That same year, German robotics pioneer KUKA built the
world’s first industrial robot with six electromechanically driven axes, known as
FAMULUS. A year later (1974), Sweden’s ASEA introduced its fully electrically
driven robot (IRb6).
• In 1986, Honda launched the “Humanoid Robot Research and Development Pro-
gram.” The result was the humanoid robot versions P1 to P3. Honda introduced a
11 Robotics 149
further development in 2004 in the form of the ASIMO humanoid robot. In 1997,
the first mobile robot landed on Mars (Sojourner).
• The toy industry has not closed its mind to robotics either. Examples of
such products are Lego Mindstorms, iPitara, Robonova, or Sony’s robot dog
Aibo (Fig. 11.2).
According to futurologists and philosophers, the ever-increasing automation and
digitization, combined with likewise growing collection and exchange of data,
requires fundamental questions about the role of humans in this process and in these
contexts. As early as 1942, Asimov, for example, formulated a corresponding code,
the “robot laws.”
• Alternative techniques to the wheel as a means of locomotion in the human
environment are now being researched, such as walking on six, four, two, or
even one leg. While industrial robots usually perform manual or handling tasks
in an environment adapted to them, such service robots are intended to provide
services for and to humans. To do this, they must be able to move and find their
way around in the human environment.
• Like a game, but with serious scientific research as a background are robot soccer
games between teams of similar robots. The researchers’ goal is to develop a
soccer team of autonomous bipedal robots by 2050 that can compete against the
soccer world champion.
• Industrial robots are mostly used in environments that are too dangerous or
unreasonable for humans. Today, modern robots do mindless assembly line work
faster and much more accurately than a human and can replace them in more
and more areas. Cars are now built with heavy robotic involvement, and even
a modern microprocessor would be impossible to manufacture without a robot.
Service robots have been used for some time to make people’s daily lives easier or
to entertain them, such as the RoboSapien. There are household robots that are
able to vacuum, mop the floor, or mow the lawn. They may specialize in only one
task, but they can perform it relatively autonomously. Research robots explore
distant planets or disaster zones, among other things, and penetrate volcanoes
150 11 Robotics
or sewage pipes. They are also used for a wide variety of detection missions in
the marine environment. There are concepts and first prototypes for cryobots
and hydrobots that will be used in space travel in the future. There are also
considerations to use robots for sample retrieval missions and asteroid mining.
• In medicine, robots are used for examinations, operations, and rehabilitation and
perform simple tasks in everyday hospital life.
A prototype for tiny nanorobots that can move in the bloodstream was already
tested on an eye at ETH Zurich in 2004. They are controlled by magnetic fields
from the outside. The FRIEND assistance robot, developed at the Institute of
Automation Technology at the University of Bremen, is designed to support
disabled and elderly people in the activities of daily living (e.g., preparing a meal)
and enable them to reintegrate into working life.
• First entertainment robots such as Sony’s robotic dog Aibo are a step toward
electronic pets. In addition to Aibo, there are other robot products from the toy
and entertainment industry that can be programmed with a computer in a mostly
simple language to, for example, follow a light source or a line on the floor or
sort colored building blocks.
• An interesting hobby is to build your own robots. This can be done supported by
prepared robot kits or by free imagination. In this case, for example, you have
to construct a car-like vehicle yourself, determine distances to the target or the
color of the ground with suitable sensors, and determine a course for the vehicle
to follow from these measurement results. The actual task is to link the sensor
data with the speed and direction of the vehicle. This is done in a microcontroller
that has to be programmed by the user. The required electronics are offered in
different designs. Well-known but also very elaborate role models are the Rovers.
• Robots are also a popular science fiction subject. There, there are humanlike
robots that have artificial intelligence.
An additional variation of the robot, already realized in a very simple form, is
the cyborg as a fusion of robotic technology with human anatomy. Androids,
artificial humanlike beings, can be robots, but robots need not be androids. A first
widely developed approach is Honda’s ASIMO robot.
• Robots are also increasingly an important topic in education. There are robots for
elementary school, robots for secondary school or for secondary schools, robots
for universities, and robots for vocational training. A special form of robots for
education are rovers, which are developed and tested, for example, as part of
space education at institutions in Germany. Usually, these specialized robots are
intended as rovers for a specific goal or competition. At the Maker Faire 2016
in Berlin, a rover called “EntdeckerRover” ER2 was presented, which is suitable
for education and leisure and can also be adapted for different educational areas.
Other systems are mostly available in plastic from other manufacturers and
projects.
• Robots and the special form rover mostly support education in the area of STEM
subjects in Germany and Austria. So it is also about the promotion of science and
technology education or technology knowledge as well as the topics of computer
11 Robotics 151
to the state of the art. The state of the art comprises all knowledge which has been
made available to the public by written or oral description, by use, or in any other
way before the date relevant for the priority of the application; cf. section 3 (1), 2nd
sentence, Patent Law. In the case of robots, the patent applicant must therefore show
that his robot has new functions that do not belong to the state of the art (e.g., on the
ability of robots to run).
Furthermore, it must be an invention. Patentable inventions are technical teach-
ings for planned action which reproducibly bring about a causally foreseeable
success using controllable natural forces without the interposition of intellectual
activities (BGH, March 27, 1969 - X ZB 15/67 = BGHZ 52, 74; NJW 1969, 1713;
GRUR 1969, 672). A technical further development of a robot is a patentable
invention only if it is not obvious from the prior art to “the average person skilled in
the art who knows the entire prior art” (a legal fiction, not a real person); cf. section 4
p. 1 Patent Act, Art. 56 p. 1 EPC. That is, there is a lack of inventive step if the
person skilled in the art can be expected to have come up with this solution promptly
and with a reasonable effort, based on the prior art, without being inventive. Thus,
in the field of robotics, only inventions that represent a significant advance in the
development of robot technologies are patentable. However, this does not have to
refer to the robot as a whole, but can also refer to individual components, such as a
robot arm or a mode of operation for locomotion.
In addition, the invention must be applicable in some industrial field according
to section 5 (1) Patent Act, Art. 57 EPC. The term “industrial applicability” is
interpreted broadly by the European Patent Office and is of secondary importance
in practice. It is sufficient that the invention can be manufactured or otherwise used
in a technical business. It is also not important whether one can “make money” with
the device or process; the only decisive factor is that the claimed subject matter
can be used outside the private sphere. Most inventions in the field of robotics are
aimed at commercial success, for example, in the creation of household helpers or
robots for operations. This is already in the nature of things, since the inventions of
robotic technologies require enormous investments and these are reclaimed by the
investment providers at a profit.
The maximum term of a patent is 20 years from the day after filing, according
to §16 PatG and Art. 63(1) EPC. However, according to section 16a PatG, Art. 63
(2) (b) EPC in conjunction with Regulation (EEC) No. 1768/92, a supplementary
protection certificate can be granted for inventions that can only be commercially
exploited after extensive approval procedures, which then extends the patent term by
a maximum of 5 years. Due to the long development cycles in robotics, this should
be applied regularly.
According to section 1 (2) and (3) Patent Act and Art. 52 (2) and (3) EPC,
scientific theories and mathematical methods, such as construction plans for a robot,
cannot be protected as patents. The same applies to the design and appearance of a
robot, since aesthetic creations of form cannot be protected by patent.
A robot’s misbehavior, whether it stems from the desire for autonomy or any
other reason, always entails a number of liability issues. These can arise, on the one
hand, from a contractual breach of duty pursuant to section 280 (1) of the German
11 Robotics 153
Civil Code, on the other hand from tort law pursuant to section 823 of the German
Civil Code vis-à-vis third parties, or also from the Product Liability Act. If a robot is
used in the context of a contractual relationship (e.g., rental) by another contractual
party and the robot causes damage to this party, this is certainly a breach of duty
in the sense of §280 BGB. A case that has become known through the media is the
use of the ROBODOC from Integrated Surgical System, which has led to numerous
claims for damages (BGH, June 13, 2006 - VI ZR 323/04 = BGHZ 168, 103; NJW
2006, 2477).
According to §249 p. 1 BGB, the debtor who is obligated to pay damages must
restore the condition that would exist if the circumstance obligating him to pay
damages had not occurred. In doing so, the tortfeasor shall compensate for all
damage caused by the result obligating to compensation (so-called total reparation).
Apart from the rule of total reparation, another principle of the law of damages is
expressed in §249 p. 1 of the Civil Code, namely, the principle of production
or natural compensation. Here, the tortfeasor is to produce the condition in
money, which would exist without the damaging event.
A question that will certainly become increasingly important in the future is who
is liable for the decision made by a robot based on artificial intelligence. Thus, it
is certainly arguable that the person who uses the robots must be liable, since he is
responsible for the road safety of the robot used and must ensure appropriate safety
measures. In a contractual relationship, this certainly arises from the general duty
of care of the obligation relationship (cf. §280 Para. 1 BGB) toward third parties
certainly from the tort law, §§823 ff BGB. In principle, the manufacturer could be
liable under the Product Liability Act (ProdHaftG). The prerequisite for product
liability according to §1 para. 1 p. 1 ProdHaftG is, among other things, that there
was a defect in the object causing the damage (i.e., in the robot). Such a defect could
possibly exist if the manufacturer has not incorporated suitable safety measures in
the programming of the robot’s control software. In any case, the manufacturer is
not liable if the robot did not have the defect causing the damage at the time it
was placed on the market (Palandt Sprau Kommentar zum BGB 69. Auflage 2009
§1 ProdHaftG Rn. 17) and if the defect could not have been detected according
to the state of the art in science and technology at the time the manufacturer
placed the product on the market; cf. §1 para. 2 no. 5 ProdHaftG. Nevertheless,
the manufacturer of robots must build safeguards into a robot (and especially into
the software) so that no damage can occur, even after an AI learning process. In
science fiction literature, for example, the three laws of robotics were developed
by Isaac Asimov for this purpose (Asimov All Robot Stories 3rd edition 2011,
short story Herumtreiber (English Runaround), pp. 276–295). Whether such rather
philosophical laws are sufficient cannot be judged today, but it is certain that the
manufacturer and developer of robots have corresponding duties to ensure traffic
safety. However, the maintenance of these traffic safety obligations is then no longer
the responsibility of the manufacturer, but of the keeper or owner of the robot.
The principles for dealing with dangerous goods apply here. A dangerous object is,
for example, a motor vehicle that poses a certain operating risk. The manufacturer
produces a car that meets the relevant requirements for the registration of a motor
154 11 Robotics
vehicle, while the owner must ensure that the vehicle is constantly in a roadworthy
condition (BGH, October 14, 1997 - VI ZR 404/96 = NJW 1998, 311). In particular,
this applies in the case of a guarantor position vis-à-vis third parties (BGH, April
24, 1979 - VI ZR 73/78 = NJW 1979, 2309). The same should also apply to the
manufacture and use of robots.
So, as a manufacturer of robots and as their operator, you have to consult a lawyer
every now and then and discuss any problems that arise.
Employees of the Scientific Center in the Republic of Kabardino-Balkaria have
developed a caterpillar robot in the shape of a turtle shell, reports the news agency
RIA Novosti. This is primarily intended to help the Russian National Guard
Rosgvardiya disperse crowds at rallies. According to the developers, the vehicle
should move faster than pedestrians or mounted police.
The robot is about half the size of a human and armed with stun guns, net
throwers, and fast-setting foam. Its streamlined shape is designed to avoid injury
to humans while preventing protesters from tipping it over.
The scientists propose two options. The first envisions the robot moving at 60
kilometers per hour and cutting through crowds. In the second option, the robot acts
as a shield bearer to erect barriers and divide the rowdy crowd. The invention is
also said to be able to network and act together based on an algorithm modeled on
state-forming insects such as waspseren.
Most accidents with robots occur during maintenance or programming of the
robot, not during controlled operation. On July 21, 1984, the first human was killed
by an industrial robot in Michigan, USA. The robot was moving workpieces on a
die-casting machine. The 34-year-old factory worker already had 15 years of die
casting work experience and had completed a 1-week robot training course only
3 weeks before the accident. He was crushed to death between the supposedly
safe back of the robot and a steel post when, against all warning, he climbed into
the robot’s danger zone to remove scattered production debris. The US “National
Institute for Occupational Safety and Health (NIOSH)” provides guidelines for
robot design, training, and guidance for workers.
It can be seen from this that the manufacture and use of robots has become an
integral part of the engineering world. This is where the benefits of using these
systems and the problems associated with them become apparent.
• In many countries, children, young people, and students have the opportunity
to participate in robotics programs. They form teams, each of which is faced
with the task of programming a robot equipped with motors and sensors so that
it can solve predetermined tasks, such as sorting objects and moving them to
specific locations, on a playing field in a given time frame, either autonomously
or by remote control. In some of the programs, the task includes designing and
building the robot; in others, prefabricated robots are used. The teamwork results
in competitions, many of which are held at an international level.
• Man-made artificial beings have a long tradition in literature, well-known exam-
ples being the Golem created by magic or alchemically generated Homunculi.
11 Robotics 155
• Jack Williamson described in 1939 with “After World’s End” (“Beyond Space
and Time”) a humanoid robot that energetically turns against humanity. Jack
Williamson then began in 1947 with “The Humanoids” or “With Folded Hands”
and in 1952 with “Wing 4” a series of novels about humanoid robots that carry
out their task of serving and protecting humans a little too thoroughly, severely
limiting human freedom.
Philip K. Dick described non-humanoid and humanoid killer robots in his
1953 short story “Second Variety” (Variant Two, filmed as Screamers) and in his
1955 short story “Autofac” (Autofab or War of the Automata), a self-sustained
evolution of robot factories competing for raw materials into nanorobots unde-
sired by humans.
The Polish author Stanislaw Lem published the story cycle “Kyberiade”
starting in the 1950s. The main characters are Klapaucius and Trurl, two
humanlike robotic beings capable of thinking and feeling, who are themselves
constructors of machines and exist in a world populated predominantly by robots.
• Robot characters also found their way into children’s literature. In 1967, the
book “Robbi, Tobbi und das Fliewatüüt” was published, in which the robot ROB
344–66/IIIa and the inventor Tobias Findteisen experience adventures together.
Ellis Kaut published the children’s book “Schlupp vom grünen Stern” in
1974 about a little robot from the planet Balda 7/3, which unwantedly develops
a soul and feelings and is therefore supposed to be shot to a garbage planet,
but ends up on Earth instead. The Augsburger Puppenkiste filmed Schlupp as a
puppet play in 1986.
• In the novel “Machines Like Me” by British author Ian McEwan, published by
Diogenes Verlag in May 2019, a pair of young lovers enter into a tangled love
triangle with an android named Adam. In the novel “The Night Was Pale, the
Lights Flashed” by author Emma Braslavsky, published by Suhrkamp in August
2019, a robotics company sells humanlike robots as artificial life partners to
combat civilizational problems such as social loneliness and lack of relationships.
In the novel “Klara and the Sun,” a terminally ill girl makes friends with
a female android named Klara. The book, published by Blessing Verlag in
March 2021, was written by the author and Nobel Prize winner for literature
Kazuo Ishiguro. The novel is written from the first-person perspective of
the robot, as the android Klara sees the world of humans. The AI novel “The
Invention of Disobedience” (Unionsverlag) by Martina Clavadetscher,
which won the 2021 Swiss Book Prize, features a woman who works in a
sex doll factory in China where artificial female bodies are produced. For her
novel, the writer was inspired by the life’s work of the mathematician Ada
Lovelace, whom she sees as the foremother of programming and the computer.
The novel “Dave” by the writer Raphaela Edelbauer, which was awarded
the Austrian Book Prize 2021, is about a computer that scientists want to give a
human consciousness.
There is still a lot of scope here for the further creation of fantastic figures
(Fig. 11.3).
11 Robotics 157
• The 1897 short film “Gugusse et l’Automate” by film pioneer Georges Méliès
was probably the first to show a robot on the cinema screen. The film, now
lost, was about a confrontation between the clown Gugusse and an “automaton.”
In 1911, the 10-minute short film “The Automatic Motorist” by Walter R.
Booth was released, featuring a robot chauffeur. In the episodic film “The
Master Mystery” released in 1919 with circus performer Harry Houdini,
there is another early example of the appearance of a robot. The 1921 Italian
film “L’uomo meccanico” by André Deed is about crooks who take control
of a remote-controlled robot in order to commit crimes with its help. Among
the most famous robot portrayals of the silent film era is the female machine
man in Fritz Lang’s “Metropolis” from 1927. The first robot to appear on
television was I Tobor in the science fiction series “Captain Video and His
Video Rangers” (from 1949).
• Well-known robot characters of the 1950s are the giant guardian Gort from the
film “The Day the Earth Stood Still” (1951) and the robot Robby from “Alarm
in Space” (1956). Robby subsequently appeared in a number of movies and
television shows and also served as the model for many toy robots. The Star Wars
saga (1977–2005) features robots in a comedic role with R2D2 and C3PO. The
two characters are at the same time characteristic examples of different types
of robots: while the “protocol droid” C3PO is a humanoid robot, R2D2 is an
“astromech droid” without humanlike appearance. In the series “2Star Trek: The
Next Generation” (1987–1994), the android Data is a command officer who often
grapples with the question of his own human-likeness. In “Number 5 Lives!”,
a military robot develops a naive personality. Films such as “Terminator” and
“I, Robot” offer other well-known examples of the portrayal of androids and
humanoid robots that are hostile to humans or assist them.
• Industrial robots (non-humanoid robots) are also finding their way onto the big
screen. KUKA industrial robots have been featured in films such as “James Bond
007: Die Another Day,” “Thunderbirds,” “Tomb Raider – The Cradle of Life,”
and “Sacrilege.” The documentary film “Plug & Pray” addresses the effects of
the increasing use of robots on the self-image of humans. It presents robots such
158 11 Robotics
Maschine. For their stage show, the musicians had special remote-controlled
robots developed with the faces of the band members replicated in plastic. In
1983, the song “Mr. Roboto” by the band Styx reached number 8 in Germany
and number 3 in the US charts. It is about a prison guarded by robots.
Music-making and dancing robots, under the direction of a mad scientist, can
be seen in the video clip for the 1984 track “Zoolook” by French musician
Jean-Michel Jarre. In Herbie Hancock’s video clip for his 1983
electronic instrumental Rockit, robotic legs dance without trunks, mechanical
dolls gather around a breakfast table, and a humanoid being twitches restlessly
under a bedspread. It was directed by the duo Godley & Creme, and the
puppets and robots were designed by artist Jim Whiting, who enjoyed great
success with his machine theater Unnatural Bodies beginning in 1988. The
instrumental Rockit is on Hancock’s album Future Shock.
• In October 2018, the writer Thomas Melle staged the play “Unheimliches
Tal/Uncanny Valley” at the Münchner Kammerspiele, in collaboration with
Stefan Kaegi from the theater group Rimini Protokoll, in which the author
converses about existential questions with a robot sitting on stage, a mechanical
replica of himself.
• Robots in sports As part of the 2020 Summer Olympics, automaker Toyota
unveiled Cue3, a basketball robot that has a 100 percent scoring rate.
You certainly can’t read, watch, or listen to it all here. But it shows that there are
no limits to imagination and that robotics technology is evolving in all directions.
This is certainly one of the most interesting and important fields of application of
robotics, which is developing extraordinarily fast and can satisfy more and more
requirements.
An essential requirement for the introduction of new technologies is always
that the costs for the acquisition of such devices are covered and that appropriate
training opportunities are provided. The possibility of manufacturing such devices
should be available, as well as the IT specialists for programming and maintaining
these systems. It should also be pointed out that there are already courses of study:
“medical informatics” or “digital medicine.”
For example, the Faculty of Informatics at the University of Augsburg announces
the following program:
“The interprofessional degree program Medical Informatics (Medical
Information Sciences) builds on the bachelor’s degree program of the same name
and is located at the interface between informatics, medicine, and the application of
these two disciplines in the field of medical informatics.”
Medical informatics uses modern information technologies to provide organi-
zational and technical support for the healthcare system. To this end, it develops
160 11 Robotics
models and simulations for individualized therapy and diagnosis. Therefore, the
study program provides the necessary skills to develop, select, and apply suitable
methods, systems, and information technology tools for concrete medical issues.
At most universities, the course of study has a strong practical orientation in order
to optimally prepare students for their later work [1]. Important study contents of
medical informatics are:
• Medical documentation
• Imaging techniques such as computed tomography
• Scientific modeling
• Computer-assisted detection
• E-health
• Biosignal processing
• Biometrics
Depending on the university, the study of medical informatics also special-
izes for certain fields, for example, dental technology. The aim of dental
informatics is to develop IT applications to actively support the dentist during
treatment. Dental informatics is the application of computer and information sci-
ences to improve dental practice, research, education, and administration. Numerous
applications have been developed to support clinical care, education, and research.
Dental informatics is beginning to show the characteristics of a discipline: a core
literature exists, and trained specialists and educational programs are already in
place.
• Structure and content of the bachelor’s degree in medical informatics: The spe-
cialty enables graduates to independently develop IT—solutions in the medical
field—to realize imaging in connection with diagnostic procedures or to organize
the management of patients’ data. For this, the basics of computer science
are taught. These include programming, mathematics and statistics, algorithms
and data structures, or the development of databases. Planning skills for the
development of customized programs, software engineering, are also taught. The
basic education also includes fundamental knowledge from medicine, such as its
most important sub-areas, but also diagnostic and therapeutic systems.
• The bachelor study analytical and digital forensics: Tragic fate or crime? Natural
death or murder?—As early as the nineteenth century, Carl Remigius
Fresenius solved a sensational murder case that didn’t look like one at all
at first, by determining poisoning with arsenic. The founder of the Fresenius
University of Applied Sciences felt obliged to “stand up for law, truth, and
science”—and his work had a decisive influence on forensic analysis.
Today, the questions have become more diverse and analytics is also much
more diverse than it was back then: The modern analyst is not only familiar
with scientific basics, but is especially at home in the digital world. Whether
it’s designer drugs or cybercrime, environmental pollution or hacker attacks,
or genetic analysis or encryption technologies, modern analytics encompasses
11.2 Robots Help in Care 161
all steps from the right question to data collection, processing, and analysis to
interpreting the results in light of various issues.
Analytical thinking and action, embedded in a scientific approach, are still
the basis for fighting crime today. In the process, the real and digital worlds
are growing ever closer together and require investigators who are at home in
both worlds and can move between them. You can become that investigator:
Study Analytical and Digital Forensics (B.Sc.) at Fresenius University of Applied
Sciences—where forensics has been at home since 1848!
In [2], 28 study programs are displayed.
When you talk to interested people about this problem, you often meet with
skepticism or rejection. But this often disappears when you explain that a doctor
is always present. And of course, it is convincing if you then simply explain that the
robot does not get tired, that it can work 24 hours without a break, and that it does
not tremble, does not overlook anything, and much more.
Science fiction movies often feature humanoid robots that help elderly people in
their daily lives. They move freely around the home, do the dishes, cook, clean,
and assist those in need of care with personal hygiene and can even carry on
conversations, much like human caregivers.
In these fictional scenarios, the impression is usually given that technological
progress will result in the loss of humanity and that at some point everything will
be done by robots. In fact, the exact opposite is true: digitization is bringing more
humanity back into nursing.
Apart from the repeatedly criticized poor pay for skilled workers, the biggest
problem in care today is the time factor. For older people in particular, interpersonal
relationships are enormously important. It’s one thing to be helped to take a shower
in the morning. But to be able to talk to the caregivers afterward about one’s
own worries and problems (or simply about the weather) is something completely
different. And it is precisely for this interpersonal communication that time is
usually lacking.
This problem can be significantly reduced by digitizing care. It starts with
simple things like electronic patient records, automated nursing documentation, or
networked route planning. It is precisely such administrative tasks that take up a lot
of time in practice. For people who need care but still live in their own homes, there
is also a lot of potential in telecare. After all, a 15-minute face time call is better
than no contact at all.
However, the idea is not for robots to replace human caregivers, but rather to
support them. This starts with simple robots that can clean the floor, for example.
The makers of the therapy seal PARO, which was specially developed for dementia
patients, are taking a completely different approach. The care robot Pepper, on
162 11 Robotics
the other hand, supports professionals in care facilities during day care. It can
tell jokes, read fairy tales, play music, demonstrate simple fitness exercises, etc.
Meanwhile, there are even solutions for private homes, such as Medisana’s Home
Care Robot, which enables video calls, regularly reminds people to check their
health, and can quickly summon help in an emergency.
“Movement disorders such as tremors, paralysis, or muscle tension disorders
affect many patients suffering from strokes or neurodegenerative diseases such
as Parkinson’s or multiple sclerosis. Recording the mobility restrictions precisely
and reproducibly is a major challenge in diagnostics and therapy monitoring,
as this requires experienced physicians who are not always available,” explains
Mathias Bähr, director of the Neurological Clinic at the University Medical Center
Göttingen.
This is where the “Deep Movement Diagnostics” project comes in. We will
combine our expertise in the fields of body and eye movements in humans and
monkeys, neurophysiology, and clinical neurology, as well as prosthetics and
robotics, says Alexander Gail.
Using the latest digital methods, walking and grasping movements are to be
measured and modeled with previously unfeasible precision in order to use them as
the basis for diagnostic tools for individualized therapy approaches, for example, in
Parkinson’s or stroke patients. In this context, motor function studies in monkeys
play an important role; they are the basis for later application in humans. “Our
goal is to develop a low-cost, easy-to-use system that can be used across the board
for diagnosis and therapy monitoring in movement disorders,” says project leader
Alexander Gail.
“Ease of use and low effort to perform the examinations are important factors
to increase patient acceptance,” said Melanie Wilke, director of the Institute of
Cognitive Neurology. “We expect the new video-based methods to provide a
qualitative leap over current clinical examination techniques.”
In addition to pure diagnostics, the research team plans to study complex
movement patterns in healthy subjects and in monkeys to better understand the
neurophysiological basis of movement disorders.
especially atrial fibrillation, are significant here because they increase the risk of
stroke.
Several projects show that AI is not always about whether the machine is better
than the human, but that the amount of work involved in diagnosing by doctors can
be reduced by almost 20 percent. For example, in August 2016 at the University of
Tokyo Medical Institute, the IBM Watson computer program was able to correct
a misdiagnosis made by doctors. The doctors diagnosed the patient with acute
myeloid leukemia. The therapy was unsuccessful, so they asked Watson for help. It
took the AI 10 minutes to match the woman’s DNA with 20 million cancer studies.
Watson identified a very rare form of leukemia that has only affected 41 patients and
is curable. However, IBM Watson’s treatment recommendations can also be flawed,
such as when too little training data is available. Corresponding reports of faulty
recommendations, the application of which put patients at risk, were published by a
medical portal in 2018. The malfunction is said to have been fixed in a later version,
according to IBM. In January 2018, scientists at Stanford University presented a
program that can calculate with a probability of 90 percent whether terminally
ill patients will die within the next 3 to 12 months from their disease data. This
could help terminally ill patients live out their final months with dignity, without
aggressive treatment methods, and possibly at home under palliative care, it said.
The wristwatch Apple Watch records a person’s heart rate, among other
things. Apple announced that programs have an 85 percent probability of detecting
diabetes mellitus in the wearer of the wristwatch from the analysis of heart
rate. The idea is based on the Framingham Heart Study, which recognized back in
2015 that diabetes can be diagnosed using heart rate alone. Apple had previously
succeeded in detecting abnormal heart rhythm from heart rate with 97 percent
probability, sleep apnea with 90 percent, and hypertension (high blood pressure)
with 82 percent.
Researchers at Mount Sinai School of Medicine demonstrated in
January 2018 how psychological interview transcripts with adolescents can tell
if they will develop psychosis in the next 2 years. Natural language processing
helped achieve up to 83 percent accuracy on standardized tests, such as those based
on disorganized thought processes, awkward phrasing, unclear associations, or
reduced language complexity. The subtle differences were apparent after training
with many such conversations, he said.
MIT researchers unveiled a program in September 2018 that can diagnose
depression in patients based on spoken text or written text. By themselves, doctors
and psychologists ask patients questions about lifestyle habits, behaviors, and sen-
sitivities to diagnose depression from answers. After training with such interviews,
the program also detected depression from everyday conversations with a hit rate of
83 percent—and in classifying the severity of depression on a scale of 0 to 27 with
a hit rate of 71 percent. The AI could assist doctors or permanently monitor them as
an app user to alert in an emergency. The researchers also want to detect dementia
from speech in the future.
According to the manufacturer, the health app Babylon Health is supposed
to be able to use a voice system (chatbot) based on AI to make a diagnosis in
11.3 Other Interesting Applications 165
conversation with patients that is about ten times more accurate than diagnoses made
by a GP. The development of the app was also co-funded by the British healthcare
system. The aim was to use it to reduce costs. Although the app is supposed to
significantly reduce doctor visits, patients quickly found out how to use the app to
get doctor appointments faster by describing symptoms incorrectly.
The app Ada from the Berlin-based company Ada Health uses a program
to assist with diagnoses based on symptom descriptions. According to the manu-
facturer, this should correspond to the quality of well-trained Western doctors. The
app Ada sent unauthorized marketing companies such as Amplitude and Adjust,
headquartered in San Francisco (USA), and in the course of app use regularly
Facebook.com personal data, even if you do not have a Facebook account.
The app has won awards from MIT and is funded by the Bill & Melinda
Gates Foundation. In 2019, Ada Health announced a collaboration with
Sutter Health. Especially in developing countries where there is a shortage
of medical personnel, the app can help build a health system.
Devices or software that use artificial intelligence must be CE marked as medical
devices in Europe and approved by the FDA in the USA. A comparison between
the USA and Europe from 2015 to 2020 yielded, a rapidly increasing number of
approvals, with CE marking slightly dominating in number (2019: US, 79; EU, 100).
Often, CE marking occurs before FDA approval, which may indicate a less rigorous
process. The emphasis is on radiology. Only 15 % of the products are aimed directly
at individuals (patients), the rest at professionals (physicians). During this period,
only 1 % of approvals are for the highest risk classes, e.g., those for the diagnosis of
breast cancer [46].
Even more so in Europe than in America, the authors of the study find a lack of
transparency in the device description and the process of evaluation. It reflects the
ethical responsibility of regulators as much as manufacturers. A publicly accessible
database of CE-marked devices and software is also urged.
Devices or software that use artificial intelligence must be CE marked as medical
devices in Europe and approved by the FDA in the US. A comparison between
the U.S. and Europe from 2015 to 2020 yielded, a rapidly increasing number of
approvals, with CE marking slightly dominating in number (2019 U.S. 79, EU 100).
Often, CE marking occurs before FDA approval, which may indicate a less rigorous
process. The emphasis is on radiology. Only 15 % of the products are aimed directly
at individuals (patients), the rest at professionals (physicians). During this period,
only 1 % of approvals are for the highest risk classes, e.g. those for the diagnosis of
breast cancer [46].
Even more so in Europe than in America, the authors of the study find a lack of
transparency in the device description and the process of evaluation. It reflects the
ethical responsibility of regulators as much as manufacturers. A publicly accessible
database of CE-marked devices and software is also urged.
The user wants to understand the reasons for an algorithmic decision. In artificial
intelligence and machine learning, algorithms are often completely opaque (“black
box”), typically in neural networks and the corresponding learning methods. To
counter this, the field of Explainable Artificial Intelligence is
166 11 Robotics
References
1. https://www.studycheck.de/studium/Medizinische-informatik
2. https://www.studieren-studium.com/studium/Medizinische-Informatik
Chapter 12
Artificial Intelligence in Finance
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024 169
C. Posthoff, Artificial Intelligence for Everyone,
https://doi.org/10.1007/978-3-031-57208-1_12
170 12 Artificial Intelligence in Finance
are many AI pilot projects at financial companies—but only very few of them
succeed in transferring these ideas into daily, operational business. Even companies
that already have AI expertise themselves often do not know how to approach the
topic in a meaningful way [2].
Currently, insurers and banks in the DACH region (Germany, Austria, Switzer-
land) are looking at the possible use of AI solutions primarily from a conventional
business perspective: 79% of those surveyed want to make business processes more
efficient digitally, almost three quarters generally want to save costs (73%), and
every second company expects to use AI to ensure compliance with all existing
regulations (50%). But also for new fields, such as chatbots, automation, and
predictive marketing, just over half of the respondents (55%) are already using the
new technology. However, many other opportunities remain untapped: for example,
the complexity of risk assessments and decision support in controlling can be greatly
reduced through automation if the underlying data is really analyzed intelligently.
In order to catch up with the European average, many financial companies are
currently examining which new projects are particularly suitable for the use of AI.
But even once these have been identified and defined, the road to implementation
is often still long: 69% of the companies surveyed identify the lack of available
data as an obstacle to adaptation. A good two-thirds of the companies surveyed are
also struggling with budget restrictions and insufficient funding for corresponding
projects (67%), and 64% of the companies simply lack employees with expertise to
answer questions about establishing AI:
• Which business unit provides an appropriate entry point for establishing AI
projects in operations?
• Which division provides funding for the integration process?
• Are AI-related projects to be regarded as a subdivision of IT or, as a strategically
relevant topic, dependent on independent management structures? [2]
Data is more than abundant in some subfields for almost 100 years. The problem
here is to get it all into one system, in a reasonable amount of time, and with a
reasonable amount of effort. Whether a body can be created to work across different
banks on these problems is not clear at this time. Normally, the various banks
are in a competitive situation. However, cooperation would clarify the problem of
resources. Independent management structures are recommended in any case, since
the IT departments are already busy as it is with all the problems of online banking.
However, if a large bank succeeds in overcoming these problems, it would have a
major advantage over its competitors.
The study also shows that using AI for day-to-day business and in established
processes, for example, with regard to personalization or new business models,
has so far been a comparatively low priority for financial service providers. In
addition, the often still low general comprehensibility in the financial sector is a
hurdle that should not be underestimated. Classic mathematical applications can
still be mapped with a comparatively simple algorithm, but closed models such as
deep neural networks are significantly more demanding and thus more difficult to
penetrate. This results in another problem: the financial services sector is highly
12 Artificial Intelligence in Finance 171
regulated. Companies are obliged to explain their processes and decisions in detail
to supervisory authorities and internal auditors. Here, AI is still considered a black-
box technology, so many companies act with restraint here, and not just in banking.
Another big area that could be revolutionized by AI is risk assessment [3]. This
is hugely important in many areas of finance. AI systems can draw on a wide range
of information to calculate the risk of default on loans. Algorithms can learn from
cases of credit decisions and customers’ repayment behaviors, identify patterns from
them, and create profiles. When a new credit request is made, the system compares
the customer’s data, checks for patterns, and ranks his or her creditworthiness
based on the analysis. The check can be used to make lending decisions. Machine
verification not only provides insights and greater certainty. It can also be used as
an argument to reject a loan. The final decision remains with the lender. However,
the lender is given the opportunity to make a quick and data-based decision.
Artificial intelligence can also be a powerful tool in market research. News from
all over the world, media activities, and studies can be evaluated to make forecasts
and identify investment trends at an early stage. With the help of such analysis
tools, geopolitical events can be taken into account and the stability of markets
can be estimated. Valuations for securities can be created from the analyses and
made available to employees for their decisions. Powerful systems can make this
assessment in real time, which holds great potential for high-frequency trading on
the stock market.
Know Your Customer programs can use artificial intelligence to screen
new customers. The algorithms detect suspicious patterns and activity and can alert
to money laundering and white-collar crime. Pre-screening customers is a critical
step for many companies and institutions in the financial sector. Therefore, a lot of
time and energy is invested in research and verification. Another example is claims
settlement in the insurance business. AI systems can be trained with data from a
large number of insurance cases, especially fraud cases. Conspicuous cases can be
flagged and presented to employees for further review.
On June 15, 2021, the German Federal Financial Supervisory Authority pub-
lished supervisory principles for the use of algorithms in decision-making processes
of financial companies. They are intended to lead to responsible use of Big Data
and Artificial Intelligence (BDAI) and to enable the associated risk to be controlled.
Background to the principles paper: technologies such as BDAI are increasingly
also being used by companies in the financial market. In 2018, BaFin had already
pointed out in its study “Big Data meets Artificial Intelligence” that this would
result in opportunities for companies, but also for consumers, but that it was also
important to manage the risks that BDAI applications would bring [4]. In this regard,
a fundamental problem underlies all of the regulatory issues surrounding BDAIs:
it is still difficult to distinguish BDAI procedures from procedures of classical
statistics. From a risk perspective, however, three characteristics can be identified
that are of particular importance in modern BDAI methods:
• First, the algorithms used are often particularly complex compared to classical
statistical methods. This complicates or prevents their traceability.
172 12 Artificial Intelligence in Finance
• Second, shorter and shorter recalibration cycles can be observed. This is due
to the combination of constantly learning algorithms and the fact that new data
is available almost daily. As a result, the boundaries between calibration and
validation are becoming increasingly blurred. Third, the use of BDAI methods
increases the degree of automation. This makes it increasingly easy to scale up
processes, and the impact of the individual algorithm increases.
In order to be able to formulate the principles as precisely as possible, the decision-
making process based on algorithms has been divided into two phases in a highly
simplified manner: the development phase and the application phase.
The development phase is about how the algorithm is selected, calibrated, and
validated. For this, there are principles for data strategy and for documenting
internal and external traceability. In the application phase, the results of the
algorithm must be interpreted and integrated into decision-making processes. This
can be done automatically, but also by involving experts on an ongoing basis. In
any case, functioning rules of procedure must be established, including sufficient
control mechanisms and appropriate feedback to the development phase. These two
phases are flanked by overarching principles, for example, on the need for a clear
responsibility structure and adequate risk and outsourcing management (Fig. 12.1).
Here you can see the transition to the theory of fractal systems. The DAX curves
are self-similar. The curves always look the same, no matter whether the values are
plotted daily, weekly, monthly, or annually. This transfer of the theory of fractals
to finance goes back to Benoît Mandelbrot (Fig. 12.2). In a founding series for
the renowned journal Quantitative Finance, he summarized his view of markets.
Here he described a fractal recursive construction process for price time series that
recursively decomposes trends into smaller trends (Fig. 12.1).
Fig. 12.1 This information regarding the DAX is offered in real time
12 Artificial Intelligence in Finance 173
Answering the question of what a trend is does not seem difficult. Mathemati-
cally, however, it is not so simple. If you want to define trends, you have to specify
two things: the exact measurement methodology and the scale used to measure.
The moving average (also moving average) is a method of smoothing time or
data series. The smoothing is done by removing higher-frequency components. As
a result, a new data point set is created consisting of the averages of equally sized
subsets of the original data point set:
1
n−1
m(t) =
. x(t − i).
n
i=0
For many values from everyday stock market life, the values .n = 38 are chosen
for 1 month and .n = 200 for half a year. Strong swings of the values up and down
can compensate each other, if they occur equally often and have approximately the
same size. The structured monthly investment in stocks that have historically risen
in price hereby appears clear. Obviously, this effect exploits overlong trends in real
data [5].
Analogously, stock prices may also be assumed to be fractal. In 1968, Mandel-
brot, together with van Ness, generalizes the model of a random walk, which is
fundamental for financial mathematics, in which he introduces a further parameter,
the so-called Hurst exponent, in addition to the growth of returns (drift) and their
fluctuation (volatility). This parameter can be used to describe the scaling behavior
of prices. Here, values <0.5 can be interpreted as mean reversion, while values
>0.5 can be understood as trending. Only in the limiting case, where the Hurst
exponent is equal to 0.5, one may speak of efficient markets. In this case,
the time series would be statistically trend-free, i.e., it would have no relation to the
past.
174 12 Artificial Intelligence in Finance
Toward the end of his career, Mandelbrot outlines a construction principle for
trends that follows the construction principle of fractals. Starting from a main trend,
he applies a subdivision rule that divides this main trend into three smaller trends. In
a next step, these smaller segments are again further decomposed into even smaller
trends. The whole leads to a fractal in which trends can be described by smaller
trends (Fig. 12.3).
In an article [6], it is shown that well-known factor strategies, such as momentum
(strongly rising stocks), low volatility (stocks that fluctuate weakly), and
value (stocks that are considered favorable due to economic key figures), can
be described and replaced by systematic trend-based strategies. For this purpose,
only the last visible trend is evaluated for all stocks and then sorted by slope. The
differences in the strategies can be described by different wavelet scales, which are
used to calculate the last visible trends.
This view of investment styles (factor investments) is also almost complete,
i.e., one is able to describe market returns with high statistical quality by these
investment styles. These styles are additionally dominant in different market
regimes, i.e., they generate excess returns in different phases. For example, investing
in defensive stocks is attractive when the market is in a down phase, while value
often generates very high returns when markets are considered cheap and market
participants expect prices to recover. With momentum, on the other hand, corporate
earnings revisions play a dominant (but not exclusive) role.
• The rise of artificial intelligence requires new thinking and new content for
finance education. AI experts are already a shortage; firms in finance are com-
12 Artificial Intelligence in Finance 175
peting with technology-based firms, startups, and other fields. The application of
efficient algorithms leads to profits that are higher than the general market rate:
• Algorithmic credit scoring can support or fully automate credit decisions.
• Algorithms can detect unusual patterns, helping to identify fraud or even prevent
it.
• Algorithms can be trained to optimally execute derivative trading.
• Algorithms can be used to design and manage funds intended for retirement.
The algorithms work better and more effectively than humans due to the
following factors:
• It is possible to cope with large amounts of data.
• The instability of financial markets does not follow constant laws.
• Many relations that occur are nonlinear. Tiny changes in the initial data can lead
to significant changes in the market.
• The high dimensionality of many problems can be handled by programs.
• Traditional econometrics dealt primarily with regression problems. Classification
problems, however, can be much better handled by learning procedures.
• Artificial intelligence learning techniques can handle structured and unstructured
data simultaneously.
A major advancement in hardware has been the development of graphics
processing units (GPU). GPUs handle the calculations for 2D and 3D graphics. This
relieves the CPU (central processing unit). This takes care of all computing tasks that
have nothing to do with graphics. Compared to the GPU and CPU, a GPU is usually
built on a lot of computing cores to be able to handle many tasks at the same time.
The calculation of single pixels is not very complex in comparison, but millions of
pixels have to be calculated at the same time. The GPU is either installed directly
on the CPU, on the motherboard, or as a plug-in card for the computer. Several
GPUs can also be installed on a plug-in card—for example, the PowerColor Radeon
R9. Today, three major manufacturers share the market for the production of GPUs:
Intel, AMD, and Nvidia.
For the future, three possibilities are mentioned.
• The monopoly: a financial institution achieves a dominant position because it
is the first to be able to apply high-quality AI methods first.
• The oligopoly: a market form characterized by few market participants.
People are currently working hard on the following problems:
• Privacy: the large-scale application of AI methods requires the use of private
data. This brings the risk of data being stolen or used improperly.
• Bias: the algorithms may take on these characteristics if it is already present in
the data.
• Explainability of the results is a major problem. In some cases, it is required
by investors; in others it is required by law. Although this area is being studied
intensively, no progress is currently seen.
176 12 Artificial Intelligence in Finance
• Habituation effect: this effect can occur when many do the same thing. In this
regard, one can study the materials on the 2008 financial crisis. A similar situation
can occur when many companies use the same or similar programs.
• Jurisprudence and financial policy must keep up with these developments and not
lag behind them.
References
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024 177
C. Posthoff, Artificial Intelligence for Everyone,
https://doi.org/10.1007/978-3-031-57208-1_13
178 13 Artificial Intelligence and Education
they have acquired and apply it in their own day-to-day business. Upon successful
completion of the examination, participants receive a Bitkom personal certificate.
The part-time certificate course “Training as AI Manager” consists of five mod-
ules in which participants are taught the fundamentals of artificial intelligence and
hybrid value creation through heuristics, knowledge-based systems, and machine
learning. Application examples on artificial intelligence and practical experiences
from research and development as well as from companies and startups are an
integral part of the training. The participants test AI services on the Internet, reflect
on them against the background of their own business fields, and are enabled to
design and critically reflect on use cases for generating AI-based hybrid value
creation for their own company.
The five modules of the AI Manager training are the following:
• What is Artificial Intelligence: history, examples, society, ethics (2 days).
• Basics and prerequisites of AI: methods, tools, hands-on, machine learning, and
data mining (2 days)
• Hybrid value creation through AI: fundamentals and methods. Applications
Industry 4.0 and Smart Services (1 day)
• Artificial Intelligence in Research and (Business) Application. Cognitive AR,
Retail Solutions, Watson, Leonardo (2 days)
• Development of business areas, change of corporate culture, and design of
transformation processes (1 day)
The Chamber of Industry and Commerce (IHK) has a range of offerings for
technical training and continuing education available through the IHK Academy
in Munich and Upper Bavaria, as well as through the various regional locations—
including AI and topics related to data handling. The Stuttgart Region IHK has
already invited participants to a number of work camps and other events on AI in
2019.
The Rhine-Neckar Chamber of Commerce and Industry has addressed AI and
analytical platforms as well as “From Big Data to Smart Data.” The Artificial
Intelligence Working Group (AKKI) of the Lübeck Chamber of Commerce and
Industry has held several lecture series on the topic together with the local university
and MEVIS. A look at the respective further education and training measures of the
local IHK can therefore also be worthwhile for smaller companies.
If you have a higher-level qualification for the relevant area of the company—
a degree in business administration, for example—a full-fledged computer science
degree with a specialization in AI is hardly worthwhile, but it might be for one
or two junior staff members in the IT department. WirtschaftsWoche has compiled
a list of German universities offering such courses. Berlin’s Beuth University of
Applied Sciences even advertises its new Humanoid Robotics course on YouTube.
The Technical University of Munich (TUM) recently introduced the interdisci-
plinary master’s program “Robotics, Cognition, Intelligence.” Bergische Universität
Wuppertal has launched an “Interdisciplinary Center Machine Learning and Data
Analytics,” or IZMD for short. Bachelor’s degree programs with an AI focus are
available in Stuttgart, Marburg, Göttingen, Karlsruhe, and Deggendorf and at the
13.2 The Application of Artificial Intelligence in Education 179
Beuth University of Applied Sciences in Berlin. Master’s degree programs are avail-
able at LMU Munich, Darmstadt University of Applied Sciences; the universities of
Marburg, Potsdam, and Bielefeld; Chemnitz University of Technology; and Stuttgart
Media University.
Seminars and support in the area of AI and data science are also offered by
consulting firms such as All for One Group, headquartered in Filderstadt
near Stuttgart. The leading IT and consulting group in the DACH region includes
a number of subsidiaries and branches, including B4B Solutions and Allfoye Man-
agementberatung GmbH, which support corporate customers in terms of change
management in order to master the challenges of digital transformation.
This area is hotly debated and extensive. Only a few examples will be given that
relate directly to education in a particular subject.
3D printing of body parts is ushering in a new era in anatomy education.
Especially in countries where exhumation of cadavers is a problem, the 3D printer
can fundamentally change medical education. The first product for this application is
likely to be Monash University’s 3D Printed Anatomy Series’ [1]. The series of body
parts created with a 3D printer works without body parts from deceased people, but
still provides all the important parts of a body needed for scientific teachings in
anatomy (tendons, ligaments, muscles, bones, etc.).
According to Professor Paul McMenamin, director of the Centre for Human
Anatomy Education at Monash University, the use of the low-cost anatomy kit
will improve the knowledge of future doctors and other medical professionals. He
says the technology could even lead to the development of new surgical treatment
procedures.
The anatomy kit is likely to generate increased interest, particularly in developing
countries. There, the handling of cadavers is often problematic for religious or
cultural reasons. The 3D-printed body models in the anatomy kit are quick and
relatively inexpensive to produce. In addition, they are not subject to constant
change. This means that with the acquisition of a 3D printer and the digital CAD
files, new models can be produced at any time worldwide (Fig. 13.1).
In a first step, real body parts are scanned by computer tomograph or a surface
laser scanner. Then the body parts are printed out of a plaster-like powder or plastic
with lifelike colors in higher resolution.
The combination of artificial intelligence and 3D printing can also help expand
the range of compatible materials to meet the needs of industries such as aerospace,
which mostly require high-temperature materials. The “futureAM” project launched
by the Fraunhofer Institute IWS in 2017 is a perfect example of this. Professor
Dr. Ing. Frank Brückner, head of the business unit “Generate and Print” at the
Fraunhofer IWS and AMCD (Additive Manufacturing Center Dresden) explains:
180 13 Artificial Intelligence and Education
“Aircraft engines could operate much more efficiently at higher temperatures if most
materials did not fail at temperatures above 1,200 degrees.”
So where does AI come into play here? Prof. Dr. Ing. Frank Brückner explains
this to us as follows: “The processing of new high-performance materials is very
complex and requires fine-tuning of all process parameters. That is why we monitor
the 3D printing process with a large number of different sensors. Using AI, we then
evaluate this flood of data and identify hidden correlations that are not recognizable
to humans.” This is precisely the advantage of artificial intelligence: it can process
many times more data than humans—and all of it much faster. Thanks to this
work, researchers can process complex alloys and obtain the exact properties of
the materials.
AI can also help improve the 3D printing process. For example, the printability
of an object can be analyzed before it is printed to ensure that it is suitable for 3D
printing. But also quality prediction as well as quality control of the printing process
can avoid errors in the final product and thus lead to a better result (Fig. 13.2).
The startup Printsyst is trying to implement this with the help of their
AI engine—in this case specifically in the aerospace industry. The goal of the
company’s patented AI algorithm is to identify part functionality and improve
References 181
printing success rates. In doing so, Printsyst’s integrated solution basically learns
from experience gained from previous projects in which objects were additively
manufactured. To meet the very stringent industry specifications that apply in
aerospace, it suggests specific printing parameters that have been proven in the
past on previously printed parts and are highly likely to deliver a successful
first print in the future. Also based on accumulated experience data, Printsyst’s
solution can accurately estimate the cost of required components, increasing not
only productivity but also delivery readiness, at the lowest possible cost.
Eitan Yona, co-founder of Printsyst explains with an example: “We reduce the
average 3D print preparation process from 30 minutes to 5 seconds for each job.
By reducing this time, we increase printer utilization, and by eliminating errors, we
reduce iterations.” So we see that due to the high complexity that a manual process
can bring, the use of AI makes perfect sense. For example, all parameters can be set
and optimized by the algorithm in a matter of seconds—significantly better results
can thus be achieved with the help of AI compared to manual processing.
Another field of application in the area of quality optimization is the printability
analysis (part evaluation) already mentioned above, which takes place before
printing. The company AMFG, mentioned above, has implemented a feature in their
comprehensive software package that ensures that a 3D-printed part is ultimately
really suitable for 3D printing. This is especially important for companies that
work efficiently with additive manufacturing. Here, too, previously obtained data
is processed and thus the printability of a part can be predicted. For example, if the
part has many weak points in terms of stability, shape, or even resilience, preventing
printing can save enormous costs.
There are many smaller systems for training in various subjects (chemistry,
physics, biology, foreign languages). These are essentially used in self-study by
individuals. There is often a lack of general application in schools from grade 1 to
grade 12 or 13. There is an urgent need to work out appropriate didactics and bring
the systems into schools.
References
1. https://www.3d-grenzenlos.de/magazin/zukunft-visionen/3d-druck-anatomie-ausbildung-
chance-dritte-welt-2755963
2. https://www.bitkom-akademie.de/lehrgang/ausbildung-zum-ki-manager
Chapter 14
Artificial Intelligence in the Arts
The application of intelligent software to painting and drawing has already been dis-
cussed in detail in the chapter on image processing. Therefore, only the applications
of artificial intelligence methods in music and architecture will be discussed in this
chapter.
Music is the most vivid and sensual of all human arts. So it’s no wonder that
computer science tried very early on to use machines to create sounds or teach
artificial intelligences to compose melodies. The use of computers and algorithms
to produce music has a long history.
As artificial intelligence and machine learning continue to evolve, so do the
capabilities of machine-generated music. Computers can derive rules and guidelines
from real, human music examples.
The Illiac Suite, later renamed String Quartet No. 4, is a 1957 composition that is
considered the first score composed by an electronic computer. It was programmed
by researcher Lejaren Hiller in collaboration with Leonard Isaacson,
both professors at the University of Illinois. They used the ILLIAC I (Illinois
Automatic Computer) computer for this purpose.
The Illiac Suite consists of four movements corresponding to four experiments:
the first is about the generation of Cantus Firmi; the second generates four-part
segments with different rules; the third deals with rhythm, dynamics, and playing
instructions; and the fourth with different models and probabilities for generative
grammars or Markov chains. On paper, the Illiac Suite was a real masterpiece, but
in reality the piece sounds very tortured and not quite fully developed.
• In 1965, inventor Ray Kurzweil premiered a piano piece created by a
computer that was able to recognize and analyze patterns in various compositions
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024 183
C. Posthoff, Artificial Intelligence for Everyone,
https://doi.org/10.1007/978-3-031-57208-1_14
184 14 Artificial Intelligence in the Arts
and create new melodies. This computer was first used on Steve Allen’s CBS
game show “I’ve Got a Secret.”
• In 1980, professor and composer David Cope at the University of California,
developed a system called EMI (Experiments in Musical Intelligence). It was
based on the idea of Markov chains, that is, on giving probabilities for the
occurrence of future events. iCope used a program to analyze existing passages
of music and create new pieces from them, which was considered a real
breakthrough. By analyzing various works, EMI was able to generate unique
structured compositions within the framework of different genres. In total, the
system created over a thousand works based on the works of 39 composers with
different musical styles.
• Aiva Technologies is one of the leading companies in the field of music
composition. It was founded in 2016 in Luxembourg and London by Pierre
Barreau, Denis Shtefan, Arnaud Decker, and Vincent Barreau.
Their AI is named “Aiva” (Artificial Intelligence Virtual Artist), which has been
taught to compose classical music. Having already released her first album called
“Genesis” as well as numerous individual tracks, Aiva has officially received
worldwide composer status. She was registered with the French and Luxembourg
Society of Copyright (SACEM), where all her works are registered with a
copyright in her own name.
• Shimon is a marimbaplaying robot from Georgia Tech professor Gil
Weinberg that sings, dances, writes lyrics, and can even compose some
melodies. Weinberg and his students trained Shimon with datasets of 50,000
lyrics from jazz, prog rock, and hip-hop.
• Noah 9000 is a music project by Michael Katzlberger, CEO of the
Viennese creative agency TUNNEL23. 9000 is a reference to the neurotic arti-
ficial intelligence “HAL 9000” from Stanley Kubrick’s 1968 masterpiece,
“2001: A Space Odyssey.”
For this project, the program was trained with piano sonatas by Ludwig van
Beethoven, among others. In order to create the perfect piece, it can sometimes
take many hundreds or thousands of iterations before correct, melodious sounds
are produced for the human ear. After all, probably the most interesting challenge
in producing modern music lies not only in the composition itself but also in the
instrumentation and sound design. Katzlberger, as curator, selected ten program-
generated generated tracks for the album Spring 9000 from around 2000 AI
compositions. It is an attempt to use artificial intelligence to create music that
feels human.
Anyone can play the steelpan via Virtual Piano. You don’t need to be able to read
music, and you don’t need to have played steelpan before. Virtual Piano’s sheet
music contains letters that correspond to the keys on a QWERTY keyboard. So it is
quick and easy to learn. You can start playing steelpan [1] immediately.
The platform virtualpiano.net allows you to play the piano on your
computer keyboard, cell phone, or tablet [2].
14.2 Artificial Intelligence and Architecture 185
There are many ways that artificial intelligence is changing the field of architecture
[3].
• Programs can now make building calculations and environmental analysis a
simple task. Today, so much data is available to an architect that retrieving
information such as temperature and weather data, material ratings, etc., which
would otherwise take a lot of time to compile, is much easier.
• The application of intelligent systems can significantly reduce the time
required to plan and design a structure through building information
modeling (BIM).
• computer-aided design (CAD) has been instrumental in creating 2D
and 3D models of buildings. BIM goes a step further, integrating product
information, time, and cost to give an architect the full scope of a project. BIM
works with other design software programs to provide a complete picture of a
structure, including conceptual design, detailed design, analysis, documentation,
fabrication, construction, operation and maintenance after the building is built,
and renovation or demolition plans.
• Through appropriate use, programs can leverage this data and identify trends
or inefficiencies. This allows architects to design better, longer-lasting projects
without spending a lot of time analyzing data.
• Smart technology also gives architects the ability to incorporate smart lighting
or smart stormwater management systems into plans. None of this was possible
a decade ago. The proliferation of technology is allowing smaller practices to
take on larger projects, as they can use the Internet and all the resources at their
disposal to develop proposals.
• About seven percent of the world’s workforce is employed in the construction
industry, but it has traditionally been one of the least technologically advanced
sectors. However, there is great potential for integrating AI into construction, and
it could reduce construction costs by up to 20 percent.
• Video security has long been used in businesses, offices, and campuses to monitor
who comes and goes. However, due to the high volume of footage collected,
things can get overlooked if no one is watching the video. An intelligent
security system integrated into a building, on the other hand, can quickly scan,
automatically detect suspicious activity, and send alerts to the building owner.
• AI-based energy management platforms can set usage patterns to create ideal
conditions for tenants, saving them both energy and money. The Nest Thermostat
is an example of such a system that is popular with consumers—it adjusts to
keep a building at a safe and optimal temperature and can alert the user if the
temperature drops or rises to a dangerous level. AI devices can also take data
from sensors and analyze it to monitor for leaks or malfunctions. They make it
easier than ever to keep an eye on a building’s performance and efficiency.
AI can also be used in the form of intelligent locking systems that can restrict
access to certain areas unless the user has a key card or code. In addition to using
186 14 Artificial Intelligence in the Arts
References
1. https://virtualpiano.net/virtual-steelpan
2. https://virtualpiano.net
3. https://blog.dormakaba.com/de/sieben-wege-wie-kuenstliche-intelligenz-die-architektur-
revolutioniert
Chapter 15
Artificial Intelligence in Law
The legal system is also beginning to include artificial intelligence in its con-
siderations, relatively hesitantly and cautiously in Germany, although research
also already exists in this field, which began more than 30 years ago and is
associated with the person of Professor J. van den Herik from the Netherlands. At
present, Leiden University can be considered as a center of connection of artificial
intelligence and law.
There are the following institutes in Leiden:
• Law
• Institute for the Interdisciplinary Study of the Law
• Van Vollenhoven Institute
• Constitutional and Administrative Law
• Institute of Private Law
• Child Law
• eLaw—Center for Law and Digital Technologies
• Europa Institute (Fig. 15.1)
• What opportunities and threats arise from institutional and normative diversity
in the field of fundamental rights for effective protection of these rights in a
pluralistic world? [2]
The research program “Effective Protection of Fundamental Rights in a
Pluralistic World” (EPFR) was launched in its current form in 2015. It builds on a
strong tradition of stimulating and high-quality human rights research at Leiden
University and seeks to further expand and strengthen this line of research. The
program is unique in the Netherlands, as it is the only stand-alone research
program on fundamental rights at a Dutch law school.
The EPFR program explores the dynamics of institutional and normative
diversity in the field of fundamental rights protection against the backdrop of
sociocultural, political, and economic pluralism, which is a prominent feature of
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024 187
C. Posthoff, Artificial Intelligence for Everyone,
https://doi.org/10.1007/978-3-031-57208-1_15
188 15 Artificial Intelligence in Law
today’s world, both globally and locally. The program will focus on the following
topical and intersecting issues:
• Many handwritten and illustrated archives contain a wealth of information, but
are largely unexplored because they are complex and difficult for computers to
decipher. The goal of this project is to develop a digital environment that solves
this challenge and connects heterogeneous archival content with other digital
sources [4].
The project will focus on one of the Naturalis Biodiversity Center’s most
important collections: the Natuurkundige Commissie archive and collection,
which contains a rich verbal and pictorial record of the nature, cultures, and
economies of the Indonesian archipelago (1820–1850). Researchers will use an
advanced handwriting and image recognition system (Monk) supplemented by
contextual information on species, locations, and habitats. Naturalis’ taxonomic
expertise will be used in combination with methods from the history of science to
further refine the system. The outcome of the project will enable Brill to offer the
system as an online heritage service. This will serve both curators of illustrated
manuscript archives and researchers seeking to further the understanding of these
collections.
As digitization continues, the legal work environment continues to change [1],
from the introduction of electronic records in government to sophisticated software
solutions to the use of specialized algorithms [3].
According to a survey conducted by Freie Fachinformation Verlag in 2020, the
greatest motivation among lawyers to deal with legal tech is to increase efficiency
while reducing costs and maintaining competitiveness. At the same time, around
30 percent of respondents specifically stated that they were planning to introduce
individual legal tech solutions. This shows that there is no way around legal tech.
During the corona crisis in particular, some of the legal tech startups were able to
record considerable growth. Canceled flights, train rides, or package tours led to
more and more consumers wanting to enforce their legal rights—with as little cost
and risk as possible.
15 Artificial Intelligence in Law 189
Legal tech generally stands for IT-supported applications that support legal
work processes. The aim is to make workflows more efficient and thus more cost-
effective. The use of various software programs and databases is now commonplace
in many law firms. The software works largely with if-then rules. Artificial
intelligence goes one step further: artificial intelligence deals with the automation of
intelligent behavior and machine learning. Here, algorithms also initially work with
rules, but in contrast to software, they repeatedly adapt them based on problems
and data that have already been solved and try to recognize patterns. In this way,
even complete legal services can be mapped and human or legal decision-making
processes can be emulated.
A large number of young companies have discovered this topic and know how to
exploit it commercially. Platforms such as flightright.de, which specializes
in airline passenger compensation, or wenigermiete.de, whose core business
is unlawful rent increases, are flooding onto the legal market. The lockdown with its
consequences has done its bit to make legal portals highly popular with consumers.
The company RightNow, for example, is advertising claims for refunds of gym
fees during the pandemic-related shutdown or compensation claims for data mishaps
at the vaccination center.
The advantages for consumers are obvious: they receive a quick assessment of
their chances of success in enforcing their claims at low cost or even completely
free of charge. The critical point here is that these programs cannot replace an
initial consultation with a lawyer, but customers are not necessarily aware of
this. At the same time, they are significantly cheaper and therefore also tempting.
Consumer protectionists also criticize the lack of cost transparency in some cases.
The commissions due in the event of success are not always transparent on the
websites and may even be higher than with a law firm.
The IT skills of lawyers will continue to develop in line with technological
developments. IT skills will go far beyond operating a computer and will be an
essential part of the legal profession. Lawyers are just as technology-savvy as other
professions. The young people currently studying at universities belong to the group
of digital natives and have grown up with the new technologies. Education
should be more interdisciplinary. The law faculty in particular must integrate the
topic of digitization directly into the curriculum.
Ultimately, not all regulatory issues have been clarified for the use of deep
learning systems. Only in April of this year did the EU Commission publish a
draft for a regulation establishing harmonized rules for AI. The German Federal
Bar Association sees this as a first important step, but also expresses a need for
further clarification.
Meanwhile, on the legal market, the dispute between legal tech companies and
bar associations continues to smolder, as consulting services in particular may only
be provided by lawyers to protect against unqualified legal advice. The fact that
this is not always clear is shown by the recent ruling of the Federal Court of
Justice on online contract offers by Smartlaw. The BGH ruled in favor of the
online contract provider, while the plaintiff bar association still sees a risk of false
advice for consumers. And while the ruling may be a small victory for providers
190 15 Artificial Intelligence in Law
of automated contract offers, the BGH made it clear that this was no substitute for
legal advice from a lawyer. Even if some legal tech companies are also consolidated
or partially disappear in the future, it brings movement into the market.
In the long term, AI and algorithms will certainly be able to compete with law
firms in certain legal matters. However, they will not replace them. It is a matter of
actively shaping the development of AI in use in the legal profession—technically,
ethically, regulatory, and culturally. During the corona crisis in particular, many
processes in the legal sector were digitized, which would otherwise have taken much
longer. This effect must be further leveraged, because ultimately, it will come down
to a combination of humans and machines that complement each other.
To achieve this goal, experts and specialists in this field are needed above all
to support law firms on this path. So well-trained personnel and know-how will
remain the drivers of innovation in the future—only the way in which this is used
will change.
LawBot, the chatbot for legal issues launched by a group of Cambridge
University law students, is expanding its capabilities by launching in seven countries
and adding a case outcome prediction feature that can be used to estimate the odds
of winning a legal case analyzed by the bot.
Bister, LawBot-X focused on English criminal law, but the new version now
supports the USA, Canada, Hong Kong, Singapore, Australia, and New Zealand, as
well as a Chinese language option. It can also analyze information provided by users
before using data science techniques to calculate a lawsuit’s chances of winning.
According to CEO Ludwig Bull, the system currently predicts the outcome of the
lawsuit with 71% accuracy.
“We are proud to make this service available to the public. LawBot-X is
important both for basic research in the application of AI to legal services and for
advancing the rule of law.” This means that once the AI learns the user’s situation,
it will be able to select actual lawyers in the right country or jurisdiction who can
provide assistance.
LawBot’s head of marketing, Rebecca Agliolo, said, “Following the success
of LawBot’s first pilot, we are excited to bring to market a platform to analyze
the quality of users’ claims. We have evolved from decision trees to data-driven
intelligence.” LawBot was founded to explain the complex law surrounding sexual
offenses. It was then expanded to include a total of 26 serious crimes.
In the USA, there is also an extensive effort to make AI reside in the legal sector.
Earlier this year, Ginni Rometty, chair, president, and CEO of IBM, told CNBC’s
@Work Talent + HR Conference: “I expect AI to change 100 percent of jobs within
the next five to 10 years. This is going to require skills development on a massive
scale, and companies are going to have to do it in a way that is inclusive - including
people who may not have a four-year college degree. The last time we saw such a
big paradigm shift was when the Internet was introduced. The evolution is not really
a question of whether, but rather when and how.”
Many don’t like change. Lawyers, in particular, are notorious for it. In the two decades that
I have been working with lawyers from all over the United States, I have found that among
all professional service providers, lawyers are among those most resistant to change. I find it
15 Artificial Intelligence in Law 191
fascinating that a group of trained logicians do not always recognize the difference between
optional and mandatory changes. Whether they like it or not, law firms are businesses, and
the most successful law firms pay attention to business fundamentals such as return on
investment (ROI), marketing, and so on [5].
Recognizing and adapting to change has always been—and will continue to be—
the hallmark of successful companies. Take e-discovery, for example. There
was a time when the use of e-discovery tools was rare in organizations. It was simply
unthinkable that a firm that had a veritable warehouse of employees sifting through
documents could ever be overtaken by a firm that used e-discovery technology.
Today, the exact opposite is true: a firm that reviewed large volumes of documents
and did not use technology would be an underdog. The e-discovery pattern has
changed relatively quickly.
In June 2007, Apple released its first iPhone. Until then, Apple was known
primarily for computers and iPods. Steve Jobs, however, recognized early on
the value of combining computer technology with smartphones. In fiscal 2007,
Apple generated $24.6 billion in revenue, with the fledgling iPhone accounting
for a relatively small share. In contrast, in the fourth quarter of 2019 alone, Apple
generated nearly $63 billion in revenue. Today, just 12 years after its launch, the
iPhone accounts for nearly 70 percent of Apple’s revenue (estimated at more than
$400 billion). The ability to recognize and proactively respond to the coming change
has made Apple a powerhouse company. Apple could have buried its head in the
sand and continued to focus on iPods and Macs. Had it done so, Apple might not
even be in business today. The iPhone was first an iPod in a phone. Now it gives
users access to Siri, a close application of AI.
From services like time tracking to contract review to brief analysis and research,
new technology companies are emerging all the time to meet the needs of lawyers.
Some of these tech companies are quickly going under, and that’s despite the fact
that their ideas are often ahead of very promising ones. Add to this the widespread
fear of change among lawyers. The result is that the advances the legal field could
make to increase its overall efficiency are almost nonexistent.
LawGeex is an automated contract review platform that uses AI to answer
the simple question “Can I sign this?”. The idea is simple: a contract is sent to
a company and uploaded to LawGeex’s AI, which then reviews the contract. If
problems are found, the contract is forwarded to a legal team that highlights the
questionable wording. The time saved is immense. More importantly, the accuracy
is unparalleled. In a study of nondisclosure agreements (NDAs), five contracts were
analyzed by 20 experienced lawyers.
There are over two and a half thousand startups in Hong Kong, employing nearly
10,000 people. Companies like Zegal are tapping into this large pool of talent and
changing the game for companies that have previously shied away from complex
legal documents [7].
“Historically, law has been an intimidating area for individual businesses, espe-
cially small- and medium-sized businesses and law firms,” says Tai Hung-Chou,
CTO of Zegal. “We provide a place where clients and lawyers can collaborate to
draft and execute contracts - all online. In the future, we will use advanced natural
192 15 Artificial Intelligence in Law
References
1. https://www.perconex.de/de/perconex-personaldienstleister.html
2. https://www.universiteitleiden.nl/en/research/research-projects/law/effective-protection-of-
fundamental-rights-in-a-pluralist-world
3. https://www.universiteitleiden.nl/en/research/research-projects/law/the-legitimacy-and-
effectiveness-of-law--governance-in-a-world-of-multilevel
4. https://www.universiteitleiden.nl/en/research/research-projects/science/making-sense-of-
illustrated-handwritten-archives
5. https://growpath.com/future-of-law-artificial-intelligence-for-lawyers
6. https://abovethelaw.com/law2020/how-artificial-intelligence-is-transforming-legal-research
7. https://zegal.com/en-hk
Chapter 16
The Development in South and Central
America and the Caribbean
Based on my work at the University of the West Indies, I would like to make a few
more remarks about artificial intelligence in this region. This is another feature of AI
that has not yet been mentioned, but which is of paramount importance, especially
in countries as small as the Caribbean islands. You develop a high-quality system in
one place and can immediately pass it on to others with ease. With a good division
of labor, you can bring several countries up to a high level.
“Some of the push in AI adoption in these countries has come from academics
and researchers, like the ones at the University of Sao Paulo who are developing AI
to determine the susceptibility of patients to disease outbreaks, or Peru’s National
Engineering University where robots are being used for mine exploration to detect
gases, or Argentina’s National Scientific and Technical Research Council where AI
software is predicting early onset pluripotent stem cell differentiation.
These and other truths were revealed to me at a Latin America and Caribbean
(LAC) Workshop on AI organized by Facebook and the Inter-American Devel-
opment Bank in Montevideo, Uruguay, in November this year. I was the lone
Caribbean participant in attendance, presenting my paper entitled: AI & The
Caribbean: A Discussion on Potential Applications & Ethical Considerations, on
behalf of the Shridath Ramphal Centre (UWI, Cave Hill)” [3].
So what is happening here in the Caribbean?
AI remains, at best nascent, with limited R&D. Reports of its limited application
in The Bahamas, Belize, and Guyana and the absence of policy discussions belie the
significant potential for AI here.
We could dream of a day when Fedo, a risk stratification AI system for
predicting patient susceptibility to noncommunicable diseases (NCDs), is used in
the Caribbean’s health sector, where NCD mortality is highest in the Americas, or
when Dragon Medical Virtual Assistants assist the region’s critical nurse shortage,
which in 2010 measured 1.25 nurses for every 1000 people. How about See &
Spray, a weed control AI robot that could reduce herbicide expenditure by 90%?
Or AI harvest robots replacing 30 workers in the Caribbean’s agricultural sector,
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024 195
C. Posthoff, Artificial Intelligence for Everyone,
https://doi.org/10.1007/978-3-031-57208-1_16
196 16 The Development in South and Central America and the Caribbean
where the food import bill is expected to reach USD 8–10 billion by 2020? Could
we ever see AI systems developed by Google, Harvard, and NASA which predict
earthquake aftershocks, flooding, and hurricanes as part of the Caribbean’s disaster
management and mitigation efforts, to save lives and mitigate potential loss?
Instead of dreaming, I propose the following three steps that the Caribbean can
take to better position itself to harness AI’s potential.
• First, we must develop an appetite for such technologies.
South America’s engagement in this field is a testament of the region’s innovative
capabilities and appetite for such technologies. This cannot be done without
firms and governments that are willing to adopt and utilize these systems
in their provision of goods and services. In addition, we need research and
studies that demonstrate how can AI be leveraged to solve some of the region’s
developmental challenges. It falls to the region’s academia and private sector to
find innovative AI solutions and spur demand for their subsequent development
and adoption.
• Second, we must form strategic partnerships.
Google is developing an AI system to predict and send flood warnings in India;
Unilever is testing various AI solutions in South America; and MIT and Harvard
are hosting AI R&D conferences in Uruguay, but, who are we partnering with in
the Caribbean?
Recognizing the importance of strategic partnerships and taking steps to reach
out to organizations like the IDB to fund such initiatives, or companies like
Facebook and Google to develop and test AI solutions in the region, or AI R&D
centers and universities to partner with are all potential avenues for overcoming
existing financial and resource constraints that hinder our progress in this field.
• Third, we must initiate AI-related policy discussions.
Realizing the wider ethical and legal considerations arising from the application
of AI, we must ask probing questions like the following: are the existing
frameworks capable of addressing our concerns? And how can we mitigate risks
and instill public confidence in such technologies?
Beyond technologists like engineers and developers, the discussion must involve
policy-makers who must be on the front lines in developing adaptive and
anticipatory frameworks. Similar to Mexico’s move toward an AI strategy, which
aims to transform the country from an observer to a key actor, we must look
toward the development of holistic approaches.
While not exhaustive, this list of recommendations is a start to riding the AI wave.
It is now up to us to either learn how to ride it, like our South American neighbors
are doing, or we get washed ashore.
16.1 Guyana 197
16.1 Guyana
v75 Inc. a local technology company has teamed up with Professor Jason Mars, creator
of Jaseci – a leading artificial intelligence programme – to create Guyana’s own Silicon
Valley. Once established, Guyana’s talented young minds will be able to build their own AI
products to be marketed and sold to top companies around the world. Already a team of five
Guyanese engineers who recently concluded an apprenticeship programme are employed
with Jaseci and are creating their own AI products. Part of the wider intention is to upskill
Guyanese in AI and create tangible opportunities to retain those persons here. [1]
So, not only can computer vision eliminate a large amount of the difficult,
manual labor involved in crop and soil monitoring; in many cases, it does it more
effectively than humans can.
• Finding bugs with code
And say you’d like to know not only if your crops have pests, but how many
there are, computer vision systems for insect detection has that covered as well
(Fig. 16.1).
This works for flying insects too. They are certainly not the most fun to capture
and count by hand.
Researchers first set up a sticky trap to capture six different species of flying
insect and collect real-time images. They then based the detection and coarse
counting method on YOLO object detection and the classification and fine
counting on support vector machines (SVM) using global features.
When all was said and done, their computer vision model was able to identify
bees, flies, mosquitoes, moths, chafers, and fruit flies with an accuracy of 90.18%
and count them with 92.5% accuracy.
These studies show that the future of AI computer vision for monitoring
the health of our food systems is promising. Not only can it reduce labor
inefficiencies, but it can do so without sacrificing reliability of the observations.
• Livestock health monitoring
So far we have focused mainly on plants, but there is more to agriculture than
wheat, tomatoes, and apples.
Animals are another major component of our agriculture systems, and they tend
to need a bit more tracking than plants. Can computer vision keep up with cows,
chickens, and pigs on the move?
16.1 Guyana 199
All this pretty well introduces BoniRob, an agricultural robot that uses camera
and image recognition technology to find weeds and remove them by driving a
bolt into the earth.
It learns to distinguish between weeds and crops through image training on leaf
size, shape, and color. That way, BoniRob can roll through a field eliminating
undesirable plants without the risk of destroying anything of value.
And while our AI friends are out in the fields anyway, perhaps there are other
jobs they could do.
A group of scientists is working on making this a reality with designs for
agricultural robots that detect weeds as well as soil moisture content. This way,
it can move through a field, removing weed and delivering appropriate amounts
of water to the soil as it goes.
Experimental results for this system show that its plant classification and weeding
rates are both at or above 90%, all the while keeping deep soil moisture content
at 80 ± 10
AI-driven agriculture bots are developing quite the resume!
• Aerial survey and imaging
At this point, it is probably unsurprising that computer vision also has some
terrific applications for surveying land and keeping an eye on crops and livestock.
But that does not make it any less significant for smart farming.
AI can analyze imagery from drones and satellites to help farmers monitor crops
and herds. That way they can be notified immediately if something looks amiss
without having to constantly observe the fields themselves.
Aerial imaging is also useful for boosting the precision and efficiency of pesticide
spraying. As mentioned previously, ensuring that pesticides only go where
they’re intended saves money as well as the surrounding environment.
• Produce grading and sorting
Finally, AI computer vision can continue to help farmers even once the crops
have been harvested.
Just as they are able to spot defects, disease, and pests as the plants are growing,
imaging algorithms can also be used to sort “good” produce from the defective
or just plain ugly.
By inspecting fruits and vegetables for size, shape, color, and volume, computer
vision can automate the sorting and grading process with accuracy rates and
speed much higher than even a trained professional.
Take carrot sorting, for example. It is laborious and usually done by hand.
However, researchers have developed an automated sorting system that uses
computer vision to pick out carrots that have surface defects or are not the correct
shape and length. A good carrot, then, is one that is the right shape (a convex
polygon) and does not contain any fibrous roots or surface cracks.
On these three criteria, the computer vision model was able to sort and grade
carrots with accuracy rates of 95.5%, 98%, and 88.3%, respectively.
Further, bringing us back to the classic tomato, another study found that AI with
machine learning was able to use image data with seven input features to grade
tomato quality with 95.5% accuracy.
16.2 The Caribbean 201
In both cases, the amount of painstaking manual labor saved is enormous. And it
is all thanks to a bit of AI training on what a good carrot or tomato looks like.
• The future of AI in agriculture: farmers as AI engineers?
Throughout human history, technology has long been used in agriculture to
improve efficiency and reduce the amount of intensive human labor involved
in farming. From improved plows to irrigation, tractors to modern AI, it’s an
evolution that humans and agriculture have undergone since the invention of
farming.
The growing and increasingly affordable availability of computer vision stands
to become another significant step forward here.
With considerable changes occurring in our climate, environment, and global
food needs, AI has the ability to transform twenty-first century agriculture by:
Increasing efficiency of time, labor, and resources
Improving environmental sustainability
Making resource allocation smarter
Providing real-time monitoring to promote greater health and quality of produce
Of course, this will require some shifts in the agricultural industry. Farmers’
knowledge of their field will need to be translated into AI training, and this will
depend on greater technical and educational investments within the agricultural
sector.
But then again, innovation and adaptation are nothing new in agriculture.
Computer vision and agricultural robotics are just the latest way farmers can
adopt new technology to meet growing global food demands and increase food
security.
If you are interested to learn more about AI applications across other industries,
check out: v7labs
Many Caribbean Small Island Developing States (SIDS) have not yet developed
strategies and national policies to manage the impact of AI on their societies. The
Caribbean Artificial Intelligence Initiative is led by the UNESCO Cluster Office for
the Caribbean and aims to develop a subregional strategy on the ethical, inclusive,
and humane use of AI in the Caribbean SIDS. The Caribbean AI Initiative raises
awareness, informs, engages in open discourse, strengthens capacities, produces and
shares knowledge, and gathers expertise from various stakeholders from different
sectors in the English- and Dutch-speaking Caribbean SIDS to identify the way
forward in developing regional and national AI strategies. Against this background,
the Caribbean AI Initiative was launched on the occasion of the Sub-Regional
Consultation on the Draft Recommendation on the Ethics of AI, which took place
on August 5, 2020 [2].
202 16 The Development in South and Central America and the Caribbean
16.3 Belize
In this latest installment of Oxford Insights’ AI strategy series, we turn our attention
to the Latin American region to take a closer look at two neighbors’ approaches to
artificial intelligence: Argentina and Uruguay [5].
In this year’s Government AI Readiness Index, Uruguay topped the Latin
American region, with an overall score of 55.57. Meanwhile, Argentina came in
a regional fourth, behind Chile and Colombia, with an overall score of 50.75. As
two regional leaders in this sphere, how these neighboring governments tackle AI
is likely to have an impact on the future digital and economic landscape in Latin
America more widely.
There are a number of parallels between Argentina and Uruguay’s AI strategies.
They were published only months apart in 2019, and both were designed by previous
administrations that have since been defeated in national elections. Yet, as we
explore in this blog, each is underpinned by seemingly very different priorities.
204 16 The Development in South and Central America and the Caribbean
For each area considered, we have determined which country’s strategy has the
comparative advantage according to our analysis.
Argentina has a broader AI strategy, which was published in 2019 under the
government of the then-president Mauricio Macri. The document seeks to promote
AI in the private sector, minimize ethical risks, and develop talent, among other
objectives. Uruguay’s AI’s strategy, also published in 2019 under the administration
of a former president, Tabaré Vázquez, is much briefer, dealing specifically with
promoting the use of AI within public administration. Nonetheless, Uruguay’s
Agenda Digital 2020, another output of the previous government, takes a much
broader view of digital transformation in the country and is in many ways more
comparable to Argentina’s strategy in this sense. As such, this blog also considers
the documents, reports, and agendas that accompany the AI strategies. In doing this,
we hope to achieve a fairer comparison, which accounts for discrepancies between
the ways in which each country structures its strategic documents.
• Background
Before diving into the analysis of each strategy, it’s worth acknowledging their
political backdrops.
As of May 2020, a report published by the Banco Interamericano de Desarrollo
(BID) claimed that Argentina’s AI strategy can be classed as to be continued, i.e.,
not currently implemented. In October, Mauricio Macri’s incumbent conservative
government was defeated by Alberto Fernández’s left-wing coalition Frente
de Todos. Although the Fernández administration did initially show some
willingness to commit to digital transformation by vowing to re-establish the
Argentine Ministry of Science and Technologies—which had been demoted to
a secretariat under the previous administration—Fernández’s mandate has since
been dominated by the coronavirus response.
During its presidential elections in 2019, Uruguay experienced something of a
reverse of the Argentine situation, when the incumbent socialist government was
replaced by a center-right President, Luis Lacalle Pou. Pou’s presidency seems
to have run more smoothly than his Argentine counterpart’s to date; as opposed
to Argentina, which was hit relatively hard by the virus, registering more than
44,000 deaths from COVID-19, Uruguay has only recorded 231 deaths at the time
of writing. Digitalization remains a government priority, and unlike Argentina,
the BID considers Uruguay’s strategy to be a completed document which is
currently in implementation. Uruguay immediately looked to technology in its
virus response, creating online resources through which people could seek advice
and report potential cases. This caught the attention of Apple and Google, and
subsequently, Uruguay was the first country in Latin America to implement their
track and trace technology.
While it’s possible that the Argentine government may disregard or modify the
2019 strategy, it’s nonetheless the most up-to-date vision of artificial intelligence
in Argentina and therefore our main point of reference for this blog.
• Vision: draw Both the Argentinian and the Uruguayan AI strategies set out
ambitious goals for a digital future and received full marks in the vision
16.4 Argentina and Uruguay 205
However, public sector reform hardly shines through as a central focus of the
strategy, as in the case with Uruguayan AI strategy, which is primarily focused
on AI reforms in the public sector. From the outset, the Uruguayan strategy
explicitly sets out to “consolidate a closer relationship between the people and
the state” and “provide more efficient and innovative services” to citizens, by
increasing digital capacity within government. To achieve this, the Uruguayan
strategy proposes to:
• Create a training program around the use of AI in public services.
Train all government departments using the aforementioned program.
Define standards, guidelines, and recommendations for auditing of the decision-
making algorithms used in government, among other initiatives.
• Feasibility and implementation: Uruguay Having a high-reaching vision for
change is one thing, but as our initial ‘What makes a good AI strategy?’ blog
sets out, AI strategies need to set out measurable goals if they are to translate
ambitious ideas into practical change.
This is something the Uruguayan strategy and its accompanying initiatives
do particularly well. Uruguay Digital’s website allows visitors to search for
particular digital initiatives proposed in the Agenda Digital 2020 and see how
close they are to full implementation, measured in percentage points. Elsewhere,
AGESIC, the ministry for digital government and society responsible for creating
Uruguay’s AI strategy, places significant emphasis on “actually making things
happen” and “improving the improvable”—suggesting that the Uruguayan gov-
ernment’s digital departments have considered the importance of a culture of
tangible change.
Example progress markers on Uruguay Digital’s website.
Source: https://uruguaydigital.gub.uy.
The Argentine government lacks a similar tool, making it hard to track progress
on some of the claims made in the AI strategy. To give credit where it’s due,
many of the goals proposed in the strategy do have clear markers. For instance,
the previous Argentine government sought to measure their investment in human
capital and digital education by monitoring tracking the number of degrees
related to AI available, scholarship programs abroad, and papers published,
among other indicators. Yet the government does not seem to have published any
data surrounding whether these goals are being met, likely due to institutional
discontinuity.
• Final takeaways
Many question marks hang over both Argentina and Uruguay’s AI strategies.
Are they simply vestiges of old administrations, as the apparent abandonment of
Argentina’s strategy seems to suggest? How will each country’s new government
address the challenges and opportunities associated with AI?
To date, it seems like Luis Lacalle Pou’s relatively new Uruguayan administration
has been quicker to take up the baton on AI than its Argentine equivalent,
integrating AI into a coronavirus response which has been praised internationally.
Meanwhile, following a change in administration, very little information is
available at all surrounding the Fernández administration’s approach to AI;
References 207
Argentina’s AI strategy, released by the ArgenIA group in 2019, has not yet been
approved by the new government in a resolution.
As the Uruguayan case demonstrates, AI can be a powerful tool for governments
coordinating a public health response to the virus. Yet, looking further forward,
the digital sector will continue to be crucial for administrations as they seek to
revitalize economies in the wake of the crisis. As inflation in Argentina nears
40% and the Argentina Central Bank runs low on dollars, how the Argentine
government addresses AI in the coming years has the potential to define the
nation’s economic trajectory. At Oxford Insights, we’re eager to see how both
administrations will make their mark on national AI strategies in Argentina and
Uruguay as they approach 1 year in office [5].
References
1. https://www.jaseci.org/the-future-of-artificial-intelligence-in-guyana/
2. https://en.unesco.org/caribbean-artificial-intelligence-initiative
3. Chelceé Brathwaite: Artificial Intelligence (AI) in South America—Setting the Bar for the
Caribbean, Caricom Today, JULY 27, 2023
4. https://unichrone.com/bz/artificial-intelligence-expert-training/belize-city
5. Jasmine Kendall: Oxford Insights’ AI Strategy Series: Argentina and Uruguay, 13/January/2021,
https://www.oxfordinsights.com/insights/2021/1/8/oxford-insights-ai-strategy-series-
argentina-and-uruguay
Chapter 17
Future Developments
Already, we are seeing major changes across the board. These developments will
continue, leading to sweeping changes in all areas of life.
In light of these developments, it is not surprising that currently 62 percent of HR managers
expect AI to fundamentally change working life in the next five years. Our working world
and job descriptions will change dramatically. For example, research by Erik Brynjolfsson,
director of the Massachusetts Institute of Technology (MIT) Initiative on the Digital
Economy, found that machine learning rarely replaces entire professions. Instead, it takes
over or transforms specific work processes within a job description. Individual jobs will
certainly disappear. Predominantly, however, many tasks will be performed jointly by
humans and AI, entirely in the sense of AI as a personal assistant. Machines can indeed take
over tasks with a high workload and routine tasks. However, they work less well on abstract
processes where skills such as empathy, judgment, creativity, inspiration and leadership are
required. These include areas such as innovation and personnel management [3].
Given these developments, the importance of people being flexible and open to
changing for new tasks, as well as to continuous learning, becomes clear. As AI
continues to evolve, the competence of employees must also follow these advances.
It is therefore necessary to develop further training strategies in order to also be
able to use the potential of AI. This is not just a matter for IT staff, but also for the
specialist departments. Salesforce, for example, offers numerous free training
opportunities via a learning platform Trailhead. The more these technologies
can best support our efforts to create better jobs and living conditions, the more
successful companies and progressive societies will exist.
The effects of the further development of artificial intelligence can be very
diverse and will be assessed very differently depending on the economic and social
context in the coming decades. In most industrialized economies, an (unconditional)
basic income is very likely—not least because it is even being pushed by the current
economic elite. Who pays the taxes? Well, the robots and algorithms! [4]
“The picture of the future in the technological realm is often referred to as
‘transmodern liberty’.” The term is composed of the elements of a transmodern
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024 209
C. Posthoff, Artificial Intelligence for Everyone,
https://doi.org/10.1007/978-3-031-57208-1_17
210 17 Future Developments
new xenobots can also swim around, find cells, and build copies of themselves.
This process can be repeated as often as desired, as the scientists describe in their
work, which was published in the journal Proceedings of the National Academy of
Sciences.
Joshua Bongard, a robotics expert and one of the study leaders, explained that
with the right design, the xenobots can spontaneously self-replicate. Normally, these
embryonic stem cells would develop into the skin layer of a clawed frog (Xenopus
laevis). These cells would then normally sit on the outside of the tadpole, fending
off pathogens and secreting mucus. But in the experimental design of their trials, the
researchers expose the cells to a new environment and context. Michael Levin, one
of the researchers, explained that this opens up the possibility for the stem cells to
remember genetic code, allowing them to develop into other cells. Sam Kriegman,
another scientist involved in the study, explained:
These are frog cells replicating in a way that is very different from the way frogs do it. No
animal or plant known to science replicates in this way.
For a long time, it was believed that all possibilities of reproduction of life-
forms had been researched and were known. But according to the scientists, this is
something that has never been observed before. The cells have the genome of a frog,
but since they do not develop into tadpoles, they can use their collective intelligence
to accomplish something so extraordinary. In previous experiments, researchers
were already amazed at the possibility of being able to engineer xenobots to perform
simple tasks. Now the researchers were all the more amazed to find that this
computer-designed collection of cells began to multiply spontaneously. According
to Levin, there was no indication from the complete frog genome that the cells could
work together and condense functional copies of themselves from the collected
cells.
A milestone toward organ cultivation was achieved in China. Researchers grew
human-monkey embryos for the first time [2]. The parent, consisting of about 3,000
cells, is shaped like a sphere and forms offspring on its own. Typically, this system
dies off quickly, so it is usually difficult to get it to continue reproducing. But
the researchers use an artificial intelligence program on a supercomputer cluster
and used an evolutionary algorithm to test billions of different body shapes in a
simulation, from spheres to pyramids to the shape of starfish. In doing so, they found
a shape that allowed cells to be as effective as possible in kinematic replication based
on their movement. As Kriegman, one of the study authors, explains, this allowed
them to figure out how to match the shape of the original parents:
“The AI, after months of work, came up with some strange designs, including
one reminiscent of Pac-Man. It’s very unintuitive. It looks very simple, but it’s not
something a human engineer would come up with.” But in practice, he said, that very
system has been shown to work: the parent xenobots were able to build children,
who then assembled grandchildren, who in turn produced great-grandchildren, and
who then produced great-great-grandchildren. With the right design, the number of
generations could be greatly expanded. Until now, kinematic replication was only
known at the level of single molecules, but was never observed at the level of whole
17 Future Developments 213
cells or organisms. With the researchers’ work, this has now changed. They have
discovered, they say, that there is a vast, previously unknown space within living
systems. According to the researchers, life harbors surprising behaviors just beneath
the surface.
Stem cell researchers grew embryos composed of parts of humans and monkeys
that survived in the culture dish for an astonishingly long time. The breeding of such
mixed creatures sparked a debate about ethical issues in research on such chimeras.
Almost 2 years ago, stem cell researcher Juan Carlos Izpis.úa Belmonte had already
made an announcement that caused some controversy: In July 2019, Ispizúa told
the Spanish newspaper El Pa.ís that he had worked with Chinese scientists to breed
embryos that were part human and part ape. Now, the associated publication has
appeared in the journal Cells, which is likely to reignite the debate about ethical
issues in this type of research.
With his team, the stem cell researcher had grown embryos from Javanese
monkeys. After 6 days, they implanted these with highly transformable human stem
cells. The cells then actually fused, and 132 embryos were created from parts of
humans and monkeys. After 10 days, 103 of the human-monkey chimeras were still
alive, and after 19 days, three were still alive.
The experiment is astonishing not only because of the ethical issues involved
but also that it worked at all. Until now, chimera embryos have never survived this
long. Earlier attempts at mixed human-mouse and human-pig embryos had failed.
The reason for this was probably the high species barrier. The chimera of monkey
and human, on the other hand, survived for quite a long time. But of course this
also raises the fear that this discovery could be misused for ethically more than
questionable purposes: if the chimera embryo were implanted in a human or animal
surrogate mother, a new, independent life-form could possibly develop on it.
Izpis.úa assured that he did not intend to do so; moreover, ethics committees had
previously reviewed his experiments. With his experiments, he pursues the goal of
better understanding embryonic development. He has also long been working on
growing human organs and tissue in pigs for sick people, he said:
It is our responsibility as scientists to think through our research well and follow all ethical,
legal and social guidelines.
people is sufficient justification for the creation of such chimeras is still an open
question. This would have to be decided by each individual, by society as a whole,
and then finally by the legislature in a legally binding manner, Behr said.
But Michael Coors, head of the Institute for Social Ethics at the University of
Zurich, demands that this decision-making process must finally begin now, because
there is an urgent need for legal regulation of this type of research. In view of
the fact that Izpis.úa already announced his attempt 2 years ago, the facts now
do not exactly come as a surprise. It is all the more astonishing that the process
has hardly progressed. It is not only about mixed creatures but also about the
instrumentalization of the animals used and the potential suffering that might be
inflicted on such chimeras.
Some people may look with horror at the idea of self-replicating nanorobots, but
the scientists explained that they are primarily concerned with understanding the
processes involved. The millimeter-sized living machines, which are exclusively in
the lab, are easy to eliminate once the experiments are over and have also been
reviewed by ethics experts at various levels. Bongard said this is not something
that keeps him up at night. On the other hand, he said, one has a moral obligation
to understand the conditions of self-replicating systems under which they can be
controlled, directed, or contained. In this context, he also refers to the corona crisis:
The speed with which we can develop solutions is of great importance. If we can develop
technologies that learn from xenobots, and quickly tell the AI: We need a biological tool
that does X and Y and suppresses Z, then that could be very useful. Today, that takes a long
time.
and machines do not have free will; it always remains between human leaders who
can eventually impose their will.
Regarding autonomous supporting functions, several participants referred in their
response to human-machine teaming issues, which is believed relevant because
unmanned systems are assumed unable to share combat experiences and to com-
municate military issues. Therefore, many respondents think it is still too early to
discuss acquisition of AWS, especially if they cannot replace soldiers or demand
more human resources to keep unmanned machines operating. Moreover, soldiers’
inner will and patriotic mentality to fight for freedom was underlined, which was
explained as inconceivable for an autonomous system, so unmanned platforms can
hardly ever be equipped with patriotical algorithms.
Unmanned warfare is a relatively new topic, for which reason I recognize that
there were limitations to this study. Participants had to imagine AI implications
for warfare, leading to very diverse representations even in the same field of
expertise, which renders results challenging to interpret. Although expert interviews
allow for in-depth examination, it is not enough to comprehensively capture the
essence of a novelty subject with a questionnaire alone. However, concerns about
key categories were asked explicitly; participants should be presented with case
studies that include these categories, which facilitate setting a focus on a particular
issue. Moreover, such designed vignettes might encompass cultural differences
and capture distinctive perspectives for this field of research. As a result, study
outcomes intend to reveal relevant aspects of AI to understand how it can be applied
responsibly in a military domain and the implications for the following research. In
that respect, it may help researchers and system engineers understand AWS’s main
benefits and principal vulnerabilities from an end-user point of view.
The datasets generated and analyzed during the current study are not publicly
available due the fact that they constitute an excerpt of research in progress but
are available from the corresponding author on reasonable request. The study was
accepted by the ethics committee of the University of Tartu. Informed consent was
obtained from all individual participants included in the study.
References
1. https://transinformation.net/ein-bericht-ueber-ausserirdische-rassen-die-mit-der-menschheit-
in-interaktion-stehen
2. Meilenstein in Richtung Organzucht: Forscher züchten in China erstmals Mensch-Affen-Embryo
https://de.rt.com/asien/90867de.rt, August 2019
3. Frank Engelhard. Die Zukunft der KI—und ihr Einfluss auf die Geschäftswelt, Silicon.de, 2019
4. Kai Gondlach. https://www.kaigondlach.de/artikel/wie-wird-kuenstliche-intelligenz-die-
zukunft-veraendern, 2021,
5. https://algorithmenethik.de/
6. https://news.mit.edu/2022/new-programming-language-high-performance-computers-0207
7. Kai-Fu Lee, Qiufan Chen, KI 2041—Zehn Zukunftsvisionen, Campus Verlag Frankfurt/New
York, 2021, 978-3-593-51549-6
Index
A Coloring problems, 59
Algorithm, 20, 27 Competence, 9
Alphazero, 76 Complexity, 41, 44
Alzheimer, 127 butterfly effect, 41
Analytic geometry, 136 exponential complexity, 41, 44
Axioms, 49 linear search, 44
logarithmic complexity, 44
polynomial complexity, 41, 44
B Computability, 30
Backpack problem, 62 Computer algebra, 37
Big Data Mathematica, 39
Alipay, 97 Microsoft Mathematics, 37
Apache Software Projects, 95 Conjunction, 16
Association Analysis, 94 CorelDRAW, 143
Clustering, 94 Correlation coefficient, 92
Corvid-19, 96 Covariance, 92
Hadoop, 95 Covid-19, 128
MapReduce, 95 Cybernetics, 63, 64
Predictive analysis, 94 control, 64
Spark, 95 feedback, 64
Terark, 97
WeChat, 97
Bog data D
regression analysis, 94 Dartmouth Conference, 19
Boole, G., 15 Database, 83
Boolean Algebra, 16 Brockhaus, 83
Brain research, 6 database language, 83
Browser, 26 dictionary, 83
management software, 83
management system, 83
C relational database, 84
Church-Turing thesis, 30 SQL, 83
Circuits, 17 Digital camera, 113
Cloud computing, 85 Digital Revolution, 113
© The Editor(s) (if applicable) and The Author(s), under exclusive license to 217
Springer Nature Switzerland AG 2024
A. C. Posthoff, Artificial Intelligence for Everyone,
https://doi.org/10.1007/978-3-031-57208-1
218 Index
F M
Fisheye projection, 136 Magnetic Resonance Imaging (MRI), 125
Mersenne’s prime, 2
Moore’s Law, 1
G
Game theory, 63
minimax theorem, 63 N
n-person games, 63 Negation, 16
Gravitational wave, 117 Netflix, 89
Network
feed-forward network, 73
H feedback, 73
HTML, 26 fully-connected, 73
HTTP, 26 recurrent network, 73
short-cuts, 73
Neural network, 69, 74
I folding network, 74
Image processing Neuron, 69
brain tumor, 125 action potential, 72
pixel, 133 membrane potential, 72
Information theory, 63 threshold, 69
complete information, 63 weight, 72
Intelligence, 5
Artificial Intelligence, 5
collective intelligence, 7 P
definition, 5 Pearson’s rank correlation, 92
emotional intelligence, 6 Perceptron, 72
environmental intelligence, 6 single perceptron, 72
human intelligence, 6 Polar coordinates, 136
Index 219