01 Digital Transformation PDF
01 Digital Transformation PDF
Digital
Transformation
Schlüsseltechnologien
für Wirtschaft & Gesellschaft
Digital Transformation
Fraunhofer-Forschungsfokus:
Reimund Neugebauer
Digital Transformation
1st ed. 2019
Reimund Neugebauer
Zentrale der Fraunhofer-Gesellschaft,
Munich, Germany
Springer Vieweg
© Springer-Verlag GmbH Germany, part of Springer Nature, 2019
This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or
part of the material is concerned, specifically the rights of translation, reprinting, reuse of illus-
trations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and
transmission or information storage and retrieval, electronic adaptation, computer software, or by
similar or
dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this
publication does not imply, even in the absence of a specific statement, that such names are exempt
from the relevant protective laws and regulations and therefore free for general use. The publisher,
the authors and the editors are safe to assume that the advice and information in this book are
believed to be true and accurate at the date of publication. Neither the publisher nor the authors or
the editors give a warranty, express or implied, with respect to the material contained herein or for
any errors or omissions that may have been made.
This Springer Vieweg imprint is published by Springer-Verlag GmbH Germany and is a part
of Springer Nature. The registered company is Springer-Verlag GmbH Germany, Heidelberger
Platz 3, 14197 Berlin, Germany
Contents
V
VI Contents
5 Audio Codecs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
5.1 Introduction: The dream of high fidelity. . . . . . . . . . . . . . . . . . . . . . 64
5.2 Hi-fi technologies from analog to digital . . . . . . . . . . . . . . . . . . . . . 64
5.3 Current research focus areas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
5.3.1 The ear and the brain . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
5.3.2 From audio channels to audio objects . . . . . . . . . . . . . . . . . . 68
5.3.3 Audio objects in practice . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
5.4 Outlook . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
6 Digital Radio . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
6.1 Introduction. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
6.2 Spectrum efficiency allows more broadcasts . . . . . . . . . . . . . . . . . . 80
6.3 Program diversity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
6.4 Innovative services: from traffic alerts to emergency management . 82
6.5 Non-discriminatory access. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
6.6 Hybrid applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
6.7 Outlook . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
18 Blockchain . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 301
18.1 Introduction. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 302
18.2 Functioning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 303
18.3 Methods of consensus building . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 304
18.4 Implementations and classification . . . . . . . . . . . . . . . . . . . . . . . . . . 306
18.5 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 307
19 E-Health . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 311
19.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 311
Contents XI
The digital era began relatively slowly. The first programmable computer using
binary code was the Zuse Z3, designed and built by Konrad Zuse and Helmut
Schreyer in Berlin in 1941. In 1971 the first microprocessor was patented; it con-
tained 8,000 transistors. Within ten years, nearly ten times as many transistors were
being used; by 2016, the number was around 8 billion.
This exponential growth in complexity and power of digital computers was
predicted by Gordon Moore in 1965. Moore’s Law, a rule of thumb to which he gave
his name, is generally held to mean that the number of transistors that fit into an
integrated circuit of a specified size will double approximately every 18 months.
While the law’s interpretation and durability may be the subject of debate, it never-
theless provides a fairly accurate description of the dynamics of the development of
digital technology.
Among the reasons for this vast acceleration in progress is the fact that digitiza-
tion, as well as changing various technical and practical fields of application, has
also changed the work of research and development itself. A processor featuring
transistors running in the billions, for example, can only be developed and manu-
factured using highly automated and digitized processes. Complex computer pro-
grams are themselves in turn designed, realized and tested in whole or in part by
computers. The immense quantities of data generated by research projects, produc-
tion facilities and social media can only be analyzed with massive computer assis-
1
2 Reimund Neugebauer
tance. The insights we gain from this analysis, however, were practically unattain-
able just a few years ago. Now, Machine Learning is becoming standard: Artificial
systems gather experiences and are then able to generalize from them. They produce
knowledge.
But the development momentum is also reinforced by the fact that technological
fields of application are expanding equally quickly. The need for digital systems
appears inexhaustible, because they contribute to improvements in nearly every area
regarding performance, quality and (resource) efficiency of products and processes.
The developmental leap is so all-encompassing that we can justifiably speak of a
“digital revolution”.
Machines need instructions in order to function. For simple processes this can be
achieved by manual operation, but this no longer meets the demands of modern
production machinery and facilities. Numerous sensors provide vast quantities of
data which the machine stores and interprets and responds according to rules stored
in a digital code. Such systems add up the knowledge of their developers, the results
of Machine Learning and current data.
Biological organisms, too, gather numerous data from their environment and
interpret it. The blueprint for their construction – the genetic code found in each and
every cell – represents the aggregated knowledge gathered over the whole course
of the organism’s evolution.
Ideally, for both organisms and machines, the information collected and stored
holds the answers to every conceivable challenge. For this reason, and in spite of
numerous differences at the level of detail, the digital code’s function and operation
is reminiscent of that of the genetic code of biological systems. In each case:
• The code contains the information for the actions, reactions and characteristics
of the organisational unit, whether it is a cell or a machine.
• Complex information is compressed into a comprehensive sequence of a limited
number of ‘characters’. DNA manages with four characters, the nucleotides
adenine, guanine, thymine and cytosine; the digital code uses just two, the num-
bers 0 and 1.
• These codes can be used to store not only the construction and behavioural
frameworks of smaller units, but of overall structures, too – the structures of
entire organisms in the case of the genetic code, and of a production plant or
factory in the case of the digital code. Both systems have at their core the capac-
ity for flexibility and learning.
1 Digital Information – The “Genetic Code” of Modern Technology 3
• The stored information can be copied limitless and practically lossless. Identical
DNA replication is done by splitting of the double helix and completing the free
binding sites by attaching new complementary DNA-bases.
• Digital information copying happens through lossless reading and renewed re-
cording of the information on a storage medium.
• The information is retained during duplication. It may however be modified if
there is a need for adaptation: In the case of the genetic code, this takes place via
mutation, for example, or through the recombination of existing partial informa-
tion with subsequent selection; in the case of the digital code, parts of the scripts
can be replaced or expanded.
In this way, both the digital and genetic codes carry the potential for conservation
as well as innovation, and both are combinable – changes can thus build on already
existing achievements. This is the explanation why digitization has led to such a
boost in innovation. It leverages the potential of evolutionary progress through re-
search and development onto an unprecedented level in the field of technology.
What reinforces the effectiveness of the digital code as a driver of evolution and
innovation even further – in comparison to the genetic code – is the fact that on the
one hand new digital information can be integrated in a very targeted manner and,
on the other hand, be transported worldwide in real time via the Internet. Evolution-
ary improvements of technologies can thus happen quickly and be made available
immediately, hindered only by restrictions of patent law and politics.
the other end of the world upon the click of a button. Efficient and flexible produc-
tion technologies allow for individual product design and the manufacturing of
many products via personal 3D printers.
The decentralized production of content or goods, as mentioned above, is a re-
markable side-effect of digitization. The fast publication of individually created
books, images, films, music, objects, ideas and opinions – often without control and
censorship – has become commonplace through the Internet. This creates new and
rapid opportunities for commercialization and self-actualization, but also leads to
societal dangers that we still must learn to deal with.
These new possibilities also generate expectations and demands: We, for exam-
ple, not only get accustomed to the convenient status quo, but also to the dynamics
of development. Tomorrow we expect even more from what today’s digitally-based
products and media already offer. And that means that the international markets for
technological products and the technologies themselves are subject to the same
massively increased dynamics of change and growth.
This is also the reason why digitization, far from having a depressive effect on
the job market, is instead a driver of it. In contrast to the commonly-held fear that
digital technology is destroying jobs, it has first and foremost led to changes in re-
gard to job profiles, and even led to an overall increase in job- and income oppor-
tunities. Current and future employees are expected to be more flexible, have the
willingness to learn and possibly also show a certain degree of professional curios-
ity. Companies have to respond faster and more flexibly to market changes, both
with new products but also with new, disruptive business models. They need to
anticipate the future requirements of the market so they can offer the right products
at the right time.
Lifelong learning has become an indispensable reality for everyone who partic-
ipates in the economic process, and digitization – along with the associated dynam-
ization of developments – is the principal cause.
Because of digital technology, nearly all areas that are essential for business, sci-
ence, public or private life are nowadays controlled technologically: security,
healthcare, energy supply, production, mobility, communications and media. The
more areas we leave to the control of data technology, the greater the importance
for its reliability. This applies to individual systems such as cars and airplanes as
well as to complex structures like supply systems and communications networks.
With the advance of digitization, resilience, i.e. the ability of a system to continue
1 Digital Information – The “Genetic Code” of Modern Technology 5
2.1 Introduction
For most readers of this book, digitization has already become a natural part of their
everyday life. In essence, “digitization” refers to the binary representation of texts,
images, sounds, films and the properties of physical objects in the form of consec-
utive sequences of 1s and 0s. These sequences can be processed by modern com-
puters at exceptionally high speeds – billions of commands per second.
Digitization acts as a kind of “universal translator”, making data from various
sources processable for the computer and thus offering a range of possibilities that
would otherwise be unthinkable. These include, for example, carrying out complex
analyses and simulations of objects, machines, processes, systems, and even human
bodies and organs – as is the case with the 3D body reconstruction technology re-
alized by the Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute,
HHI. Digitized data from sensor captured brain signals can even be used to control
computers and robots. The inverse is now also possible, using digital signals to
produce haptic sensations in prosthetics. Here, digitization acts as a direct link be-
tween the biological and cyber-physical worlds.
Although digitization can be used in nearly all fields, this introductory chapter
focuses on the following areas of application, with especially significant impacts on
people’s lives:
• Data analysis and data transfer
• Work and production
• Security and supply
© Springer-Verlag GmbH Germany, part of Springer Nature, 2019
Reimund Neugebauer, Digital Transformation
https.//doi.org/10.1007/978-3-662-58134-6_2
9
10 Sophie Hippmann • Raoul Klingner • Miriam Leis
Also in medicine the digitization of data – e.g. of medical images, textual informa-
tion or molecular configurations – plays a considerable role. Effective, safe and
efficient processes for analyzing large datasets play an important role here. The
analysis of multimodal data – i.e. computer-aided comparison of different images
(x-rays, MRIs, PETs, etc.), in combination with lab results, individual physical
parameters and information from specialist literature – can pave the way for more
accurate diagnoses and personalized treatments that cannot be achieved by any
specialist alone to this extent. Digitization in medicine is a specialist field of the
Fraunhofer Institute for Medical Image Computing MEVIS. Here, as in other fields,
big data needs to be transformed into smart data. Likewise, digital assistants are
expected to gain the ability to automatically recognize supplementary, missing or
contradictory information, and to close information gaps through targeted enquiries.
The immense spread of digitization has led to an exponential growth in data volume
so that despite capacity increases, data transfer bottlenecks still occur. Some projec-
2 Digitization – Areas of Application and Research O
bjectives 11
tions predict that the volume of data produced globally could multiply tenfold by
2025, in comparison to 2016, reaching as much as 163 zettabytes (a 163 followed
by twenty-one zeros, equal to forty-one thousand billion DVDs).
Today, music and video streaming make up a large part of global data transfers.
Lossless data compression thus has an important role to play in reducing the size of
digital data packages, thereby shortening transfer times and requiring less memory.
Since the groundbreaking development of MP3 coding based on research by the
Fraunhofer Institute for Integrated Circuits IIS and Fraunhofer Institute for Digital
Media Technology IDMT, Fraunhofer has been constantly continuing to further
develop audio and video data compression processes, in order to achieve ever better
transmission quality with the lowest possible data volumes. Uncompressed audio
files, for example, are up to ten times larger than MP3 files with identical sound
quality.
The Fraunhofer IIS also played a key joint role in another accomplishment in the
field of audio technologies – digital radio – having developed its basic, broadcasting
and reception technologies. Digital radio offers huge advantages over the analog
process. Terrestrial wireless reception can be received even without Internet con-
nection and is free of charge. Energy-efficient transmission, interference-free recep-
tion, high sound quality, the option of additional data services and extra space for
more broadcasters as well, are among its clear benefits. Digital radio also allows for
real time communication of traffic and disaster alerts, with high reliability and
reach, even when no Internet connection is available. In Europe, digital radio will
largely replace analog systems over the coming years. Even in emerging economies
such as India the transition is in full swing.
Fast data transfer with minimal delays or latency, often referred to as the “tactile
Internet”, is the basis for a range of new technological applications. These include:
connected machines; autonomous vehicles and objects that can communicate with
people and each other in real time; augmented reality applications that feed in up-
to-the-minute updates; or specialists who are able to carry out highly complex
surgery safely from the other side of the world via telerobots. In the future, global
12 Sophie Hippmann • Raoul Klingner • Miriam Leis
data usage will shift from video and audio data to (sensor) data from industry, mo-
bility and the “Internet of things”. The new 5G cellular communications standard
– with Fraunhofer playing a key role in promoting its development, testing and
distribution – promises data transfer rates of 10 Gbp per second, ten times faster
than today’s 4G/LTE. The requirements for the industrial Internet are high, demand-
ing above all scalability, real-time capabilities, interoperability and data security.
Work on developing and testing new technologies for the tactile Internet is on-
going at the four Berlin-based transfer centers – the Internet of Things Lab (Fraun-
hofer Institute for Open Communication Systems FOKUS), the Cyber Physical
Systems Hardware Lab (Fraunhofer Institute for Reliability and Microintegration
IZM), the Industry 4.0 Lab (Fraunhofer Institute for Production Systems and Design
Technology IPK) and the 5G Testbed (Fraunhofer Institute for Telecommunications,
Heinrich Hertz Institute, HHI). In addition to 5G technologies, cloud computing and
edge computing also play a key role here. With the latter, a large part of computing
is performed within the individual machines and sensor nodes, thus reducing laten-
cy, since not all data has to be processed in the cloud.
Digitization has fundamentally changed our working world and will continue to do
so. Today, email (or chat) has nearly completely replaced the classical letter for
day-to-day written communications. Nowadays engineers design prototypes on the
computer instead of the drawing board, and in future robots and interactive digital
assistance systems will stand by our sides to help with everyday tasks. Technologies
for recognizing speech, gestures and emotions enable people to communicate with
machines intuitively. Discoveries from neuroscience are also helping to identify
what people are focusing on when they use machines, thus helping with the devel-
opment of better, safer and more user-friendly designs and interfaces. The Fraun-
hofer Institute for Industrial Engineering IAO, for example, is analyzing what hap-
pens in the brains of users of technological devices in order to be able to optimize
interaction interfaces. In the workplace of the future, interactive cooperation be-
tween people and machines will develop further, while humans nevertheless will
increasingly take center stage. This will not only change production work but will
also enable new processes and services; research in this direction is ongoing at the
Fraunhofer IAO’s Future Work Lab.
2 Digitization – Areas of Application and Research O
bjectives 13
The connection between the digital and physical worlds becomes especially appar-
ent with additive manufacturing, also known as “3D printing”. This allows for the
transfer of information about the configuration of objects via computer, so that the
object can be “materialized” in another location. Similar to the way how living
creatures obtain their physical form from the information stored in their DNA, ad-
ditive manufacturing automatically instantiates objects physically on the basis of
digital data.
The diversity and quality of materials and the resolution are constantly improv-
ing in additive manufacturing. This process facilitates a relatively inexpensive pro-
duction of prototypes, spare parts, customized products or tailored personal pros-
thetics. Benefits of additive manufacturing processes also include efficient use of
materials (since the object is constructed directly from a mass of material rather than
being “carved” from an existing piece by removing material), and the ability to
produce very small or complex geometries. Additive manufacturing processes are
thus going to be firmly integrated into Industry 4.0 concepts. The Fraunhofer Insti-
tute for Electronic Nano Systems ENAS, together with other Fraunhofer institutes,
is addressing these challenges in a major large-scale project.
14 Sophie Hippmann • Raoul Klingner • Miriam Leis
Data is the DNA and fuel of our modern world. As in the example of additive manu-
facturing, information about the structure and composition of matter gives the object
its function and thus its value. Anyone with the data for a 3D model could in principle
produce it anywhere. Similar is the case of pharmaceuticals: they are made of common
atoms, but their structure – the arrangement of the atoms – is key to their efficacy. Still,
the correct configuration and mode of synthesis can take years to identify. Anyone
who possesses and controls relevant data and information holds a competitive advan-
tage. The Fraunhofer Big Data and Artificial Intelligence Alliance helps to unearth
such data treasures, without losing sight of quality and data protection issues.
The more complex a technological system is, the more susceptible it is to mal-
function and the ensuing effects. This is why complex technological systems have
to be tuned for resilience; that is, they need to be able to resist malfunctions and, in
the case of damage, still be able to function with a sufficient degree of reliability.
2 Digitization – Areas of Application and Research O
bjectives 15
IT-, data- and cybersecurity are essential to the functioning of digital societies. Key
issues include the prevention of unauthorized access to and manipulation of data
and data infrastructures as well as securing personal and person-specific data. When
it comes to cars (which may now contain more lines of code than an airplane), en-
ergy and water provision (managed by computers), highly-networked Industry 4.0
facilities or smart homes, cyber threats can cause serious problems and thus need to
be identified and defended against early on.
Digital system outages can lead to the breakdown of entire supply networks,
affecting power, mobility, water or food distribution, among others. The Fraunhofer
Institutes for Secure Information Technology SIT and for Applied and Integrated
Technology AISEC do not only provide technological innovations for the early
detection of potential cyber threats with the aid of Machine Learning, but also cy-
bersecurity courses and training labs.
For the sake of security is it important that IT- and cybersecurity applications are
designed to be user-friendly and easy to use. If their usage is too complex and labo-
rious then they will not be used at all, thus increasing the risks. This is why the
Fraunhofer Institute for Communication, Information Processing and Ergonomics
FKIE is investigating ways to maximize user-friendliness of information technolo-
gy and cybersecurity systems and how they can be designed to be as ergonomic as
possible. The new “Usable Security” research project aims at extending the current
limits of computer system usability. Technology should be people-centered, not the
other way around, as has often been the case so far. Only when people are in the
2 Digitization – Areas of Application and Research O
bjectives 17
Pulling back from digitization is unthinkable for a high-tech society, and would
likely be tantamount to a catastrophe. But there are still a lot of new applications
ahead of us: automated driving, cooperative robots and assistance systems, telemed-
icine, virtual reality, and digital public services. The Fraunhofer Institutes for Ma-
terial Flow and Logistics IML and for Transportation and Infrastructure Systems
IVI are developing driver assistance systems for safe and reliable automated driv-
ing, in the fields of road traffic, agriculture and logistics.
The continued progress of the digital world also poses challenges like the pro-
tection of digital data and infrastructures; efficient, the effective and intelligent
management of big data; faster data transfer and reduction of latency as well as the
further development of processor technologies and computation methods.
The next developmental phase may be characterized by the linking of digital and
biological concepts, since genetic and binary codes are similar. Learning-enabled
robotic systems, swarm intelligence in logistics, biosensors, 3D printing, and pro-
grammable materials already all point towards this direction. The Fraunhofer-Ge-
sellschaft is dedicated to innovation and solutions for challenges in order to improve
and drive forward the process of digitization, with humans always remaining at the
center.
Virtual Reality in Media and Technology
Digitization of cultural artifacts and industrial
3
production processes
Summary
Virtual and Augmented Reality technologies have by now become established
in numerous engineering areas of application. Also in the cultural and media
fields interactive three-dimensional content is being increasingly made availa-
ble for information purposes, and used in scientific research. On the one hand,
this development is accelerated by current advances in smartphones, tablets
and head-mounted displays. These support complex 3D applications in mobile
application scenarios, and enable us to capture our real physical environment
using multimodal sensors in order to correlate it with the digital 3D world. On
the other hand, new automated digitization technologies such as CultLab3D
of the Fraunhofer Institute for Computer Graphics Research IGD allow the
production of the necessary digital replicas of real objects, quickly, economic-
ally and of high quality.
To allow the best possible retention and documentation of our shared cultural her-
itage, digital strategies were formally established at the political level worldwide.
New initiatives such as the iDigBio infrastructure or Thematic Collections Net-
works in the USA promote the advanced digitization of biological collections. EU
19
20 Dieter W. Fellner
member states have also been called on by the European Commission to bolster their
digitization efforts. This executive priority is part of the Digital Agenda for Europe
and underlines the necessity of facilitating large-scale improvements of the online
accessibility of historical cultural artifacts [1]. The Digital Agenda identifies the
long-term retention of our cultural heritage as one of the key initiatives of the Eu-
rope 2020 strategy, with the goal to provide improved access to culture and knowl-
edge via improved usage of information and communications technologies [2].
These initiatives are closely tied to Article 3.3 of the European Union Treaty of
Lisbon [3] which guarantees that “Europe’s cultural heritage is safeguarded and
enhanced”.
Despite the legal framework conditions that recognize its significance for soci-
ety, our cultural heritage is at risk of all sorts of dangers. Recently, a range of nat-
ural and man-made catastrophes have highlighted just how fragile our cultural
heritage is. Events such as the deliberate destruction of the ancient Semitic city of
Palmyra in Syria or of archeological finds in the museum of Mosul, Iraq underscore
the necessity for new and faster methods of documentation, leading to a reapprais-
al of high-definition facsimiles. Moreover, the fact that only a small proportion of
all artifacts in exhibition collections is publicly accessible provides further moti-
vation for improving access to information about our cultural heritage [4]. Innova-
tive documentation methods for cultural heritage items are thus gaining ever-in-
creasing significance. This arises, on the one hand, from the desire to enable im-
proved access to unique items, to make collections more easily accessible for re-
search purposes or to a wider audience for example; and on the other hand, from
the latent threat of losing them forever through catastrophes or other environmen-
tal factors.
Against this backdrop, and in times of digital transformation, the use of 3D
technologies in the cultural sphere is becoming increasingly important. This is
because they offer a so far unexploited potential use, whether for documentation
or retention purposes, for innovative applications in fields as diverse as education
and tourism, for optimal accessibility and visualization of artifacts, or as a basis for
research and conservation. They also enable physical copies to be produced as a
result of highly precise 3D models. The increasing demand for information and
communications technologies demonstrates the growing need for research across
the whole value chain, from digitization and web-based visualization through to
3D printing.
This research produces tools that stimulate the development of new technologies
for digitally processing and visualizing collections, enabling the safeguarding of
cultural heritage. Today, the digital capture of two-dimensional cultural treasures
such as books, paintings or “digitally-born” collections such as films, photos and
3 Virtual Reality in Media and Technology 21
In order to make collection items accessible to various user groups in 3D, too, and
meet the need of easy-to-use, fast and thus economical 3D digitization approaches,
the Fraunhofer Institute for Computer Graphics Research IGD is developing the
CultLab3D modular digitization pipeline [5]. This enables three-dimensional ob-
jects to be digitized in 3D with micrometer accuracy, via a completely automated
scanning process. This is the first solution of its kind to take account of the aspect
of mass digitization. The intention is to increase speeds in order to reduce the cost
of 3D scans by between ten- to twentyfold. The project is also striving for true to
original reproduction at high quality, including geometry, texture and the optical
properties of the material.
The modular scanning pipeline (see Fig. 3.1) currently consists of two scanning
stations, the CultArc3D and CultArm3D. The digitization process is fully automat-
ed, using industrial materials handling technology and autonomous robots to convey
items to the corresponding optical scanner technologies. By decoupling the
color-calibrated capture of an object’s geometry and texture from its final 3D recon-
struction via photogrammetry, the scanning pipeline achieves a throughput of just
5 minutes per object. Additionally, with sufficient computing power, a final 3D
model can be produced every five minutes. In most cases, little or no post-process-
22 Dieter W. Fellner
ing is required. The scanning pipeline currently digitizes items of 10–60cm in height
or diameter. The entire scanning pipeline is operated by a tablet PC, which all of the
system’s components log in to.
CultArc3D
The CultArc3D can either be operated alone or in conjunction with other scanners.
The module captures both geometry and texture as well as optical material proper-
ties according to previous works [6][7]. During the digitization process, a conveyor
belt fully automatically moves the objects to be scanned through the CultArc3D, on
glass tablets.
The CultArc3D (see Fig. 3.2) consists of two coaxial semicircular arches. These
turn around a common axis. Both arches cover a hemisphere that is centered on an
object placed in the midpoint. Each arch is moved by its own actuator so that a
discrete number of stop positions is possible. The arches have different radiuses so
they can be moved independently. The outer arch (hereinafter referred to as the
“camera arch”) holds nine equidistant cameras. Nine additional cameras beneath
the object conveyor surface capture the artifact’s underside.
3 Virtual Reality in Media and Technology 23
Fig. 3.2 CultArc3D: two hemispherical rotating arches, one with cameras, the other with
ring light sources (Fraunhofer IGD)
As with the outer camera arch, the inner arch (hereinafter “light arch”) includes
nine light sources mounted equidistantly. At the moment, all objects are captured in
the visible light spectrum. However, for capturing more optically-complicated ma-
terial, multispectral sensors or laser sensors can easily be integrated into the system,
as can volumetric data capture sensors, x-ray tomography or MRI. One additional
strength of the CultArc3D is the capture of optical material properties. To this end,
both arches can be positioned anywhere in relation to one another, allowing every
possible combination of light direction and photographic perspective for capturing
an object’s upper hemisphere.
CultArm3D
Few fields present objects as varied in material and form or demands as high in
terms of the precision and color-accuracy of digital replicas as the cultural heritage
field. In order to guarantee the complete and accurate 3D reconstruction of any
object it is thus important to position the scanner carefully with regard to the object’s
surface geometry and the scanner’s measuring volume. The CultArm3D (see
Fig. 3.3) is a robotic system for this kind of automated 3D reconstruction. It consists
24 Dieter W. Fellner
Fig. 3.3 CultArm3D: lightweight robotic arm with a 24-megapixel camera mounted to
the end effector next to a turntable with a white two-part folding studio box
(Fraunhofer IGD)
of a lightweight robotic arm with a camera on its end effector and a turntable for the
object to be scanned. It can operate either independently or in conjunction with the
CultLab3D; in independent mode it is capable of capturing geometry and texture
completely and independently.
The robotic arm selected has five degrees of freedom and is able to stably move
loads of up to 2kg. It is collaborative and safe to operate in the presence of people,
within a quasi-hemisphere of around 80cm. The robotic arm is equipped with a
high-resolution (24 megapixel) camera and is synchronized with a diffuse lighting
structure (static soft box setup). A two-part white studio box on the turntable ensures
a consistent background, which is important for proper photogrammetric 3D recon-
struction and the avoidance of incorrect feature correspondences.
Together with the CultLab3D, the CultArm3D serves as a second scanning sta-
tion to the CultArc3D, which is the first scanning station on the conveyor. Unlike
the CultArc3D with its large field of view on a fixed hemisphere around the object,
3 Virtual Reality in Media and Technology 25
the camera optics of the CultArm3D system are optimized for adaptive close-up
detail images. These views, based on the 3D preview model of the first CultArc3D
scan, are then planned so that additional features of the object will be captured. In
this way, any remaining gaps or undercuts can be resolved and the quality and com-
pleteness of the scan can be improved locally. The number of additional planned
views to meet specific quality criteria largely depend upon the complexity of the
object’s surface and texture. The automated planning of different numbers of views
and the associated dynamic scanning time, depending on the complexity of the
object, result in high throughputs of the CultLab3D digitization pipeline. When
operating the CultArm3D in standalone mode separately from the CultLab3D as-
sembly, the first round of scanning is carried out using generic camera angles that
are only planned relative to the object’s external cylindrical measurements such as
height and diameter.
Given the limited workspace of the lightweight robotic arm, the task of capturing
any kind of object up to 60cm in height and diameter is demanding. In some cases,
camera views planned to provide optimal scan coverage and quality cannot be cap-
tured exactly, for reasons of safety or reachability. Nevertheless, in most cases it is
possible to identify slightly modified, practical views (or series of views) that equal-
ly contribute to the quality of the scan.
of the 3D models with data from the broadest possible range of scans including CT,
ultrasound, confocal microscopy, terahertz imaging, and mass spectroscopy. This
approach enables objects to be comprehensively investigated in their entirety, inside
and out, and thus opens up new possibilities for monitoring, analysis and virtual
presentation extending beyond the cultural sphere.
Fig. 3.5 3D reconstruction of a replica of a Chinese warrior; CultLab3D also captures the
undersides of objects (Fraunhofer IGD)
3 Virtual Reality in Media and Technology 27
Today, Virtual andAugmented Reality technologies (VR/AR) are scaled for different
computing capacities, operating systems and input and output options, from com-
plete cloud infrastructures right through to head mounted displays. Along with
scalability and platform diversity, mobile systems also bring completely new secu-
rity requirements with them, since confidential data should be transmitted wireless-
ly, or CAD data should be visualized on mobile systems but not saved to them.
Virtual and Augmented Reality technologies will only fulfill these requirements in
future if they are based on web technologies, which are platform-independent and
optimized for security. Against this backdrop, current research and development
work in the VR/AR fields is closely linked to web technologies that provide enor-
mous benefits, particularly for industrial uses.
Virtual Reality has been successfully utilized in European industry to make digital
3D data tangible for over 25 years. In the automotive industry and in aircraft and
shipping construction in particular, digital mock-ups (DMUs) are replacing real and
physical mockups (PMUs) in numerous areas of application.
Industrial VR applications use general manufacturing and DMU data for digital
validation processes (e.g. assembly/development validation) in spite of the relative-
ly high hardware and software costs of VR solutions.
integration capability
industrial
VR
applications
VR
media
dynamic configurability
Fig. 3.7 “Input data sensitive” visualization issues with current standard VR processes;
the image refresh rate increases with the size of the data, an unacceptable situation for VR/
AR application. (Fraunhofer IGD)
Over the last five years, overall improvements in LCD and OLED technology
have reduced the costs for VR headset to just a few hundred Euros. The associated
VR games and media are available in the established online stores (e.g. Stream) and
are undergoing a fierce pricing war. For industrial VR applications, no universal
solutions have yet become established since the applications require a significantly
higher capability of being integrated and dynamically configured. Classical solu-
tions from the gaming market are not directly transferrable since the processes and
data need to be adapted via manual configuration processes.
Dynamic configurations require the “smart” fully-automated adaptation of pro-
cesses in order to be able to cope with the ever-widening distribution and explosion
of data volumes. Current standard systems for industrial 3D data visualization use
established document formats, and the processes used are “input data sensitive”. As
the volume of data grows, so does the visualization overhead.
Despite the constant increase in the processing power of modern graphics hard-
ware, visualization of massive 3D data sets with interactive frame rates cannot be
solved simply by increasing graphics memory and rasterization speeds. Problems
such as visibility calculations for scene elements have to be accelerated by spatial-
ly hierarchical data structures. Instead of processing the highly complex geometry
of individual elements in a scene, these data structures focus only on their bounding
volume. Here the “divide and conquer” approach applies, based on recursive anal-
ysis of the tree structures produced.
Various concepts from different lines of research are available for structuring
these kinds of hierarchies, for example collision detection or ray tracing. Whereas
binary k-d trees, for example, split the allocated space in each node with a hyper-
plane, the structure of a bounding volume hierarchy is based on the recursive group-
ing of the bounding volumes of the elements in the scene. Hybrid processes such as
bounding interval hierarchies connect these approaches in order to combine the ad-
vantages of each. These irregular approaches can be contrasted with structures such
as regular grids or octrees where the space division strategy largely or completely
3 Virtual Reality in Media and Technology 29
Fig. 3.8 Smart Spatial Services (S3) use a constant frame rate for globally budgeting visi-
bility calculations and image generation. As with film streaming, a constant frame rate is
the goal. (Fraunhofer IGD)
ignores the density of the space being divided. These different approaches are con-
tinuously being developed and optimized within the various fields of research.
The uniqueness of so-called “out-of-core technologies” lies in the fact that they
do not save all their data in the main memory but load it on demand and on the fly
from the secondary storage. In contrast to the automated virtual memory manage-
ment of operating systems—which are normally not influenced by applications—
programs here require complete control of memory management and data caching.
Ideally, applications are supported in this by output-sensitive algorithms.
The focus of visualization here is less on 100% accuracy and/or completeness of
the representation and more on guaranteeing a certain level of desired performance
in terms of hard real-time requirements.
This can be achieved by the use of spatial and temporal coherence. Whereas
spatial coherence is represented via the hierarchical data structures, temporal cohe-
sion across frames has to be captured within the render algorithm. Hidden surface
determination of the elements within the scene (“occlusion culling”) may for exam-
ple be achieved via algorithms such as coherent hierarchical culling.
In the Industry 4.0 context, simulation and production processes are parallelized
with the aim of guaranteeing optimal production quality via the comparison of
target and actual outputs (cyber-physical equivalence). In the same way, the digital
twin is used to feed the actual data back into an agile production planning process.
Augmented Reality (AR) processes are pertinent here for registering the target/
actual differences in real time and visualizing them in superimposition to the cap-
tured environment. Augmented Reality processes have proven themselves in nu-
merous areas of application in this context and are already finding applications in
routine planning and monitoring processes:
30 Dieter W. Fellner
• Gesture-based interaction
Gesture-based interaction via the HoloLens is designed for Augmented Reality
games and is not always suitable for industrial applications. For this reason, it
should be possible to control the interaction with a tablet while the Augmented
Reality visualization is executed on the HoloLens.
These restrictions in using HoloLens can only be compensated for if the algorithms
are distributed across client server infrastructures. The key here is that the complete
3D models – which represent significant IP in industrial organizations – do not leave
the PDM system and are saved exclusively on the server. Then the client only dis-
plays individual images transmitted via video streaming technologies. Alternative-
ly, G-buffers are streamed, processed for rendering the current views. VR/AR tech-
nologies will only be able to meet these demands in future if they are built on web
technologies and service architectures. The development of VR/AR technologies
based on service architectures is now possible, as libraries are now available such
as WebGL and WebCL, which facilitate powerful and plugin-free on-chip process-
ing on the web browser. Web technologies as a basis for VR/AR applications in
particular offer the following benefits for industrial applications:
• Security
If VR/AR applications are run as web applications on the user’s end device
(smartphone, tablet, PC, thin client), then, in the best case scenario, native soft-
ware components – which always entail potential insecurities – do not need to
be installed at all.
• Platform dependence
On the whole, web technologies can be used independent of platform and with
any browser. In this way, platform-specific parallel development (for iOS, An-
droid, Windows etc.) can be avoided as far as possible.
3 Virtual Reality in Media and Technology 33
One example of industrial relevance in this context is the linking to the PDM system
that centrally manages and versions all the relevant product data (e.g. CAD data,
simulation data, assembly instructions). Central data storage is designed to ensure
that the most current and suitable data version is always used for planning and de-
velopment processes. In the field of VR/AR applications for target/actual compar-
isons in particular, correct versioning needs to be guaranteed. Using service archi-
tectures, VR/AR applications can be created that pull the current version from the
PDM system when the application is started and, during data transfer, code it into
geometric primitives that can be visualized in the web browser. To do this, 3D data
is organized into linked 3D data schemas, which permit flexible division and use of
the data in the service architecture.
Fig. 3.11 Left: 3D data from CAD systems is usually depicted as a linked hierarchy of
resources; right: from the point of view of the application, the 3D data is made up of a
linked network of 3D data. (Fraunhofer IGD)
34 Dieter W. Fellner
A typical PDM structural description may, for example, be mapped in the stand-
ardized STEP AP242/PLM XML formats, which reference geometries that can
themselves be saved in the JT format. (These in turn reference additional JT sub-
components and may contain additional 3D data.) Development work is currently
being carried out at Fraunhofer IGD on the instant3Dhub web-based service infra-
structure for visualizing optimized data. Here, 3D data is stored in a linked 3D data
network (“linked 3D data” or “L3D”) that provides the complete structure and ge-
ometry transparently to all resource formats and is scalable and extensible via ad-
ditional links (see Fig. 3.11 right).
Conversion between the resource description and the linked 3D data takes place
via a transcoding process. Since the infrastructure was designed for fast, adaptive
data reload it is equipped with corresponding delivery strategies. Here, the data is
stored in a distributed 3D cache that is populated via transcoding jobs and transmit-
ted via regular services.
The client-side application controls the display of the linked 3D data. instant3D-
hub provides this access via a client-side API: webVis JavaScript API for integration
into a browser application. The API communicates interactions and changes in
camera position, initiating server-side visibility analysis and new links.
When the data is accessed, application-managed transcoding happens transpar-
ently; that is, the client only uses the linked 3D data, and the associated cache entry
is only automatically generated on initial access. In order to speed up the access, the
data can be completely transcoded at the point of publication, through successive
access to the corresponding entries.
Fig. 3.12 Client/server architecture for 3D components: the client-side application uses
the webVis API to access cache entries. The instant3Dhub service manages the transmissi-
on of cache entries and/or any transcoding required. A data gateway provides the connec-
tion to original 3D resources. (Fraunhofer IGD)
3 Virtual Reality in Media and Technology 35
Individual cache entries are identified via a naming scheme using Uniform Re-
source Name (URN) coding, so that they can be permanently identified on the ap-
plication side regardless of location. The advantage of the 3D data network is that
the storage location for data delivery can be changed without also changing the
client application.
On the client side, the 3D components consist of a JavaScript application running
on the web browser. The client-side JavaScript library, for managing the 3D com-
ponents (webVis), offers a lightweight API for
• Adding or removing structural elements to and from the scene displayed,
• Reading and changing properties,
• Carrying out measuring functions, and
• Setting the visibility of (sub-)components.
When adding an element higher up the hierarchy, the entire hierarchy beneath is
automatically displayed. Properties that may be changed include visibility (for
showing or hiding elements) and color, for example. Alongside the direct 3D view,
additional UI components (as web components) such as a tree view of the data
structure or a bar for saving and loading current views (“snapshots”) are also avail-
able.
The 3D component responds to user inputs (e.g. mouse clicks on a visible ele-
ment of the 3D view) via events, i.e. a callback method is registered for the selected
36 Dieter W. Fellner
ronments. Indeed, this also raises the question of how CAD models can be efficient-
ly distributed to the output units for AR applications such as tablets, smartphones
or HoloLenses, and how reference data can be generated for individual client appli-
cations.
off-screen 3D-rendering
image processing
model-based tracking
Fig. 3.14 In model-based tracking, edge shaders are used to render hypotheses from the
CAD data that are then compared with the camera images and applied to the edge detec-
tors. (Fraunhofer IGD)
38 Dieter W. Fellner
not require user interaction for initialization and have no drift. The CAD geometry
is permanently geared to the identified geometry. In addition, CAD nodes – which
can be separated from one another in the structure tree, for example—can be tracked
independently of one another. These are precisely the properties that are fundamen-
tal for Industry 4.0 scenarios in the area of target/actual comparisons and quality
control.
Alongside the continuous tracking of the objects (frame-to-frame), the mod-
el-based tracking processes need to be initialized. During initialization, the real and
virtual 3D objects are converted into a shared system of coordinates. To do this,
initial camera positions are determined from which the initialization can be carried
out. These positions for initialization are specified using the 3D models.
Just like the visualization functionalities, the tracking services are integrated into
the instant3Dhub infrastructure. This enables new forms of efficient Augmented
Reality usage to be implemented, via the same philosophy of data preparation and
transmission. To this end, services are implemented for inferring and producing
tracking reference models from CAD data that are transmitted via L3D data con-
tainers (see Fig. 3.15). Incorporating this into the instant3Dhub infrastructure will
Fig. 3.15 Incorporation of tracking infrastructure into the VCaaS (Fraunhofer IGD)
3 Virtual Reality in Media and Technology 39
enable new forms of load balancing for tracking tasks (distribution across client-side
or server-side processes). However, the most important benefit of using the instant-
3Dhub infrastructure is that the reference models for the model-based tracking are
generated directly from the CAD data at full resolution. This completely removes
the need for expensive model preparation and reduction. To allow this, the instan-
t3hub infrastructure provides efficient server-based rendering processes for large
CAD models.
The model-based tracking processes in the instant3Dhub infrastructure use off-
screen rendering processes to render the objects for tracking from the current cam-
era position (tracking hypotheses). These off-screen renderings are intended to be
used not only for tracking but also for identifying the state of construction. Which
components have already been used in the current state, and which have yet to be
added? These states of construction are to be displayed in screenshots and recoded
each time for the tracking reference model.
With the help of the instant3Dhub infrastructure, the tracking service can be used
in the following configurations according to parameters (network quality, client
system performance, etc.):
• Hybrid tracking
In this approach, image processing takes place on the client, while the tracking
hypotheses are rendered on the server. The advantage of this approach is that the
elaborate image processing is executed on the client, enabling minimal tracking
latency. In this approach, however, a native tracking component must be installed
on the client. The advantage of using the server as opposed to a purely native
process is that the model data does not need to be saved on the client and the
models for the model-based tracking do not need to be reduced.
Fig. 3.16 Incorporation of tracking infrastructure into the VCaaS (Fraunhofer IGD)
40 Dieter W. Fellner
• Server-side tracking
For a server-side tracking process, the video data needs to be transmitted to the
server infrastructure with minimal latency. The image processing takes place on
the server and the camera positions calculated need to be transferred back to the
client at a high frequency. This requires a video streaming component that must
be implemented via a dedicated transmission channel (e.g. WebSockets). Serv-
er-based tracking requires a very good network connection and implementation
in an asynchronous process so that positions can also be calculated asynchro-
nously if there are bottlenecks in the communications infrastructure.
[7] Noell T., Koehler J., Reis R., and Stricker D. (2015): Fully Automatic, Omnidirectional
Acquisition of Geometry and Appearance in the Context of Cultural Heritage Preserva-
tion. Journal on Computing and Cultural Heritage – Special Issue on Best Papers from
Digital Heritage 2013 JOCCH Homepage archive, 8 (1), Article No. 2.
[8] Behr J., Jung Y., Franke T., Sturm T. (2012): Using images and explicit binary container
for efficient and incremental delivery of declarative 3D scenes on the web, in ACM
SIGGRAPH: Proceedings Web3D 2012: 17th International Conference on 3D Web
Technology, pp. 17–25
[9] Brüderlin B., Heyer M., Pfützner S. (2007): Interviews3d: A platform for interactive
handling of massive data sets, in IEEE Comput. Graph. Appl. 27, 6 (Nov.), 48–59.
[10] Behr J., Mouton C., Parfouru S., Champeau J., Jeulin C., Thöner M., Stein C., Schmitt
M., Limper Max, Sousa M., Franke T., Voss G. (2015): webVis/instant3DHub – Visual
Computing as a Service Infrastructure to Deliver Adaptive, Secure and Scalable User
Centric Data Visualisation, in: ACM SIGGRAPH: Proceedings Web3D 2015 : 20th
International Conference on 3D Web Technology, pp. 39-47
[11] Nvidia grid: Stream applications and games on demand. http://www.nvidia.com/object/
nvidia-grid.html
[12] Limper M., Thöner M., Behr J., Fellner D. (2014): SRC – A Streamable Format for Ge-
neralized Web-based 3D Data Transmission, in: ACM SIGGRAPH: Proceedings Web3D
2014 19th International Conference on 3D Web Technology, pp. 35-43
Video Data Processing
Best pictures on all channels
4
Dr. Karsten Müller · Prof. Heiko Schwarz · Prof. Peter Eisert ·
Prof. Thomas Wiegand
Fraunhofer Institute for Telecommunications,
Heinrich-Hertz-Institute, HHI
Summary
More than three quarters of all bits transmitted today over the consumer
Internet are video data. Accordingly, video data is of major importance for the
digital transformation. The related field of digital video processing has played
a key role in establishing successful products and services in a wide range
of sectors including communications, health, industry, autonomous vehicles
and security technology. Particularly in the entertainment sector, video data
has shaped the mass market via services like HD and UHD TV or streaming
services. For these applications, efficient transmission only became feasible
through video coding with methods of efficient compression. In addition,
production systems and processing techniques for highly realistic dynamic 3D
video scenes have been developed. In these key areas, the Fraunhofer Institute
for Telecommunications, Heinrich-Hertz-Institute, HHI is playing a worldwi-
de leading role, in particular in the key areas of video coding and 3D video
processing, also through successfully contributing to video coding standardi-
zation.
43
44 Karsten Müller • Heiko Schwarz • Peter Eisert • Thomas Wiegand
Video data has become a central topic in the digital world. This was achieved
through major contributions in the related research field in video processing during
the last decades, e.g. by establishing a number of successful systems, business sec-
tors, products and services [35], including:
• Communications: Video telephony, video conferencing systems, multimedia
messenger services such as video chats are now used in both business and home
environments.
• Health services: Digital imaging techniques allow displaying and processing
high-resolution computed tomography scans. Here, virtual models for preparing
and supporting surgery are reconstructed from medical data, and automatic im-
age analysis methods of such data are carried out to assist in diagnosis.
• Industry: For automated process monitoring, quality and production control,
cameras are used as sensors. Here, video analysis methods derived from pattern
recognition are used to quickly and automatically verify whether products in a
production line exactly meet the specifications. Digital video data has become
even more important since the introduction of the Industry 4.0 postulate, leading
to increased production automation and modeling of production processes in the
digital world.
• Vehicles and logistics: Video data from camera sensors is being utilized for au-
tomatic traffic routing. In logistics, this method has been used for a long time to
implement fully automated cargo roads with computer-guided vehicles. For ve-
hicles, self-driving systems are being developed for some years, which analyze
data from all sensors for automated vehicle guidance All areas of security tech-
nology use optical surveillance systems and thus video data also plays an impor-
tant role here.
• And finally, the entertainment industry as a mass market is being shaped by the
primary role of video data. Here, entire business areas have been successfully
developed, including television broadcasting in high-definition (HD) and ul-
tra-high-definition (UHD), with resolutions of 3840 x 2160 pixels and above;
mobile video services; personal recording and storage devices such as camcord-
ers; optical storage media such as DVDs and Blu-ray discs; video streaming
services as well as Internet video portals [57].
For achieving global distribution, international standards play a major role for the
definition of video formats, image acquisition, efficient video coding, transmission,
storage, and video data representation. This global role of digital video data is un-
4 Video Data Processing 45
derlined by a study from Cisco [5], which shows that 73% of all bits transmitted
over the Internet in 2016 were video bits, and it is further expected that this increas-
es to 82% by 2021.
Digital video data represents projections of the visual world in digital form. The
creation of this data requires image capturing or recording, followed by digitization.
In real-world-capturing, similar principles as in the human eye are applied: Light
enters the camera through the aperture, is bundled by a convex lens and hits a
light-sensitive medium – a film or digital photo sensor. By this procedure, the real
physical world is perceived in its familiar form, as the camera aperture only allows
photons from the direction of the objective with small angular deviations. The sub-
sequent light bundling produces a visible representation of the real world, similar
to what we see with our eyes. With some simplifications, the brightness level of the
resulting image is determined by the respective number of entering photons, and the
color from the respective wavelengths. In order to advance from individual pictures
to moving pictures or videos, pictures must be taken at particular time intervals. Due
to the visual persistence of the human eye, a series of individual images displayed
at a high enough rate (more than 50 images per second) is perceived as a moving
picture/video.
In order to transform a recorded film to digital video data, the individual images
are first discretized, i.e. spatially sampled. For analog recordings, this is done ex-
plicitly; for recordings using a digital camera sensor, the maximum number of light
sensors in the sensor array already provides implicit discretization, e.g. an HD
sensor records an image resolution of 1920 x 1080 pixels. Next, the color and bright-
ness levels of the individual pixels are discretized in order to be represented in
digital form. Typically, the three color values red, green and blue are discretized or
quantized into 256 different values for each pixel. Each color value is then repre-
sented in binary format by 8 bits (256 = 28) and finally, the entire digital video can
be easily processed, saved and coded by computer systems.
As a result, video data can be specified by formats that describe the properties
and settings of the recording and digitization stages. E.g., the format “1920 x 1080
@ 50fps, RGB 4:4:4, 8bit”, specifies a digital video with a horizontal and vertical
resolution of 1920 and 1080 pixels respectively, a refresh rate of 50 images per
second, in RGB color space, with the same resolution of all color components
(4:4:4) and each quantized into 8 bits.
Video data has achieved global significance and worldwide distribution through
digitization and the rise of the Internet. Digitization led to video data of new previ-
ously unseen quality, since video signals could now be processed efficiently and their
data volumes drastically reduced via coding methods. Across the various distribution
channels such as television and Internet broadcasting, via wired as well as wireless
46 Karsten Müller • Heiko Schwarz • Peter Eisert • Thomas Wiegand
mobile channels, video data has become the most transmitted and consumed type of
data. The overall volume of video data transmitted is growing faster than the capac-
ity of transmission networks, meaning that – within the video processing chain from
production to display and consumer playback – compression plays a predominant
role. For the latter, international video coding standards are developed by the ITU
VCEG (International Telecommunications Union – Visual Coding Experts Group,
officially ITU-T SG16 Q.6) and ISO/IEC-MPEG (International Organization for
Standardization/International Electrotechnical Commission – Moving Picture Ex-
perts Group – officially ISO/IEC JTC 1/SC 29/WG 11) in particular. Both standard-
ization bodies often cooperate in joint teams and integrate improved compression
methods into each new generation of standards. This enables video data to be trans-
mitted at the same levels of quality with significantly lower data rates [32]. Notably,
some applications have only become possible due to efficient video compression.
E.g., the H.264/MPEG-4 AVC (Advanced Video Coding) standard [1] [23], enabled
the wide distribution of high-resolution television (HDTV) and video streaming
services via DSL (Digital Subscriber Line). Furthermore, DVB-T2 could only be
commenced in Germany due to the successor standard H.265/MPEG-H HEVC
(High-Efficiency Video Coding) [17] [47], as it enabled transmitting HD television
with the same visual quality at the available (lower) data rate.
The digitization of video data has also initiated new areas of application, which
extend two-dimensional video data into the third dimension in various ways. Here,
fields of research have developed that target 3D object and scene modeling, using
natural data captured with several cameras. In the field of computer graphics, pow-
erful computers also enabled purely synthetic scene modeling, where entire animat-
ed films are modeled on the computer. Collaboration between the two fields has
created mixed reality scenes that contain computer-animated graphics such as wire-
frames as well as natural video data. Key aspects here are accurate 3D perception
of scenes and easy navigation. Accordingly, VR headsets were developed, which
depict scenes in high-quality 3D – here, pairs of stereoscopic images, separately for
the left and right eye – and which allow viewers to navigate within the scene much
more naturally by moving or turning their head. In recent years, the field of virtual
reality has further developed towards augmented reality (AR) and is likely to spawn
application in various fields. These include, for example, medical technology, where
computed tomography video data is combined with synthetic 3D models in order to
improve diagnoses and surgery techniques. In architecture, building plans are in-
creasingly produced virtually and can thus also be inspected virtually, allowing for
more efficient planning of the real building process. Accordingly, a global market
is currently also developing in this field, and all these areas add to the global impor-
tance of digital video data.
4 Video Data Processing 47
The growing significance of digital video data has resulted in new technological
challenges and market opportunities. In response, a video processing research sec-
tion was formed at the Heinrich-Hertz-Institute (HHI) in 1989 within the Image
Processing Department, in order to carry out early basic research, technological
development and video standardization. In the last 15 years in particular, the video
processing research section, now within the Heinrich-Hertz-Institute as an institute
of the Fraunhofer Gesellschaft, has been playing a leading role in the field of video
coding worldwide. In the academic field, this is expressed in a variety of renowned
publications, guest lectures and editorial responsibilities for international journals
and conferences. In addition, the institute has published an extensive two-part mon-
ograph on source coding [56] and video coding [57]. Important technologies for
video coding have been developed and integrated into various standards. These
include H.264/MPEG-4 AVC [59] along with its extensions SVC (Scalable Video
Coding) and MVC (Multi-View Video Coding), successor standards H.265/MPEG-H
HEVC [47] and their extensions for scalability, multi-view and 3D (SVHC [4], MV-
HEVC [50] and 3D-HEVC [28][50]). Besides the technological co-development of
standards, Fraunhofer HHI has been continuously involved in the management of
relevant standardization committees. The acquired expertise is exploited commer-
cially in a number of public and private projects. In parallel to the development
of video coding technology, Fraunhofer HHI has also developed efficient trans
mission methods [16][39][40].
Based on its longstanding experience, Fraunhofer HHI has acquired an equally
prominent position in the field of computer graphics and computer vision. Here,
new methods for movie and television technology have been developed, allowing
three-dimensional video content to be created, integrated and displayed in mixed
reality scenes. Fraunhofer HHI has also carried out pioneering work in the field of
synthetic and natural 3D video and computer graphics compression, as well as
standardization work for dynamic wireframe models for MPEG-4 AFX (Animation
Framework Extension).
In the video processing section, systems for 3D production such as the Stereo-
scopic Analyzer (STAN) for supporting depth control in 3D video content produc-
tion have been created. Furthermore, technologies for video panorama recording
with video camera systems, and automatic real-time stitching systems were devel-
oped in order to avoid visible transitions between the individual recording cameras
in the panorama. By combining methods from the areas of computer vision, com-
puter graphics and visual computing, new solutions for a broad range of applications
48 Karsten Müller • Heiko Schwarz • Peter Eisert • Thomas Wiegand
in the fields of multimedia, augmented reality, medical imaging, and security could
be developed. One example is the processing of video-based animated 3D models
that are captured and reconstructed from real people and led to virtual 3D interactive
films. With this, Fraunhofer HHI is also taking a leading role in international re-
search and development in the new field of immersive 360° and VR technologies.
Section 4.1 described that video services such as UHD TV and Internet streaming
are only possible by coding standards with efficient video compression methods
such as H.265/MPEG-H HEVC. As an example, we first consider the required rate
for an uncompressed digital video signal with a high-enough temporal resolution of
50 images per second, each sent at UHD resolution of 3840 x 2160 pixels. This
equals around 414 million pixels per second (50 images/s x 3840 x 2160 pixels/
image). Each pixel represents a color value composed of the three components red,
green and blue. At a moderate color resolution/quantization of 8 bits per color value,
24 bits are required for each pixel. This results in a bit rate of 10 Gbit/s for an un-
compressed UHD video signal. In contrast, the available data rate for a UHD video
is typically 20 Mbit/s. Uncompressed UHD signals are thus 500 times larger than
the available transmission channel permits at maximum.
The above example thus requires a video compression method that compresses
the original video to a size of 1/500 while maintaining a high video quality with no
visible distortion. This leads to the formulation of a key requirement: An effective
video compression method must provide the highest possible video quality at the
lowest possible bit rate. This requirement can be found in every generation of video
coding standards and is achieved via the general rate distortion optimization (RD
optimization) in Eq. 4.1 [48] [58]:
Jmin = min(D + λ R). Eq. 4.1
Here, the required rate R is added to the distortion D, where the additional param-
eter λ weights between the two variables [8]. Here, D is the deviation of a recon-
structed video segment from the original segment and is inversely proportional to
video quality. That is, the smaller D, the higher the video quality. Thus, the optimi-
zation requirement means achieving the lowest possible value of R with the lowest
possible distortion D. This is achieved by minimizing the weighted sum in Eq. 4.1,
and thus by finding the optimal Euler-Lagrange functional Jmin. This general formu-
lation is used as a specific optimization for the choice of optimal coding mode, as
described below. According to the area of application, a maximal video/transmis-
4 Video Data Processing 49
sion rate or maximal distortion/minimal quality may be specified. In the first case,
an optimal video compression method would aim for the best possible quality (i.e.
lowest distortion) at a given rate, in the second case, a minimal rate results from a
given level of quality.
In order to achieve the required compression ratio for video signals, fundamen-
tal statistical analyses of image signals were first conducted [57], in particular in-
cluding research on human perception [33]. The found properties were analyzed and
applied to video coding. First, pixels are no longer displayed in RGB color space
but in a transformed color space, better adapted to the human perception. Here, the
YCbCr color space is used in particular, where the Y components contain the lumi-
nance information and the Cb and Cr components contain the chrominance infor-
mation as a color difference signal between B and Y, and R and Y respectively. With
respect to the different sensitivity of the human eye for luminance and chrominance,
a data reduction can be achieved by subsampling the Cb and Cr components. In this
way, a YCbCr 4:2:0 color format can be used for video coding, where the Y com-
ponent of a UHD signal still contains 3840 x 2160 luminance pixels, while the
chrominance resolution is reduced to 1920 x 1080 pixels for both Cb and Cr. For
image representation and backward transformation into the RGB color space, one
chrominance pixel from the Cb and one from the Cr signals are assigned to a quad-
ruple of 2 x 2 luminance pixels from the Y signal. Subsampling the Cb and Cr sig-
nals already produces a reduction of the uncompressed data rate by a factor of 2,
since each Cb and Cr only contain a quarter of the number of luminance pixels. In
the case of video coding, it has also been shown that all texture details and fine
structures are located in the luminance signal, while the chrominance signals have
much less detail and can thus be higher compressed.
Additional statistical methods concentrated on natural video, i.e. camera recorded
video data. Such videos can be rather different, as they show a great variety of various
scenes and thus also feature different colors, color distributions as well as different
motions and motion distributions. Thus, a filmed scene of a fleeing group of animals,
with additional camera panning and zooming, shows a completely different pattern
of color and motion in comparison to an anchor person in front of a static camera.
Despite the differences in content of natural video sequences, natural video have some
key similarities: within the different objects in a scene, neighboring pixels share a
high local color similarity. This similarity exists in the spatial neighborhood of each
frame, as well as in the temporal neighborhood between successive images. In the
latter case, the local motion of an object must be considered in order to identify image
areas between temporally neighboring images with high similarity [37].
Due to the similarity between neighboring pixels, lower data rates can already
be achieved by taking the difference between neighboring pixels or blocks of pixels.
50 Karsten Müller • Heiko Schwarz • Peter Eisert • Thomas Wiegand
intra coding modes from the spatially neighboring blocks as well as different inter
coding modes of temporally neighboring and motion-compensated blocks. To iden-
tify the optimal mode, D(p) is calculated as the mean squared error between s[x, y,
t] and ŝ[x, y, t], and thus as the variance of u[x, y, t]. For the corresponding rate R(p),
the number of bits required for coding the block with the corresponding mode p is
calculated (both for coding the transformed and quantized residual error signal and
for the motion and signaling information [36]). Finally, the rate is weighted with the
Lagrange parameter λ, which depends on the quantization selected [54], and finally
guarantees the optimal coding mode popt for different bit rates.
In the next step within the coding loop, transform coding of the residual signal
u[x, y, t] is conducted, e.g. via integer versions of the discrete cosine transform
(DCT) [2]. The subsequent quantization and scaling of the transform coefficients is
controlled via a quantization parameter (QP) that controls the rate point of the vid-
eo encoder. Finally, the resulting quantization indices, motion vectors and other
control information are coded losslessly using entropy coding. For this, CABAC
(context-adaptive binary arithmetic coding) has been developed [22][49] as a pow-
erful tool of entropy/arithmetic coding and integrated into both families of standards
(AVC and HEVC).
As shown in the lower section of the coding loop in Fig. 4.1, the coded signal is
reconstructed block-by-block as s′[x, y, t], to obtain a new prediction signal for the
next loop cycle. To do this, the transformed residual signal is inversely scaled and
transformed in order to obtain the reconstructed version of the residual signal u′[x,
y, t]. Subsequently, the current prediction signal ŝ[x, y, t] is added. For improved
picture quality, a filter is applied to avoid visible block boundaries in the recon-
structed image (in-loop filtering). Finally after this filtering, the reconstructed image
is created. The encoder also contains the decoder (shaded gray in Fig. 4.1), thus
knows the quality of the reconstructed video and can use it to optimize the compres-
sion method.
This describes the basic principles and working methods of modern video coding
methods, in particular those of the AVC and HEVC standards. For details and pre-
cise descriptions of all the tools, please refer to the overview and description liter-
ature for AVC [23][45][46][52][59] and HEVC [12][15][20][21] [30][31][38][43]
[47][49][51].
Although the same basic techniques have been used in all video coding stand-
ards since H.261 [60], the standards differ in key details, which led to a continuing
increase in the achievable coding efficiency from one standard generation to the
next. The majority of improvements here can be accounted to an increase in the
supporting methods for coding an image or block. These include, among others,
the number of supported transform sizes; the number of partitioning options and
4 Video Data Processing 53
block sizes for intra prediction and motion-compensated prediction; the number of
supported intra prediction modes; the number of usable reference images; the ac-
curacy of the motion vectors coded, etc. Additional coding efficiency gains were
achieved via an improvement in entropy coding and introduction of different in-
loop filters.
As an illustration of the development of video coding, Fig. 4.2 shows a com-
parison of the coding efficiency of the most recent video coding standard H.265/
MPEG-H HEVC [17] versus its preceding standards H.264/MPEG-4 AVC [1],
MPEG-4 Visual [6], H.263 [61] and MPEG-2 Video [14], for two test videos. For
fair comparison, the same encoding control concept was used for all standards
based on the Lagrange technique described in Eq. 4.2 above. For the first test
video, Kimono, all encoders were configured such that an entry point at every
second was available in the bitstream, from which decoding could start as required
for streaming and broadcasting applications. The second video, Kristen & Sara,
represents an example for videoconferencing applications; all of the images were
coded in original recording order to provide as little delay as possible between
70%
40
60%
39 MPEG-2 50%
38 MPEG-4 40%
37 H.263 30%
36 AVC 20% versus MPEG-2 versus MPEG-4
35 HEVC 10% versus H.263 versus AVC
34 0%
0 2000 4000 6000 8000 10000 34 35 36 37 38 39 40 41 42 43
Bitrate [kbit/s] Quality in PSNR [dB]
Kristen & Sara, 1280x720, 60Hz Kristen & Sara, 1280x720, 60Hz
45 100%
44 90%
43 80%
Bitrate Saving of HEVC
Quality in PSNR [dB]
42 70%
41 60%
40 MPEG-2 50%
39 MPEG-4 40%
38 H.263 30%
37 AVC 20% versus MPEG-2 versus MPEG-4
36 HEVC 10% versus H.263 versus AVC
35 0%
0 1000 2000 3000 4000 5000 35 36 37 38 39 40 41 42 43 44 45
Bitrate [kbit/s] Quality in PSNR [dB]
Fig. 4.2 Coding efficiency of international video coding standards for two selected test
sequences. Left: reconstruction quality as a function of the bit rate; Right: bit rate savings
of current HEVC standard in comparison to the different predecessor standards, source [32]
(Fraunhofer HHI)
54 Karsten Müller • Heiko Schwarz • Peter Eisert • Thomas Wiegand
sender and receiver. For the results shown, PSNR (peak signal-to-noise ratio) was
used as quality measure, defined as the mean squared error (mse) between original
and reconstructed images, and calculated as PSNR = 10 log10 (2552/mse). While
the curves in Fig. 4.2 left compare the quality/bit rate functions, the graphs in Fig.
4.2 right show bit rate savings achieved with HEVC versus the predecessor stand-
ards for a given video quality (PSNR). In order to process a video at a particular
video quality, the HEVC standard adopted in 2014 only requires 20–30% of the bit
rate, necessary for the MPEG-2 standard from 1995. The bit rate savings for the
same subjective quality, i.e. the quality as perceived by a viewer, are generally even
larger [32].
In parallel to the highly successful standards for 2D video coding, extensions to
AVC and HEVC were also specified, facilitating scalability [4][39][40][42], larger
bit depths [10], larger color dynamic ranges [11], efficient coding of multiple cam-
era views [26][50][52], and additional use of depth data [25][28][50][52].
With the global distribution of digital video data, not only new markets for classi-
cal two-dimensional video data developed. Moreover, new research areas and
technologies for producing and displaying 3D video scenes [7][24][27][44] have
emerged. In contrast to classical video, where one dimension of the actual scene
disappears due to projection onto the 2D image plane of the camera, the entire
geometry of the environment is captured and represented via suitable 3D models.
Thus, the choice of a viewing point for the scene is no longer determined by the
recording method but can actually be freely selected by the user afterwards. By
rendering the 3D scene to a virtual camera position, arbitrary camera and viewing
trajectories become possible, as well as interactive navigation within the scene,
where the viewer can freely select regions of interest. Additionally, by producing
separate images for the viewer’s left and right eye, a 3D impression of the observed
scene can be created and thus the perception of the real spatial scene structure is
improved.
The demand for producing three-dimensional video scenes has been driven re-
cently both by the development of new virtual reality (VR) headsets that achieve
improved immersion and natural viewer navigation in the virtual scene as well as
by higher-quality 3D images. Users can now navigate scenes far more naturally by
means of their own motion and head movements and better merge with the scene.
This facilitates new multimedia applications such as immersive computer games,
virtual environments (such as museums and tourist sights), and even innovative
4 Video Data Processing 55
movie formats such as interactive films. Instead of moving in a purely virtual world,
the connection of virtual 3D objects with the real scene in augmented and mixed
reality applications is also possible, driven by technological advances in AR head-
sets (e.g. Microsoft HoloLens) or see-through applications using smartphones and
tablets. Here, new content such as supplementary information and virtual 3D objects
are inserted into the viewer’s field of vision and registered in the real scene with the
correct perspective. This technology will create new assistance systems and user
support for sectors like health (endoscopy and microscopy), industry (production
and maintenance), and mobility (driver assistance).
Currently, the high-resolution capture of three-dimensional dynamic environ-
ments is a challenge. The development of 3D sensors is further progressing, how-
ever the sensors are often less suitable for dynamic objects (e.g. scanners) or offer
limited spatial resolution (e.g. time of flight sensors). Passive photogrammetric
approaches, on the other hand, have gained importance due to increased camera
resolutions, falling prices, ubiquitous availability and are capable of delivering
high-quality results [3]. In multi-camera 3D systems, the scene is captured using
two (stereo) or more (multi-view) cameras simultaneously from different viewing
angles. Then, for each pixel in a reference image, corresponding pixels in the other
images are identified, which originate from identical surface position of the real
object. The displacement of the corresponding pixels between the two images is
called “disparity” and can be converted directly to depth information from the cam-
era plane to the real object point, since both variables are inversely proportional.
The larger the displacement between two pixels, the closer the object point to the
camera. As a result, a depth value can be assigned to each camera pixel. From this,
a 3D point cloud is produced that represents the part of the object surface that is
visible from the direction of the camera. By fusing the surfaces from different cam-
era pairs, the entire scene can be reconstructed.
Fraunhofer HHI has developed numerous methods for analyzing and estimating
3D video objects in the sectors of multimedia/film [13][44], telepresence [53][18],
health [41] and industry [35]. In the following, we describe 3D person reconstruc-
tion to illustrate the production of dynamic 3D video objects using multi-camera
systems, as shown in Fig. 4.3 [7]. The aim is to produce high-quality content for
virtual reality applications and interactive film formats.
To capture 3D video objects, the first step is to arrange a number of synchronized
high-resolution camera pairs that capture the scene from several directions. The
small baseline within each stereo pair avoids any masking and view dependent
surface reflections, thus enabling the estimation of robust depth maps. In addition,
the usage of several stereo pairs, enables global object capture via fusion of the
individually recorded object surfaces.
56 Karsten Müller • Heiko Schwarz • Peter Eisert • Thomas Wiegand
Depth maps are estimated from the stereo pairs using a patch-based matching
approach. Here, the dependencies are minimized in such a way that the matching
between corresponding patches can be evaluated independently, permitting effi-
cient GPU parallelization and real-time implementation [53]. Propagating depth
information from neighboring and temporally preceding patches ensures spatial
and temporal smoothness and improved robustness. Lighting and color compensa-
tion reduces differences between the images from different cameras. Depending on
the degrees of freedom of the patch used, the accuracy of the depth estimation is
much below 1 mm, as shown in Fig. 4.4. In the next stage, the point clouds for all
stereo pairs are fused, and points with incorrectly estimated 3D positions are elim-
inated via consistency tests across the individual point clouds. As a result, a 3D
point cloud of the entire 3D video object is produced with approx. 50 million points
(see Fig. 4.3 left).
Using the color information, a dynamic 3D object representation can be pro-
duced at this stage, however only as a colored point cloud without a (closed) surface.
Hence, in the next step, the point cloud is triangulated, i.e. all points are connected
by triangles. For consistent triangulation, Poisson surface reconstruction is used
[19], placing a smooth surface through the point clouds and normals. Then, the
triangulated 3D model can be simplified, e.g. via successive point elimination in
smooth surface regions. The aim here is to retain surface details while reducing the
4 Video Data Processing 57
Fig. 4.4 Accuracy of iterative depth estimation, in comparison to known reference objects
(Fraunhofer HHI)
number of 3D points and triangles at the same time. The result is a compact wire-
frame as shown in Fig. 4.3, center. Here, the original 50 million points were reduced
to approximately 10,000 points [7].
In order to guarantee a high quality of texture and color of the 3D model, the
original video data from the cameras is now projected onto the wireframe. To
achieve the highest detail information, the most suitable camera view in terms of
spatial resolution is identified for each surface triangle of the 3D model. Then, color
information is primarily projected from that camera. Interpolating between neigh-
boring triangles avoids visible breaks and inconsistencies in the surface texture. The
resulting textured object is depicted in Fig. 4.3 right.
This method was developed by Fraunhofer HHI as a fully-automated method for
producing highly realistic, dynamic 3D video objects and integrating them into
virtual worlds [7]. These objects are characterized in particular by their natural
appearance and realistic movements, thus representing a cornerstone of interactive
3D video worlds.
When integrating the 3D video objects into new virtual environments, they are
displayed in the same way as they were captured during the recording method.
Often, the lighting, body pose or motion require customization in order to tailor
them to the environment or to enable interaction with the video objects. To do this,
semantic body models (avatars) are customized to the individual 3D geometry and
motion patterns are learnt from measured data [9]. Using the skeleton assigned to
the model, corrections to the body pose can be carried out that are then translated to
the individual geometry, e.g. to represent interactions in the virtual world. Moreo-
ver, individual motion sequences can be recombined and seamlessly superimposed
[34] in order to realize complex representations or new interactive forms of media
viewing.
58 Karsten Müller • Heiko Schwarz • Peter Eisert • Thomas Wiegand
In this chapter, we have shown the development of digital video data towards a
globally dominant data form, and explained the development of video processing.
Two of the most successful technologies, with Fraunhofer HHI strongly involved
in the development, are video coding and production of three-dimensional dynam-
ic video objects. Both fields will continue developing, leading to common topics for
research and production, and shape a number of fields in digital transformation.
In video coding, a successor standard will be developed in the coming years that
once again provides improved compression for video data. This is particularly nec-
essary as in the process towards higher video resolutions from HD via UHD/4K, 8K
has already been announced and will once again multiply the uncoded video data
rate. In the case of 3D video object reconstruction, further improvements in 3D
display technology together with fast, high-quality production of natural 3D video
scenes will allow for interactive 3D video worlds to be created where reality and
virtuality will merge seamlessly. The successful distribution of 3D video scenes also
requires standardized formats and compression processes that will be developed in
the coming years. The first steps have already been taken with extensions to depth-
based 3D video coding and standardization plans for omnidirectional 360° video
within the new coding standards. Nevertheless, further research is required, in order
to internationally standardize methods for efficient, dynamic 3D scene coding and
thus to establish globally successful systems and services for interactive 3D video
worlds.
[1] Advanced Video Coding, Rec. ITU-T H.264 and ISO/IEC 14496-10, Oct. 2014.
[2] N. Ahmed, T. Natarajan, and K. R. Rao, “Discrete Cosine Transform”, IEEE Transac-
tions on Computers, vol. C-23, issue 1, pp. 90–93, Jan. 1974
[3] D. Blumenthal-Barby, P. Eisert, “High-Resolution Depth For Binocular Image-Based
Modelling”, Computers & Graphics, vol. 39, pp. 89-100, Apr. 2014.
[4] J. Boyce, Y. Yan, J. Chen, and A. K. Ramasubramonian, “Overview of SHVC: Scalable
Extensions of the High Efficiency Video Coding (HEVC) Standard,” IEEE Transactions
on Circuits and Systems for Video Technology, vol. 26, issue 1, pp. 20-34, Jan. 2016
[5] Cisco Systems, Inc. Cisco visual networking index, “Forecast and methodology, 2016-
2021”, White paper, June 2017. Retrieved June 8, 2017, from http://www.cisco.com/c/
en/us/solutions/collateral/service-provider/visual-networking-index-vni/complete-
white-paper-c11-481360.pdf .
[6] Coding of audio-visual objects – Part 2: Visual, ISO/IEC 14496-2, 2001
4 Video Data Processing 59
[22] D. Marpe, H. Schwarz, and T. Wiegand, “Context-based adaptive binary arithmetic co-
ding in the H.264/AVC video compression standard”, IEEE Transactions on Circuits and
Systems for Video Technology, vol. 13, issue 7, pp. 620–636, July 2003
[23] D. Marpe, T. Wiegand and G. J. Sullivan, “The H.264/MPEG4 Advanced Video Coding
Standard and its Applications”, IEEE Image Communications Magazine, vol. 44, is. 8,
pp. 134-143, Aug. 2006.
[24] T. Matsuyama, X. Wu, T. Takai, T. Wada, “Real-Time Dynamic 3-D Object Shape Re-
construction and High-Fidelity Texture Mapping for 3-D Video”, Transaction on Circuits
and Systems for Video Technology, vol. 14, issue 3, pp. 357-369, March 2004.
[25] P. Merkle, K. Müller, D. Marpe and T. Wiegand, “Depth Intra Coding for 3D Video based
on Geometric Primitives”, IEEE Transactions on Circuits and Systems for Video Techno-
logy, vol. 26, no. 3, pp. 570-582, Mar. 2016.
[26] P. Merkle, A. Smolic, K. Müller, and T. Wiegand, “Efficient Prediction Structures for
Multiview Video Coding”, invited paper, IEEE Transactions on Circuits and Systems
for Video Technology, vol. 17, no. 11, pp. 1461-1473, Nov. 2007.
[27] K. Müller, P. Merkle, and T. Wiegand, “3D Video Representation Using Depth Maps”,
Proceedings of the IEEE, Special Issue on 3D Media and Displays, vol. 99, no. 4, pp.
643 – 656, April 2011.
[28] K. Müller, H. Schwarz, D. Marpe, C. Bartnik, S. Bosse, H. Brust, T. Hinz, H. Lakshman,
P. Merkle, H. Rhee, G. Tech, M. Winken, and T. Wiegand: “3D High Efficiency Video
Coding for Multi-View Video and Depth Data”, IEEE Transactions on Image Processing,
Special Section on 3D Video Representation, Compression and Rendering, vol. 22, no.
9, pp. 3366-3378, Sept. 2013.
[29] K. Müller, A. Smolic, K. Dix, P. Merkle, P. Kauff, and T. Wiegand, “View Synthesis for
Advanced 3D Video Systems”, EURASIP Journal on Image and Video Processing, Spe-
cial Issue on 3D Image and Video Processing, vol. 2008,Article ID 438148, 11 pages, 2008.
doi:10.1155/2008/438148.
[30] T. Nguyen, P. Helle, M. Winken, B. Bross, D. Marpe, H. Schwarz, and T. Wiegand,
“Transform Coding Techniques in HEVC”, IEEE Journal of Selected Topics in Signal
Processing, vol. 7, no. 6, pp. 978-989, Dec. 2013.
[31] A. Norkin, G. Bjøntegaard, A. Fuldseth, M. Narroschke, M. Ikeda, K. Andersson, M.
Zhou, and G. Van der Auwera, “HEVC deblocking filter”, IEEE Transactions on Circuits
and Systems for Video Technology, vol. 22, issue 12, pp. 1746–1754, Dec. 2012
[32] J.-R. Ohm, G. J. Sullivan, H. Schwarz, T. K. Tan, and T. Wiegand, “Comparison of the
Coding Efficiency of Video Coding Standards – Including High Efficiency Video Coding
(HEVC)”, IEEE Transactions on Circuits and Systems for Video Technology, Dec. 2012.
[33] G. A. Østerberg, “Topography of the layer of rods and cones in the human retina”, Acta
Ophthalmologica, vol. 13, Supplement 6, pp. 1–97, 1935
[34] W. Paier, M. Kettern, A. Hilsmann, P. Eisert, “Video-based Facial Re-Animation”, Proc.
European Conference on Visual Media Production (CVMP), London, UK, Nov. 2015.
[35] J. Posada, C. Toro, I. Barandiaran, D. Oyarzun, D. Stricker, R. de Amicis, E. Pinto, P.
Eisert, J. Döllner, and I. Vallarino, “Visual Computing as a Key Enabling Technology
for Industrie 4.0 and Industrial Internet”, IEEE Computer Graphics and Applications,
vol. 35, no. 2, pp. 26-40, April 2015
[36] K. Ramchandran, A. Ortega, and M. Vetterli, “Bit allocation for dependent quantization
with applications to multiresolution and MPEG video coders”, IEEE Transactions on
Image Processing, vol. 3, issue 5, pp. 533–545, Sept. 1994
4 Video Data Processing 61
[37] F. Rocca and S. Zanoletti, “Bandwidth reduction via movement compensation on a model
of the random video process”, IEEE Transactions on Communications, vol. 20, issue 5,
pp. 960–965, Oct. 1972
[38] A. Saxena and F. C. Fernandes, “DCT/DST-based transform coding for intra prediction
in image/video coding”, IEEE Transactions on Image Processing, vol. 22, issue 10, pp.
3974–3981, Oct. 2013
[39] T. Schierl, K. Grüneberg, and T. Wiegand, “Scalable Video Coding over RTP and MPEG- 2
Transport Stream in Broadcast and IPTV Channels”, IEEE Wireless Communications
Magazine, vol. 16, no. 5, pp. 64-71, Oct. 2009.
[40] T. Schierl, T. Stockhammer, and T. Wiegand, “Mobile Video Transmission Using Scala-
ble Video Coding”, IEEE Transactions on Circuits and Systems for Video Technology,
Special Issue on Scalable Video Coding, vol. 17, no. 9, pp. 1204-1217, Sept. 2007,
Invited Paper
[41] D. Schneider, A. Hilsmann, P. Eisert, “Warp-based Motion Compensation for Endosco-
pic Kymography”, Proc. Eurographics, Llandudno, UK, pp. 48-49, Apr. 2011.
[42] H. Schwarz, D. Marpe, and T. Wiegand, “Overview of the Scalable Video Coding Exten-
sion of the H.264/AVC Standard”, IEEE Transactions on Circuits and Systems for Video
Technology, Special Issue on Scalable Video Coding, vol. 17, no. 9, pp. 1103-1120, Sept.
2007, Invited Paper
[43] R. Sjöberg, Y. Chen, A. Fujibayashi, M. M. Hannuksela, J. Samuelsson, T. K. Tan, Y.-K.
Wang, and S. Wenger, “Overview of HEVC High-Level Syntax and Reference Picture
Management”, IEEE Transactions on Circuits and Systems for Video Technology, vol.
22, no. 12, pp. 1858‒1870, Dec. 2012.
[44] A. Smolic, P. Kauff, S. Knorr, A. Hornung, M. Kunter, M. Müller, and M. Lang, “3D
Video Post-Production and Processing”, Proceedings of the IEEE (PIEEE), Special Issue
on 3D Media and Displays, vol. 99, issue 4, pp. 607-625, April 2011
[45] T. Stockhammer, M. M. Hannuksela, and T. Wiegand, “H.264/AVC in wireless environ-
ments”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 13, no.
7, pp. 657–673, July 2003.
[46] G. J. Sullivan and R. L. Baker, “Efficient quadtree coding of images and video”, IEEE
Transactions on Image Processing, vol. 3, issue 3, pp. 327–331, May 1994
[47] G. J. Sullivan, J.-R. Ohm, W.-J. Han, and T. Wiegand, “Overview of the High Efficiency
Video Coding (HEVC) Standard,” IEEE Trans. Circuits Syst. Video Technol., vol. 22,
no. 12, pp. 1649‒1668, Dec. 2012.
[48] G. J. Sullivan and T. Wiegand, “Rate-distortion optimization for video compression”,
IEEE Signal Processing Magazine, vol. 15, no. 6, pp. 74–90, June 1998
[49] V. Sze and M. Budagavi, “High throughput CABAC entropy coding in HEVC”, IEEE
Transactions on Circuits and Systems for Video Technology, vol. 22, issue 12, pp. 1778–
1791, Dec. 2012
[50] G. Tech, Y. Chen, K. Müller, J.-R. Ohm, A. Vetro and Y. K. Wang, “Overview of the
Multiview and 3D Extensions of High Efficiency Video Coding”, IEEE Transactions
on Circuits and Systems for Video Technology, Special Issue on HEVC Extensions and
Efficient HEVC Implementations, vol. 26, no. 1, pp. 35-49, Jan. 2016.
[51] K. Ugur, A. Alshin, E. Alshina, F. Bossen, W.-J. Han, J.-H. Park, and J. Lainema, “Mo- tion
compensated prediction and interpolation filter design in H.265/HEVC”, IEEE Journal
of Selected Topics in Signal Processing, vol. 7, issue 6, pp. 946–956, Dec. 2013.
62 Karsten Müller • Heiko Schwarz • Peter Eisert • Thomas Wiegand
[52] A. Vetro, T. Wiegand, and G. J. Sullivan, “Overview of the Stereo and Multiview Video
Coding Extensions of the H.264/AVC Standard”, Proceedings of the IEEE, Special Issue
on “3D Media and Displays”, vol. 99, issue 4, pp. 626-642, April 2011, Invited Paper.
[53] W. Waizenegger, I. Feldmann, O. Schreer, P. Kauff, P. Eisert, “Real-Time 3D Body Recon-
struction for Immersive TV”, Proc. IEEE International Conference on Image Processing
(ICIP), Phoenix, USA, Sep. 2016.
[54] T. Wiegand and B. Girod, “Lagrange multiplier selection in hybrid video coder control”,
In Proc. of International Conference on Image Processing (ICIP), vol. 3, pp. 542–545,
2001.
[55] T. Wiegand, M. Lightstone, D. Mukherjee, T. G. Campbell, and S. K. Mitra, “Rate- distor-
tion optimized mode selection for very low bit rate video coding and the emerging H.263
standard”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 6, issue
2, pp.182–190, April 1996
[56] T. Wiegand and H. Schwarz, “Source Coding: Part I of Fundamentals of Source and
Video Coding”, Foundations and Trends in Signal Processing, vol. 4, no. 1-2, pp. 1-222,
Jan. 2011, doi:10.1561/2000000010
[57] T. Wiegand and H. Schwarz, “Video Coding: Part II of Fundamentals of Source and
Video Coding”, Foundations and Trends in Signal Processing, vol. 10, no. 1-3, pp 1-346,
Dec. 2016, doi:10.1561/2000000078
[58] T. Wiegand, H. Schwarz, A. Joch, F. Kossentini, and G. J. Sullivan, “Rate-constrained
coder control and comparison of video coding standards”, IEEE Transactions on Circuits
and Systems for Video Technology, vol. 13, issue 7, pp. 688–703, July 2003
[59] T. Wiegand and G. J. Sullivan, “The H.264/AVC Video Coding Standard {Standards in a
nutshell}”, IEEE Signal Processing Magazine, vol. 24, no. 2, March 2007.
[60] Video codec for audiovisual services at p × 64 kbits, Rec. ITU-T H.261, Nov. 1988.
[61] Video coding for low bit rate communication. Rec. ITU-T H.263, Mar. 1996.
Audio Codecs
Listening pleasure from the digital world
5
Prof. Dr. Karlheinz Brandenburg · Christoph Sladeczek
Fraunhofer Institute for Digital Media Technology IDMT
Summary
The development of music recording and reproduction has been characterized
by the quest for perfection ever since its inception under Thomas Alva Edison.
This is true for all fields of recording and reproduction technology, from
microphones and sound storage media to loudspeaker technology. Concert-
quality sound reproduction which matches the original with complete fidelity
remains the goal of research. This can only be achieved, however, if all of the
factors involved in listening are addressed, whether they are acoustic, psycho-
acoustic, or psychological. Fraunhofer IDMT’s further development of wave
field synthesis technology as a marketable product – already in use in highly
demanding open air and opera house productions – plays a key role here. The
uncomplicated synchronous storage and transmission of metadata provided
by the current MPEG-H 3D Audio coding method allows listeners at home to
interactively shape their listening experience.
63
64 Karlheinz Brandenburg • Christoph Sladeczek
Even Thomas Alva Edison already dreamt of perfect sound reproduction. Agents
marketing his phonograph as a consumer product travelled the world, conducting
what were perhaps the earliest sound tests: Audiences listened amazed as music was
performed in darkened halls – first as a live performance by a singer or cellist, for
example, and then as a phonograph recording of the same piece. Many listeners
found the recording’s sound quality so good that they could not tell the difference
[8]. From this we can conclude that our assessment of sound quality is particularly
related to our expectations of a medium. Background noises such as hisses or crack-
les were not part of the music and were thus not heard.
Ever since then, research on how to deliver the most perfect sound reproduction
possible has been ongoing. High fidelity has been the common term used to describe
this for many years now. If we were to assess Edison’s test by modern standards,
there would be little in common between the phonograph’s wax cylinder and our
current understanding of hi-fi, even though considerable sound quality was already
achieved in those days. Today, a good reproduction setup could probably pass the
test for individual musical instruments, but not for larger ensembles such as a string
quartet or a symphony orchestra. Sound reproduction that creates a perfect illusion
within the room still remains just out of reach. In recent decades, however, signifi-
cant progress has been made in terms of both loudspeaker and headphone playback.
The goal of complete immersion – diving into an alien sound environment – is thus
getting ever closer. Keywords here are stereophonics, surround sound, wave field
synthesis, ambisonics, and object-based audio signal storage.
In the early years of hi-fi technology, the research emphasis lay on the different
elements of the processing chain: microphones, analog recording processes (tapes,
vinyl records), playback devices for tapes and records, amplifiers, loudspeakers and
headphones. At that, analog audio recording processes such as tape recorders have
fundamental limitations at the level of sound quality they are able to provide. Ever
greater progress has been made over the decades in optimizing the components:
whereas the tube technology amplifiers display nonlinearity of transmission (output
voltage over input voltage) – audible even to less practiced listeners – modern am-
plifier technology (implemented in analog integrated circuits or as so-called class-D
amplifiers using mixed analog/digital circuit technology) is close to perfect. Micro-
phones, too, are near to physical limits for professional applications. Loudspeakers
5 Audio Codecs 65
and headphones still remain the weakest link in this transmission chain. If our
hearing were not as good as it is at adapting to the deviations from the ideal frequen-
cy response curve, a lot of electronically amplified music would sound off-key and
alien.
The various music recording media represented key milestones in the develop-
ment of hi-fi technology: the phonograph cylinder, gramophone records, vinyl long-
play records, reel-to-reel tape recorders, compact cassettes and compact discs. Over
time, ever-newer technologies had to be introduced in order to provide a higher
potential sound quality, increased usefulness (for example, so people could also
make their own recordings on tape recorders) and, in particular, convenient storage
medium handling. The first digital music storage medium to gain widespread ac-
ceptance was the CD. Its parameters (16-bit resolution, 44,100 Hz sampling fre-
quency, two sound channels) were the state of the art at the time and were designed,
according to detailed tests, to facilitate “perfect” sound reproduction.
The underlying optical storage medium – the CD-ROM – became the storage
medium for software distribution for several years. As a storage medium for audio-
visual applications such as short films, the CD-ROM was also one of the first inno-
vations to initiate the development of processes for video and audio coding in the
Moving Pictures Experts Group (MPEG, officially ISO/IEC JTC1 SC29 WG11). The
standards produced by this standardization committee simultaneously represented
the state of the art in their respective fields when they were adopted. From MPEG-
1 Audio Layer-3 (known as MP3) through MPEG-2/4 Advanced Audio Coding
(AAC) and further developments such as HE-AAC to today’s MPEG-H, this fami-
ly of audio coding methods is now incorporated into around 10 billion devices. Each
new generation of processes brought improved coding efficiency (identical sound
quality at lower bit rates/better sound quality at equally low bit rates) as well as
increased flexibility (for example, supporting surround sound processes). Through
all these years, Fraunhofer IIS has made significant contributions to these process-
es and their standardization.
Fig. 5.1 Selective listing of important developments in digital audio technology in recent
decades (Fraunhofer IDMT)
66 Karlheinz Brandenburg • Christoph Sladeczek
Whereas for a long time the standard for home music listening remained
two-channel stereo, cinema technology saw the introduction of surround sound,
designed to facilitate an immersive sound experience. Since MPEG-2, the audio
coding methods standardized by MPEG have also supported 5.1-channel surround
sound as a minimum. Newer methods, especially for professional applications, will
be described in later sections. If we compare the sound experience of today with the
dreams of Edison’s days then we have to accept that the perfect audio illusion has
still not been achieved. This is not due to the data reduction processes but to the
fundamental technology of sound reproduction. Today, the reproduction of materi-
al in mono or in two-channel stereo is possible to such a perfect degree – even when
using AAC data reduction, for example – that blind tests show that it is impossible
to distinguish it from a reference recording. In short, the Edison blind test is suc-
cessful as long as there is only one instrument or musician on stage. Reproducing
the sound experience of listening to a symphony orchestra, however, remains a
challenge.
Recent developments in hi-fi technologies are closely related to the scientific disci-
pline of psychoacoustics. The aim of psychoacoustic research is to discover the
relationships between stimuli – i.e. sounds such as music as well as all kinds of
other audio signals – and human perception. Two areas are particularly key to hu-
man hearing:
• The ear (including the outer ear and pinna, the auditory canal, the eardrum, the
middle ear, and the inner ear with the cochlea and sensory hairs)
• The brain (by processing the electrochemical reactions that are triggered by the
motion of the sensory hairs connected to the neurons – the so-called “firing” of
neurons), including the higher region of our brain, responsible for complex per-
ception and the identification of sound.
To simplify, we can say that the ear’s characteristics define the general parameters
of our hearing such as the frequency range of discernable sounds and the dynamic
range, from the quietest audible sounds through to sounds that can lead to damage
of the inner ear. The everyday phenomenon of masking – that is, the obscuring of
quieter sounds by louder ones – is a significant effect for audio coding in particular,
5 Audio Codecs 67
and can be explained by the mechanics of the inner ear. Sounds of one frequency
mask quieter tones of the same or neighboring frequencies. This effect is utilized in
audio coding (MP3, AAC, etc.), where frequencies with reduced accuracy are trans-
mitted in such a way that the differential signals are masked by the actual music
signal and thus no difference to the input signal is audible.
Since these masking effects are mechanically produced– and since the stimuli
transmitted from the ear to the brain are already “data reduced”, so to speak – they
are stable. Even with extensive listening training, people are unable, during blind
tests, to distinguish pieces of music coded at a sufficiently high bit rate (using AAC
for example) from the original.
Nevertheless, the processing that happens in the brain is utterly vital to the audi-
tory impression. Research into the so-called cognitive effects of hearing has been
ongoing for decades, but we have to admit that our understanding of these process-
es is significantly more limited than our understanding of the workings of the ear.
In short, feedback effects from the higher levels of the brain, and especially expec-
tations of the sound play a role. All spatial hearing is connected with these complex
mechanisms: the conjunction of signals from both ears (binaural spatial hearing)
takes place in the brain. Here, differences in time and especially phase of the sound
heard by both ears are evaluated.
A further key effect is the filtering of signals by the pinna and head. These effects
are usually described as the outer ear transfer function (technically, HRTF or Head
Related Transfer Function). This function, which varies from person to person, is
something we will keep in mind as we continue through this chapter. It is this func-
tion that enables us to roughly ascertain a sound’s direction of origin even when
using one only ear. As things currently stand, spatial hearing in hi-fi technology is
determined by how “coherent” our perception of a sound is, given where we expect
the sound to come from (especially when we can see its source) and given the play-
back system and any additional effects. The greater the divergence in our percep-
tion, the more frequently localization errors occur and – especially in the case of
headphone playback – the more confusion between sound coming from behind or
the front or sound source localization within the head (instead of outside the head)
is observed. All of these effects are highly individual, varying significantly over
time. The brain can be trained to perceive certain illusions more frequently.
Research is currently being carried out into precisely these issues:
• How does the listening environment – the room and the reflections of the sound
from the walls and furniture – affect the acoustic illusion that hi-fi technology
tries to create?
• How much are we influenced by our expectations, and to what extent do we fail
to perceive changes in the sound even though they are clearly measurable? On
68 Karlheinz Brandenburg • Christoph Sladeczek
the other hand, to what degree do we hear differences that are purely based on
our expectations?
In classical audio production, loudspeaker signals are saved to the sound storage
medium as a result of mixing. Instruments and sound sources are distributed spa-
tially in the audio signal by means of different volume weighting of the loudspeak-
ers. This process is known as panning and is set separately for each sound source
by the sound engineer so that voices, for example, are heard by the user in the
middle of the stereo panorama during playback, and instruments such as guitars and
drums are heard to the left and right. Specific guidelines need to be followed in
order to correctly perceive the spatial mix intended by the sound engineer. The
stereo speakers, for example, must be set up in the same positions as they are in the
recording studio, and the listening location must form an equilateral stereo triangle
with the speakers. The same holds for stereo playback processes such as 5.1 sur-
round sound, for example, for which supplementary speakers are added. Since the
5 Audio Codecs 69
spatial sound reproduction in each case is only correct for a single listening location,
this location is known as the sweet spot. If we look at actual playback locations such
as cinemas, theatres, and the home, for example, we see that for most listeners the
ideal playback location cannot be maintained.
Ever since the initial beginnings of the development of spatial sound playback
processes, the desire has been to record and reproduce the sound field of a virtual
sound source in such a way as to provide the correct spatial sound impression for
all listeners. One attempt uses loudspeaker arrays to synthesize a sound field in a
way that is physically correct. To achieve this goal, Steinberg and Snow in 1934
published the principle of the acoustic curtain [11], where a fixed network of mi-
crophones in the recording studio connected to loudspeakers in the playback room,
is used to record and reproduce the sound field of a source. Steinberg and Snow were
using Huygens ’ Principle here, which states that an elementary wave can be pro-
duced by the superimposition of many individual secondary sound sources.
At the time, however, it was still not technologically possible to implement this
complex arrangement in practice, which was why Steinberg and Snow limited their
system to three loudspeakers. The use of these three loudspeakers represents the
beginnings of the stereophonic recording and playback technology that was extend-
ed with additional channels in subsequent years.
In the 1980s, Guus Berkhout carried out research into acoustic holography pro-
cedures for use in seismology. In doing so, he used arrays of microphones that re-
corded the reflection patterns of special sound signals from different layers of the
earth, thus providing information about the substances contained in the ground.
Because of his personal interest in acoustics in general, he suggested that the tech-
nology he had developed could be reversed and used for loudspeaker arrays. This
marked the beginnings of the development of wave field synthesis technology,
where arrays of loudspeakers are used to synthesize – within specific limits – a
physically correct sound field for a virtual sound source [1]. The underlying princi-
ple is illustrated in Fig. 5.2.
In the years that followed, the technology was developed at TU Delft to the point
where it could be presented as a functional laboratory demonstrator in 1997 [12]. A
key characteristic of wave field synthesis technology is its object-based approach.
In contrast to classical sound reproduction processes, audio objects are stored rath-
er than loudspeaker signals as a result of the spatial sound mixing. An audio object
is defined as a (mono) signal that contains audio information – for example a violin
or a female voice – together with its associated metadata, which describes properties
such as the position, volume or type of audio object. In order to investigate this new
technology and its associated production, storage, transmission and interaction re-
quirements with respect to a potential introduction to the market, a consortium was
70 Karlheinz Brandenburg • Christoph Sladeczek
formed in 2001, consisting of industry and research and development, under the
banner of the EU CARROUSO project [2]. As a key outcome of this project Fraun-
hofer IDMT was able to present a first marketable product prototype installation at
the Ilmenau cinema in 2003.
perception via headphones [7]. This is far more difficult with a channel-based audio
approach.
One additional interesting distinguishing feature of the object-based audio play-
back concept is sideways compatibility with channel-based audio content [3]. Here,
the loudspeaker signals are reproduced via audio objects so that a virtual loudspeak-
er setup is produced. In this way, channel-based audio content is practically inde-
pendent of the actual number of playback speakers. In addition, the user has the
option of changing the virtual loudspeaker setup and can thus vary parameters such
as stereo width or spatial envelopment intuitively.
For audio objects to be able to be used intuitively in a practical application, the
technology must be capable of being seamlessly incorporated into an audio envi-
ronment that is familiar to the sound engineer [9]. To do this, it is necessary that the
technology mirrors established interfaces and interaction paradigms. Fig. 5.4 shows
one potential possibility for integration.
The object-based sound reproduction technology is thus based on a rendering PC
that is equipped with a professional sound card with standard multi-channel audio
transmission formats. A sound source of the user’s choice can then be connected via
this interface. This is shown in the illustration by a digital audio workstation (DAW)
and a mixer symbol. For the audio objects to be able to be positioned intuitively a
72 Karlheinz Brandenburg • Christoph Sladeczek
Fig 5.4 Sample integration of object-based sound reproduction into existing professional
audio production structures (Fraunhofer IDMT)
graphical user interface (GUI) is required; in the example above, this is provided by
a web-based application accessed via a web browser. In this case, the server provid-
ing the user interface also runs on a rendering PC. An example of this kind of user
interface is shown in Fig. 5.5.
The interface here is divided into two regions. On the right-hand side is the po-
sitioning region where the audio objects can be positioned. Here, audio objects are
represented by round icons. In order to provide the sound engineer with a familiar
interface, the audio objects’ properties are listed clearly on the interface’s left-hand
side in the form of channel strips. The sound engineer is thus able to shape the
playback of the individual audio objects in the same way as at a mixing desk.
Fig. 5.5 Graphical user interface for object-based audio production (Fraunhofer IDMT)
5 Audio Codecs 73
Fig. 5.6 Object-based sound reproduction for the Bregenz Festival live sound reinforce-
ment (Fraunhofer IDMT)
74 Karlheinz Brandenburg • Christoph Sladeczek
reinforcement from all spatial directions. Often, opera houses and theatres are
equipped with loudspeaker installations of 80 speakers and more, but these are
distributed loosely and in three dimensions throughout the space. In order to sup-
port these precise installations in terms of acoustics, the wave field synthesis al-
gorithm was altered so that, taking account of human perception, audio objects
can still be stably localized over a large listening area, but allowing for greater
distance and three-dimensional distribution of speakers. This SpatialSound Wave
technology has been applied in several prestigious venues such as the Zurich
Opera House shown in Fig. 5.7.
Alongside live sound reinforcement, there is an additional area of application
in the acoustic support of large-screen playback systems. Planetariums are good
examples of dome projection installations that, in recent years, have moved away
from classical visualizations of the night sky towards being entertainment adven-
ture venues. Although dome projection offers an impressive, enveloping image
from different seating positions, installations historically often only featured a few
loudspeakers using channel-based playback formats. That meant that in practice,
the image was spread across the dome, but the sound came from individual loud-
speakers below the projection surface. Plausible reproduction, however, requires
images and sound objects to be coordinated spatially, in the same way that Spatial-
Sound Wave technology does, by allowing positioning of the loudspeakers behind
the projection surface. Fig. 5.8 shows a typical SpatialSound Wave technology
application in the world’s oldest functioning planetarium, the Zeiss Planetarium in
Jena.
5 Audio Codecs 75
Fig. 5.8 Object-based audio reproduction for large-scale projection systems in-room
sound reinforcement (Fraunhofer IDMT)
MPEG-H 3D Audio
Object-based audio productions require new file formats for storage and transmis-
sion since, along with pure audio signals, metadata also needs to be reproduced
synchronously. The MPEG-4 standard already supported storing audio objects via
the so-called BIFS (Binary Format for Scenes) scene description, but this was hard-
ly ever used due to the great complexity of the MPEG-4 standard [5].
MPEG-H is a new standard, which enables both the transmission and storage of
high-quality video and audio. By means of the relevant component for audio data
– known as MPEG-H 3D Audio – it is now possible to store the most diverse audio
formats in a standardized structure. Alongside the ability to save channel-based
audio unambiguously even with a large number of speaker channels, this innovation
offers storage of audio objects and higher-order ambisonics signals. By supporting
various audio formats it becomes apparent that the standard will become accepted
over the coming years as the container format for all new immersive audio playback
methods.
The standard is currently being introduced in the field of broadcasting. Alongside
the loudspeaker system scaling advantages already mentioned, audio objects offer
new possibilities for interaction here. One example is in the transmission of sporting
events where today a sound engineer mixes commentator vocals and stadium at-
mosphere sound in the broadcasting vehicle, so that they can then be transmitted to
76 Karlheinz Brandenburg • Christoph Sladeczek
the end user in stereo or surround sound format. Depending on what end device the
consumer is using, problems can arise with the intelligibility of speech, for example,
if the commentator’s voice is obscured by the atmosphere in the stadium. Since
signal elements in a stereo signal cannot be changed retrospectively anymore (or
can only be changed very little), one potential solution is the transmission as audio
objects. This way, the user has the option, for example, to change the volume of the
commentator at home to restore speech intelligibility or to concentrate completely
on the atmosphere in the stadium [10].
5.4 Outlook
It has been a long journey from Edison’s “sound tests” to current sound reproduc-
tion, assisted by complex digital signal processing algorithms. Modern technology
enables us to listen to music comfortably and at near perfect sound quality whether
we are at home or on the road. Certain problems, such as achieving the perfect illu-
sion, have however still only been partially solved. But as fundamental research
continues and is applied to the latest audio signal storage and playback standards,
both live and pre-recorded music will, over the coming decades, deliver even more
listening pleasure.
[1] Berkhout A.W,: „A Holographic Approach to Acoustic Control“, JAES, Volume 36, Issue
12, pp. 977-995, December 1988.
[2] Brix S., Sporer T., Plogsties J., „CARROUSO – An European Approach to 3D Audio“,
110th Convention of the AES, Amsterdam, The Netherlands, May 2001.
[3] Brix S., Sladeczek C., Franck A., Zhykhar A., Clausen C., Gleim P.; „Wave Field Syn-
thesis Based Concept Car for High-Quality Automotive Sound“, 48th AES Conference:
Automotive Audio, Hohenkammern, Germany, September 2012.
[4] Brandenburg K., Brix S., Sporer T.: „Wave Field Synthesis – From Research to Applica-
tions“, European Signal Processing Conference (EUSIPCO) 2004, September, Vienna,
Austria.
[5] Herre J., Hilpert J., Kuntz A., Plogsties J.: „MPEG-H 3D Audio – The New Standard
for Coding of Immersive Spatial Audio“, IEEE Journal of Selected Topics in Signal
Processing, Volume 9, No. 5, pp. 770-779, August 2015.
[6] Lembke S., Sladeczek C., Richter F., Heinl, T., Fischer C., Degardin P.; „Experimenting
with an Object-Based Sound Reproduction System for Professional Popular Music Pro-
duction“, 28th VDT International Convention, Cologne, Germany, November 2014.
5 Audio Codecs 77
Summary
Digitization is moving ahead at full speed. For most people, the smartphone
is a constant companion; manufacturing businesses are also pushing ahead
with digitization on the factory floor in the wake of Industry 4.0. Even radio
cannot stop the trend in its tracks: step by step, digital radio is replacing its
familiar FM cousin. A common practice in numerous European nations, and
many developing nations are preparing for the switchover. Digital radio offers
numerous advantages: greater program diversity, improved reception, a wealth
of enhanced services. Digital radio will grow together with mobile telephony
in the long term. Whereas radio sends information that is of interest to everyo-
ne, mobile telephony takes on “personalized” information. In this way, the two
technologies can be the ideal complement to one another.
79
80 Albert Heuberger
6.1 Introduction
Crackling and hissing on the radio? No way! Digital radio consigns that sort of in-
terference to history. It offers listeners and radio stations alike numerous advantag-
es. Listeners gain more program diversity along with supplementary information
via data services. Reception quality, too, is better. Radio stations in turn save energy
due to increased transmission efficiency and are at the same time able to broadcast
a greater number of programs due to the more efficient use of the transmission
spectrum. Both provide economic benefits. Just like familiar FM radio, terrestrial
digital radio uses radio signals; but contrary to popular belief, there is absolutely no
need for an Internet connection, making it thus free to receive.
In most European nations, DAB+ digital radio is already part of everyday life.
Norway is relying exclusively on this new form of radio – it switched FM radio off
by the end of 2017. Switzerland and the UK, too, are actively considering moving
away from FM before 2020. Numerous developing nations are currently planning
the transition from analog short- and medium-wave radio to DRM digital radio, and
the digitization of local FM radio infrastructure has started. India is one of the pio-
neers here and is on the way to becoming the largest market for digital radio in the
world. All of the necessary technologies – from the necessary basic technologies to
transmission and reception systems for digital radio applications – have been co-de-
veloped with significant contributions by the Fraunhofer Institute for Integrated
Circuits IIS. Software built by Fraunhofer, for example, is contained in every single
DRM device in India.
One of the great advantages of digital radio lies in its spectrum efficiency. In order
to explain what lies behind this, we will first take a look at traditional FM radio.
Traditional FM radio operates at frequencies of 87.5 to 108 megahertz, giving it a
bandwidth of about 20 megahertz. This means that the available frequencies are
very limited. A lot of radio stations that would love to enter the market are shut out
because there are no more free frequencies available.
Digital radio comes to the rescue. The frequency resource is limited here, too, but
in the 20-megahertz useable spectrum there is room for up to four times as many
broadcasters. The exact number depends on the audio quality desired and the robust-
ness of transmission – if a program is broadcast at a higher quality, it requires a high-
er bit rate. The increased efficiency is first and foremost the result of compression,
that is, of the standardized xHE-AAC and HE-AAC audio codecs, developed essen-
6 Digital Radio 81
tially by researchers at Fraunhofer IIS. These codecs are responsible for reproducing
speech and audio at lower data rates with higher quality and thus form the foundation
for the high sound quality of digital radio. The second transmission quality parameter
is the robustness of the signals transmitted. Depending on how much supplementary
error-correction information is available for the transmission, reception of the signal
is better or worse, with a direct impact, for example, on the broadcaster’s range.
One additional important reason for the greater broadcast capacity is what is
known as the single frequency network. With digital radio, all stations broadcasting
the same information can do so using the same frequency; at the receiving end,
where the signals from two transmitters are received simultaneously, there is no
interference. With FM radio, this is not the case. Here, neighboring transmitters
have to work on different frequencies. Imagine taking a car journey through Ger-
many: the frequencies for a given station change as you move from one area to
another. A single program thus simultaneously requires several frequencies. With
digital radio, a single frequency suffices to cover the entire country. In this way,
significantly more radio programs can be accommodated in a given frequency
spectrum than before.
With FM radio, each program needs its own frequency. This means that if Bayern
3 is broadcasting on a certain frequency with a bandwidth of 200 kHz, this frequen-
cy range is occupied and is only available to other broadcasters if they are located
far away. Digital radio on the other hand is more diverse – with DAB+, 15 different
programs can typically be broadcast over a single frequency with a bandwidth of
1.536 MHz. So you could broadcast Bayern 1, Bayern 3, and various regional pro-
grams within the entire transmission area on the same frequency. These are known
as a multiplex or ensemble. It opens up greater program diversity to listeners. Black
Forest Radio (“Schwarzwaldradio”), for example, was only available in the region
of the same name via FM. On the nationwide digital radio multiplex it can now be
heard across the nation. People from Hamburg, for example, who like to holiday in
the Black Forest can now listen to “holiday radio” at home, too. Classic Radio
(“Klassikradio”), too, already has a place in the German nationwide multiplex – the
broadcaster is increasingly emphasizing digital radio and has already switched off
its first FM transmissions. These examples show how digital radio opens up the
possibility of offering special-interest radio to listeners.
At the moment, digital radio use in Germany is around 12% to 18%; advertising
revenues are thus still limited. Nevertheless, numerous broadcasters are already
82 Albert Heuberger
turning to digital radio, especially in order to secure a transmission slot for them-
selves for the future.
Digital radio not only transmits radio content in high quality, it also offers a diverse
range of new services. One example is traffic information services. Current naviga-
tion systems receive analog radio traffic updates via the traffic message channel
(TMC) – a narrow-band extension to the FM signal – which are then factored into
the navigation. The trouble is this: the amount of information that can be sent via
TMC is severely limited. This creates difficulties when it comes to locating traffic
jams, for example. In order to keep the content as brief as possible, instead of send-
ing specific location information via TMC, so-called location codes are sent, which
are then translated back by the receiver device from a cached list. Newly-built
freeway exit-ramps, however, are not included in this list and thus cannot be used
as locations. Any number of new location codes also cannot simply be added to the
list. Although significantly more accurate data is available on the sending side, the
limited data rate prevents it from being communicated. In short, TMC is a limited
and generally overloaded channel that is no longer a fit for the demands of modern
navigation devices. Higher quality navigation systems receive the information via
mobile telephony, but this does not always function seamlessly either. For one,
transmitting information this way is expensive, and each car also needs to be noti-
fied individually. If numerous drivers request information simultaneously, the net-
work quickly becomes overloaded.
Digital radio enables these kinds of traffic information services to operate
significantly more efficiently. The associated technology is known as TPEG, short
for Transport Protocol Experts Group. Whereas the data rate for TMC is 1 kbit/s,
for TPEG it is usually 32 to 64 kbit/s. Even higher data rates would not be a prob-
lem where required. In addition, the information arrives intact even under poor
reception conditions. This opens up new applications such as getting an alert
about the tail end of a traffic jam. Using sensor data from the freeways or via
floating data from driver smartphones, precise calculations can be made regarding
where a traffic jam ends. A driver alert must be sent during the right time period:
if the alert arrives too early, the driver will already have forgotten about it by the
time they reach the traffic jam end tail; if it arrives too late, the driver has no time
to respond. The alert should thus be issued 800m to 1500m before the tail end of
the traffic jam. In other words, transmission needs to be exceptionally timely and
6 Digital Radio 83
reliable. In addition, the tail end of the traffic jam also moves, so the data needs
to be updated every minute. All of this is possible using TPEG and digital radio
– and it is completely irrelevant how many people require this information simul-
taneously.
Digital radio can also be used to send traffic predictions to navigation systems.
If you are setting out from Munich to Hamburg, for example, then it is of little
concern to you at that point whether there is a traffic jam in Kassel. What you want
to know is what will the traffic situation be like there in three to four hours? If the
navigation system receives traffic predictions in Munich already, for the suggested
route, it can largely bypass congestion in advance. Parking availability information
would be another conceivable possibility: using digital radio, the navigation system
can announce where the most parking is available. TPEG offers all of these possi-
bilities; with regard to traffic information, it is an enormously more powerful tool
than the old TMC or mobile telephony.
New digital radio services are not at all limited only to traffic. Digital radio also
offers numerous applications beyond this. The news service Journaline for digital
radio (again developed in partnership with Fraunhofer IIS researchers) allows lis-
teners to receive comprehensive information that they can read on the radio’s dis-
play. This may be background information regarding the radio program or the music
being played, but it could also be airport departure times or soccer results. Journa-
line is like a kind of video text for radio – adapted, of course, to modern conditions.
It not only features intuitive operation but also additional functionalities such as
Internet links, images, and local information.
Digital radio also has significant potential in the field of public safety notices.
The Emergency Warning Functionality (EWF) – that Fraunhofer IIS played a key
joint role in developing – is the first to allow real-time public alerts. This is not
possible via FM since it always entails delays – information arrives first on the
presenter’s desk here and must then be passed on by them. This can sometimes take
as long as 45 minutes. And some programs are even automated – alerts here are
ignored completely. With the new EWF technology, on the other hand, the emer-
gency services control center issues the alert directly and almost immediately. Even
devices that run on standby, such as alarm clocks, can be used: they automatically
turn themselves on when an alert is issued. In these cases, the alerts are brief and to
the point, e.g. Keep your doors and windows closed. With services such as Journa-
line, additional information can be accessed simultaneously and in parallel via radio
display – with no need for an Internet connection, which may in any case no longer
be available in a disaster situation. In addition, the information is provided in vari-
ous languages. This way, the alert service is also suitable for those with difficulty
hearing and non-native speakers.
84 Albert Heuberger
Of course, the choice bet+ween mobile telephony and digital radio is not either/or.
Rather, the two technologies will continue growing together in the long term, acting
as helpful ancillaries to one another. While digital radio takes care of information
that is of interest to everyone, mobile telephony looks after “personalized” informa-
tion. The kind of form this could take is indicated by the Journaline service men-
tioned above. Current news, for example, can be transmitted via digital radio. The
user can then access additional information about events in the immediate vicinity
through Internet links as required; this is achieved via mobile telephony. The user
is unaware of the process – everything appears as a single homogenous service.
The advantages of mobile telephony and digital radio can be combined for the
ideal outcome. Mobile telephony is strong when it comes to information that inter-
ests individual users in particular. When it comes to listening to the radio, however,
mobile telephony is a poor option. In the end, cellular sites only have limited total
capacity. If a user is listening to the radio and moves from cell A to cell B, then cell
B must immediately provide the necessary capacity. Mobile service providers need
to maintain spare capacity “just in case”. Mobile telephony is thus only of limited
suitability for “one-to-many” applications – this kind of information belongs on
broadcast-type networks that are precisely tailored to these tasks.
6.7 Outlook
In Germany new regional and local multiplexes will start in various regions.
All over Europe, there is a growing number of DAB+ programs with additional
services such as the news service Journaline and slideshow images. In its Digital
Radio Report, the European Broadcasting Union (EBU) finds that more than 1,200
radio stations are currently broadcasting in Europe via DAB or DAB+, of which
6 Digital Radio 85
more than 350 are only available digitally. Digital radio thus offers increasing use-
fulness and added value to listeners, on top of the advantages already listed such as
a wider variety of programs, improved sound quality, more stable networks, as well
as cost and energy savings due to more efficient frequency use and lower transmis-
sion power. Especially in countries with poor or no Internet connection, these new
systems facilitate free and widespread access to information, entertainment and
education.
Contact details:
Prof. Dr. Albert Heuberger
+49 9131 776-1000
[email protected]
5G Data Transfer at Maximum Speed
More data, more speed, more security
7
D. Eng. Ronald Freund · D. Eng. Thomas Haustein ·
D. Eng. Martin Kasparick · D. Eng. Kim Mahler ·
D. Eng. Julius Schulz-Zander · D. Eng. Lars Thiele ·
Prof. D. Eng. Thomas Wiegand · D. Eng. Richard Weiler
Fraunhofer Institute for Telecommunications,
Heinrich-Hertz-Institute, HHI
Summary
Mobile communications have permanently changed our society and our ways
of communicating ever since the global availability of mobile speech services,
and because these communications form the basis for mobile Internet. They
have facilitated a new dimension of productivity growth and manufacturing
and service process networking since the use of the Internet. Their technical
basis is founded on a deep understanding of the relationships between radio
and telecommunications technology, beginning with radio wave propagation
and modeling, through techniques for digital signal processing and a scalable
system design for a cellular radio system with mobility support, to methods for
system analysis and optimization. The Fraunhofer Institute for Telecommuni-
cations, Heinrich-Hertz-Institute, HHI has been working in the field of mobile
telephony communications for 20 years and has made key contributions to the
third, fourth, and fifth generations. Alongside research articles and numerous
first-time demonstrations of key technological components, the institute is also
an active contributor to 3GPP standardization.
87
88 Ronald Freund • Thomas Haustein • Martin Kasparick • Kim Mahler • et al.
The furious onward march of digitization in society requires the support of flexible
and scalable mobile telephony solutions in order to meet the demands of new appli-
cations and business models. The socio-economic transformation expected is built
upon the availability of mobile connections to the data network everywhere, at any
time, and with consistent quality of service for communications between people,
between people and machines, and between machines. This requires a fundamen-
tally new vision of mobile communications. While the first generation of mobile
telephony was founded on analog radio technology, 2G (GSM) was a purely digital
communications system right from the start, which made it possible to provide
speech and the first data services (SMS) globally. UMTS, the third generation (3G),
enabled the technological basis for the mobile Internet via broadband mobile teleph-
ony connections. 4G extended the dimension of broadband data communications
significantly and reduced its complexity by means of a so-called all-IP approach
handling both speech as well as IP-based data services. At the same time, this laid
the foundation for a convergence between fixed-line and mobile telephony commu-
nications networks.
The goal for the fifth generation is to enable mobile data communications in new
fields as a platform for communications between everything and everyone, opening
up brand new possibilities for production, transport and healthcare, and addressing
issues such as sustainability, security and well-being that are relevant to modern
society. The vision of a completely mobile and connected society requires the sup-
porting conditions for an enormous growth in connectivity and traffic volume den-
sity in order to sustain the broadest possible range of use cases and business models.
The long-discussed concept of the Internet of Things will finally be made possible
via the organic embedding of mobile communications capabilities in every field of
society. A comprehensive motivation together with the goals of 5G were first out-
lined in the NGMN White Paper [1]; a current analysis of the challenges, trends and
initial field trials was set out in the IEEE Journal on Selected Areas in Communica-
tions [18]. Driven on by technological developments and socio-economic transfor-
mations, the 5G business context is characterized by changes in customer, techno-
logical and operator contexts [2].
increase significantly through new devices such as wearables and sensors. Assisted
by cloud technology, the capabilities of personal devices will be seamlessly extend-
ed to all sorts of applications such as high-quality (video) content production and
sharing, payments, proof of identity, cloud gaming, mobile television and support-
ing intelligent living in general. These devices will play a significant role in the
fields of healthcare, security and social life, as well as in the operation and monitor-
ing of household appliances, cars, and other machines.
The mobile telephony industry is expecting the first 5G devices with a limited
range of functions in 2018 and 2020 during the Olympic Games in South Korea and
Japan, with a large-scale rollout starting from 2022.
Business context: Analogous trends to those in the consumer realm will also feed
into daily company life; the boundaries will blur the line, for example, between
private and professional use of devices. Businesses thus need flexible and scalable
solutions in order to manage the security and data protection issues that arise in this
usage context. For businesses, mobile communications solutions are some of the
key drivers of increased productivity. Over the coming decades, businesses will
increasingly make their own applications available on mobile devices. The spread
of cloud-based services facilitates the portability of applications across several de-
vices and domains and offers entirely new opportunities, together with new chal-
lenges as regards to security, privacy, and performance.
Business partnerships: In many markets we see the trend of network operators en-
tering into partnerships with so-called over-the-top (OTT) players, in order to pro-
vide better integrated services to their respective end customers. For OTT players
the communication network’s quality of service profile is becoming increasingly
important; it is necessary in order to be able to provide new services in the private
and above all business spheres. Inherent synergy here between connectivity with a
guaranteed quality of service, on the one hand, and high-quality services on the
other, enables these partnerships to become the foundation for shared success.
Alongside classical broadband access, 5G is also relevant for new markets, par-
ticularly in the area of vertical industries. The combination of sales of more than
300 million smartphones per quarter, with a total of more than 10 billion smart-
phones in existence globally, combined with an expected future 50–100 billion ra-
dio-connected devices (machine-to-machine communications), means that an in-
crease in overall data traffic by a factor of 1,000 by the year 2020 can be expected.
Alongside pure peak data rates, the field will be confronted with additional demands
such as, for example, extreme ranges, extreme energy efficiency, ultra-short laten-
cies, extremely high availability or reliability, through to scaling issues with massive
90 Ronald Freund • Thomas Haustein • Martin Kasparick • Kim Mahler • et al.
user access, massive antenna systems, and heterogeneous radio access and conver-
gence requirements for the networks.
The breadth of application scenarios to be serviced demands a variety of com-
munications solutions in radio access and at the network layer, which, despite their
heterogeneity, need to be incorporated into the 5G framework via standardized in-
terfaces.
This highly dynamic environment provides Fraunhofer with opportunities to
actively contribute to the innovation process, to develop new technological solu-
tions and thus make a significant contribution to the ecosystem. Due to the shift in
the focus of applications away from people as the end customer (in the case of te-
lephony or the mobile Internet), business-to-business (B2B) communications solu-
tions will in future increasingly become the focus as the enabler of automated pro-
cesses. The demands are varied and cannot be fulfilled with a one-size-fits-all solu-
tion. New market participants from the Internet sphere are flooding into the field
and significant transformation of the existing mobile telephony ecosystem is to be
expected. Fraunhofer’s understanding of sector-specific challenges provides it with
opportunities here to develop targeted solution components for and with clients and
partners.
From pure theory through to initial field trials (proofs-of-concept), the Fraun-
hofer Institute for Telecommunications, Heinrich-Hertz-Institute, HHI has already
made significant contributions to 4G-LTE research and the continued development
of LTE Advanced and LTE Advanced Pro, as well as in the early phases of 5G re-
search. These have included extensive studies on wave forms [32][38][42][41][28],
MIMO [37][24][19], CoMP [34][35][31][36], Relaying [33][40], cognitive spec-
trum management [CoMoRa][29][30], energy-efficient network management [43]
and other key technologies [39][27][20] that today form the starting point for new
approaches to 5G.
household appliances, clocks, and industrial robots. Here, many use cases will make
specific demands on the communications network with respect to data rates, relia-
bility, energy consumption, latency etc. This diversity of applications and the cor-
responding requirements will necessitate a scalable and flexible communications
network, as well as the integration of diverse and partly very heterogeneous com-
munications solutions.
Fig 7.2 KPIs for the application categories; markings at the outer edge indicate high im-
portance, further inside lesser importance (ITU-R IMT2020, Fraunhofer HHI)
The diversity of the areas of application and scenarios for 5G that were discussed
as relevant can be categorized into three main groupings from a communications
requirements point of view, as shown in Fig. 7.1. The relevant key performance
indicators (KPIs) indicating the importance of each of the three categories are illus-
trated in Fig 7.2.
Massive MIMO
Antenna Array
Inter-cell
3. Precoding interference
2. Grouping
Inter-group
interference
Group 1
1. Clustering
a) Define input density
b) Cluster users into groups &
Selected detect outliers below density
Users Group G
as full dimension MIMO and are expected to resolve the need for ever-increasing
peak data rates in radio cells. This is facilitated via multiuser multiplexing mecha-
nisms – that is, supporting a range of users over the same time and frequency re-
sources by utilization of the spatial dimension.
Fig. 7.4 Examples of a 64-element planar (left) and 128-element circular antenna array
(right) for massive MIMO antennas for new radio bands at 3.5 GHz (Fraunhofer HHI)
96 Ronald Freund • Thomas Haustein • Martin Kasparick • Kim Mahler • et al.
In the fifth generation mobile telephony standard, the number of active antenna
elements is expected to once again be significantly increased; this is known as
massive MIMO, where data rates are significantly above 1 GBit/s. This necessitates
major changes with respect to the current standard in order to guarantee cost-effi-
cient operation. Multilevel spatial beamforming methods need to be extended so
that the large number of antenna elements can be divided into so-called sub-antenna
arrays. In each of these sub-antenna arrays, beamforming weights are then adapted
in keeping with the phase, whereas phase control of different sub-antenna arrays
must only be readjusted or coordinated on a slow time base [6]. New kinds of an-
tenna shapes play an important role in meeting the heterogeneous requirements in
the cellular system. Thus so-called planar antenna arrays can be utilized, for exam-
ple, in order to shape a range of differing beamformers in a clearly defined solid
angle [3]. In contrast, (semi-)circular antenna arrays may ideally be used in order to
Fig. 7.5 Illustration of the degrees of freedom for realizing an increase in capacity by a
factor of 10,000 compared to existing 4G solutions. A combination of network densifica-
tion (ultra-dense networks) and the use of a spectral region above 6 GHz demonstrates the
necessary potential (source: MiWEBA, joint EU/Japanese research program as part of
FP7). Within the framework of 5G standardization, a distinction is currently being made
between solutions below and above 6 GHz; within the 6–100 GHz band, the regions around
28 GHz, 60 GHz, and 70 GHz are initially being addressed for reasons of spectral region
availability and technological maturity [21]. (Fraunhofer HHI)
7 5G Data Transfer at Maximum Speed 97
Fig. 7.6 5G key solutions for radio communications in the millimeter wave spectrum; top
left: use of the spectrum above 6 GHz; bottom left: interference management between base
stations and end devices within the radio ranges; top middle: densification of networks with
small cells; bottom middle: small cells in indoor areas; right: incorporation of millimeter
wave small cells into the macrocellular infrastructure – control plane/user plane splitting
(Fraunhofer HHI)
98 Ronald Freund • Thomas Haustein • Martin Kasparick • Kim Mahler • et al.
entails using more of the spectrum. Here, the frequency range above 6 GHz to 100
GHz in particular was identified in order to enable an extension of the available
spectrum by a factor of 20 [7][8].
The use of spectral regions above 6 GHz requires new technologies and solutions
for transceiver chips, antenna design and signal processing in order to develop cost-
and energy-efficient components suitable for the mass market. Alongside techno-
logical development challenges, a deep understanding of the propagation behavior
of radio waves at high frequencies is essential for a sustainable system design.
Fraunhofer is contributing to a deeper understanding of wave propagation for rele-
vant indoor and outdoor scenarios with a range of radio field measurements and is
providing channel modeling contributions within the 3GPP standardization process
[9].
Communication in the millimeter wave range requires a high degree of beam
focusing in the direction of the communication between the end device and the base
station due to the high path loss and associated restrictions in the radio range. New
forms of compact antenna arrays using hybrid beamforming approaches are expect-
ed to facilitate the necessary gains here in terms of range, signal quality and inter-
ference limitation. The high frequencies permit compact integration in a limited
space, but also require new approaches to connection establishment, link optimiza-
tion, and link tracking, particularly in mobile scenarios.
Fig. 7.8 The 5G infrastructure investigated as part of the 5G-PPP Metro-Haul project
(Fraunhofer HHI)
ally, a SDN-based network management will also facilitate the fast configuration of
5G services: current goals are 1 min. for simple network path establishment, 10 min.
for complete installation of a new virtual network function (VNF), and 1 hour for
establishing a new virtual network slice. Fig. 7.8 gives an overview of the network
infrastructure for providing future 5G services addressed in research projects.
The broad spectrum of open research questions that remain to be answered in the
5G context has led to a variety of research programs being initiated in Germany,
Europe, and across the world in order to address the issues in good time, in a focused
manner, and in partnership between industry and research. Fraunhofer HHI is en-
gaged in numerous association and research projects related to 5G, first and fore-
most in the context of H2020 5GPP, as well as in various programs sponsored by
German federal ministries, in particular the BMBF, BMWi and BMVI. The follow-
ing sections are intended to provide a brief insight into the technical questions ad-
dressed by the Heinrich Hertz Institute in several selected 5G projects, working with
partners on solutions which will subsequently, through the standardization process,
deliver a lasting contribution to the fifth generation of mobile telephony.
IC, Fantastic 5G, 5G-Crosshaul, Carisma, One5G, and in the joint EU/Asia projects
MiWEBA, STRAUSS, 5G-MiEdge, 5G!PAGODA and 5GCHAMPION, as well as
in national research initiatives including Industrial Communication for Factories
(IC4F), 5G NetMobil, AMMCOA or SEKOM.
Fig. 7.9 5G mechanisms for supporting tactile connected driving (Fraunhofer HHI)
104 Ronald Freund • Thomas Haustein • Martin Kasparick • Kim Mahler • et al.
enabled, and foundations are laid for flexible decision-making. To support the high
mobility requirements of tactile connected driving new, efficient and scalable cog-
nitive network management concepts must be developed. To do this, learning algo-
rithms are being utilized that are able to merge and process all of the available in-
formation and measurement data together with any available contextual informa-
tion. One significant challenge here is guaranteeing the resilience and scalability of
the outlined mechanisms even when there are a large number of users and corre-
spondingly large quantities of sensor data and contextual information. The high
reliability requirements of tactile connected driving make the use of new diversity
concepts such as multi-connectivity necessary. One additional focus of studies at
Fraunhofer HHI within the 5GNetMobil project is thus the development of and
research into new diversity and network management strategies. In order to put
network management in the position of supporting the formation of mobile virtual
cells for interruption-free handovers at high mobility, new mechanisms for network
coding, in-network data processing and distributed decision-making are thus being
employed.
Fig. 7.10 Functionalities and basic principle of the BMBF’s AMMCOA project for high-
ly reliable and real-time-capable networking of agricultural and construction machinery,
based on millimeter wave radio for communications and localization (Fraunhofer HHI)
7 5G Data Transfer at Maximum Speed 105
7.5 Outlook
In this chapter, we have provided a brief insight into the current focus areas of fifth
generation mobile telephony research. Alongside the current 5G standardization
process in 3GPP, not only fundamental but also implementation- and praxis-driven
questions remain to be answered. The Fraunhofer Institute for Telecommunications,
Heinrich Hertz Institute, HHI is working in association with other Fraunhofer insti-
tutes on technological solutions that are able to be included as key components in
the overall specification for 5G, or as use case-specific or scenario related solution
modules for specific industries. Due to its close cooperation with customers from
the broadest range of sectors, Fraunhofer is able to work across sectors and disci-
plines here, making a significant contribution to the 5G ecosystem.
[9] Weiler, Richard J., et al. „Quasi-deterministic millimeter-wave channel models in Mi-
WEBA.“ EURASIP Journal on Wireless Communications and Networking 2016.1
(2016): 84.
[10] L. Grobe et al., „High-speed visible light communication systems,“ IEEE Comm. Ma-
gazine, pp. 60-66, Dec. 2013.
[11] V. Vucic et al., „513 Mbit/s Visible Light Communications Link Based on DMT-Modu-
lation of a White LED,“ Journal of Lightwave Technology, pp. 3512-3518, December
2010.
[12] D. Schulz et al., „Robust Optical Wireless Link for the Backhaul and Fronthaul of Small
Radio Cells,“ IEEE Journal of Lightwave Technology, March 2016.
[13] V. Jungnickel et al., „Software-defined Open Access for flexible and service-oriented 5G
deployment,“ IEEE International Conference on Communications Workshops (ICC), pp.
360 – 366, 2016.
[14] J. Fabrega et al., „Demonstration of Adaptive SDN Orchestration: A Real-Time Con-
gestion-Aware Services Provisioning Over OFDM-Based 400G OPS and Flexi-WDM
OCS,“ Journal of Lightwave Technology, Volume 35, Issue 3, 2017.
[15] https://www.hhi.fraunhofer.de/presse-medien/nachrichten/2017/hannover-messe-
2017-ultraschneller-mobilfunkstandard-5g.html
[16] IC4F Projekt: https://www.ic4f.de
[17] Cognitive Mobile Radio, BMBF Projekt 2012-2015, http://www.comora.de
[18] M. Shafi et. al. „5G: A Tutorial Overview of Standards, Trials, Challenges, Deploy-
ment and Practice,“ in IEEE Journal on Selected Areas in Communications, April 2017,
DOI:10.1109/JSAC.2017.2692307.
[19] M. Kurras, L. Thiele, T. Haustein, W. Lei, and C. Yan, „Full dimension mimo for fre-
quency division duplex under signaling and feedback constraints,“ in Signal Processing
Conference (EUSIPCO), 2016 24th European. IEEE, 2016, pp. 1985–1989.
[20] R. Askar, T. Kaiser, B. Schubert, T. Haustein, and W. Keusgen, „Active self-interference
cancellation mechanism for full-duplex wireless transceivers,“ in International Confe-
rence on Cognitive Radio Oriented Wireless Networks and Communications (Crown-
Com), June 2014.
[21] K. Sakaguchi et.al., Where, When, and How mmWave is Used in 5G and Beyond, arXiv
preprint arXiv:1704.08131, 2017
[22] ITU-T Technology Watch Report, „The Tactile Internet,“ Aug. 2014.
[23] J. Pilz, M. Mehlhose, T. Wirth, D. Wieruch, B. Holfeld, and T. Haustein, „A Tactile Inter-
net Demonstration: 1ms Ultra Low Delay for Wireless Communications towards 5G,“ in
IEEE INFOCOM Live/Video Demonstration, April 2016.
[24] T. Haustein, C. von Helmolt, E. Jorswieck, V. Jungnickel, and V. Pohl, „Performance of
MIMO systems with channel inversion,“ in Proc. 55th IEEE Veh. Technol. Conf., vol. 1,
Birmingham, AL, May 2002, pp. 35–39.
[25] T. Wirth, M. Mehlhose, J. Pilz, R. Lindstedt, D. Wieruch, B. Holfeld, and T. Haustein,
„An Advanced Hardware Platform to Verify 5G Wireless Communication Concepts,“ in
Proc. of IEEE VTC-Spring, May 2015.
[26] 3GPP TR 36.881, „Study on Latency Reduction Techniques for LTE,“ June 2016.
[27] R. Askar, B. Schubert, W. Keusgen, and T. Haustein, „Full-Duplex wireless transceiver
in presence of I/Q mismatches: Experimentation and estimation algorithm,“ in IEEE
GC 2015 Workshop on Emerging Technologies for 5G Wireless Cellular Networks – 4th
International (GC’15 – Workshop – ET5G), San Diego, USA, Dec. 2015.
7 5G Data Transfer at Maximum Speed 107
[28] Dommel, J., et al. 5G in space: PHY-layer design for satellite communications using non-
orthogonal multi-carrier transmission[C]. Advanced Satellite Multimedia Systems Con-
ference and the 13th Signal Processing for Space Communications Workshop (ASMS/
SPSC), 2014 7th. 2014: Livorno. p. 190-196.
[29] M. D. Mueck, I. Karls, R. Arefi, T. Haustein, and W. Keusgen, „Licensed shared access
for wave cellular broadband communications,“ in Proc. Int. Workshop Cognit. Cellular
Syst. (CCS), Sep. 2014, pp. 1–5.
[30] M. Mueck et al., „ETSI Reconfigurable Radio Systems: Status and Future Directions on
Software Defined Radio and Cognitive Radio Standards,“ IEEE Commun. Mag., vol.
48, Sept. 2010, pp. 78–86.
[31] J. Dommel, P.-P. Knust, L. Thiele, and T. Haustein, „Massive MIMO for interference
management in heterogeneous networks,“ in Sensor Array and Multichannel Signal
Processing Workshop (SAM), 2014 IEEE 8th, June 2014, pp. 289–292.
[32] T. Frank, A. Klein, and T. Haustein, „A survey on the envelope fluctuations of DFT pre-
coded OFDMA signals,“ in Proc. IEEE ICC, May 2008, pp. 3495–3500.
[33] V. Venkatkumar, T. Wirth, T. Haustein, and E. Schulz, „Relaying in long term evolution:
indoor full frequency reuse“, in European Wireless, Aarlborg, Denmark, May 2009.
[34] V. Jungnickel, T. Wirth, M. Schellmann, T. Haustein, and W. Zirwas, „Synchronization of
cooperative base stations,“ Proc. IEEE ISWCS ’08, pp. 329 – 334, oct 2008.
[35] V. Jungnickel, L. Thiele, M. Schellmann, T. Wirth, W. Zirwas, T. Haustein, and E. Schulz,
„Implementation concepts for distributed cooperative transmission,“ Proc. AC- SSC ’08,
oct 2008.
[36] L. Thiele, T. Wirth, T. Haustein, V. Jungnickel, E. Schulz, , and W. Zirwas, „A unified
feedback scheme for distributed interference management in cellular systems: Benefits
and challenges for real-time implementation,“ Proc. EUSIPCO’09, 2009.
[37] M. Schellmann, L. Thiele, T. Haustein, and V. Jungnickel, „Spatial transmission mode
switching in multi-user MIMO-OFDM systems with user fairness,“ IEEE Trans. Veh.
Technol., vol. 59, no. 1, pp. 235–247, Jan. 2010.
[38] V. Jungnickel, T. Hindelang, T. Haustein, and W. Zirwas. SC-FDMA waveform design,
performance, power dynamics and evoluation to MIMO. In IEEE International Confer-
ence on Portable Information Devices. Orlando, Florida, March 2007.
[39] V. Jungnickel, V. Krueger, G. Istoc, T. Haustein, and C. von Helmolt, „A MIMO system
with reciprocal transceivers for the time-division duplex mode,“ in Proc. IEEE Antennas
and Propagation Society Symposium, June 2004, vol. 2, pp. 1267–1270.
[40] T. Wirth, V. Venkatkumar, T. Haustein, E. Schulz, and R. Halfmann, „LTE-Advanced
relaying for outdoor range extension,“ in VTC2009-Fall, Anchorage, USA, Sep. 2009.
[41] FP7 European Project 318555 5G NOW (5th Generation Non-Orthogonal Waveforms
for Asynchronous Signalling) 2012. [Online]. Available: http://www.5gnow.eu/
[42] G. Wunder et al., “5GNOW: Non-orthogonal, asynchronous waveforms for future mo-
bile applications,“ IEEE Commun. Mag., vol. 52, no. 2, pp. 97_105, Feb. 2014.
[43] R. L. Cavalcante, S. Stanczak, M. Schubert, A. Eisenblaetter, and U. Tuerke, „Toward
energy-efficient 5G wireless communications technologies: Tools for decoupling the
scaling of networks from the growth of operating power,“ IEEE Signal Process. Mag.,
vol. 31, no. 6, pp. 24– 34, Nov. 2014.
[44] 5G Automotive Association http://5gaa.org/
International Data Spaces
Reference architecture for the digitization of industries
8
Prof. D. Eng. Boris Otto
Fraunhofer Institute for Software and Systems
Engineering ISST
Prof. Dr. Michael ten Hompel
Fraunhofer Institute for Material Flow and Logistics IML
Prof. Dr. Stefan Wrobel, Fraunhofer Institute for Intelligent
Analysis and Information Systems IAIS
109
110 Boris Otto • Michael ten Hompel • Stefan Wrobel
The digitization of society and economy is the major trend of our time. The intro-
duction of consumer technologies into businesses, the shift towards digital services
for companies’ business offerings, and the networking of things, people and data
open up new business opportunities in nearly all sectors of the economy. Many of
these have already been analyzed by the Smart Service Welt (“Smart Service
World”) working group within Plattform Industrie 4.0. There are five distinguishing
features of the smart service world [1].
• End-to-end customer process: smart services do not address particular customer
needs but rather support an entire customer process. Sensor manufacturer En-
dress+Hauser, for example, does therefore not only offer measuring systems but
supports customers in all of their questions from planning a production facility,
from advice on product selection, installation and commissioning, and configu-
ration during operation, to advice on replacement investments and plant replace-
ment [4].
• Hybrid service bundles: are no longer just the products central to the product
offering but bundles of physical products and digital services. An example would
be the sportswear manufacturer adidas, which, in addition to running shoes, also
offers the “runtastic” digital service in order to support its customers’ running
experience – the customer process – in its entirety.
• Data at the center: hybrid service bundles, which support the end-to-end custom-
er process, are only possible, if companies efficiently and effectively link their
own data with the data of business partners and contextual data – maintaining
data sovereignty above all. This contextual data often stems from public data
sources and is available as open data. Examples include location data, traffic
information and so on.
8 International Data Spaces 111
• Business ecosystems: the need to combine data from various sources in order to
support the end-to-end customer process leads to the development of business
ecosystems. These are networks of different actors, which form dynamically
around an end-to-end customer process [5]. Well-known examples are the Apple
App Store and Uber.
• Digital platforms: the interaction within the ecosystem and the sharing of a broad
variety of data in the interest of the end-to-end customer process require new
information systems. Digital platforms provide this functionality according to
the “as a service” principle.
The overall characteristics of the smart service world facilitate new business mod-
els, both for new and established companies. For the latter, the key element in dig-
itization is not simply copying new digital business models, but rather supplement-
ing existing strengths – often in product design, production or sales – with smart
service world opportunities. One example is preventative maintenance services by
plant manufacturers that draw on usage data from a number of plant operators for
the early identification of maintenance needs, thus improving the overall equipment
efficiency (OEE).
For the manufacturing industry in particular, digitization changes not only the
product offering but also the product manufacturing side. This is because the con-
sequence of the hybrid service bundle on the offering side is an increase in product
and above all process complexity on the product manufacturing side. Industry 4.0
offers a solution to managing this complexity founded on the organizational princi-
Fig. 8.1 Smart Data Management from the perspective of the focal business
(Fraunhofer-Gesellschaft/Industrial Data Space)
112 Boris Otto • Michael ten Hompel • Stefan Wrobel
The following basic assumptions are applicable when designing this smart data
management system:
• Data is a strategic and economically valuable resource.
• Companies need to share data more intensively than in the past. This applies to
the efficiency, effectiveness, and especially flexibility of conventional data shar-
ing as well as to the extent of the data to be shared. More and more, companies
need to incorporate data into the business ecosystem that was previously classi-
fied as too sensitive to be made accessible to third parties.
• Companies only enter into this data sharing, if their data sovereignty is guaran-
teed, that is, if they can decide who is allowed to use their data on which condi-
tions and for what purpose when the data leaves the company.
The IDS initiative pursues the goal of digital sovereignty in the business ecosystem.
The IDS participants’ aim is to be able to share data with business partners interop-
erably while always retaining self-determination with regards to these data assets.
This aim is to be achieved by designing an IT architecture, demonstrating its
applicability and usefulness in use-case projects. The initiative is founded on two
pillars: a Fraunhofer research project and the International Data Spaces Association
(IDSA). The research project is being carried out by Fraunhofer funded by the
Federal Ministry of Education and Research. Both the project and the association
8 International Data Spaces 113
Fig. 8.2 Smart service world requirements for the International Data Space (Smart Ser-
vice World working group)
are precompetitive and non-profit. The initiative encourages the broadest possible
dissemination and thus also welcomes commercial paths of exploitation, which are
open to all market participants.
From the overarching goal of facilitating confident data sharing in business
ecosystems (smart service world), we are able to develop key requirements for the
architecture of the IDS (see Fig. 8.2) [8]:
A1 Terms of use for data: when the data is shared, the data owner is able to
link enforceable terms of use to the data – rules – that specify on which
conditions the data may be used, by whom, and for what purpose.
A2 Secure data supply chains: the entire data value chain, from the creation/
generation of the data (e.g. via sensors) to its use (e.g. in smart services)
must be able to be secured.
A3 Simple data linking: in-house data must be able to be easily linked with
that of business partners but also with (public) contextual data.
A4 Decentralized data storage: the architecture must not necessarily require
central data storage1. Rather, it must be possible to only share the data with
trusted partners if it is required from a clearly identifiable partner in keep-
ing with the terms of use.
A5 Multiple operating environments: the software components of the IDS
Architecture, which facilitate participation in this virtual data space need
to be able to be run in conventional company IT environments, but also,
for example, on IoT cloud platforms and mobile devices or sensor PCs.
1 Two thirds of companies mistrust central data lake architectures, for example, because they
fear that third parties will have unwanted access to the data cf.[10].
114 Boris Otto • Michael ten Hompel • Stefan Wrobel
A6 Standardized interface: data sharing in the IDS must take place in accord-
ance with a pre-defined and open information model.
A7 Certification: software components and participants must be certified with
respect to keeping to the requirements of Industrial Data Space software
and its operation. The certification criteria are the responsibility of the
International Data Spaces Association.
A8 Data apps and app store: data services (data apps) provide essential func-
tions based on the reference architecture for handling data in the IDS.
Examples are data format conversion and assigning terms of use to the
data. These data services need to be made available via an app store func-
tionality.
The Reference Architecture Model defines the International Data Space’s individ-
ual components and their interaction [9] in terminologically and conceptually con-
sistent terms. It distinguishes five levels and three perspectives.
Data Owner, Data Provider and Data User are the core roles in the IDS. Data sharing
between these core roles is supported by the following intermediary roles:
• Broker: enables data sources to be published and found.
8 International Data Spaces 115
Software and services providers offer the software and services necessary for ful-
filling the different roles.
Finally, the certification authority and one or more inspection bodies ensure that all
Industrial Data space requirements are fulfilled by the software and the participants.
2 The IDS draws on VoCol technologies, software that supports the shared creation and ma-
nagement of vocabularies [9].
3 For details on consortium-based research cf. [6].
116 Boris Otto • Michael ten Hompel • Stefan Wrobel
prototypes in their use cases in order to demonstrate the feasibility and applicabili-
ty of the architecture and the associated benefits, as well as identifying additional
development needs.
Twenty months into the project, six software components are available as proto-
types:
K1 Connector: The central component is the IDS Connector that is used by
the core roles. Prototypes are available for a basic version (without func-
tionality for exercising usage control), for a so-called Trusted Connector
(with usage control), for a sensor connector, and for an embedded version
(for mobile end devices, too).
K2 Usage control: A version of the INDUCE framework [11] and processes
for data labeling are available as technology packages and are additional-
ly integrated into the Trusted Connector.
Information Model
Trusted Connector
Sensor Connector
Identity Provider
Base Connector
Usage Control
Base Broker
App Store
K3 Information model: The model has not only been conceptualized, it is also
available as a software library and can thus be used by software develop-
ers within the International Data Spaces Association.
K4 App store: An initial prototype is available.
K5 Base broker: An initial prototype of the broker with basic registry func-
tionality is available.
K6 Identity provider: An initial prototype version of an identity provider ser-
vice is available.
Table 8.1 shows to what extent the requirements for the IDS are fulfilled by six
software components.
All of the requirements are fulfilled by at least one component. Thus, the require-
ment for the terms of use is addressed both conceptually in the information model
and implemented in initial versions, in particular as the Trusted Connector. The
requirement for various operating environments is comprehensively implemented
via four versions of the Connector architecture. However, requirements A7 (Certi-
fication) and A8 (Data apps and app store) are only addressed in initial basic com-
ponents.
A growing number of models, derivatives, in-car functions, and shorter product li-
fecycles are leading to increasing supply chain complexity in the automotive indus-
try. Modern mid-size vehicles offer so many options and versions that theoretically
more than 1030 varieties of configuration are possible. Around three quarters of the
components required for this do not originate from the manufacturer (original
equipment manufacturers, OEMs) but from suppliers. This product complexity can
only be efficiently and effectively replicated in production and logistics processes,
if suppliers and OEMs work closely together, not only at the planning stage (of re-
quirements, capacities and job sequencing) but also during the execution of the
processes (during transport, manufacture, and assembly). In addition to data that has
been shared for decades through electronic data interchange (EDI) solutions, more
and more data must be shared today that was considered too sensitive to be shared
in the past. This includes:
118 Boris Otto • Michael ten Hompel • Stefan Wrobel
This data will only be shared, if the data owner can specify the terms of use. This is
where the IDS comes into play.
Fig. 8.4 shows the software systems and data flows in the use case of collabora-
tive supply chain management in the automotive industry. In the first phase, the use
case encompasses a tier 1 supplier and an OEM. Both use the Base Connector to
support the following data sharing operations during the project’s initial phase:
In step one, the tier 1 supplier informs the OEM of a supply risk for one of his
suppliers (tier 2) for a subcomponent that the OEM needs. To do this, the tier 1
supplier combines data from his risk management system with master records from
the supplier system and transmits this data to the OEM via the IDS Connector, in-
cluding the terms of use that this data is only to be used for a specific supplier
management purpose for a period of 14 days.
The OEM imports this data into its supplier management system and uses it to
calculate the updated inventory days for specific components that it obtains from
the tier 1 supplier. The OEM in turn sends this inventory days data via the IDS
Fig. 8.4 System components in collaborative supply chain management (Chair for Indus-
trial Information Management TU Dortmund University - formerly Audi Endowed Chair
Supply Net Order Management TU Dortmund University)
8 International Data Spaces 119
Connector to the tier 1 supplier, again including the terms of use, stating the purpose
for use (risk management) and the maximum days of use (three days).
The benefit for the OEM of this use case is obvious. Supply risks in production,
“breaks in the supply chain”, are recognized sooner and manufacturing downtime
therefore avoided. For the tier 1 suppliers, the benefit lies principally in improved
ability to plan their own production since the OEM’s inventory days are made
available to them.
This use case is currently in the software prototype implementation phase.
ished products and the end customer themselves – have an interest in being informed
in real time about events in the supply chain that lead to plan changes.
This use case addresses notification of delayed arrival on the inbound side of the
manufacturing company. Here, the transportation company informs the logistics
service provider, using a mobile version of the IDS Connector, that a specific ship-
ment is delayed and states the reason for this delay. The logistics service provider
calculates an updated expected arrival time and transmits this to the manufacturing
company awaiting the consignment. The manufacturing company, in turn, confirms
the new arrival time and transmits updated details in regards to the unloading point.
Data is shared according to the IDS information model. For the payload itself
“packed” in the IDS notification, GS1 EDI XML is used.
The benefit for the manufacturing business lies in the improvement of the plan-
ning of inbound logistics processes/yard management (including staff planning
within the receiving department) and production (including job controlling). For the
logistics service provider, the use case offers value-added services for customers.
For the transportation company the check-in process at the manufacturing company
is made easier because updated information regarding time slots and unloading
points is always available.
This use case is currently being piloted with thyssenkrupp and already covers
more than 500 shipments per month.
The model of roles for the IDS is fundamentally structured so as to allow for a
separation of the role of Data Owner from that of Data Provider. This allows for new
business models, specifically data trusteeship.
The processing of data across different service and process steps involves sig-
nificant requirements in terms of data protection and data sovereignty that not all
companies render themselves, but obtain as a service. The data trustee thus needs
to ensure that data does not leak out to competitor companies, that personal data is
adequately anonymized, and that services are instructed to only use data in accord-
ance with the terms, deleting the data where necessary after use.
Independent audit firms and technical testing organizations are well suited
to provide data sovereignty services in business ecosystems, including the
following:
• Reviewing rules on terms of use for any conflicts
• Monitoring observation of terms of use
• Monitoring data transactions as a clearing house
8 International Data Spaces 121
Fig. 8.6 System chain for transparency in the supply chain (Fraunhofer ISST)
The certification requirements and criteria, roles and certification levels, and audit-
ing methods in the IDS are defined in a certification schema. Every organization
that participates in the IDS is audited and certified according to this schema, as are
all of the IDS’s software components.
In addition to this, the certification body also facilitates, for example certifi-
cate-based identity management within the IDS, which is the technical basis for all
Connector implementations.
This use case is currently in the conception phase.
Industry 4.0 is an organizational principle for the industrial operations of the future
and rests on, among other things, the networking of all resources within product
manufacturing. Machines, facilities, and staff within manufacturing are able to
share information in near real time, transmitting manufacturing status data and order
data, but also receiving contextual information on individual production steps. Since
in many industries today manufacturing happens in distributed production net-
works, companies have two challenges to overcome:
• Shared semantic descriptions of manufacturing resources in the production net-
work
• Confident data sharing between the individual resources (e.g. machines)
122 Boris Otto • Michael ten Hompel • Stefan Wrobel
More than two thirds of all new products are based on new materials. In order to
safeguard the innovational strength and retain technological sovereignty and closed
value chains within Germany, universal digitization of material and product prop-
erties over their entire lifecycle (“from ore to the fridge”) is of strategic importance.
The Materials Data Space provides digitized information on materials and their
properties, components, and their alterations in manufacturing and use over the
entire value chain, thus covering their entire lifecycle from use to strategic recy-
cling. The Materials Data Space, a verticalization of the IDS architecture, is a
strategic initiative of the Fraunhofer MATERIALS group, which aims for compre-
hensive digital image within the entire business ecosystem (see Fig. 8.8).
The steel industry, for example, does not only sell steel strips themselves but also
the digital image of the steel strip (microstructure, composition, inclusions, materi-
al history, etc.). The combination of the physical product together with the digital
image is a key success factor for the future competitiveness of companies such as
Salzgitter AG, the implementation partner in this use case. In addition, combining
the data over the entire lifecycle results in shorter development times and learning
Networking in value chains is not only limited to companies or people but, with the
spread of the Internet of Things, increasingly stretches to objects, too. Load carriers,
containers, trucks, etc. are all uniquely identifiable, communicate with their envi-
ronment, and make decisions independently (with respect to shipping routes, for
example). In the course of this, these objects produce data, specifically value chain
status data. They provide information, for example, on locations (including time
stamps), atmospheric conditions (temperature, humidity), and shocks.
Agile networking in value chains requires both confident sharing of this kind of
status data by the objects themselves and the identification of relevant data sources.
This is because master records and status data from small load carriers are shared
in value chains with changing business partners across companies.
In the STRIKE use case (Standardized RTI Tracking Communication Across
Industrial Enterprises), status data is sent via the IDS Connector in a useable format
for the recipient. This status data is produced and managed using the EPCIS stand-
ard. Using the example of new load carriers equipped with RFID, a demonstration
is given of how the status data can be assigned to the data owner within the GS1
community. The data owners therefore remain sovereign over their data.
The existing GS1 partner registry, GEPIR (Global Electronic Party Information
Registry), is also being developed into an Industrial Data Space broker in order to
be able to identify providers of status data in the value chain. In addition, this ex-
ample shows how value-added services can be derived from the data with the help
of apps.
Fig. 8.9 shows the overall architecture for the STRIKE use case.
Use cases assessment mirrors the strategy for the implementation of the prototype
reference architecture model (see Table 8.2). This is because prototypes of the IDS
Connector, as the architecture’s core component in various implementations, are
being put to work or have been planned in all of the use cases. In addition, the use
cases confirm the coexistence of various Connector implementations with different
functional characteristics that nevertheless all adhere to the principles of the refer-
ence architecture model.
The prerequisite here is conformity to the IDS information model. This criterion
is being fulfilled in all cases.
This analysis also shows the step-by-step development or implementation of
important functionalities, since usage control technologies are neither in use nor
planned for all cases.
In addition, the need of action with respect to implementation and use of the app
store and broker components becomes apparent.
Table 8.3 additionally shows to what extent the strategic requirements for the
IDS architecture are a reality in the six use cases.
This analysis demonstrates that the focus in the use cases during the initial phas-
es of the initiative lies on the standardized interface, which is in use in all of the
examples. The requirement for decentralized data storage is also being fulfilled in
nearly all of the use cases.
126 Boris Otto • Michael ten Hompel • Stefan Wrobel
Embedded Connector
Information Model
Trusted Connector
Sensor Connector
Identity Provider
Base Connector
Usage Control
Base Broker
App Store
F1 Collaborative SCM X (X) X X
F2 Transparency in supply chains X X X X (X)
Specific requirements – such as those for a secure data supply chain and multiple
operating environments – feature in those use cases where these specifications are
relevant.
Areas that have been addressed only to a limited extent in the project’s initial
phase are the terms of use (only currently being utilized in F1) and the possibility
to provide data apps. The latter is no surprise since an app ecosystem can only de-
velop with the increasing dissemination of the initiative. The same applies to data
linking: this requirement will increasingly move to the fore of implementation as
use case scenarios become more complex.
The requirement to be able to attach terms of use to the data will become increas-
ingly significant during future use case implementation phases when the basic
communication is able to be conducted via the IDS and when increasingly sensitive
data elements are being shared.
8 International Data Spaces 127
Digital networking of
Agile networking in
Collaborative SCM
production lines
[1] Arbeitskreis Smart Service Welt. (2015). Smart Service Welt: Umsetzungsempfehlungen
für das Zukunftsprojekt Internetbasierte Dienste für die Wirtschaft. Berlin.
[2] Bauernhansl, T., ten Hompel, M., & Vogel-Heuser, B. (2014). Industrie 4.0 in Produk-
tion, Automatisierung und Logistik. Anwendung · Technologien · Migration. Berlin:
Springer.
[3] Hermann, M., Pentek, T., & Otto, B. (2016). Design Principles for Industrie 4.0 Scena-
rios. 49th Hawaii International Conference on System Sciences (HICSS 2016) (S. 3928-
3927). Koloa, HI, USA: IEEE.
[4] Kagermann, H., & Österle, H. (2007). Geschäftsmodelle 2010: Wie CEOs Unternehmen
transformieren. Frankfurt: Frankfurter Allgemeine Buch.
[5] Moore, J. F. (2006). Business ecosystems and the view from the firm. Antitrust Bulletin,
51(1), S. 31-75.
128 Boris Otto • Michael ten Hompel • Stefan Wrobel
[6] Österle, H., & Otto, B. (Oktober 2010). Konsortialforschung: Eine Methode für die
Zusammenarbeit von Forschung und Praxis in der gestaltungsorientierten Wirtschafts-
informatikforschung. WIRTSCHAFTSINFORMATIK, 52(5), S. 273-285.
[7] Otto, B. (2016). Digitale Souveränität: Beitrag des Industrial Data Space. München:
Fraunhofer.
[8] Otto, B., Auer, S., Cirullies, J., Jürjens, J., Menz, N., Schon, J., & Wenzel, S. (2016).
Industrial Data Space: Digitale Souveränität über Daten. München, Berlin: Fraunhofer-
Gesellschaft; Industrial Data Space e.V.
[9] Otto, B., Lohmann, S. et al. (2017). Reference Architecture Model for the Industrial Data
Space. München, Berlin: Fraunhofer-Gesellschaft, Industrial Data Space e.V.
[10] PricewaterhouseCoopers GmbH. (2017). Datenaustausch als wesentlicher Bestandteil
der Digitalisierung. Düsseldorf.
[11] Steinebach, B., Krempel, E., Jung, C., Hoffmann, M. (2016). Datenschutz und Datenana-
lyse: Herausforderungen und Lösungsansätze, DuD – Datenschutz und Datensicherheit
7/2016
EMOIO Research Project
An interface to the world of computers
9
Prof. D. Eng. Hon. Prof. Wilhelm Bauer ·
Dr. rer. nat. Mathias Vukelić
Fraunhofer Institute for Industrial Engineering IAO
Summary
Adaptive assistance systems are able to support the user in a wide range of
different situations. These systems take external information and attempt to
deduce user intentions from the context of use, without requiring or allowing
direct feedback from the user. For this reason, it remains unclear whether
the system’s behavior was in accordance with the user’s intentions – leading
to problems in the interaction between human and adaptive technology. The
goal of the EMOIO project is to overcome potential barriers of use with the
aid of neuroscientific methods. Merging ergonomics with the neurosciences
into the new field of neuroergonomics research produces enormous potential
for innovation, to make the symbiosis between humans and technology more
intuitive. To this end, brain-computer interfaces (BCIs) offer a new generation
of interfaces between humans and technology. BCIs make it possible to regi-
ster mental states such as attention and emotions and transmit this information
directly to a technological system. So-called neuroadaptive systems conti-
nuously use this information in order to adjust the behavior, functions or the
content of an interactive system accordingly. A neuroadaptive system is being
developed by a consortium of partners from research and industry as part of
the EMOIO project. The goal of the system is to recognize, based on the users’
brain activity, whether system-initiated behaviors are approved or rejected.
The system is able to use this information to provide the person with the best
possible assistance and thus adapt to individual and situational demands. To do
this, neuroscientific methods such as electroencephalography (EEG) and
129
130 Wilhelm Bauer • Mathias Vukelić
Project aim
The aim of the project is to use neuroscientific methods to reduce barriers to the use of
assistance systems. To do this, Fraunhofer IAO, together with partners from research and
industry, is developing a neuroadaptive system as part of the EMOIO project. The system
is intended to recognize, based on the users’ brain activity, whether the user approves or
rejects the system’s behavior, and adapt that behavior accordingly.
Cooperation partners
University of Tübingen – Institute of Medical Psychology and Behavioral Neurobiology at
the Faculty of Medicine and Institute of Psychology, NIRx Medizintechnik GmbH, Brain
Products GmbH, University of Stuttgart – Institute of Human Factors and Technology
Management (IAT)
Research plan
Project schedule:
Phase 1: 01/2015 to 12/2016
Phase 2: 01/2017 to 12/2017
Key findings
• Representative neuronal correlates of affective responses during interaction with techno-
logy
• Real-time classification of neuronal correlates of affective user reactions
• Miniaturization of EEG/fNIRS hardware sensors and optimized simultaneous EEG/
NIRS measurement
• Testing of real-time classification in three fields of application: smart home, human-robot
collaboration, vehicle interaction
Contact
Dr. rer. nat. Mathias Vukelić,
[email protected]
9 EMOIO Research Project 131
Digitization is having a vital impact on human’s world and everyday lives, and in-
teraction with digital technology is becoming ever more self-evident and important.
Increasingly intelligent technological products are finding their way into our every-
day working lives. Digitization fosters the networking of various technological
systems, facilitating greater ease of communication and cooperation with one an-
other as we use them in our everyday and working lives. Autonomous robots are
making industrial production easier, and architects and designers are planning and
developing their solutions in virtual reality. These kinds of solutions are facilitating
improvements in productivity and efficiency, with corresponding time savings.
Nevertheless, the increasing integration of technology in our workplaces also entails
new challenges and potential for conflict. Often, humans with their individual pref-
erences and needs find themselves overlooked in the development of technological
systems. The resulting solutions, whilst technologically advanced, may nevertheless
offer limited gains in terms of the productivity, creativity, and health of the users in
question. The challenge facing this growing use of technology is to create suitable
working environments where humans can receive the best possible support in their
tasks in the broadest range of situations.
The increasing digitization of the workplace means research findings that help us
understand and improve the interaction between human and technology – and there-
fore facilitate humans using technology products efficiently – are of key impor-
tance. Set against this backdrop, the research field of human-technology interaction
is constantly growing in importance. In addition to usability, also cognitive and
emotional user factors play a key role here, especially in the workplace. This raises
the following questions: how great is the human cognitive load when they are work-
ing? What happens to human’s emotional well-being when they are interacting with
technology? The classical psychological and engineering methods used in ergonom-
ics provide inadequate answers to these questions. There is thus a need for supple-
menting existing procedures with a component that facilitates access to the user’s
implicit mental state. This would allow the cognitive and emotional processing,
which is not immediately apparent to the conscious mind, to be made accessible to
the interaction design process.
Since 2012, under the heading of neuroergonomics, Fraunhofer IAO has been
researching the extent to which neuroscientific methods may be used to capture
132 Wilhelm Bauer • Mathias Vukelić
cognitive and emotional user states during technology use, and to facilitate new
perspectives and technological options in ergonomics. The institute is pursuing an
interdisciplinary research approach here, combining expertise from the fields of
psychology, information technology, engineering and neuroscience. Neuroergo-
nomics is considered a research field of the future with great potential for science
and economic praxis: ranging from workplace design, through virtual engineering
and vehicle ergonomics, to user-friendly IT systems. Appropriate techniques for
recording and measuring brain activity that are non-invasive, inexpensive, and se-
cure are electroencephalography (EEG) on the one hand and functional near-infra-
red spectroscopy (fNIRS) on the other. EEG directly captures the sum of the electri-
cal activity of nerve cells in the brain by recording voltage fluctuations via elec-
trodes on the scalp [1]. fNIRS measures the metabolic processes related to nerve
cell activity, specifically the amount of oxy- and deoxygenated hemoglobin in the
blood vessels and thus captures changes in blood flow in the brain [1]. These are the
two techniques that are primarily used in the institute’s “NeuroLab” to measure
brain activity, measurement techniques, which are also well suited to mobile ap-
plication scenarios. Based on the brain’s activation patterns, it is thus possible to
measure and quantify various cognitive and emotional states of the user that are
relevant in the working context. Against the backdrop of the increasing significance
of emotional aspects in human-technology interaction, an initial pilot project has
already demonstrated that basic brain processes of the “user experience” [2] during
technology use [3] can be measured as objectively as possible with neuroscientific
techniques.
One of the key issues, that Fraunhofer IAO has been working on intensively in the
context of neuroergonomics research, is adaptive and assistance systems. Ever more
frequently today we encounter interactive assistance systems that are able to act
independently in certain everyday situations or carry out tasks in the workplace
autonomously. Assistance systems that act independently can support users in a
wide variety of situations. They can be seamlessly integrated into everyday life and
can help to reduce the fear that less technologically savvy users have of interacting
with increasingly complex digital systems. Nevertheless, these systems are not al-
ways purely beneficial; conflict can also arise precisely during the interaction be-
tween user and adaptive technologies. In order to support users, these kinds of
systems have to fall back on external information and attempt to deduce the aims
and intentions of the human from the context of use. Current technological systems
9 EMOIO Research Project 133
are insufficiently equipped to observe their users without any interruption and delay
and thus draw corresponding conclusions in real time. It thus remains unclear
whether the system’s behavior was in the interest of the user. Instead of the intend-
ed support being provided by the system’s adaptive behavior, it can lead to users
feeling a loss of control and rejection towards the system. The question arises, then,
how assistance systems can be designed in future so that this potential for conflict
is reduced and user acceptance increased. Whereas intelligent systems already make
use of a range of contextual data such as devices used, environmental conditions,
and distance from the display [4–6] in order to offer optimal adaptation, the poten-
tial for emotion sensing as an input parameter for adaptive systems is as of yet
largely untapped.
In the EMOIO project, sponsored by the Federal Ministry of Education and
Research (Bundesministerium für Bildung und Forschung, BMBF), Fraunhofer
IAO and its five additional partners have set themselves the goal of developing a
brain-computer interface that captures the subjectively perceived suitability (ap-
proval/rejection) of system-initiated assistance behaviors, evaluates it, and delivers
it to an adaptive system so that its assistance functions can be optimally adapted to
the user. The following section provides information on the initial results of the
project, a project which is part of the human-technology interaction and neuroergo-
nomics research fields.
The period where we only communicated with computers using input media such
as a mouse and keyboards is over. Some technological products let us talk to them
while others respond to gestures. Technological products that are becoming increas-
ingly intelligent are being integrated into everyday working life. The brain-comput-
er interface (BCI) is currently the most direct form of an interface for interaction
and communication between user and technological systems. BCI applications have
thus far been concentrated first and foremost on the clinical environment. The in-
terface enables users with physical or perceptual limitations, such as after a stroke
or locked-in patients, to communicate with their environment, surf in the Web, or
even paint [7–11]. Furthermore, BCIs have been used in the context of neurofeed-
back training for treating psychiatric conditions such as depression and schizophre-
nia [12–14]. Users without physical impairment, too, can benefit from this kind of
interface in their everyday lives or at work. Alongside the active and intentional
control of BCIs by the user, so-called “passive” BCIs have also become established
134 Wilhelm Bauer • Mathias Vukelić
in recent years [15]. Passive BCIs do not require intentional control by the indi-
vidual. They capture cognitive and emotional states such as affect, mental load, or
surprise effects, and transmit these directly to a technological system [16][17]. If
this information is used as a basis for correspondingly adapting a system’s content
and functions or user interface during interaction, then it is known as a neuro-adap-
tive technology [19]. This kind of technology is formed essentially of three parts
(see Fig. 9.1).
The first part (afference) consists in collecting the available and observable data
about the user and the user context via neuroscientific measurement techniques (e.g.
EEG or fNIRS). In the second stage (inference), this data is classified with the aid
of machine learning algorithms, so that relevant cognitive or emotional information
from the user can be interpreted. In the final stage (efference), the decision regard-
ing the system’s adaptation behavior and the execution of the adaptation on the
basis of the user’s cognitive and emotional states (and thus transparent feedback on
system behavior) is provided to the user.
In daily life, emotions help us to assess situations we experience and thus to act
accordingly. Just like perception and movement, emotions are represented by neu-
ronal patterns in the brain. These patterns can be captured using appropriate sensors
and computer algorithms [20]. The key issue here is improving the interaction be-
tween user’s and technology. Neuroscience shows how closely the individual’s
emotional system is connected with their brain. Furthermore, cognitive and memo-
ry processes influence human emotional states such as anger, joy, or affect.
In psychology, emotional intelligence is said to play an essential role in human
decision-making. The theory is that social interaction and communication with
other individuals works because we form a mental model of the other person’s
current state (see “Theory of Mind” in [21]). This enables us to identify the intend-
ed actions of the person we are interacting with and adapt our actions accordingly.
This ability – the ability to recognize the emotional state of our human interaction
partner – has until now been completely absent in technological systems [22]. It
follows, that the interaction can be decisively improved if information about the
current emotional state of the user can be provided to the technology system via a
BCI. In this way, there should be the option, via the BCI, to optimally adapt the
assistance system’s behavior to the user. Thus, the active user feedback becomes
superfluous and the interaction is uninterrupted.
The vision is to consistently orient the increasing intelligence and autonomy of
technology towards people’s individual needs and preferences so that neuro-adap-
tive technologies support and assist them as effectively as possible. Despite the
many advantages, there is a still a gap between the potential benefit offered and the
actual value added by this technology for applications outside of the medical field
– for application-oriented human-technology interaction scenarios. The EMOIO
project is laying the foundations for closing this gap between basic BCI research
and practical applications of human-technology interaction.
136 Wilhelm Bauer • Mathias Vukelić
Up to this point, the capturing of emotional or affective states via mobile neuro-
physiological methods such as EEG and fNIRS has largely remained in the do-
main of basic neuroscientific research. Normally, the experiments take place un-
der strictly controlled test conditions where mostly purely receptive image, audio,
or video material is used to induce corresponding emotional responses in the
participants.
There are also very few studies in the field of human-technology interaction
research that have examined affective responses using neurophysiological methods
[23][24]. Studies with a plausible interaction scenario between an user and an
adaptive assistance system have until now rarely been conducted. The main aim of
EMOIO’s initial phase was thus to research and identify the brain’s pattern of ac-
tivation, that are underlying the user’s emotional state of satisfaction (positive af-
fect) or rejection (negative affect) while they are interacting with assistance sys-
tems. As part of the project, a new, interactive experimental paradigm called AF-
FINDU [23] was developed to identify affective user reactions to assistance func-
tions in realistic settings. An empirical study was conducted to research the
foundations of affect detection. The study simultaneously used EEG, fNIRS, and
various secondary psychological measuring techniques (such as measuring muscle
activity in the face, or recording the heart rate variability). In total, the affective
reactions of 60 participants (aged between 18 and 71 years) to static image stimu-
li (standardized stimulus material from neuroscientific basic research suitable for
comparative purposes) and during the interaction with AFFINDU were recorded.
Fig. 9.2 provides an insight into the work at Fraunhofer IAO’s “NeuroLab” whilst
preparing the experiments (attaching the measuring sensors to individuals) and
conducting them.
Fig. 9.2 C shows the experiment being carried out while a participant is interact-
ing with AFFINDU. AFFINDU is a prototype of an assistance system consisting of
a graphical menu interface with 16 different menu items. The system is able to in-
duce corresponding affective responses in the users during the interaction. A de-
tailed description of the functionality and procedure of the experiment is provided
in [23]. In summary, the interaction with AFFINDU represents a plausible usage
scenario in which two key system-initiated behaviors of an adaptive assistance
9 EMOIO Research Project 137
Fig. 9.2 Work at the Fraunhofer IAO NeuroLab: test preparations for conducting the ex-
periment (A, attaching the measurement techniques); measurement techniques (B, EEG
and fNIRS sensors, whole head coverage); carrying out the experiment (C, participant, left,
interacting with AFFINDU) (Fraunhofer IAO)
system are simulated. Participants were asked to use a keyboard to navigate through
the menu while they were asked to select various target menu items. This naviga-
tional task represents a very simple user goal where AFFINDU is able to support
the user appropriately (positive scenario). In the positive scenario, the user goal –
that is, the desired menu item – is recognized as correct by AFFINDU and the du-
ration of navigation to the target menu item correspondingly shortened. AFFINDU
may also misidentify the user’s goal and thus hinder the user in reaching their goal
(negative scenario). The user’s emotional assessment of the system behavior can be
plotted in two independent dimensions: valence (positive to negative) and arousal
(calm to aroused). Thus, the system’s behavior is assessed positively if it is condu-
cive to the user’s goal (the user reaches the desired menu item more quickly),
whereas it is assessed negatively if AFFINDU’s behavior is not aligned with the
user’s goal (the user needs more time to reach the desired menu item). The results
of our research show that the induction and quality of affective user reactions (pos-
itive and negative) to individual assistance functions can be successfully measured
during interaction using neurophysiological measuring methods such as EEG and
fNIRS [23].
138 Wilhelm Bauer • Mathias Vukelić
In order to be able to research the brain patterns of related to the individual’s af-
fective user reactions towards system-initiated assistance functions in a specific
manner, the interaction with AFFINDU was implemented in an event-related
experimental trial-procedure [23]. In this way, the brain’s responses within a fixed
time window of interest to a given assistance event can be identified and
interpreted. Objective quantification of the event-related activity can also be car-
ried out by analyzing the amplitudes and latencies of the time course of the EEG
and fNIRS signals over time. This provides the condition for specific examina-
tions and discriminability of neuronal correlates of emotional-affective user reac-
tions using EEG and fNIRS. The identification and localization (including the
selection of representative EEG and fNIRS positions) of reliable patterns from the
individual measuring modalities is thus a focus of the EMOIO research project.
These neuronal correlates serve as the basis for the development of an algorithm
capable of real-time evaluation and classification of EEG/fNIRS data for affect
detection.
The results from the evaluations of the EEG signal dynamics over time show that
it is possible to measure neuronal correlates, which permit a reliable discrimination
between supportive (positive affective response) and hindering (negative affective
response) assistance behaviors after just 200msec in specific EEG positions. Addi-
tional representative EEG time periods for discriminating affective user reactions
can be observed from approx. 300msec and 550msec after the system-initiated as-
sistance function.
Due to its high temporal resolution, the electrical activity measured by the EEG
can furthermore be divided into several frequency bands, so-called EEG bands that
are characterized by specific rhythmic oscillations. Here, oscillatory activities can
be observed from low frequencies in the range from 0.1 to above 4Hz (delta waves),
4 to 8Hz (theta waves), 8 to 13Hz (alpha waves), up to faster oscillations in the range
from 13 to 30Hz(beta waves) and above 30Hz (gamma waves). Analyzing the am-
plitudes of the EEG’s different frequency components may allow further conclu-
sions to be drawn regarding the user’s cognitive and emotional processes. Further-
more, with respect to assessing positive and negative system-initiated assistance
functions, the alpha, beta, and gamma EEG frequency bands have shown to be re-
liable neuronal correlates of an individual’s emotional-affective reactions during
interaction with AFFINDU.
The advantage of EEG measurements lies in their capacity for high temporal
resolution, permitting an exact chronometric allocation of cognitive and emotional
9 EMOIO Research Project 139
processes. EEG’s spatial resolution, however, is very limited and usually lies in the
region of a few centimeters.
In order to investigate the relationships between the individual’s emotional states
and brain patterns, methods are required that offer not only a very high temporal
resolution of the activity of particular brain regions but also a good spatial resolu-
tion. The requirement for a good spatial resolution is fulfilled by the fNIRS method.
Using fNIRS, changes can be recorded in the blood flow of the brain nerve cells can
be picked up in brain regions lying up to 3 cm below the surface of the scalp. Thus,
the fNIRS method is highly suited to capture the local activity in specific regions of
the cerebral cortex that are related to emotional processing. The results from the
evaluations of the fNIRS data from the AFFINDU experiment show that both the
frontal as well as regions in the back of the cerebral cortex respond sensitively to
different adaptation behaviors. It is well known that these regions are associated
with motivational aspects and the semantic meaning of emotional processing in
human beings. In particular, it was shown that activity in these regions increases in
the case of a positive support by AFFINDU, whereas the activity decreases in the
case of a negative event provided by AFFINDU. These neuronal correlates provide
additional representative neuronal signatures for user’s emotional-affective reac-
tions when interacting with technology that can be combined with the patterns found
from the EEG data analysis.
Research was also carried out within the project into the differences in brain
activity in different age groups. We know from basic neurobiological and psycho-
logical research that developmental changes (differentiation/dedifferentiation) in
cognitive and sensory functions occur with age. Emotional research has also found
that the extent of positive and negative affectivity changes in elderly people, show-
ing that subjective assessments of negative emotions appearing to diminish with
increasing age [25]. We can thus assume that older people in particular will demon-
strate differences in the neuronal correlates of emotional-affective reactions. We
were also successfully able to demonstrate within the project that this is the case by
means of additional results from the AFFINDU experiment [26]. To do this, study
participants were divided into two groups based on a median split of their age:
“young” (aged between 22 and 44 years), and “old” (aged between 48 and 71 years).
We were able to show that the older participants experienced AFFINDU’s support-
ive assistance function as more positive and its impeding function as less negative
as compared to the younger group of participants. Furthermore, , it could been
shown that the event-related EEG amplitude responses correlate with age, with dif-
ferences between the two groups especially present in the time interval starting at
around 300 msec after the self-initiated assistance event. These results suggests that
there are distinct cognitive strategies for emotional-affective processing in the aging
140 Wilhelm Bauer • Mathias Vukelić
brain. These results show that age-related differences in the neuronal correlates must
be taken into account when developing adaptive technologies towards user’s current
preferences and individual needs.
These neuronal correlates of emotional-affective user reactions found in the
project correspond extensively with results for temporal and spatial (EEG and
fNIRS positions) EEG and fNIRS patterns known from the neuroscientific litera-
ture. The results from the first part of the project thus provide an important contri-
bution to basic research into affect detection in human-technology interaction using
neurophysiological methods.
The results presented on the neuronal correlates provide the basis for developing an
algorithm that combines and evaluates EEG and fNIRS data in real-time. This data
can be used to classify user’s affective reactions during the interaction with tech-
nology. To do this, further studies in signal processing and machine learning algo-
rithms for classifying the neuronal correlates were required from EMOIO project
partners. The principal advantage of the multimodal approach lies in the fact that
the individual strengths of the one measurement techniques can be used to compen-
sate for the disadvantage of the other. Thus, the limited spatial resolution of the EEG
can be supplemented by the fNIRS measurement techniques, thus, facilitating the
localization of brain activity. On the other hand, the higher temporal resolution of
EEG can compensate the deficiencies of fNIRS in this regard. Compared with a
unimodal approach, the combination of metabolic and neuroelectrical data enables
us to achieve improved classification accuracies for estimating the user’s emotion-
al-affective state. Furthermore, the multimodal approach reduces the susceptibility
that are prone to errors of a unimodal classification approach, which might occur
due to artifacts, for example due to the temporary failure of a recording modality
alone or the recording of incorrect values such as muscle artifacts in the EEG. In
this way, the affect classification can be carried out exclusively by using the inter-
ference-free measurement modality. For applied research, a combination of the two
techniques also offers, alongside generally more reliable recording of brain activity,
the benefit of an inexpensive mobile measuring technique without any usage restric-
tions. Building on the results achieved in the foundational empirical studies, the
9 EMOIO Research Project 141
A report published recently by the BNCI Horizon 20201 project sponsored by the
EU [26] shows that there are currently 148 BCI-related industrial stakeholders in
the market. These stakeholders encompass a variety of sectors such as automotive,
aerospace, medical technology, rehabilitation, robotics, and entertainment and mar-
keting. The information and communications sector offers a large potential for the
long-term development and integration of BCIs. Especially, in the fields of multi-
modal operating systems and ambient intelligence. Companies such as Microsoft2
and Philips3 have been looking into the question of how neuroscientific methods
can be integrated into human-technology interaction for many years. In addition,
other technology-driven companies from Silicon Valley such as Facebook4 and Elon
Musk’s recently founded start-up NeuraLink5 are also investing heavily into the
development of future BCI applications. An additional field for applied research
might also be the factory of the future. Cognigame6 and ExoHand7 are a couple of
research projects by FESTO that investigate interaction concepts using BCI in the
industrial context. The automotive industry, too, would see benefit from BCI’s in
1 http://bnci-horizon-2020.eu/
2 http://research.microsoft.com/en-us/um/redmond/groups/cue/bci/
3 http://www.design.philips.com/about/design/designnews/pressreleases/rationalizer.page
4 https://www.scientificamerican.com/article/facebook-launches-moon-shot-effort-to-deco-
de-speech-direct-from-the-brain/
5 https://www.wsj.com/articles/elon-musk-launches-neuralink-to-connect-brains-with-com-
puters-1490642652
6 https://www.festo.com/cms/de_corp/12740.htm
7 https://www.festo.com/cms/de_corp/12713.htm
142 Wilhelm Bauer • Mathias Vukelić
the future. For example, BCI could model driver states for situational adaptive
driver assistance. Thus, an intelligent system could warn the driver if he or she is
tired, stressed, or distracted from information gathered via the BCI. Jaguar Land
Rover8, Nissan9 and Daimler [28] have already demonstrated such concepts of driv-
er modelling through various research projects. Capturing brain states via BCI also
offers interesting potential applications in the field of digital media for knowledge
and training software. For example, training software could adapt itself to the user’s
momentary receptiveness and ability to concentrate and thus controlling theamount
of learning material such that the user is not overwhelmed.
In summary, we can say that neuroergonomics is still a relatively new field of
research, which is why neuro-adaptive technology use is oriented around future
possibilities. Whether in the end, this kind of technology really has the potential to
be a success depends on user acceptance as well as technological feasibility. Espe-
cially, in the area of user acceptance there is still, room for improvement due to the
limitations of current sensor technology. However, neuroscientific research is al-
ready providing promising results in terms of relevant improvements in the minia-
turization and mobility of the sensor technology [29]. Furthermore, this field will
be advanced quite significantly also by the entertainment industry10. They are work-
ing on new design concepts for head-mounted sensor technology that can be used
in a broad range of applications. Neuro-adaptive technologies offer a huge potential
for use in the workplace. It is precisely in this context that diverse questions arise,
ranging from issues of user’s individual autonomy to data privacy. This means that
neuro-adaptive technologies are also the subject of ethical discussions. These re-
quire future academic debate by incorporating so-called ELSI questions (ethical,
legal, and social implications). In the context of the EMOIO project, ELSI questions
are being closely examined through accompanying research.
[1] R. Parasuraman and M. Rizzo, Eds., Neuroergonomics: the brain at work. New York:
Oxford University Press, 2008.
[2] M. Hassenzahl, „User experience (UX): towards an experiential perspective on product
quality,“ 2008, p. 11.
8 http://newsroom.jaguarlandrover.com/en-in/jlr-corp/news/2015/06/jlr_road_safety_res-
search_brain_wave_monitoring_170615/
9 http://cnbi.epfl.ch/page-81043-en.html
10 https://www.emotiv.com/
9 EMOIO Research Project 143
[20] R. W. Picard, E. Vyzas, and J. Healey, „Toward machine emotional intelligence: analysis
of affective physiological state,“ IEEE Trans. Pattern Anal. Mach. Intell., vol. 23, no. 10,
pp. 1175–1191, Oct. 2001.
[21] H. Förstl, Ed., Theory of mind: Neurobiologie und Psychologie sozialen Verhaltens, 2.,
Überarb. und aktualisierte Aufl. Berlin: Springer, 2012.
[22] T. O. Zander, J. Brönstrup, R. Lorenz, and L. R. Krol, „Towards BCI-Based Implicit
Control in Human–Computer Interaction,“ in Advances in Physiological Computing, S.
H. Fairclough and K. Gilleade, Eds. London: Springer London, 2014, pp. 67–90.
[23] K. Pollmann, D. Ziegler, M. Peissner, and M. Vukelić, „A New Experimental Paradigm
for Affective Research in Neuro-adaptive Technologies,“ 2017, pp. 1–8.
[24] K. Pollmann, M. Vukelic, and M. Peissner, „Towards affect detection during human-
technology interaction: An empirical study using a combined EEG and fNIRS ap-
proach,“ 2015, pp. 726–732.
[25] M. Mather and L. L. Carstensen, „Aging and motivated cognition: the positivity effect
in attention and memory,“ Trends Cogn. Sci., vol. 9, no. 10, pp. 496–502, Oct. 2005.
[26] M. Vukelić, K. Pollmann, M. Peissner, „Towards brain-based interaction between hu-
mans and technology: does age matter?,“ 1st International Neuroergonomics Confe-
rence, Oct. 2016.
[27] C. Brunner et al., „BNCI Horizon 2020: towards a roadmap for the BCI community,“
Brain-Comput. Interfaces, vol. 2, no. 1, pp. 1–10, Jan. 2015.
[28] A. Sonnleitner, et al., „EEG alpha spindles and prolonged brake reaction times during
auditory distraction in an on-road driving study“. Accid. Anal. Prev. 62, 110–118, 2014
[29] S. Debener, R. Emkes, M. De Vos, and M. Bleichner, „Unobtrusive ambulatory EEG
using a smartphone and flexible printed electrodes around the ear,“ Sci. Rep., vol. 5, no.
1, Dec. 2015
Fraunhofer Additive Manufacturing
Alliance 10
From data straight to highly complex products
Summary
Additive manufacturing is known as 3D printing in popular science. It refers
to a relatively new group of manufacturing techniques with unique properties
and possibilities compared with conventional manufacturing technologies.
The Fraunhofer Additive Manufacturing Alliance currently coordinates 17
Fraunhofer institutes working on additive manufacturing. It covers the entire
process chain: the development, application and implementation of additive
manufacturing techniques and processes as well as the relevant materials. This
chapter provides an overview of the technologies, applications, particular op-
portunities and further goals of applied research in the area of additive manuf-
acturing within the Fraunhofer-Gesellschaft. We make particular mention of
mesoscopic lightweight design, biomimetic structures, high-performance tools
for hot sheet metal forming, ceramic components, printable biomaterial, large-
size plastic components, integrating sensory-diagnostic and actuator therapeu-
tic functions into implants, and three-dimensional multimaterial components.
145
146 Bernhard Müller
The last two points are the key drivers today, contributing to the increasing spread
of additive manufacturing as an alternative production technique. The central chal-
lenge here is to master the competition regarding cost and quality of established
batch production processes such as machining and injection molding, and to signif-
icantly increase process efficiency (energy use, waste generation, and robustness).
This is particularly the case where there are high demands in terms of surface
quality and component failure on the application side, such as in aerospace and
mechanical engineering, and also in the case of large volumes of personalized
mass-produced products (mass customization, e.g. of glasses or shoes).
Since 2005 development, and since 2009 market activity have been strongly in-
fluenced by two trends: the increasing activities of open source communities (in
particular the RepRap project), and the Fab@Home concept (desktop printing such
as MakerBot). The fascination with additive manufacturing techniques, the desire to
participate in production processes, the opportunity to produce replacement parts on
demand, and the reintegration of consumer product manufacturing into local econo-
mies are all important drivers of this development. Whereas this development prin-
10 Fraunhofer Additive Manufacturing Alliance 147
Fig. 10.1 Graphical roadmap of additive manufacturing to the present (Fraunhofer UM-
SICHT, Fraunhofer Additive Manufacturing Alliance 2012)
The roots of the Fraunhofer Additive Manufacturing Alliance stretch back to the
year 1998 when the Rapid Prototyping Alliance was born. Formally relaunched as
the Fraunhofer Additive Manufacturing Alliance in 2008 with eight member insti-
Additive Manufacturing at Fraunhofer
148 Bernhard Müller
One topic – eighteen institutes – one alliance
Hamburg
Additive Production Technology (IAPT)
Bremen
Manufacturing Technology and Advanced Materials (IFAM)
Berlin
Production Systems and Design Technology (IPK)
Braunschweig
Surface Engineering and Thin Films (IST) Magdeburg
Factory Operation and Automation (IFF)
Oberhausen
Environmental, Safety, and Energy Technology (UMSICHT)
Dresden
Ceramic Technologies and Systems (IKTS)
Aachen Machine Tools and Forming Technology (IWU)
Production Technology (IPT) Material and Beam Technology (IWS)
Laser Technology (ILT)
Darmstadt
Computer Graphics Research (IGD)
Stuttgart
Manufacturing Engineering and Automation (IPA)
Augsburg
Industrial Engineering (IAO) Casting, Composite and Processing Technology (IGCV)
Interfacial Engineering and Biotechnology (IGB)
Freiburg
Mechanics of Materials (IWM)
High-Speed Dynamics, Ernst-Mach-Institute (EMI)
Fig. 10.2 Members of the Fraunhofer Additive Manufacturing Alliance (Fraunhofer Ad-
ditive Manufacturing Alliance 2017)
© Fraunhofer 1
tutes, it now comprises 17 (cf. Fig. 10.2) and thus reflects the dynamic global de-
velopment of this manufacturing technique.
The Fraunhofer Additive Manufacturing Alliance’s comparative global position
is measured against various criteria and was assessed as part of a competitor anal-
ysis carried out by the Fraunhofer ISI, INT, IAO, and IMW institutes on behalf of
Fraunhofer head office [1]. To do this, researchers used first and foremost certain
specific indicators to review the appeal of the research field and Fraunhofer’s rela-
tive position in comparison to other research bodies. The researchers then used fu-
ture-oriented studies and publications to assess the long-term potential for applica-
tions and research and estimated future market potential and research market dy-
namics by analyzing patents and publications. They calculated Fraunhofer’s current
position compared with other research bodies based in part on Fraunhofer’s patent
10 Fraunhofer Additive Manufacturing Alliance 149
and publication activities compared with other research institutes. In addition, they
carried out a brief survey at the Fraunhofer institutes, which are members of the
Additive Manufacturing Alliance or which have been active in publishing in this
field. The breadth of the additive manufacturing technologies researched and mate-
rials produced here can be seen in Tables 10.1 and 10.2.
IWU x x
UMSICHT x x
IGD (x)² (x)² (x)² (x)² (x)² (x)² (x)²
IGB x x x x
EMI x x
IST
IGCV x x x
IAO
Alliance
IWS x x x x x x
IPK x x
ISC x x x
publications
ICT x
of Alliance
(Previous)
IOF
IAP
(x) R&D contributions only
(x)1 Working on the mechanical and tribological characterization of additively manufactured
components, the design of components for additive manufacturing and the simulation
of process steps.
(x)² Development of algorithms and software for controlling 3D printers (no materials de-
velopment, but optimization of materials and component properties through adaptation
of process parameters)
150 Bernhard Müller
Legend:
VP – Vat Photopolymerization: selective light curing of a liquid photopolymer in a vat, e.g.
stereolithography (SLA/SL)
MJ – Material Jetting: drop-by-drop application of liquid material, e.g. multijet modeling,
polyjet modeling
PBF – Powder Bed Fusion: selective melting of regions within a powder bed, e.g. laser
sintering (LS), beam-based melting (LBM, EMB), selective mask sintering
SL – Sheet Lamination: successive layering and bonding of thin sheets of material, e.g.
layer laminated manufacturing (LLM), laminated object manufacturing (LOM) also:
stereolithography (SLA)
ME – Material Extrusion: targeted deposition of material through a nozzle, e.g. fused layer
modeling (FLM), fused deposition modeling (FDM)
BJ – Binder Jetting: selective adhesion of powdery material using a liquid binder, e.g. 3D
printing (3DP)
DED – Directed Energy Deposition: targeted welding of the material during deposition, e.g.
laser powder build-up welding (Laser-Pulver-Auftragschweißen – LPA), direct metal
deposition (DMD), laser cladding
10 Fraunhofer Additive Manufacturing Alliance 151
ISC A D/A
publications
IOF
IAP
Legend
D = Development; A = Application
² Development of algorithms and software for controlling 3D printers (no materials deve-
lopment, but optimization of materials and component properties through adaptation of
process parameters)
152 Bernhard Müller
This integrated view and assessment takes account both of the appeal of the
technological field of additive manufacturing/the individual technological sub-
themes as well as of the positioning of Fraunhofer within this field. The following
criteria were used to identify Fraunhofer’s position (largely defined by the Alli-
ance’s member institutes):
• Publically funded projects (Nationally: ranked first by a large margin according
to number of projects and total amount; EU: ranked second by total amount, first
by number of projects)
• Patent activity (ranked 21st across all additive manufacturing technologies ac-
cording to analysis of patent family registration between 2009 and 2014; ranked
first globally among research institutes; across all technologies among the top
10 research institutes)
• Publication activity (ranked first in Germany and fourth for global scientific
publications in peer-reviewed journals; ranked between first and sixth for con-
ference papers, non-peer-reviewed publications and press releases)
• Networking in the scientific community (close networks with players with insti-
tutional connections, e.g. professorships of heads of institutes; diverse network
with European players in particular, but also with American and select Chinese
players but a lack of clear and well-developed network with many players)
On the basis of these assessments, it can be concluded that Fraunhofer is the world’s
most broadly-positioned research player in the field of additive manufacturing. Its
network clearly concentrates on industry companies.
Alongside Fraunhofer’s unique position of being active in all of the technology
fields, for certain technologies (powder bed fusion, material jetting, and binder
jetting) Fraunhofer is among the leading players [1].
The scientific excellence of the Additive Manufacturing Alliance is reflected,
along with the aforementioned aspects, in the specialist international Fraunhofer
Direct Digital Manufacturing Conference DDMC, organized by the Alliance since
2012, every two years in March, in Berlin. The research findings of the Alliance
institutes presented here, the renowned global keynote and specialist speakers, and
the large number of conference delegates show Fraunhofer’s outstanding worldwide
reputation in the field of additive manufacturing.
The Alliance strives to play a leading global role in applied additive manufactur-
ing research. Its focus, then, is on combining the strengths of the Alliance’s members
and using the various complementary skills to provide an appealing offer of com-
prehensive commissioned research to industrial customers. The Alliance’s research
spectrum here stretches right across the entire field of additive manufacturing, in a
10 Fraunhofer Additive Manufacturing Alliance 153
very comprehensive way, and can essentially be divided into four key areas or re-
search focus areas:
• Engineering (application development)
• Materials (plastics, metals, ceramics)
• Technologies (powder bed, extrusion and printing based)
• Quality (reproducibility, reliability, quality management)
The following selected example projects provide an insight into the diversity of
applied research in additive manufacturing (3D printing) at Fraunhofer.
Prof. Dr. Christoph Leyens · Prof. Dr. Frank Brückner · Dr. Elena López ·
Anne Gärtner
Fraunhofer Institute for Material and Beam Technology IWS
The AGENT-3D joint research project coordinated by the Fraunhofer Institute for
Material and Beam Technology IWS aims to develop solutions to existing scientif-
ic and technical, political and legal, and socioeconomic challenges for additive
manufacturing together with more than 100 project partners, mostly coming from
industry.1
Following the completion of the strategy phase and the construction of a meas-
uring and testing center with modern equipment for, among other things, optical and
x-ray examination and measurement of additively manufactured components (e.g.
using scanners or computer tomography), the consortium is working nowadays on
implementing the strategical roadmap by means of basic projects and over 15 tech-
nology projects (further technology projects will complete the strategical roadmap
until the year 2022). These projectsfocus, for example, on integrating functionalities
into components, combining conventional and additive manufacturing, allowing the
processing of multimaterials and enhancing the material portfolio for additive pro-
cesses, and last but not least, quality management along the whole process chain.
Additive manufacturing allows complex components to be constructed lay-
er-by-layer directly based on digital data (cf. Fig. 10.3). The principle of reverse
engineering, on the other hand, allows a scaled-up reproduction of an original part
Fig. 10.3 Demonstrator from the Fig. 10.4 Reproduction following scan-
AGENT_3D_Basis subproject, manufactured ning of a bird skull in original size and at
using laser beam melting, to illustrate chal- ten times actual size (Fraunhofer IWS)
lenging shapes (Fraunhofer IWS)
Also redesign of the part to fulfill special requirements can be easily done in this
way. Scans are used to provide data from which identical or optimized parts in terms
of i.e. design can be printed (cf. Fig. 10.4). In addition, a function integration can
be included, for instance in terms of electronic functionalities such as conductor
tracks (cf. Fig. 10.5) or sensors that can be directly printed into three-dimensional
components. In this way, digitization enables completely new possibilities for de-
signing and manufacturing products.
Contact:
Dr. Elena López, [email protected], +49 351 83391-3296
www.agent3D.de
10 Fraunhofer Additive Manufacturing Alliance 155
Fig. 10.6 Honeycomb structure adapted to a freeform surface, and honeycomb with load
application elements (Fraunhofer IGCV)
example (cf. Fig. 10.6). By using powder bed-based additive manufacturing pro-
cesses, the honeycomb structures produced in this way were able to be generated in
both plastic as well as metal. Further potential for reducing weight and increasing
rigidity lies in the combination of additively manufactured honeycomb structures
with a covering layer of carbon fiber-reinforced plastic.
Contact:
Dr. Christian Seidel, [email protected], +49 821 90678-127
Together with our project partners Industrial Design, Folkwang University Essen,
Fraunhofer UMSICHT, Sintermask GmbH, rapid.product manufacturing GmbH and
Authentics GmbH, a numeric tool for design, analysis and optimization was devel-
oped at Fraunhofer IWM.2 The tool fills specified external shapes with a cellular
structure based on trabecular cells, similar to cancellous bone. For additive manu-
facturing to be cost-effective, it is important to be able to assess the mechanical
properties of products without having to produce additional exemplars for mechan-
2 Bionic Manufacturing, DLR, part of the BMBF’s Biona program (BMBF-FKZ 01RB0906)
10 Fraunhofer Additive Manufacturing Alliance 157
Fig. 10.7 Cellular Loop, a mechanically developed and manufactured designer cantilever
chair. Photo by Natalie Richter (Folkwang University of the Arts)
ical testing. Due to the regularity of the cellular structure, this approach facilitates
the advance calculation of mechanical properties such as load-bearing capacity and
rigidity. Experimental data from just a few representative samples is the only input
parameter needed to characterize the material and the process for finite element
models. Any additive manufacturing technique and material can be used here.
In order to optimize the component’s mechanical properties, the trabecular cell’s
microstructure can be adapted to a given load. This is done by locally anisotropical-
ly increasing the diameter of the trabecular rods. In this way, the load-bearing ca-
pacity of the component can be significantly raised, with minimum material use and
production time.
The tool presented can be used on a large number of components and allows
mechanical properties to be calculated and improved. Due to its visual properties,
the biomimetic cellular structure also leads to esthetically pleasing products, as
shown by the product described in what follows.
As a demonstrator, a bionic cantilever chair was developed by the group around
Anke Bernotat at the Folkwang University of the Arts. The loads produced by an
individual sitting on it were calculated at Fraunhofer IWM. Next, the microstructure
158 Bernhard Müller
was adapted to this load. The shape was divided into producible segments and the
chair was then manufactured in selective-laser-sintering by our partners at rpm-fac-
tories. The chair provided the expected load-bearing capacity and also corresponds
to the highest esthetic standards.
Contact:
Dr. Tobias Ziegler, [email protected], +49 761 5142-367
Mathias Gebauer
Fraunhofer Institute for Machine Tools and Forming Technology IWU
Manufacturing complex sheet metal parts from high-strength steel places large
demands on cold forming. High pressing forces and the relatively high springback
represent huge challenges. Very rigid tools made of expensive material are required
that are nevertheless subject to increased wear. An alternative to cold forming is
sheet metal hot forming or press hardening. Here, the sheet metal blank is heated
above the austenitizing temperature (above 950 °C) and rapidly cooled to below 200
ºC during forming, creating a martensitic structure.
The structure of a hot forming tool is more complex than that of a conventional
one. The reason for this is the necessary integration of cooling channels into punch
and die. The channels, generally produced by deep drilling, are limited in their
minimally representable diameters, which has a direct impact on the contour dis-
tance achievable. For this reason, targeted tempering of individual regions confor-
mal to the contour of the tool is only achievable with great effort and significant
limitations for hot forming tools. This often causes insufficient target temperature
achievement and too little heat dissipation in the tool’s critical regions.
As part of the additively manufactured HiperFormTool3 project, research has
been carried out into how sheet metal hot forming can be made more efficient by
means of additively manufactured active tool components. To achieve this, the
thermal behavior of the tools and of the forming process was precisely analyzed via
simulation and various cooling channel geometries compared. Based on the results
of the simulation and by using the geometric freedom of additive laser beam melt-
ing, an innovative and contour-close cooling system was developed. The primary
goal of the research studies was to significantly shorten the cycle time. In addition,
a concept for sensor integration during the additive manufacturing process was
developed and implemented.
3 HiperFormTool, high-performance sheet metal hot forming tools using laser beam melting,
ERA-NET joint project, MANUNET HiperFormTool (BMBF-FKZ 02PN2000)
160 Bernhard Müller
Contact:
Mathias Gebauer, [email protected], +49 351 4772-2151
Fig. 10.10 Additively manufactured aluminum oxide components for applications as heat
exchangers or mixers for two fluids (Fraunhofer IKTS)
10 Fraunhofer Additive Manufacturing Alliance 161
before the ceramic particles are sintered, which generally involves a significant
reduction in volume, at temperatures above 1000 ºC. It is only during the sintering
phase that the component gains its final properties.
For manufacturing ceramic components additively, different processes are used
that can generally be divided into powder bed- and suspension-based or indirect
processes (areal application of the material and selective solidification) and direct
processes (selective application of the material).
With suspension-based processes, the base materials are in the form of suspen-
sions, pastes, inks, or semi-finished products such as thermoplastic feedstocks,
green sheets, or filaments. Compared with powder bed-based processes, higher
green densities are achieved with suspension-based additive manufacturing pro-
cesses, which then lead to a dense microstructure in the sintered part and decreased
surface roughness. New, complex ceramic structures illustrate the potential of ad-
ditive manufacturing for ceramic (cf. Fig. 10.10).
One current focus is on the development of processes for manufacturing mul-
ti-material compounds (e.g. ceramic/metal) and components with a gradient of
properties (e.g. porous/dense). The direct processes in particular offer huge potential
here due to the selective application of different materials. In this way, in future it
will be possible to manufacture components with highly complex inner and outer
geometries that will also combine the properties of various materials (e.g. electri-
cally conductive/non-conductive, magnetic/non-magnetic).
Materials, equipment, process, and component development issues are being
worked on together with national and international partners in several BMBF pro-
Contact:
Dr. Hans-Jürgen Richter, [email protected],
+49 351 2553-7557
Fig. 10.12 Bioinks made of modified biomolecules are designed for digital generation of
biological tissue replacements. The modified biomolecules at Fraunhofer IGB can be for-
mulated to form low or high viscous fluids and composed to host different cell types. Left:
viscous ink for bone cells, right: soft ink for sensitive fat cells.
10 Fraunhofer Additive Manufacturing Alliance 163
Contact:
Dr. Achim Weber, [email protected], +49 711 970-4022
Fig. 10.13 Via the formulation of differently modified biomolecules, hydrogels with the
same biopolymer concentration and composition can be produced with different combina-
tions of properties (Fraunhofer IGB).
164 Bernhard Müller
Fig. 10.14 Additive manufacturing of large components using universal industrial ro-
botics and simulation tools to assist development (Fraunhofer IFF)
ponent volumes of 1000 x 1000 x 1000 mm³ and maximum component weights of
25 kg.
A key element of the overall technological concept is the simulation of complex
production processes associated with development. To do this, the VINCENT uni-
versal simulation tool developed by Fraunhofer IFF is used (cf. Fig. 10.14). The
results of the simulation are directly incorporated into the constructive development
of the manufacturing plant. The program facilitates process visualization and reach-
ability testing of all paths for the layer-by-layer component construction as well as
a collision detection in the workspace. In this way, a geometric and functional test
of the plant is possible even before commencing the manufacture of its components
so that a significant reduction in the plant development and commissioning time is
achieved.
The new 3D printing process combining extruders and robots opens up new
possibilities in the production of large, complex, plastic parts. Since expensive
molds or tools are not required, component production is subject to hardly any
spatial limitations. With the significantly shorter process chain, large components
will in future be able to be produced economically and flexibly, leading to a variety
of new or improved products in a large range of market segments.
Contact:
Dr. Uwe Klaeger, [email protected], +49 391 40 90-809
166 Bernhard Müller
Both diagnostic as well as therapeutic functions are expected from theranostic im-
plants. In engineering terms, then, implants should have sensory and actuator com-
ponents integrated into them. The advantage of this kind of strategic approach is
that treatment-relevant information can be gathered where it arises so that biological
treatment effects can be achieved locally precisely there. This strategy was imple-
mented in the Fraunhofer Theranostic Implants lighthouse project in terms of a
form-fit, force-fit bonded embedding of actuators and sensors into a compact addi-
tively manufactured hip implant. By means of this complete integration in the im-
plant, the measurement of forces or stresses can thus take place directly in the region
where they occur within the implant.
For therapeutic functions, the corresponding actuator module, hermetically en-
capsulated in the interior of the implant, can ensure a partial or total excitation of the
implant close to the desired surface area for biomechanical, electrical, or chemical
Fig. 10.15 Hip stem implant with integrated sensor-actuator unit; top right: CT image
(Fraunhofer IWU)
10 Fraunhofer Additive Manufacturing Alliance 167
excitation of the interface between the implant and the tissue. As a result of this
project, it was possible to integrate thermally sensitive functional components into
a titanium hip stem implant produced additively via laser beam melting. To do this,
both the sensor/actuator and inductor for the wireless energy and data transmission
were incorporated into an additively manufactured carrier structure that is welded
with the main body of the implant later in the process. Therefor, the inherent proper-
ties of the additive manufacturing process are used to apply thermal energy into the
material in a spatially and temporally highly limited and highly controlled manner
using the laser beam. Combined with a suitable laser beam process control and a
specially developed additively manufactured ceramic-metal multilayer protective
coating system (ceramic metallic covering – CMC) for the sensors/actuators, it was
possible to ensure that the functionality of the sensors/actuators was retained in spite
of the high melting temperatures of the TiAl6V4 titanium alloy of almost 1700 °C.
The process chain developed for the form-fit, force-fit, firmly bonded integration of
the sensors/actuators can be transferred to additional applications and used for com-
ponent-integrated condition monitoring or actuator functionalization for example.
Contact:
Dr. Holger Lausch, [email protected], +49 341 35536-3401
Thomas Töppel, [email protected], +49 351 4772-2152
Fig.10.16 Multi-material structures with 1.2709 tool steel and copper alloy CuCr1Zr
(Fraunhofer IGCV)
layer here. To manufacture these parts, it is necessary to adapt the powder applica-
tion mechanism in order to facilitate the deposition of a second material in the
powder layer. For this reason, at Fraunhofer IGCV, a new application mechanism
was integrated into an LBM plant via software and hardware so that the construction
of 3D multi-material components is now possible in a laser beam melting system.
An initial application of the modified laser beam melting system focused on the
production of structures made of 1.2709 tool steel and a copper alloy (cf. Fig. 10.16)
within the project ForNextGen supported by the Bavarian Research Foundation.
The project consortium consisting of six academic partners and 26 industrial com-
panies has the goal of laying manufacturing science foundations for the use of ad-
ditive manufacturing processes in mold and tool making, and is being supported by
the Bavarian Research Foundation. The classifying and subsequent introduction of
these processes is intended to lead to significant improvements in the complexity of
shapes, strength, and production time and cost of tools in primary shaping and
forming. The multi-material processing researched by Fraunhofer IGCV thus offers
great potential for tool shapes and uses. Using the example of a sprue bushing for a
die casting mould, a base body of 1.2709 tool steel is constructed and equipped with
CuCr1Zr (copper alloy) in two different component regions for improved heat dis-
sipation. By means of these internal cooling structures made of highly thermal-
ly-conductive material, the heat balance can be improved and the cycle time thus
reduced.
Beyond the work carried out as a part of the ForNextGen project, Fraunhofer
IGCV could already show that the current laser beam melting system can even be
used to produce multi-material components of metal alloy and a technical ceramic
(AlSi12 and Al2O3).
Contact:
Christine Anstätt, [email protected], +49 821 90678-150
10 Fraunhofer Additive Manufacturing Alliance 169
Summary
The Future Work Lab – an innovation laboratory for work, people, and technolo-
gy – provides companies, associations, coworkers and labor unions with exten-
sive opportunities to experience future-oriented work concepts. The laboratory
combines demonstrations of specific Industry 4.0 applications with competency-
development offers and integrates the current state of work research. In this way,
it facilitates holistic developmental progress in the field of work, people, and
technology. Taken as a whole, the Future Work Lab provides a significant contri-
bution to long-term increases in companies’ competitiveness through participa-
tive design of sustainable working environments..
171
172 Wilhelm Bauer • Moritz Hämmerle • Thomas Bauernhansl • Thilo Zimmermann
Project overview
Partner organizations
Fraunhofer Institute for Industrial Engineering IAO
Fraunhofer Institute for Manufacturing Engineering and Automation IPA
Institute of Human Factors and Technology Management (IAT) at the University of
Stuttgart Institute of Industrial Manufacturing and Management (IFF) at the University of
Stuttgart
Contact
Dr. Moritz Hämmerle
Fraunhofer IAO
Tel. +49 711 970 -2284
[email protected]
www.futureworklab.de
Our society – and with it, many others in the world – is faced with new and extensive
challenges as a result of demographic changes, shortages in specialist staff, and the
onward march of digitization. The Internet and digital technologies, first and fore-
most the mobile use of data and artificial intelligence, are not only reshaping our
everyday lives, they are also leading to far-reaching changes in the economy and
the workplace.
Following the invention of the steam engine, industrialization, and the beginning
of the age of computers, we are now in the midst of the fourth industrial revolution
with the Internet of Things and Services. The ongoing development of information
and communications technology (ICT) has ensured that in large sections of industry,
powerful and competitively priced embedded systems, sensors and actuators are
now available. Industry 4.0 is the current buzzword used to describe developments
towards a production environment that consists of intelligent and autonomous ob-
jects temporarily and deliberately networking with one another to carry out tasks.
Cyber-physical systems (CPS) and cyber-physical production systems (CPPS) are
additional features spoken about in this context [11][15]. CPSs are systems that link
11 Future Work Lab 173
Fig. 11.1 The work of the future – between people, technology and business (Fraunhofer
IAO/Shutterstock)
the real and virtual worlds within an Internet of things, data, and services. A broad
field of applications is beginning to emerge in the areas of automation, production
and robotics, for example, as well as in healthcare and energy distribution [1][9][5].
Successful development and integration of digital technologies within processes
in the industrial application sectors is key for Germany’s competitiveness [7]. This
entails identifying successful responses to new challenges: how can we harness the
opportunities for industry, public administration and society afforded by digitiza-
tion, and how can we overcome the challenges together? How do we want to live,
learn, and work in a digital world? How can the possibilities provided by new
technologies be reconciled with the demands of demographic change and work-life
balance? How can the competitiveness of companies and at the same time the qual-
ity of work be positively influenced and continue to increase?
The increasing use of digital technologies in production and related fields of work
is giving rise to new forms of socio-technical working systems, leading to massive
changes in the organization and structure of work. Volatile markets as control mech-
174 Wilhelm Bauer • Moritz Hämmerle • Thomas Bauernhansl • Thilo Zimmermann
anisms of our work are leading to an increasing degree of flexibility in terms of time
and space. The mobility requirements on employees grow ever greater; new forms
of employment are increasingly making inroads alongside regular working relation-
ships. Training in digitization, IT, and intelligent technical systems is ever more
becoming the “entry ticket” an employee requires for numerous roles [3][15]. In
future, human-technology interaction, workplace flexibility, and the necessary com-
petency and training requirements will need to be actively taken into account when
designing work.
In spite of all the changes brought about by digitization and continuing automation,
future work systems in the office and factory will still nevertheless remain so-
cio-technical systems. In this flexible and connected environment, staff members
will take on different roles. Their cognitive abilities will enable them to close the
sensory gaps of technology, grasping complex situations quickly and comprehen-
sively. As decision-makers, they will resolve the conflicts between networked ob-
jects and use digital tools to intervene in time-critical processes. As actors in their
own right, employees will complete irregular and highly complex tasks. In their role
as innovators and process optimizers, they will continue to be actively involved in
the further development of industrial value creation in the future [15]. Equipped
11 Future Work Lab 175
with mobile devices, staff will thus work on different tasks with a high degree of
flexibility in terms of time, space, and content matter. The scope of future jobs will
shift between executive and supervisory or management activities. Mobile forms
and content of work will supplement the existing structure of work. In the manufac-
turing sector, too, the degree of flexibility will increase such that even in production,
flexible working locations and flextime will in future become relevant topics.
Alongside massive increases in market-side requirements for flexible staffing,
from the employee side new demands for flexible working have also grown in recent
years. Here the requirement for temporary and self-determined changes to working
hours because of concerns for a healthy work-life balance, for empowerment, and
the trend towards self-management are especially at the fore[13][4][12].
The enhanced features of CPS provide the opportunity to redesign the industrial
work of the future. When it comes to workplace design, the key elements which
build upon one other on the user side are networking, context sensitivity, assistance,
and intuitivity. Intelligent workplaces and work systems will be enhanced by digi-
tized work organization. In what follows, we will go into more detail in the work-
place level only.
The goal of connected workplaces is to connect systems and data from products and
processes thoroughly instead of using existing IT silo systems. To do this, workplac-
es need to be equipped with sensors that capture the accruing data. The transmission,
Fig. 11.2 Future Work Trends change work design. (Fraunhofer IAO/Fotolia)
11 Future Work Lab 177
processing and sharing of data between objects in production must also be facilitat-
ed. Connected data sharing here should take place both horizontally along the pro-
cess chain as well as vertically within the company, and bi-directionally – between
systems and users, and between users and systems. It is only in this way that the
information gained from the data can be made available to systems and staff and
utilized by them.
In an Industry 4.0 environment, connected objects and real-time utilization of
relevant production parameters find application, for example, when production
events are fed back into the production control system (e.g. for incident manage-
ment). Within the field of connectedness, the design tasks are developing new
Production 4.0 concepts (beyond lean) and supporting staff acceptance of work in
transparently generating work systems. Furthermore, staff must be qualified for
work within highly connected systems and must be able to interpret accruing data
in order to optimize their work processes.
Traditional production structures only take limited account of the diversity of em-
ployees and operations, as well as their requirements in the work process. The high
number of variants in production makes this necessary however. Context sensitive
work systems allow this challenge to be met and form the basis for the vision of
batch size 1 production. For the production context to be systematically integrated,
the system, or the information to be provided, must be adapted to specific changing
environmental conditions. The system continuously monitors the work situation and
also knows its user. It identifies the current process sequence by comparing it with
(de)centrally stored data and provides the user with data related to the current pro-
cess step in personalized form.
In an Industry 4.0 environment, the production context can be utilized for work-
space personalization, for example by adapting working heights, lighting, and pro-
viding operation-specific information to the employee. Defining rules for personal
data use and identifying the usable personalization space within which work can be
designed to suit the employee or process are current design focus areas.
Assisting work systems represent the next stage of development. These systems
primarily provide support in mastering the extensive diversity arising from the in-
178 Wilhelm Bauer • Moritz Hämmerle • Thomas Bauernhansl • Thilo Zimmermann
Today, a broad spectrum of applications for assisted work systems is already con-
ceivable. A distinction needs to be made here between digital assistance systems
(e.g. augmented reality/virtual reality) and physical ones (e.g. lightweight robots,
exoskeletons). These systems are able to use feedback to adaptively support em-
ployees’ learning during the process. Planning the distribution of control between
the person and the technology here is the key determining factor of their successful
introduction.
More and more production tasks are being supported by IT systems or carried out
by them entirely. For staff, the result is that these process accompanying systems
are becoming increasingly complex to use. Intuitive digital work system design and
utilization thus represents an important lever for increasing efficiency. Ergonomic
design of the human-technology interaction both physically as well as with respect
to information ergonomics is a key requirement here. Control systems for the work-
er also need to be reduced to process-relevant parameters and override values. In
this way, intuitive interaction concepts based on gestures, speech, touch and in fu-
ture even brain function, combined with mobile devices and wearables, will be able
to provide productive support to the work process. For these enablers to be proper-
ly utilized, the necessary competencies for managing new tools and interaction
concepts need to be identified and formulated – combined with new mobile device
use concepts such as “bring your own device” (BYOD).
11 Future Work Lab 179
The variety of opportunities, enablers, and new demands for successful work design
in industry must be made adaptable in a practical environment and tangible for all
stakeholders. In the Future Work Lab, a new innovation laboratory for work, people,
and technology is being developed in Stuttgart, dedicated to these issues. The Future
Work Lab functions as an interactive shop window and ideas center for sustainable
and people-centered work design in production and related fields. In the Future
Work Lab, the design of future industrial work in Germany is being discussed,
participatively advanced, and made tangible in close coordination with the relevant
parties. The foundation for this is already being laid via pilot and soon-to-be-imple-
mented digitization and automation solutions.
In order to achieve this, the Future Work Lab is divided into three areas [10]:
• A central demo world with Workplace 4.0 Tracks that introduce visitors to the
workplace of the future via hands-on exhibits of different forms of digitization
and automation,
Fig. 11.3 Structure of the Future Work Lab (Fraunhofer IAO/Ludmilla Parsyak)
180 Wilhelm Bauer • Moritz Hämmerle • Thomas Bauernhansl • Thilo Zimmermann
• The Fit for Future Work learning world providing information, qualification
and discussion opportunities on the developmental trends of future workplac-
es
• The Work in Progress world of ideas, a think tank for research into work and a
safe space for the design and development of new and unforeseen or hard-to-
foresee solutions
The close relationship between these three areas ensures that the latest solutions for
the work of the future can constantly be discussed, developed and demonstrated in
an open and adaptable environment on the basis of current solutions in the context
of technological transfer.
The Workplace 4.0 Tracks – demonstrators with content related to one another in
the demo world – are designed to tangibly show short- and medium-term changes
in the workplace as well as long-term developmental trends. For this, 50 different
demonstrators are being put together in the Future Work Lab. The layout is oriented
around today’s typical work profiles across the operational value chain.
The Today+ track shows operational use cases demonstrating industrial work
over the period from 2016 to 2018. The Future Work Lab thus illustrates develop-
ments within the industrialized and modern medium-sized enterprise – lean produc-
tion, lean systems, and integrated production systems.
The long-term developmental trends demonstrate operational use cases for the
digitization and intelligent automation of industrial work of the time horizon up to
2025. They illustrate diverse demonstrators at the crossroads between technolo-
gy-centered automation and human-centered specialization that may be standard in
the manufacturing industry in 2025. In this way, both developmental scenarios of the
Industry 4.0 currently discussed are referred to: the automation and specialization
scenarios [16]. In order to illustrate these, potential Industry 4.0 operational use cas-
es are combined: it is possible to demonstrate and experience how work in the future
can on the one hand be designed to be more technology-oriented and on the other hand
more strongly people-focused. In the Future Work Lab, different developmental
trends and their consequences are highlighted in this way as they relate to work de-
sign, technology integration, competency or qualification requirements, and so on.
The available demonstrators serve as the basis for the Future Work Lab. They
serve the other elements of the laboratory – the learning world and world of ideas
– as an interactive work, learning and research environment.
11 Future Work Lab 181
The formats developed in this part of the laboratory enable different target groups
(e.g. management, planners, team leaders, works committee members, and employ-
ees) to be informed and advised regarding design options and to develop them
participatively.
The Work in Progress world of ideas reinforces the academic character of the Future
Work Lab, establishing a think tank as a safe space for research, innovation and
dialog around the future of work and the human-technology interaction it entails.
The world of ideas thus stands for sharing about and developing new solutions for
the workplace of the future based on cyber-physical systems.
182 Wilhelm Bauer • Moritz Hämmerle • Thomas Bauernhansl • Thilo Zimmermann
When complete, the Future Work Lab will offer more than 50 different demonstra-
tors: they will show what the changes in work prompted by Industry 4.0 might look
like across the value chain. The demonstrators are being produced for operational
fields of work such as machine operation, assembly, factory logistics with receiving
and shipping departments, quality assurance, scheduling, maintenance, and indus-
trial engineering. The first demonstrators implemented can be seen in the illustra-
tions. In what follows, we will take a closer look at two Future Work Cases by way
of example.
Fig. 11.4 Demonstrators in the Future Work Lab (excerpt) (Fraunhofer IAO/Fotolia)
11 Future Work Lab 183
Robotic systems have been taking on a central role in industrial work for several
decades now – a role that will further expand significantly in future. New robotic
systems – in particular lightweight robots as well as improved safety technologies
and correspondingly adapted operational processes – ensure that robots are more
and more becoming the physical assistants of humans in a range of different pro-
cesses.
One of the demonstrators in the Future Work Lab shows what this kind of work-
place for human-robot collaboration (HRC) might look like. The unique feature here
is that it shows that HRC is not only achievable with lightweight robots. While these
small, compact systems with partially integrated force/torque sensors are inherent-
ly safer than heavy-duty robots, their load-bearing capacities are equally limited by
virtue of their construction.
In the aforementioned workplace, the workspace is open, and the worker can
directly monitor and coordinate the progress of the work. In this way, the specialist
knowledge and dexterity of the individual can be combined with the strength and
endurance of the robot. The results are workplaces with improved ergonomics and
11 Future Work Lab 185
higher productivity and quality. This is made possible by the workplace safety-cer-
tified SafetyEye camera system that watches over the robot’s working space from
above. The system recognizes when people are approaching the robot’s working
space. The robot then either reduces its speed or stops altogether in order to guar-
antee the individual’s safety. The robot can also be switched to a manual operation
mode.
Human-robot collaboration thus not only offers benefits by virtue of improved
ergonomics due to the robot taking over physically demanding tasks. Specific qual-
ity-critical processes can also be safely carried out by robots. In addition, the scal-
ability and personalization of production also become increasingly important in the
Industry 4.0 context. Since industrial robots can be employed universally, they
generally provide good conditions for implementing versatile production.
This is all the more the case if the “rigid monuments” of worker safety railings
that have thus far been so common in factories are in future done away with.
11.6 Outlook
Following the construction of the Future Work Lab, the innovation laboratory will
enter its launch and operational phase, to be designed in cooperation with its users,
for example its social partners. In addition, the more than 50 demonstrators will
continue to be actively developed and incorporated into new research assignments.
The training and advice formats will receive new impetus based on constant use,
and research on the demo world will permit new scientific findings, in the context
of user experience for example.
186 Wilhelm Bauer • Moritz Hämmerle • Thomas Bauernhansl • Thilo Zimmermann
Summary
Digitization is the defining innovation driver for value creation in the modern
global industrial society. At the forefront stand the increase in efficiency for
flexibilization and improved resource utilization provided by the self-optimizing
automation of processes. Digital technologies must become inherent components
of the production system.
A cyber-physical system represents the sought-after unity of reality and its digital
reproduction, and is the next stage in development of mechatronics into a symbi-
otic systems approach based on the IT networking of all components. IT together
with non-technical disciplines have produced a range of methods, techniques and
processes by which sensors, actuators, and cognition can be integrated into tech-
nical systems so they demonstrate functionalities that have only been fulfilled
by biological systems until now. In this way, the evolution prompted by Industry
4.0 technologies leads to a genuinely disruptive paradigm change. Production,
suppliers, and product developers enter a new quality of innovative cooperation.
189
190 Welf-Guntram Drossel • Steffen Ihlenfeldt • Tino Langer • Roman Dumitrescu
12.1 Introduction
Digitization is the defining driver of innovation for value creation in the modern
global industrial society. This is the reason for a range of activities that are often
grouped together today under the heading of Industry 4.0 or IoT – Internet of Things.
All of these approaches have in common that they prioritize efficiency gains re-
quired for enhancing flexibility and improving utilization of resources in produc-
tion, via the self-optimizing automation of processes.
Nevertheless, connecting production with the latest IT and communications
technology via Internet technologies – often described as the fourth industrial rev-
olution –poses enormous challenges for numerous companies. In addition to the
technical controllability of much more flexible production and supplier networks,
it also entails a far-reaching economic revolution. The enhanced technological pos-
sibilities also come with a change in traditional customer-supplier relationships as
well as global market access. The global flexibilization within supplier networks
means a disruption of the usual sharing of risk in traditional supplier chains. For the
supplier industry in particular, which is characterized by small and mid-size enter-
prises, this in turn holds huge economic risks.
New opportunities, and in particular new potentials for employment, will only
arise as results if digitization becomes a business model not only for developers and
suppliers of software. The traditional sectors of German industry, such as machinery
and plant engineering, and automobile manufacturing, have to obtain the capability
of using digital technologies to produce new products and services, which also
implies new business models. Digital technologies have to become an intrinsic
component of the production system and production facility.
A cyber-physical system (CPS, see Fig. 12.1) represents the sought-after unity
of reality and its digital reproduction – a further stage in the development of syner-
getic approaches to mechatronics (combining the best of all disciplines) towards a
symbiotic systems approach based on the IT networking of all components.
By now CPSs have become a key trend in product development. Initial applica-
tions include, for example, intelligent electricity meters in smart homes or self-ori-
enting logistics systems in smart factories. Autonomous vehicles are another exam-
ple of a future system that will be based on these same principles.
The information technology and non-technical disciplines such as cognitive
science or neurobiology have given rise to a range of methods, techniques, and
processes by which sensors, actuators, and cognition can be integrated into techni-
cal components. These components then demonstrate functionalities that have only
been fulfilled by biological systems thus far. As a result, CPSs are significantly more
than connected mechatronic structures. They provide the basis for fascinating per-
spectives on technological systems [1]:
Making these kinds of scenario projections for manufacturing is unique since man-
ufacturing is not a technically homogeneous system. Alongside an enormous tech-
nological diversity, there are also large variations in how technically advanced
production facilities are, together with a great range of organizational forms. One
key reason for this, alongside economic boundary conditions such as business size
or position in the value chain, lies in the extreme variation of innovation and invest-
ment cycles. The lifecycle of production equipment, for example, may range from
several years to several decades, while the innovation cycle in the software industry
often equates to just a few weeks.
For this reason, it is not only the design of cyber-physical systems that is of
paramount importance, but also the development of methods for transforming the
structure of production systems into the CPS system architecture. The migration of
existing production plants is only possible through the implementation of intelli-
gent, connected subsystems. Their collaboration can only be guaranteed, if com-
munication between all of the subsystems can be ensured. However, the basic
12 Cyber-Physical Systems 193
structure of production systems will change very little in the near future. Compo-
nents such as sensors, drive units, or frames will largely retain their respective core
functionality, but will require software components with functional and structural
models as well as communications hardware and software in order to upgrade them
into CPSs. As shown in Fig. 12.2, three domains can be distinguished at the basic
level:
• CPSs for mechanical engineering and electrical engineering include, for exam-
ple, intelligent machine frames with integrated sensors (e.g. to measure force) in
order to send measured data to a superior monitoring level. An additional exam-
ple could be integrated sensor nodes that measure temperatures and accelera-
tions, combine them on single-board computers, and pre-process them.
• CPSs for mechatronics include, for example, integrated active attenuators with
built-in sensors, actuators, computing unit and the ability to communicate.
• The third aspect at the component level is CPSs for IT/data, with the Internet of
Things (IoT) being the main enabler for the remaining domains. These store
recorded data and calculate simulation models and digital reproductions of sys-
tems and system components, either using master computers near the machines
or applying cloud systems.
194 Welf-Guntram Drossel • Steffen Ihlenfeldt • Tino Langer • Roman Dumitrescu
In keeping with the CPS definition [8], the components of a system are connected
via the hierarchy levels (e.g. the automation pyramid). Thus, the communications
level in Fig. 12.2 gains a particular significance. Once again, two classes of com-
munications are distinguished here: communications between CPSs (according to
specific functionality) and between CPSs and people, e.g. via control panels, smart
glasses, cellphones, or tablets.
Further, these production systems also permit enhanced value creation since it is
no longer merely the manufactured products that represent value for the customer,
but increasingly also the data recorded. This is a key benefit of intelligent connect-
ed systems, for the automotive and aerospace industries in particular, due to their
greater requirements for documentation.
Cyber-physical systems from various domains can be combined into an overar-
ching system (“system of systems”). For production systems use cases, they become
cyber-physical production systems (CPPS). These offer potential for
• New business models and ecosystems such as leasing models and tailored pro-
vision of certain functions for processing units,
• A changing orientation for companies towards information- and data-driven
service products, e.g. by integrating service planning and predictive maintenance
for their own products as well as
• Customized production and process control based on additional information
gained from the process, and the knowledge thus generated.
Fig. 12.3 Production system with extensions necessary for being turned into a CPPS
(with elements from IWM, TU Dresden)
196 Welf-Guntram Drossel • Steffen Ihlenfeldt • Tino Langer • Roman Dumitrescu
1. Data is acquired from the machine control (SPS, CNC, motion control) and from
the drive systems of all of the actuators,
2. Additional sensors are attached to the production system and to the utilized tools;
their signals are also recorded,
3. The properties of the semi-finished product are identified – here, application-spe-
cific sensor units often have to be developed and installed.
Information from the drive systems can be acquired via software extensions of the
machine control system (in this specific case, a CNC) and via the fieldbus [12]. The
data transmission from the control to a management system may take place by
means of the OPC unified architecture (OPC UA) interface, for example, something
which current observations suggest is crystallizing into an unofficial standard for
Industry 4.0 [13].
Once the data is available, it can be processed and analyzed using a process data
management tool. The goal is the representation of the process in the form of a
model (mathematical model, black box model, or system simulation) which can be
used for carrying out extensive analyses. This requires a decentralized knowledge
base for material and component properties together with the interactions in ques-
tion, available in cyberspace. It is then used to identify the optimum process param-
eters that ensure a stable and reproducible process according to current input crite-
ria and boundary conditions. An important aspect in developing the knowledge base
is a procedure for process description and experimental design that is as efficient as
possible. Fig. 12.4 illustrates the individual steps from process description through
to its reconfiguration.
For manufacturing the spring dome, a graphical process description (input,
processing, output, cf. Fig. 12.5) was found to be highly suitable. This encom-
passes
The steps of pre-shaping, sensor placement, consolidation, and component qual-
ity assessment. The interrelationships and interactions between the following as-
pects are registered and mathematically modeled:
• The properties of the initial materials,
Based on the graphical process description and the registered interrelationships, the
mathematical model is able to identify situational optimal process parameters for
the production system. Thus it ensures that production systems have the desired
greater adaptability. This implies that it represents a core functionality of cy-
ber-physical production systems.
Based on the registered graphical process model, the necessary experiments for
identifying the required characteristic values are calculated via statistical design of
experiments.
In this specific case, a D-optimal design was selected since this permits contin-
ually changing variables (e.g. temperature) and discrete states (true/false) to be
equally incorporated into the planning process. The main influencing parameters
from Fig. 12.5 were systematically varied and corresponding experiments carried
out. The respective results were logged in terms of component quality and recorded
in the mathematical model.
Based on the model, it was possible to test various process settings and to assess
their effects on component quality. In doing so it is possible to map destabilizing
settings of the process. Beginning with a desired level of quality for the parts pro-
duced, it is also possible to deduce the required parameterization of the process/
machine. This represents a basis for auto-tracking of the control parameterization.
A cloud solution is best for calculating in a parallelized manner variant analyses and
model, which must supply a new set of parameters to the control after each cycle,
and must be calculable within the control cycle. Furthermore, it limits the repre-
sentable model complexity and requires fast communication, e.g. via a fieldbus.
More limited variations in component quality and higher time constants in the
process can be managed via “part-to-part regulation”. This does not require re-
al-time model calculation and communications, but nevertheless poses challenges
e.g. for the measurement of initial material parameters. The least demanding reg-
ulation strategy allows for a parameter adaptation only after changing the semi-fin-
ished product (“batch-based regulation”). Although there is a range of processes
with stable behavior throughout a batch of the initial material, they still require
re-parameterization after changing the semi-finished product. Due to longer calcu-
lation periods possible even very complex process models can be calculated in this
case, and alternative considerations carried out in order to achieve suitable param-
eterization.
Enhanced data acquisition, networking with data transport, and intelligent (sub-)
components are all core functions of cyber-physical production systems. Develop-
ments in cyber-physical systems can be observed in the various domains. The ex-
ample selected demonstrated that CPPS can be equipped to react autonomously to
changes in the process and ensure production of OK parts by using data acquisition,
modeling, and feedback to the machine control.
Cyber-physical production systems will only reach their full potential when data
and information are used beyond the flexibilization, control, and quality assurance
of the internal technological process. The increasing networking of machines and
logistics systems is driving the volume of “data as raw materials”. The availability
of ever greater data volumes offers huge potential for the targeted analysis of the
information contained.
Many companies strive for constant availability of all relevant data and informa-
tion on key processes and procedures. The goal is to provide details on the current
state of production quickly and easily, with a reliable outlook on the near future by
means of suitable forecasting approaches where necessary. In order to meet the
existing requirements, suitable systems for information and communications are
required for recording and providing the relevant data, and in particular the infor-
mation inferred from it. The heterogeneity of the data sources here requires an ap-
propriate, flexible IT infrastructure in order to later generate knowledge and to
200 Welf-Guntram Drossel • Steffen Ihlenfeldt • Tino Langer • Roman Dumitrescu
#$+ (& '% %&$%&('1
#$+ (& %&"(&'/%(!1
(&#'%&#(/!,!/&%!!1
&( 3$2&''$)&%(*1
$"")#(/#(&(/1
$%("!&''$)& $#')"%($#1
$%&( )($#$"$)'1
#()(*/&!("/$#(,(3'
'")!($#/%&($#3' *!($#1
'$)&'0&)#$&
support decision-making using data and information, using semantics and cognitive
algorithms.
The overall system must integrate both hardware and software solutions into
production. These solutions have to be flexibly established and decentralized. With
each new system, the complexity increases, as does the risk of diminishing technical
and indeed organizational availability. For this reason, the focus lies on ensuring a
high level of resilience at all levels, from the sensors through communications and
IT infrastructure to the selected algorithms in order to achieve the synergistic pro-
tection of the overall production system. .
At present, the data captured is often only analyzed and processed in line with
its original reason for capture. By correlating data that has previously been managed
in individual systems, the use of suitable methods of analysis facilitates the infer-
ence of new information [14] (cf. Fig. 12.6). Here, the mass of data should be linked
and consolidated in such a way that humans can make the correct decisions in
production. This in turn is a precondition for agility and productivity [15].
Fully establishing this kind of approach requires new, modular component solu-
tions for heterogeneous production systems to be made available. In order to achieve
this, the following research questions need to be answered, among others:
• How can the heterogeneity of production and information technology be man-
aged and minimized economically?
12 Cyber-Physical Systems 201
Fig. 12.7 LinkedFactory – Fraunhofer IWU holistic research approach for the factory of
the future (Fraunhofer IWU)
The realization of a robust overall system was achieved by means of a holistic ap-
proach where the modules supplement one another synergistically in the Linked-
Factory (cf. Fig. 12.7), and can be combined with one another as tasks require. The
goal is the real-time synchronization of material and information flow to allow for
agile planning and efficient forecasting.
/"#$"%$%"
Fig. 12.8 Modular digitization toolbox – key questions and solution modules
(Fraunhofer IWU)
Selected sustainable solutions for the modular elements highlighted for produc-
tion were developed within the SmARPro SmARt Assistance for Humans in Pro-
duction Systems [17] research project sponsored by the BMBF.
In addition to flexible solutions for linking machines (Brown and Greenfield) as
data sources (smart systems), the research focused on the SmARPro platform as a
central data hub (entity of the LinkedFactory), SmARPro wearables as innovative
solutions for visualization, and solutions for location identification.
Using CPS components, machines can be specifically incorporated into the ex-
traction of data and information – sensors and communication elements allow the
registration of data directly at the machine and in the process, enabling the transfer
of the collected data to the LinkedFactory as a data hub. As shown in the transfor-
mation strategy depicted in Figure 12.3.1, particular attention is paid to being able
to provide these to existing as well as new machines and plants as low-cost compo-
nents. The goal is to facilitate pre-processing near the point of data origin. The
solutions utilized in the context of standardized communication and data provision
include both web technologies and solutions from the environment of ma-
chine-to-machine communications.
The focus of the overall LinkedFactory concept is the information hub as the
central platform for data and services. It forms an integral element on the way to
implementing innovative solutions to support flexible production structures. It in-
tegrates and links data beyond domains, for example with regard to
• Structure and design of existing machines/production systems and their relation-
ships,
• Targets for controlling operation of the factory (PPS, ERP),
• Indicators and sensor information from current processes, and the processing
status and outcomes of finished products (MES),
• Resource consumption of production components and of production and build-
ings infrastructure (control systems).
The aim is to link data managed specifically by domain and link it with other simi-
lar data in accordance with given specifications in order to infer new information or
requested knowledge. This data is made available contextually via specified inter-
faces, depending on the role of the requester in question. By using the data thus
linked and information generated on the basis of it, various services can be provid-
ed and combined. At Fraunhofer IWU, semantic web principles are being used for
implementing these approaches in software [18]. One important property of this
process is the formal representation of information using defined vocabularies,
making it comprehensible to computers.
204 Welf-Guntram Drossel • Steffen Ihlenfeldt • Tino Langer • Roman Dumitrescu
Fig. 12.9 Data as the basis for deducing information (Fraunhofer IWU)
Currently, the following support solutions, among others, are being implement-
ed:
• Manufacturing a control that uses linked information from data sources within
and outside of the company,
• Mobile contextual assistance systems for increasing product quality,
• Solutions for monitoring, controlling, and visualizing production processes,
• Mobile solutions for providing support in terms of servicing and maintenance.
A key aspect is that the data that is now becoming increasingly available may con-
tain interrelationships that that have previously been unknown – they especially
represent a large part of a company’s technical manufacturing knowledge. These
“hidden interrelationships” and the manufacturing-related knowledge, together with
employees’ experiential knowledge, represent an important basis for decision-mak-
ing and planning processes (cf. Fig 12.9).
Further developments in information and communications technologies and new
methods of data analysis, for example in the context of data mining or machine
learning, may help to unearth “treasures” in the data stock generated within produc-
tion-related IT systems during plant operation. Then the potential can be exploited
for production-related savings or improvements [19].
The data flow entering the LinkedFactory need to be processed in a number of
different ways in order to infer relevant information for creating added-value. The
use of linked data technologies proves to be of great benefit since these technologies
allow the connection of data flows from a range of resources – which are themselves
occasionally subject to change over time. The basis for data processing here is
12 Cyber-Physical Systems 205
Fig. 12.10 Role-specific and person-specific visualizations (plant overview, job details,
notifications, directions) [20] (TYP4 Photography + Design, Phillip Hiersemann,
www.typ4.net)
formed by a complex event processing engine (CEP engine). Rules are to be set by
staff who understands the processes, bearing in mind the operational requirements
and the available data flow. A key requirement in this context is that rules can easi-
ly be changed, which makes it possible to respond flexibly to process changes.
Using mobile terminals, information is provided to staff members contextually
and depending on their current position, for example in the form of an augmented
reality image.. The goal is to provide information that is directly relevant to the
object in question. Work guidelines and information relevant for production can be
received by employees without their workflow being interrupted. This changes the
way information is displayed fundamentally. Information appears precisely when
and where the individual needs it, without having to actively request it. The research
here is focused on a broad range of different devices from tablet computers of var-
ious sizes, smartphones, up to smart watches and data goggles. Fig. 12.10 shows
example interfaces for the contextual provision of information for assembly facili-
ties. Here staff is provided with precisely the information required to better carry
out their individual tasks for creating value,
206 Welf-Guntram Drossel • Steffen Ihlenfeldt • Tino Langer • Roman Dumitrescu
Due to the aforementioned heterogeneity of the overall system (caused by the ar-
chitecture of a CPS system extending across domains and instances), the develop-
ment of cyber-physical systems cannot be carried out from the perspective of a
single specialist discipline. It requires a perspective that sets the multidisciplinary
system at the center. One such comprehensive perspective, which extends beyond
individual specialist disciplines, is provided by systems engineering.
Scarcely any explicit research has yet been carried out into how these kinds of
intelligent systems are successfully developed, balancing requirements of time,
cost, and quality. At that, the increasing intelligence within and networking between
systems in particular as well as the accompanying multidisciplinary nature of ap-
proaches required pose new challenges for product development. Cyber-physical
systems thus do not necessarily have firm system boundaries. Their functionality
changes over the course of the product lifecycle; it often depends on application
scenarios which occur during product usage on an ad hoc basis, so developers can
only anticipate and take responsibility for it in limited measure. The often cited
autonomous driving provides an example here where both individual vehicles as
well as entire convoys can operate as autonomous systems with differing function-
alities. Strictly speaking, however, only the convoy is a cyber-physical system for
which however, as a rule, no one is directly responsible. Critical analysis is clearly
lacking here as to whether companies in Germany will not only be able to invent
and produce intelligent systems, but also successfully develop them in future in the
face of global competition [21].
12 Cyber-Physical Systems 207
Systems engineering (SE) seems to be the right approach to overcoming the chal-
lenges described. Systems engineering is understood as the general interdisciplinary
school of technical systems development which takes account of all the different
aspects. It places the multidisciplinary system at the center and encompasses all of
the different development activities. SE thus claims to orchestrate the actors in the
development of complex systems. It addresses the product (and associated services,
where applicable), production system (and value-added networks where applica-
ble), business model, project management, and the operational structure. Systems
engineering is thus extremely multi-faceted [22].
One particular focus of SE is the general and interdisciplinary description of the
system to be developed, resulting in a system model. This includes an external
presentation (such as diagrams) and an internal computer representation in the form
of a digital model (the so-called repository). While data only appears once in a re-
pository, it can be used multiple times and with various interpretations in external
representations in order to generate specific views of the system. Model-based
systems engineering (MBSE) places a cross-disciplinary system model at the heart
of development. In the process, it does not exclude the existence of other models of
the system, in particular those that are specific to one discipline, but it incorporates
these via appropriate interfaces. Various languages (e.g. SysML), methods (e.g.
CONSENS, SysMod), and IT tools (e.g. Enterprise Architect) are available for
producing the system model, and they can even be variously combined with one
another. So far there has not been any unified and recognized methodology in terms
of an established school of MBSE [23].
This form of digital specification is simultaneously the basis for the ongoing
seamless virtualization of the project activities when the project management is also
founded on the information that has been generated correspondingly. The ongoing
development of peripheral devices such as AR glasses permits new forms of work-
place design for developers and an associated change in working methods in favor
of efficiency and job satisfaction.
somehow integrated into one single standard (which in any case would be highly
unrealistic). Instead, a new kind of exchange format will be required similar to the
STEP format in the field of CAD, to be specified by an industry-led consortium.
Digitization approaches for production and logistics within existing factory struc-
tures are currently only effective in isolated cases, specifically with respect to in-
creases in efficiency, quality, and flexibility. In order to overcome this deficit, robust
overall systems need to be realized where material and information flow synchro-
nization facilitates production system structures with greater agility and flexibility,
while retaining existing levels of productivity. In this context, possible starting
points include achieving greater agility by means of event-based production control,
the involvement of employees in specific real-time decision-making, and the result-
ing increase in flexibility of production processes. This can only be achieved if
modular, task-oriented, combinable solution modules are available for implement-
ing suitable support systems in production and logistics such that complexity can
be mastered by means of a step-by-step process.
Nevertheless, for all the euphoria surrounding digitization, it is clear that in fu-
ture people will still remain the guarantors of success and the central component of
industrial value creation. This requires, among other things, improving knowledge
of how to implement Industry 4.0 concepts in SMEs, developing suitable qualifica-
tion offerings for (further) training, and planning sustainable factory structures,
organizational forms, and processes to make the most of the opportunities provided
by digitization.
All of the elements in the value chain will in future be changed by digitization.
In cyber-physical systems, a universal structural approach is available that in future
will also become established in smart, connected products. The structural equiva-
lence of products and production systems will open up completely new opportuni-
ties in the design of value creation processes. Cyber-physical systems are the core
element, characterized by the networking of smart production systems and smart
products across the entire lifecycle.
This symbiosis allows for a range of new technical options such as using a prod-
uct’s sensory capabilities
• During its manufacture for process monitoring and control,
• During its use cycle for property adaptation and for diagnosing and assessing its
condition, or
• For creating a data basis for improving and optimizing the design process.
Functional improvement will only be achieved by passing on knowledge and expe-
rience. The production optimization control loop is not only closed via information
from the processes and plant of the production itself – data from the product’s uti-
lization phase also feeds into production optimization, e.g. for quality assurance in
12 Cyber-Physical Systems 211
Fig. 12.11 Potential of the interaction between smart products and smart production sys-
tems (Fraunhofer IWU, Fraunhofer IEM)
the case of sensitive product properties. In the same way, data from the production
process is combined with data from the product lifecycle to optimize product design,
e.g. to facilitate the efficient manufacturing of the next product generation (cf.
Fig. 12.11). These design and optimization processes are only efficient if they can
be automated. This is where machine learning algorithms gain particular signifi-
cance.
In this way, the evolution caused by Industry 4.0 technologies actually gives rise
to a disruptive paradigm change. Production, equipment suppliers, and product
developers enter a new quality of innovative cooperation. Production sites become
co-developers of products, and equipment suppliers become the designers of the
infrastructure for information and value streams.
212 Welf-Guntram Drossel • Steffen Ihlenfeldt • Tino Langer • Roman Dumitrescu
[19] Sauer, O. (2011): Informationstechnik in der Fabrik der Zukunft – Aktuelle Rahmenbe-
dingungen, Stand der Technik und Forschungsbedarf. ZWF Zeitschrift für wirtschaftli-
chen Fabrikbetrieb. 12, 955-962.
[20] Stoldt, J., Friedemann, M., Langer, T., Putz, M. (2016). Ein Systemkonzept zur durch-
gängigen Datenintegration im Produktionsumfeld. VPP2016 – Vernetzt Planen und Pro-
duzieren. 165-174
[21] WIGEP: Positionspapier: Smart Engineering, Wissenschaftliche Gesellschaft für Pro-
duktentwicklung, 2017
[22] ACATECH – DEUTSCHE AKADEMIE DER TECHNIKWISSENSCHAFTEN: Smart
Engineering. acatech DISKUSSION, Berlin (2012)
[23] Querschnittsprojekt Systems Engineering im Spitzencluster „Intelligente Technische
Systeme OstwestfalenLippe (it’s OWL)“, BMBF/PTKA(2012-2017)
[24] Gausmeier, J,; Dumitrescu, R, Steffen, D, CZAJA, A, Wiederkehr, O, Tschirner, C: Sys-
tems Engineering in der industriellen Praxis. Heinz Nixdorf Institut; Fraunhofer-Institut
für Produktionstechnologie IPT, Projektgruppe Entwurfstechnik Mechatronik; UNITY
AG, Paderborn (2013)
[25] Verbundprojekt „IviPep – Instrumentarium zur Gestaltung individualisierter virtueller
Produktentstehungsprozesse in der Industrie 4.0“, BMBF/PTKA (2017-2020)
“Go Beyond 4.0” Lighthouse Project
Individualized mass production
13
Prof. Dr. Thomas Otto
Fraunhofer Institute for Electronic Nano Systems ENAS
Summary
Industry’s need for new technologies to provide differentiation and efficiency
gains in production is the driving force of the Fraunhofer-Gesellschaft to pool
competencies in order to provide technologies for success. The thus far rigid
mass production will in future gain new impetus through digital manufacturing
technologies such as inkjet printing and laser-based techniques, in particular. The
integration of digital manufacturing technologies into a range of mass production
environments will permit individualized production with zero setup times and
only slightly increased cycle times.
Project overview
215
216 Thomas Otto
Project partners
The following Fraunhofer institutes are involved in the project:
• Fraunhofer Institute for Electronic Nano Systems ENAS
• Fraunhofer Institute for Manufacturing Technology and Advanced Materials IFAM
• Fraunhofer Institute for Laser Technology ILT
• Fraunhofer Institute for Applied Optics and Precision Engineering IOF
• Fraunhofer Institute for Silicate Research ISC
• Fraunhofer Institute for Machine Tools and Forming Technology IWU
Research schedule/sponsorship
€8 m. (+ €1 m. as required) via the Fraunhofer-Gesellschaft
Contact
Prof. Dr. Thomas Otto, Fraunhofer ENAS
Tel. +49 371 45001 100
13.1 Introduction
Across industries, the demand for innovative, individualized components for the
future markets of production equipment, automotive and aerospace engineering,
and lighting is continuously growing. Special functional materials are used to pro-
vide the corresponding components with the necessary highly qualified functional-
ities, with applications focused clearly on electronic and optical functionalities. The
anticipated efficiency stemming from the use of high-quality organic and inorganic
materials is the driving force for the development of customized process chains for
the manufacture of intelligent components with high diversification.
This demand is a global phenomenon; currently, in highly developed industrial-
ized nations, comprehensive economic structural measures are taken. Of particular
note here are the revitalization of industrial manufacturing in the USA and the de-
velopment of flexible manufacturing technologies in China.
Mass production facilities are designed to produce the largest possible numbers of
standardized components and products with the lowest cycle times. In order to
achieve this, tool-based manufacturing technologies are lined up in assembly lines.
The scope and complexity of the assembly line vary depending on the complexity
of the specific component or product. In the case of changeover to another product,
time-consuming machine resetting is required, which in turn reduces production
efficiency.
Market demands for individualized parts and products inevitably lead to smaller
volumes (down to batch size one), which in turn necessitate more frequent changes
of process chains. This additional effort significantly reduces the efficiency of mass
218 Thomas Otto
In the 500 years since the invention of the printing press by Johannes Gutenberg,
generations of technicians have developed the picture-by-picture transfer of ink
onto substrate to such a technical degree that today the human eye generally per-
ceives printed images as halftone objects just like the natural original. Despite the
fact that printed products consist of a high-definition cloud of microscopic halftone
dots. Traditional printing processes (letterpress, gravure print, screen printing, and
lithography) use hard-wearing printing blocks that are equipped with the image or
text content for reproduction one-time. From there identical copies are mass pro-
duced in high numbers during the actual printing process. Production of the printing
blocks is technically elaborate and only becomes economically beneficial if this
work can be allocated across as large a number of copies as possible.
In order to avoid this effort, two alternative pathways have been pursued in the
last 100 years towards producing images on a substrate that the human eye perceives
as a visual whole. In both techniques, the picture elements are generated each time
a copy is produced. Printers refer to this production strategy as “single edition
printing” while production technicians rather refer to it as “batch size one”. This
describes the fundamentals of digital printing.
This promising approach is based on the idea of producing picture elements
(halftone dots) by using small droplets of colored liquid. Nevertheless, the real
challenge of this printing process, nowadays known as inkjet, is the precise pictori-
al placement of the droplets on the substrate and the reproducible production with-
in narrow boundaries.
220 Thomas Otto
While in traditional printing processes the printing block is imaged once and
subsequently produces many copies (image one – print many), in the digital Inkjet
printing process printing blocks are no longer required; the ink image is transferred
once onto the substrate (image one – print one).
This results in a unique geometric distribution of ink on the substrate every time
the printing process is run. It is crucial, that for each individual illustration cycle a
new unique dataset can be taken as the basis. Exactly this is facilitated by modern
digital data processing systems, which allow for cycle times within the digital print-
ing process with substrate speeds of up to several meters per second. These digital
data systems were the force behind the conceptualization of digital printing.
Digital printing processes essentially fulfill the technological requirements for
low-volume manufacturing down to batch size one. So far, the geometrical require-
ments of the substrates have addressed the human sense of sight (i.e. the human eye)
with their functionality of color, now they must be able to address additional func-
tionalities such as electrical conductivity or insulation. These functionalities are also
addressed by producing an “image”, but instead of a landscape, the image may be
a conductor track capable of carrying electric current, for example. The discovery
that digital printing can produce material patterns with different functionalities at
high rates of productivity gave rise to functional printing over the last 25 years. The
approach is pursued by equipping ink systems with new properties such as electrical
conductivity in order to form systems of layers, according to the principles of print-
ing, which then can be used as electronic components (resistors, capacitors, diodes,
transistors, and sensors) or simple circuits. Another central challenge is, aside from
the corresponding adaptation of the printing processes, the preparation of the inks.
Today, as a result of the multifaceted development work conducted, functional
printers are able to choose between two kinds of functional inks that are commer-
cially available.
The first kind is suspensions of nanoparticles, which determine the functionality
of the ink. The second kind is solutions of functional molecules. Commercially
available inks permit nowadays, among other things, the manufacture of corre-
sponding material patterns made of conductive and semi-conductive organic poly-
mers.
Today, the field of digital functional printing technology is dominated by inkjet
systems. Due to the ink formulation benefits the so-called Drop-On-Demand (DoD)
inkjet is preferentially used; there again, printing systems based on piezoelectric
actuators (MEMS technology) dominate, since they avoid extreme thermal loads on
the inks compared to bubble jet printing.
Inkjet technology has experienced an enormous boost from the graphical indus-
try in the last decade. Today, all notable manufacturers of printers have inkjet print-
13 “Go Beyond 4.0” Lighthouse Project 221
ing systems in their portfolios or already integrated them into their traditional print-
ers. All optimizations of the inkjet process will aid to establish and further develop
digital functional printing, permitting an economical manufacturing of batch size
one in mass production environments.
No other tool can be metered and controlled nearly as precisely as the tool of light.
Currently, lasers are used in a wide range of fields ranging from telecommunications
and measurement engineering, to production of electronic microchips to ships.
Alongside classical laser-based techniques such as cutting, drilling, marking, and
welding, laser technology has also facilitated new production processes. These in-
clude selective laser beam machining for prototype construction, laser structuring
and laser polishing of component surfaces, laser deposition welding e.g. in turbo
engine production, selective laser etching, and EUV lithography; leading to the
opening of new markets. Exemplary innovative applications are the laser-based
generative manufacture of functionally- and resource-optimized metal components
using “3D printers” and the extensive direct micro structuring of functional surfac-
es via high-performance short-pulse lasers (photovoltaic, OLED, friction- and
wear-optimized surfaces).
In contrast to conventional processes, lasers allow the cost-effective manufacture
of both small quantities as well as complex products. The to a large extent existing
independence of production costs from quantity, wide variety, and product complex-
ity offers huge economic advantages. This enables high-wage nations such as Ger-
many to remain globally competitive by means of innovative products with de-
mand-optimized properties such as functional integration. For this purpose, laser
sources and processes in the field of surface technology have been the subject of
continuous research to accelerate the progress of industrial implementation of indi-
vidual, additive functional integration in components, realized by combining digital
printing techniques and laser-based pre- and post-processing techniques.
In laser beam processing, the optical energy emitted by the laser is converted into
thermal energy by absorption in the work piece (in this case, component surface or
printed functional layers on a component). Alongside many other dependencies, this
absorption is primarily dependent on the wavelength and material. Thus, not all
laser types are equally suitable for e.g. surface structuring via laser ablation or
222 Thomas Otto
13.4 Demonstrators
André Bucht
Fraunhofer Institute for Machine Tools and Forming Technology IWU
In the automotive industry, the market-based trend for individual products encoun-
ters rigid manufacturing chains. Production is strongly tool-related in order to guar-
antee efficient manufacturing of large batch sizes, but offers only limited opportuni-
ties for individualization. The functional individualization is mainly achieved by
installing mechatronic systems such as actuators, sensors, and control units. First,
this differential construction method leads to significantly increasing assembly effort
as, for example, the installation of the cable tree is nowadays one of the most exten-
224 Thomas Otto
Fig. 13.3 Finite element analysis of the printed ultrasound transducer (left); printed insu-
lation layers and conductor tracks after forming (right) (Fraunhofer IWU)
sive steps in vehicle assembly. Secondly, adding further flexibility to the already
existing manufacturing structure results in exponentially increasing complexity costs
in the fields of logistics, development, and production. Thirdly, thinking in terms of
single components leads to increased installation space and weight requirements.
Thus, there are limits set on the desire for increased individualization and func-
tional integration.
As mentioned in previous paragraphs, digital manufacturing steps offer a poten-
tial solution. The integration of digital process steps into analog tool-based manu-
facturing chains allow the individualization of batch sizes down to one.
With the aid of the technology demonstrator SmartDoor, these possibilities are
developed and demonstrated. The design and functionality of this demonstrator is
based on a real vehicle door. Functional elements such as ultrasound transducers,
conductor tracks, and controls are applied using printing measures on both, the
exterior component manufactured via forming technology as well as the fairing
component manufactured via injection molding. The functional elements are man-
ufactured in a hybrid process chain as a combination of analog and digital process
steps.
The design of the functional elements to be printed is based on actual require-
ments of a vehicle door. For this purpose, typical industry-related requirements as
well as the necessary physical parameters were taken as a basis. It could be demon-
strated, that hitherto mechatronic sensors and actuators such as ultrasound transduc-
ers are basically implementable as printed layer construction. But to achieve the
performance level of currently used systems, further optimizations with regard to the
design of the system and the properties of the printed layers are necessary. A compari-
son to the current state of the art pointed out, that increased resilience and improved
physical properties of the printed layers are necessary. Furthermore, the integration in
analog process chains – in this case forming technology –requires a significant leap
in the productivity of digital processes. In the further course of the project, the focus
will mainly be put on these aspects.
13 “Go Beyond 4.0” Lighthouse Project 225
Carbon fiber- and glass fiber-reinforced plastics (CFRPs and GFRPs) are character-
ized by their high specific stiffness and strength while simultaneously maintaining
low weight. Thus, their application in lightweight structures increased. However,
the advantages of these lightweight materials cannot yet be fully exploited: on the
one hand, quality fluctuations occur due to partially manual process chains; on the
other hand, components made of fiber-reinforced plastics (FRPs) cannot be analyz-
ed using the usual nondestructive testing methods. As in contrast to metallic struc-
tures, for example, damages caused by impacts during operation cannot be clearly
recognized or detected.
For this reason, structural components made of FRP are construed with large
safety margins or require short service intervals while in use to guarantee sufficient
reliability. Both lead to increased costs. Structural health monitoring (SHM) during
operation – e.g. by equipping fiber-reinforced structures with sensors- can reduce
safety factors and thus costs and energy use during operation. This means that dam-
age can be identified not only at the surface but ideally also inside the component,
provided that the stability will not be compromised by the FRP’s construction. The
integration of very thin film-based sensors, however, is already problematic since
in extreme cases these can lead to delamination within the FRP component and thus
to the breakdown of an FRP structure.
In the course of the subproject B “Smart Wing”, digital printing processes are
used to apply sensors on components to monitor the occurring load stresses, and
further, to integrate these sensors at relevant points during manufacture inside the
FRP components. Integrated sensors offer the possibility to monitor the fiber-re-
inforced components permanently, detecting high loads and damages early. Ad-
ditionally, sensors to detect icing are integrated as well as heating structures to
remove them. Digital printing and laser processes permit to print and functional-
ize sensor structures directly and locally onto FRP surfaces. An integration of
electrical, sensor, or capacitive functions right into the fiber composite is also
possible: structures made of functional materials can be applied with high resolu-
tion directly onto non-wovens or fabric via digital printing processes and further
be used as an impregnable textile layer in the manufacturing process of the fib-
er-reinforced composite. This way, in the vacuum infusion process printed poly-
ester and glass fiber non-wovens are processed into functionally integrated GFRP.
Carbon fibers, however, need to be electrically insulated first. For this step, how-
ever, printing processes are also suitable, enabling insulating and barrier materials
to be applied directly onto the fibers. In general, the materials applied during the
printing process require subsequent thermal treatment. A subsequent local thermal
treatment of the printed structures using lasers or high-energy UV radiation is
suitable.
The demand for custom LED lighting systems is constantly growing in, among others,
the automotive, medical technology, industrial manufacturing, and interior and exte-
rior architectural sectors. This demand requires optical components and systems that,
according to the application, illuminate specific areas and generate within these areas
defined, individual lighting patterns. These lighting patterns can serve as information
and interaction purpose, but also support the wellbeing of the user.
In subproject C “Smart Luminaire”, this demand is addressed. Based on standard
optics, individual optical components manufactured via laser and digital printing
processes are investigated. Using the inkjet printing process, the optical hybrid
polymer ORMOCER® is applied in layers onto a standard optical system or any
other substrate and subsequent hardened by UV lighting or infrared laser exposure,
respectively. This results in three-dimensional refractive optical structures with
13 “Go Beyond 4.0” Lighthouse Project 227
dimensions in the millimeter or centimeter range that, with optimum process param-
eters, are comparably transparent to optical bulk materials made of glass or polymer.
The transparency of the printed optical structures, as well as the required shape
accuracy and surface roughness represent particular challenges in the process de-
sign.
These three-dimensional refractive optical structures are combined with dif-
fractive structures, which are printed on the surface of the inkjet-printed optic
using two-photon absorption. Furthermore, printed electrical conductor tracks are
integrated into the optical structure to allow the integration of hybrid optoelectron-
ic components such as LEDs or photodiodes via precision assembly and contact-
ing. These are partially or completely embed into the optical structure in the
subsequent printing process. LEDs and photodiodes permit the interaction of the
lighting components with their environment. This takes place via visual and sensor
functions that display system states or measure environmental parameters. The
approach to manufacture the optical component digitally thus not only addresses
their application-oriented individualization but also the optoelectronic system
integration.
This allows the manufacture of completely new, individual, and highly integrat-
ed components and LED-based systems for structured lighting without reconfigur-
ing times of the machine during production. The possibility to address individual
customer requests opens new application fields for modern LED light sources and
supports the further distribution into fields of consumer and industrial lighting
technology.
228 Thomas Otto
Exploitation plan
The exploitation of the project’s results is based on three models:
• Direct exploitation of the results in industry via research services, technology
transfer, and licenses
• Exploitation via an online technology-atlas provided to industry
• Exploitation via a Fraunhofer-led Application Center
The direct exploitation of the results in industry started simultaneously to the project
work. Here, individual Fraunhofer technologies are transferred, for example, to
efficiently manufacture components for aerospace and automotive engineering and
equip them with new functional properties. Aim is to establish Fraunhofer as a
technology brand for individualized mass production over the course of the follow-
ing years. The technology exploitation will take place by means of suppliers and
original equipment manufacturers (OEMs).
13 “Go Beyond 4.0” Lighthouse Project 229
Summary
Cognitive systems are able to monitor and analyze complex processes, which
also provides them with the ability to make the right decisions in unplanned
or unfamiliar situations. Fraunhofer experts are employing machine learning
techniques to harness new cognitive functions for robots and automation solu-
tions. To do this, they are equipping systems with technologies that are inspired
by human abilities, or imitate and optimize them. This report describes these
technologies, illustrates current example applications, and lays out scenarios for
future areas of application.
231
232 Christian Bauckhage • Thomas Bauernhansl • Jürgen Beyerer • Jochen Garcke
14.1 Introduction
Recently, we have seen rapid progress in the field of machine learning that has led
to breakthroughs in artificial intelligence (AI). Above all, this development is driv-
en by deep neural networks. Neural networks are complex mathematical deci-
14 Cognitive Systems and Robotics 233
In order to understand why this equation works out and what future developments
it leads us to anticipate, we want to answer the following questions here: what are
artificial neural networks? How do they work? Why have they suddenly become so
good? What should we expect from this field in future?
y 1
f (z) 0
w0
−4 −2 0 2 4
w0 wm
w1 w2
x0 x1 x2 ... xm −1
positive number; otherwise it produces a negative one. The closer this output is to
either of the extreme values 1 or -1, the more certain the neuron is that its weighted
input lies above or below the threshold value.
Individual neurons are then typically arranged in layers and connected with one
another. This produces neural networks such as that shown in Fig. 14.2 which we can
imagine as essentially calculating a very large number of parallel or consecutive yes/
no decisions. If neural networks are large enough and their weight parameters are set
appropriately then they can solve practically any problem imaginable in this way.
20 21
As this was already shown mathematically in the 1980s [9][10], the question of
course arises as to why neural networks have only recently begun to be used uni-
versally with great success. In order to answer this question, we first need to con-
sider how a neural network learns to solve a desired task and what learning means
in this context.
The best way to do this is to take a look at a simple, concrete example. Let us
assume that a neural network like the one in Fig. 14.2 is designed to recognize
whether a dog or a cat is shown in an image with a resolution of 256x256 pixels. To
do this, the network’s input layer must first and foremost be significantly larger than
that shown above, containing a minimum of 256² neurons. It would also certainly
have to consist of more than six layers in order to be able to solve this apparently
simple but actually very demanding problem. The output layer however could still
consist of two neurons as in Fig. 14.2 because we have the freedom as developers
to specify which subject should produce which output. The following output coding,
for example, would make sense here: [1,-1] for images of dogs, [-1,1] for images of
cats, and [0,0] for images containing neither dogs nor cats.
In order for the neural network to adequately solve this classification problem,
the weight parameters of its neurons must be set according to the task. To achieve
this, the network is trained using examples. The training requires a dataset that con-
tains pairs of possible inputs and corresponding desired outputs, in this case images
of dogs and cats together with the corresponding output codes. If this kind of training
data is available, the training can be carried out using an algorithm that proceeds as
follows: starting with randomly initialized weights, the network calculates an output
for each training input. At the start of the training, this typically shows significant
deviation from the desired output. Using the difference between the calculated and
target outputs, however, the training algorithm is able to calculate how the network’s
weights need to be set so that the average error is as low as possible [11]. The weights
are thus automatically adjusted accordingly, and the next round of training begun.
This process is repeated until the network has learned to produce the desired outputs.
In practice, this training methodology has for a long time been faced with two
fundamental problems. On the one hand, the calculations that are required to adjust
the weights of a neural network are extremely elaborate. In particular large or deep
neural networks with numerous layers of neurons could thus not be trained within
reasonable timescales on earlier generations of computers. On the other hand, one
of the basic theorems of statistics states that statistical learning processes only
function robustly if the number of training instances is significantly greater than the
number of parameters of the model [12]. To train a neural network with 1 million
weight parameters, for example, you would need at least 100 million training ex-
amples so the network can learn its task correctly.
236 Christian Bauckhage • Thomas Bauernhansl • Jürgen Beyerer • Jochen Garcke
One area where machine learning (ML) techniques can be used is robotics. When
we are talking about these machines, there is one traditional image that generally
still comes to mind: a variety of robotic arms carrying out precise and strictly de-
fined movements in large production facilities, for example for handling or welding
purposes. This is where robots demonstrate their strengths, including repeat accu-
racy, precision, and the ability to handle heavy loads. Once programmed, they may
continue carrying out this one activity for many years, paying off the significant
initial investment in time and money by the long period of operation. Nevertheless,
changes to the production process or parts entail programming outlay since the robot
generally does not “know” how to interact with procedural changes. Its actions are
thus limited in their autonomy.
Increasing autonomy is the goal currently being pursued via developments in
so-called cognitive robotics: the aim is for robots, according to the term cognitive,
to be able to demonstrate perception, recognition, and implement corresponding
action by means of improved technologies. The basic principle here is that the sys-
tem perceives something via its sensors, processes the data, and uses this to derive
an appropriate response.
In the production environment, this allows users to respond flexibly as well
as economically to the growing requirements of short product lifecycles and in-
creasing numbers of variants. Capturing and analyzing large volumes of data in
near real time with the aid of the machine learning processes explained in Chap-
ter 14.2 enables systems to learn from the actions carried out and their success
or failure.
For many of these systems, the basis for the degree of autonomy required is
formed by cognitive functions since service robots must often navigate dynamic,
potentially unknown environments. They need to be able to identify, pick, and pass
objects, and even identify faces. The behavior for these actions cannot be prepro-
grammed, at least not in its entirety, and hence sensor information has to form the
basis for adaptive decision-making. Only with cognitive functions robots are able
to leave the production environment and become part of service provision in the
commercial or home environments. A number of new areas of application have
developed here in the last 20 years or so, including, in particular, agriculture, logis-
tics, healthcare, and rehabilitation.
238 Christian Bauckhage • Thomas Bauernhansl • Jürgen Beyerer • Jochen Garcke
The primary domains for industrial robots (IRs) are in work piece handling (50% of
all IRs) and welding (25% of all IRs) [source IFR]. In mass production, robot pro-
grams are mostly specified via manual teach-in processes and then carried out
blindly, so to speak. For years now, image processing systems (computer vision)
have been utilized e.g. for picking random objects such as in bin picking. Current
solutions for object recognition utilize recognizable features and fixed programmed
model-based object recognition methods. This means it poses no problem to the
robot that the work pieces are randomly placed since the system can use its cognitive
functions to identify where each work piece lies and how best to pick it.
Algorithm configuration as well as, for example, training for new work pieces is
carried out manually by experts. Machine learning can now be utilized so that the
bin picking algorithms – for object recognition and position estimation, for picking,
or for manipulation for example – can be optimized autonomously on the basis of
the analyzable information: the calculation times for picking shorten while the rate
of successful picks rises. Process reliability thus increases with each picking at-
tempt.
Fig. 14.3 Machine learning processes optimize robot-based systems for bin picking.
(Fraunhofer IPA/photo: Rainer Bez)
14 Cognitive Systems and Robotics 239
Fig. 14.4 With the welding robot software, welding even very small batches become
economically viable using a robotic system. (Fraunhofer IPA/photo: Rainer Bez)
14 Cognitive Systems and Robotics 241
For more than 20 years now, Fraunhofer IPA has been working on the vision of a
service robot that can be used in numerous environments including hospitals and
care facilities, warehouses, and even hotels. In 2015, researchers from Stuttgart
presented the fourth generation model, developed in partnership with Schunk und
Phoenix Design. Care-O-Bot® 4 offers diverse options for interaction, is highly
agile, and allows for various system enhancements with its modular approach. This
might start with the mobile platform for transportation, and then the torso-like
structure can be used with one or two arms, or none at all. Care-O-Bot® 4 is equipped
Fig. 14.5 Sales clerk Paul accompanies customers to the product they are looking for in a
Saturn store. (Image rights: Martin Hangen, right of reproduction: Media-Saturn-Holding)
242 Christian Bauckhage • Thomas Bauernhansl • Jürgen Beyerer • Jochen Garcke
with several sensors with which it is able to recognize its environment as well as
objects and people and orient itself in space and even navigate freely. Researchers
also placed an emphasis on a clear and appealing design.
The robot thus uses numerous cognitive functions with which it is able to
perceive its environment and operate autonomously. This is illustrated by an ac-
tive project with Media-Saturn-Holding: Care-O-Bot® 4 is being utilized in Sat-
urn stores as a sales assistant, a “digital staff member” called Paul. It greets cus-
tomers upon arrival, asks them which product they are looking for and accompa-
nies them to the correct shelf. It is able to recognize objects and its environment,
orienting itself and moving freely within it. Thanks to speech recognition soft-
ware, it is able to conduct conversations with customers. To do this, Paul is
equipped with domain-specific knowledge, that is, it is able to understand typical
terms and topics within the specialist electronics trade (e.g. products, services,
online order collections, etc.) and supply relevant information. In addition, it is
able to appraise its dialog with customers and “understand” satisfied or critical
feedback. And since it recognizes faces, it can tailor its communication to the
customer’s age and mood, for example. The goal of using the robot is to offer
customers an innovative retail purchasing experience in store. It is additionally
able to relieve staff by serving customers as their first port of call and assisting
with finding products. For detailed queries, it then calls staff from customer ser-
vices.
In future, it will be essential for cognitive robotics that knowledge acquired at
one time is also made available centrally and can thus be used by several systems.
The Paul in Ingolstadt, for example, will in future be able to share its acquired
knowledge with the Paul in Hamburg via a private cloud. In as much as the com-
plexity of what a robot needs to know is constantly increasing. In the same way,
individual systems are already able to utilize knowledge available from online
sources. Paul the robot is thus linked to Saturn’s online store so it can utilize the
product information already provided.
14 Cognitive Systems and Robotics 243
produce a map of the environment so the robot can explore an unknown environ-
ment by itself. In order to constantly improve the localization and map via fusion
with current sensor data, probabilistic methods for simultaneous localization and
mapping are employed.
To allow collision-free motion planning, the sensor data is continuously used to
generate a current obstacle map. It includes details on navigability and soil condi-
tion. The planning integrates both the map as well as the kinematic and dynamic
properties of the platform in order to generate an optimal trajectory with guaranteed
navigability. During planning, a variety of additional optimization criteria can be
incorporated such as driving with maximum efficiency or preferably driving on
even paths based on the data of the soil condition.
Due to the modular concept of the Algorithm Toolbox, a variety of different robot
platforms can be equipped with the relevant autonomous abilities for their respec-
tive uses without requiring major adaptation efforts. The Algorithm Toolbox is thus
utilized both for all-terrain platforms like the IOSB.amp Q2 (see Fig. 14.6) as well
as for larger trucks.
The methods of cognitive robotics can also be used for mobile work machines.
Examples include diggers and other construction machines, forklift trucks, and
agricultural and forestry machines. These are usually operated manually and gener-
ally need to be converted for autonomous operation first. This includes facilities for
electronically controlling driving and operational functions but also sensors for
capturing the current state of the machine e.g. of the joint angle of the digger arm.
It is also important to produce a computerized model of the work machine so that
the joint angle measurement can be used to calculate where exactly the grab/exca-
vation bucket or any other part of the machine is situated.
Once these conditions are in place, a mobile work machine can be equipped with
autonomous capabilities from the Algorithm Toolbox like a robot. In comparison to
mobile robots, additional methods for carrying out dedicated tasks with the manip-
ulator are required here. A manipulator is any moving part that is capable of fulfill-
ing handling tasks, such as the digger arm for example.
When using 3D environmental mapping, the fact that the manipulator may be in
the sensor’s field of view, requiring algorithms for distinguishing the manipulator
and obstacles in the sensor data, must be considered. Based on the 3D environmen-
tal mapping and the geometric model of the machine, collision-free manipulator
movements can be planned for the handling of tasks. The control of hydraulic driv-
14 Cognitive Systems and Robotics 245
The vehicle carries the usual sensor systems for both mapping and sounding the
seafloor simultaneously so that the various sensor data can be recorded without the
need for surfacing, refitting, and submerging again. In this way, fewer missions are
required and ship costs are reduced. Pressure-independent construction of suitable
components reduces the vehicle weight and saves space that would otherwise be
required for large pressure hulls. The new design of the rectangular payload section
permits the application of current sonars and optical systems (required by potential
customers). Water samplers and new systems for future applications can also be
easily integrated here.
Extensive mission planning support is provided to the user by means of semiau-
tomatic and automatic functions as well as libraries of driving maneuvers for the
DEDAVE vehicle. DEDAVE’s modular guidance software allows the user to change
the AUV’s behavior depending on sensor events. Since the patented system allows
energy and data storage to be exchanged without the use of tools or lifting gear, very
short turn-around times of approx. one hour are achievable between missions.
14 Cognitive Systems and Robotics 247
14.4.4 Summary
Prof. Dr. Jochen Garcke · Dr. Jan Hamaekers · Dr. Rodrigo Iza-Teran
Fraunhofer Institute for Algorithms and Scientific Computing SCAI
among other factors. Efficient software solutions are in place to assess several
simulation results as long as only simple parameters such as bends, intrusion of
the firewall at selected points, acceleration, etc. are being investigated. For de-
tailed analysis of individual crash simulations, specialized 3D visualization soft-
ware is utilized.
In order to analyze this large amount of complex data, we use machine learning
(ML) approaches for so-called nonlinear dimensionality reduction. With this a
low-dimensional representation is calculated from the available data. By visually
arranging the data with regard to this small number of key quantities calculated
using ML approaches, a simple and interactive overview of the data (in this case,
the simulation results) is facilitated. In particular, we have developed a method
which calculates a small number of elemental and independent components from
the data and thus enables the representation of a numerical simulation as their com-
bination. This data-based decomposition can be understood as a kind of elemental
analysis of component geometries and facilitates a very compact and efficient de-
piction. Elemental modes obtained when examining crash simulations may include
for example the rotation of a component or its global or local deformation in one of
the component’s areas, which especially also enables a physical interpretation of the
analysis results. In this way, a study can be carried out efficiently since all of the
simulations can be described with the aid of these elemental components and there-
by compared.
This reduction in data dimensionality facilitates the intuitive interactive visual-
ization of a large number of simulations. An interpretable arrangement in three
coordinates with respect to selected elemental analyses demonstrates the differenc-
es between the simulations, for example the various geometric deformations during
the crash. As an example, we will take a look at a digital finite element model of a
pickup truck, which we studied in the BMBF big data project, VAVID [14]. A fron-
tal crash is simulated, with the component plate thicknesses being varied. The
deformations of the main chassis beam are the subject of analysis. Due to the new
representation by the calculated elemental decomposition, it is possible to represent
the various deformations as the sum of the elemental components in a manner that
is compact and interpretable.
For each simulation, we examine around a hundred intermediate steps that are
then visualized simultaneously in a graphic, something that was impossible with
previous analysis methods. By means of these components, our ML methods allow
the progression of the crash behavior over time to be neatly illustrated (see Fig.
14.9). Each point here represents a simulation for a specific time step. It can be
clearly seen how all of the simulations begin with the same geometry and over time
produce two variations of the crash behavior, illustrated in each case by means of
14 Cognitive Systems and Robotics 249
typical deformations of the main chassis beam under observation. In addition, the
point in time at which this division occurs can be approximately identified. On the
basis of this kind of analysis of the simulation results, the development engineer is
better able to decide how design parameters should be chosen.
In addition to this, new digital measuring techniques have also been developed
and made available in recent years that enable high-definition, time-dependent 3D
data to be obtained from actual crash tests. The techniques that we have newly
developed facilitate comparisons to be carried out for the first time between simu-
lations and the precise measurement data from an actual experiment [14]. In this
way, the correct numerical simulation for an actual crash test can be identified,
something that was previously not feasible at this level of quality. This also makes
it possible to obtain an overview of all of the simulations and identify whether an
actual experiment is proceeding along the left or right deformation path as per
Fig. 14.9.
These recently developed ML methods are able to significantly simplify and
accelerate the virtual product development R&D process since the development
engineer requires less time for data preparation and analysis and can concentrate on
the actual core technical engineering tasks.
250 Christian Bauckhage • Thomas Bauernhansl • Jürgen Beyerer • Jochen Garcke
The full range of materials, chemicals, and active agents possible is absolutely vast.
The number of active agent molecules alone, for example, is estimated at 1060, with
the number of molecules with 30 atoms or less lying between 1020–1024. By con-
trast, fewer than 100 million known stable compounds are currently accessible in
public databases. The difference between known and potential compounds suggests
that a very large amount of new materials, chemicals, and active agents likely re-
main to be discovered and developed. On the other hand, the huge extent of this
scope represents an enormous challenge for the design of new materials and chem-
icals since exploring it is usually very costly. Fraunhofer SCAI is particularly re-
searching special ML methods here to substantially accelerate these kinds of design
and optimization processes for materials and chemicals.
For example, numerous materials and molecule databases are currently being
built worldwide, with the results of quantum chemical simulations in particular.
These can be used to develop efficient ML-based models for predicting properties.
A model developed by Fraunhofer SCAI can thus be trained on a multitude of
smaller molecules, but utilized especially for predicting molecules of arbitrarily
greater sizes. Here, the techniques developed allow chemical accuracy to be
achieved for many properties [13]. In this way, the costs of an elaborate quantum
chemical calculation can be substantially reduced by several orders of magnitude
(typically from a few hours to a few milliseconds). The descriptors and distances
specially designed by Fraunhofer SCAI to be suitable for molecules and materials
can also be used for ML and analysis methods to identify interesting and promising
areas within the range of all compounds for more precise exploration.
Overall, by means of ML techniques and numerical simulations the development
and design of new materials and chemicals can be made significantly simpler, fast-
er, and more cost effective.
[1] Bengio Y (2009) Learning Deep Architectures for AI, Foundations and Trends in Ma-
chine Learning 2 (1) 1-127
[2] Krizhevsky A, Sutskever I, Hinton G (2012) ImageNet Classification with Deep Convo-
lutional Neural Networks, in Proc. NIPS
[3] Hinton G, Deng L, Yu D, Dahl G, Mohamed A and Jaitly N (2012) Deep Neural Net-
works for Acoustic Modeling in Speech Recognition: The Shared Views of Four Re-
search Groups, IEEE Signal Processing Magazine 29 (6) 82-97
14 Cognitive Systems and Robotics 251
[4] Mikolov T, Sutskever I, Chen K, Corrado G and Dean J (2013) Distributed Representa-
tions of Words and Phrases and their Compositionality in Proc. NIPS
[5] Levine S, Wagener N, Abbeel P (2015) Learning Contact-Rich Manipulation Skills with
Guided Policy Search, Proc. IEEE ICRA
[6] Esteva A, Kuprel B, Novoa R, Ko J, Swetter S, Blau H, Thrun S (2017) Dermatologist-
level Classification of Skin Cancer with Deep Neural Networks, Nature 542 (7639)
115-118
[7] Silver D, Huang A, Maddison C, Guez A, Sifre L, van den Driesche G, Schrittwieser J
(2016) „Mastering the Game of Go with Deep Neural Networks and Tree Search“ Nature
529 (7587) 484-489
[8] Moracik M, Schmidt M, Burch N, Lisy V, Morrill D, Bard N, Davis T, Waugh K, Jo-
hanson M and Bowling M (2017) DeepStack: Expert-level Artificial Intelligence in
Heads-up No-limit Poker, Science 356 (6337) 508-513
[9] Hornik K, Stinchcombe M, White H (1989) Multilayer Feedforward Networks Are
Universal Approximators, Neural Networks 2 (5) 359-366
[10] Cybenko G (1989) Approximation by Superpositions of a Sigmoidal Function, Math-
ematics of Control, Signals and Systems 2 (4) 303-314
[11] Rumelhart D, Hinton G and Williams R (1986) Learning Representations by Back- pro-
pagating Errors, Nature 323 (9) 533-536
[12] V. Vapnik and A. Chervonenkis (1971) On the Uniform Convergence of Relative Fre-
quencies of Events to Their Probabilities, Theory of Probability and its Applications 16
(2) 264-280
[13] Barker J, Bulin J, Hamaekers J, Mathias S, LC-GAP (2017) Localized Coulomb Descrip-
tors for the GaussianApproximation Potential, in Scientific Computing andAlgorithms in
Industrial Simulations, Griebel, Michael, Schüller, Anton, Schweitzer, Marc Alexan- der
(eds.), Springer
[14] Garcke J, Iza-Teran R, Prabakaran N (2016) Datenanalysemethoden zur Auswertung von
Simulationsergebnissen im Crash und deren Abgleich mit dem Experiment, VDI-Tagung
SIMVEC 2016.
Fraunhofer Big Data and Artificial
Intelligence Alliance 15
Mining valuable data
Summary
Big data is a management issue across sectors and promises to deliver a competi-
tive advantage via structured knowledge, increased efficiency and value creation.
Within companies, there is significant demand for big data skills, individual busi-
ness models, and technological solutions.
Fraunhofer assists companies to identify and mine their valuable data. Experts
from Fraunhofer’s Big Data and Artificial Intelligence Alliance demonstrate how
companies can benefit from an intelligent enrichment and analysis of their data.
The data revolution will bring lasting and massive changes to many sectors. As far
back as 2013, this development was identified by the German Federal Ministry for
Economic Affairs when they commissioned Fraunhofer to conduct an initial analy-
sis of the use and potential of big data in German companies [14]. The aim was to
highlight potential courses of action for economy, politics, and research. Extensive
web research, a survey and several sector workshops soon revealed that companies
could be more efficiently managed through real-time analysis of their process and
business data, that analyzing customer data could facilitate increasingly personal-
ized services, and that connected products could be equipped with a greater level of
253
254 Stefan Wrobel • Dirk Hecker
Fig. 15.1 Data forms a new pillar of company value. (Fraunhofer IAS)
direct access to the diverse skills provided by Fraunhofer experts from more than
30 institutes. It thus combines unique sector-specific know-how across Germany
with deep knowledge of recent research methods in intelligent data analysis. The
benefit for companies? Best practices and use cases from other sectors can easily be
adapted and utilized for creative solutions.
The box that follows provides an overview of the Alliance’s core areas of business.
Fraunhofer Big Data and Artificial Intelligence Alliance core business areas
Production and industry:This is concerned with making better use of the grow-
ing volumes of data in production. By utilizing machine learning methods, pro-
cesses can be optimized, errors recognized sooner by means of anomaly detec-
tion, and human-machine interaction made safer by means of improved robotics
technology.
Logistics and mobility: The core issue of this business area is optimizing the
whole of logistics across all modes of transport. This helps to reduce empty runs,
waiting times, and traffic jams. Autonomous vehicles improve traffic flow and
provide occupants with opportunities to make good use of travel time.
Life sciences & health care: Connected devices monitor patients during their
everyday lives, intelligent systems evaluate medical data autonomously, and
telepresence robots facilitate distance diagnoses. The use of modern data analy-
sis in life sciences and health care offers numerous options and opportunities for
the future.
Energy & environment: Data can be used to predict noise in cities and identify
sources of noise, to analyze flora and fauna, and to optimize the energy manage-
ment of large buildings.
Companies that turn to the Fraunhofer Big Data and Artificial Intelligence Alliance
are at different stages in their journeys to digitization and benefit from a modular
offering divided into four levels, as shown in Fig. 15.2.
In the beginning, the focus is on getting to know the potential in your own sector,
generating enthusiasm among staff, and coming up with initial ideas for your own
company. This is facilitated by excite seminars and innovation workshops. The
guiding principle here is to learn from the best.
At the next level, the aim is for the most promising ideas to gain momentum
quickly and efficiently. A data asset scan is used to identify all of the relevant inter-
nal data, and a search is conducted for complementary publicly available data. The
most important tools and algorithms are integrated within a Fraunhofer starter
toolkit so that even large datasets can be analyzed quickly. Fraunhofer institutes
provide support where needed in developing concepts, demonstrators, and scalable
prototypes.
Data is a production and competitive factor and thus an asset. For this reason, it is
worthwhile for companies to collect data or buy it in, to curate it, to protect it, and
to intelligently combine and analyze it. A modern car, for example, is not merely a
vehicle for transport, it can also be used as a weather station. From integrated ther-
mometers and rain sensors or by logging the activation of windshield wipers, from
light sensors, hygrometers and GPS, important data can be collected and transmitted
via the mobile telephony network. By aggregating the data from a large number of
vehicles and comparing this with satellite images, a very high-resolution local im-
age of the current weather can be produced without substantial investment in infra-
structure. Initial pilot trials have shown that there is significant willingness to pay
for these kinds of local weather reports: traditional weather services may use them,
for example, to optimize weather forecasting for agriculture or to estimate the pow-
er injection from photovoltaic and wind power facilities. A less obvious and new
group of customers are facility managers, for example, who are able to use the data
to optimize the air conditioning for large buildings and thus improve the indoor
climate and energy efficiency.
Other forms of monetization are more subtle, as can be seen from a number of
technology start-ups in the financial services sector – so-called fintechs. A research
15 Fraunhofer Big Data and Artificial Intelligence Alliance 259
project conducted by Gimpel et al. [4] on the business models and service offerings
of 120 fintech start-ups showed that for fintechs the rich data generated at the cus-
tomer interface is an important production and competitive factor. Alongside data
from individual customers, this increasingly includes comparisons with other cus-
tomers (peers) and combinations with publicly available data.
Finally, monetization addresses the question of revenue model. There are various
“currencies” that the customer may pay with. In only around a third of the cases they
pay with money. In the largest number of service offerings observed, the users “pay”
with loyalty, attention, or data. Loyalty to a brand or firm may lead to indirect mon-
etization in other areas of operations. Both, selling data and advertising lead to fi-
nancial payments from business partners. Where monetization via business partners
is concerned, the service user is no longer the customer but the product itself.
The key dimensions for designing data monetization are thus immediacy (direct
vs. indirect sales), the user’s currency (money, data, attention, loyalty), and the role
of the user (customer or product). Additional dimensions are the billing increment
(subscription or transaction-based), price differential (none, segment-oriented, ver-
sioned, personalized), and price calculation (based on marginal costs that tend to-
wards zero, or value for the user and customer).
Not all of the options that could ultimately be imagined for the use and mone-
tization of data are legal and legitimate. There are substantial legal restrictions on
the use of data (e.g. in Germany the Teleservices Act, Unfair Competition Act, EU
General Data Protection Regulation), there are limits of acceptance on the part of
users, customers, and partners, and there are ethical boundaries to be observed.
Value-centered corporate leadership requires balancing these boundaries with the
technological and economic possibilities and finding a viable middle way that
creates value from data as a production and competitive factor and asset. In the
future, users and regulators will also increasingly become aware of the value of
data and its systematic analysis. More and more, users will demand appropriate
recompense (free services, fees) for their data. Discussions are ongoing in special-
ist legal and political circles as to whether and how the value of data should be
accounted for and assessed, and whether payments with data for allegedly free
services should not be seen as payment as much as money. This would have clear
consequences for the liability of providers of allegedly free services.
260 Stefan Wrobel • Dirk Hecker
A few years ago, many companies were able to increase their efficiency simply by
combining their data from previously separate “data silos” and analyzing it in a new
way.
But the new possibilities of predictive and prescriptive analysis for early detec-
tion and targeted actions were soon recognized. Now, the trend is increasingly to-
wards self-regulating – i.e. autonomous – systems [3]. Here, machine learning
methods are being utilized that are particularly well applicable to large datasets.
These methods identify patterns from historical data and learn complex functions
that are also able to interpret new data and make proposals for action. These are
mostly functions that could not be programmed explicitly due to the large number
of different cases.
This becomes most evident in speech and image processing. Here, deep learning
methods have recently achieved spectacular results by training networks with many
layers of artificial neurons. They enable intelligent machines to speak with us in any
language of choice and to perceive and interpret our shared environment [12]. Ar-
tificial intelligence creates a new communications interface with our home, car, and
wearables, and is replacing touchscreens and keyboards. Major technology giants
are spending up to $27 billion on internal research and development for intelligent
robots and self-learning computers, and leading Internet giants are positioning
themselves as companies for artificial intelligence [2]. Germany, having won a
leading role for itself in the field of Industry 4.0, is now seeing the next wave of data
rolling in with the Internet of Things [9]. IDC estimate that there will be 30 billion
connected devices by 2020 and 80 billion by the year 2025 [7]. By using data gen-
erated during industrial manufacturing and over the entire product lifecycle, “indus-
trial analytics” reveals potential for value creation along the entire production chain
– from construction through logistics and production to reengineering. In the digital
factory, where smart and connected objects constantly map the current state of all
processes at all levels of production and logistics to a digital twin, demand-driven
autonomous control of production processes, data-driven forecasting and deci-
sion-making become possible [5]. Work in the area of production is becoming more
user-friendly and efficient because of intelligent assistance systems [1]. The use of
machine learning techniques, especially in industry, requires high levels of do-
main-specific knowledge. This begins with the selection of the relevant data in the
15 Fraunhofer Big Data and Artificial Intelligence Alliance 261
flood of sensors and notifications, its semantic enrichment, and the interpretation of
the patterns and models learned. Here we are increasingly concerned with the com-
bination of engineering and machine knowledge, with comprehensibility and liabil-
ity, with the controllability of autonomous intelligent machines and collaborative
robots, with data protection, with security and certifiability, with the fear of jobs
losses and the need for new staff qualifications.
The arrival of big data gave rise to the new occupational profile of data scientist
[12]. Ideally, these data scientists are not only familiar with up-to-date software for
Fig. 15.3 Teams of data scientists bring together a number of talents. (Fraunhofer IAIS)
262 Stefan Wrobel • Dirk Hecker
big data systems and predictive data analysis, but also understand the respective
business sector and have the relevant domain knowledge.
As far back as 2011, McKinsey already warned that the need for specialists in
big data analytics would become a bottleneck in companies in view of the rapid
Increase in data volumes [11]. The growing need for processing images and videos,
speech and text for digital assistants and intelligent devices is making the situation
even more acute. Data scientists who master current tools such as Spark and Python
are therefore well remunerated [8]. In practice, therefore, all-rounders are less com-
mon than teams with a mix of specialists.
Although many British and American universities offer masters courses for data
scientists, and a number of German universities have now followed suit, this is not
enough. Existing staff need to be trained for new methods, tools, languages, librar-
ies and platforms, and sector specialists need to be able to collaborate in teams with
data scientists. Here, too, we cannot wait until data science courses become estab-
lished in the universities’ curricula in the application sectors.
Fraunhofer Big Data and Artificial Intelligence Alliance has thus offered cours-
es in big data and data analysis for professionals ever since it was founded. More
specific training courses in the different tasks of data scientists followed: potential
analysis for big data projects, security and data protection for big data, and – most
recently – deep learning. Sector-specific and application-specific training courses
followed, e.g. for smart energy systems, smart buildings, linked data in enterprises,
data management for science, and social media analysis. In 2015, a three-level
certification program for data scientists was started.
The Fraunhofer Big Data and Artificial Intelligence Alliance also advises com-
panies that have their own professional development academies. Here, support
ranges from the development and adaptation of modules to training in-house train-
ers in open Fraunhofer training courses. In 2016, for example, the Verband Deutscher
Maschinen- und Anlagenbau (The Mechanical Engineering Industry Association,
VDMA) produced plans for future machine learning in machinery and plant build-
ing [9] and is now offering online training materials developed for engineers, both
with the support of Fraunhofer [15].
15.6 Conclusion
As far as value creation from data is concerned, US Internet giants followed by their
Chinese counterparts are in the lead. They have the data, the platforms for big data
processing and learning, they develop new methods, and by now they can draw on
extensive trained models for speech and image comprehension. Industrial analytics
15 Fraunhofer Big Data and Artificial Intelligence Alliance 263
and machine learning for intelligent machines and devices, however, offer a field
where German industry and research can and must develop unique selling points
with secure and transparent solutions. If production companies can use qualified
staff and the right technologies to strategically mine their valuable data, then they
will also be able to secure a lead in the face of international competition.
business-functions/digital-mckinsey/our-insights/big-data-the-next-frontier-for-innova-
tion (abgerufen: 07/2017)
[12] Neef A (2016) Kognitive Maschinen. Wie Künstliche Intelligenz die Wertschöpfung
transformiert. Z_punkt. URL: http://www.z-punkt.de/de/themen/artikel/wie-kuenstli-
che-intelligenz-die-wertschoepfung-treibt/503 (abgerufen: 07/2017)
[13] Patil D, Mason H (2015) Data Driven – Creating a Data Culture. O’Reilly. URL: http://
www.oreilly.com/data/free/data-driven.csp (abgerufen: 07/2017)
[14] Schäfer A, Knapp M, May M et al (2013) Big Data – Perspektiven für Deutschland.
Fraunhofer IAIS. URL: https://www.iais.fraunhofer.de/de/geschaeftsfelder/big-data-
analytics/referenzprojekte/big-data-studie.html (abgerufen: 07/2017)
[15] University4Industry (2017) Machine Learning – 1.Schritte für den Ingenieur. URL:
https://www.university4industry.com (abgerufen: 07/2017)
Safety and Security
Cybersecurity as the basis for successful digitization
16
Prof. Dr. Claudia Eckert
Fraunhofer Institute for Applied and Integrated Security AISEC
Prof. Dr. Michael Waidner
Fraunhofer Institute for Secure Information Technology SIT
Summary
Cybersecurity is the basis for successful digitization and for innovation in all
sectors, e.g. in digital production (Industry 4.0), smart energy supply, logistics
and mobility, healthcare, public administration, and cloud-based services, too.
The role of cybersecurity [13][11] is to protect companies and their values and to
prevent damage or at least limit the impact of any potential damage. Cybersecu-
rity encompasses measures to protect IT-based systems (hardware and software)
from manipulation and thus safeguards their integrity. Furthermore, it includes
concepts and processes that guarantee the confidentiality of sensitive information
and the protection of the private sphere as well as the availability of functions
and services. Guaranteeing integrity, confidentiality, and availability are the
familiar safety objectives already pursued by traditional IT security, but achie-
ving them has become increasingly difficult and complex with digitization and
networking and the accompanying connection between the digital and physical
worlds.
The article that follows provides an insight into current trends and develop-
ments in the field of application-oriented cybersecurity research and makes use
of selected example applications to outline challenges and potential solutions.
265
266 Claudia Eckert • Michael Waidner
Cybersecurity is essential for sustainable value creation and for viable future busi-
ness models. Current studies confirm this:
In mid-2017, the Bundesdruckerei published the findings of a representative
survey [1] on IT security that highlights the importance of IT security for digitiza-
tion: “Nearly three quarters of respondents view IT security as the foundation for
successful digitization.” In a survey of specialists conducted in early 2017 [2], the
majority of respondents assessed the current significance of IT security for value
creation in Germany as high, and nearly 90% were of the opinion that this signifi-
cance would further increase over the next five years. The study commissioned by
the German Federal Ministry for Economic Affairs and Energy stresses that IT se-
curity is not only an important market in itself, it is also the prerequisite for the
development of additional viable future business models.
For the Bitkom industry association, IT security is also one of the two top issues
in 2017 [3]: “IT security is becoming even more important since digitization is
leading to more and more critical systems such as vehicles, medical technology, and
machines becoming digitally connected,” stated Bitkom General Manager Dr. Bern-
hard Rohleder. “At the same time, the attacks of criminal hackers are becoming
increasingly refined. Regular security tools such as virus scanners and firewalls are
often no longer enough for companies.”
sors, cloud platforms or apps, as well as the machine learning processes utilized
[14][15].
The problem, however, is that networking and digitization not only drastically
raise the potential impact of successful cyberattacks, but that fundamental rethink-
ing about our interaction with cyberthreats is also required, since the attack land-
scape, too, has changed dramatically in recent years. Cybercrime and cyberespio-
nage have become professionalized [16]. Attacks are increasingly targeted at spe-
cific organizations and individuals and evade regular protective mechanisms such
as firewalls, antivirus programs, and intrusion detection systems.
Today’s attackers generally use human error, weak points in IT or in IT-based
processes, and inadequate precautions in order to penetrate systems, manipulate
them, or gain unauthorized access to sensitive information. One problem here is the
ever-increasing number of hackers carrying out targeted attacks for their own ben-
efit or to the disadvantage of a third party. A key aim of these kinds of attacks is to
infiltrate systems with malicious software, so-called trojans, that gathers informa-
tion such as passwords or login details unnoticed and makes them available to the
attacker. The ease with which attackers with even limited technical knowledge can
carry out these kinds of attacks represents an enormous threat for current connected
blacklisted
Android
top 200 apps sorted
by categories
Fig. 16.1 Vulnerabilities of Android apps by application area according to the Appicaptor
Security Index (in illustration of [6]) (Fraunhofer SIT)
268 Claudia Eckert • Michael Waidner
blacklisted
Android
top 200 apps sorted
by categories
Fig. 16.2 Vulnerabilities of iOS apps by application area according to the Appicaptor Se-
curity Index (as an illustration of [6], Fraunhofer SIT)
systems – and an even greater one for the systems of tomorrow. Also increasingly
of note are attacks by white-collar criminals, secret service surveillance, and organ-
ized crime. Alongside forms of extortion, e.g. by means of so-called ransomware
[13], targeted attacks are being conducted on company managers who usually pos-
sess multiple authorizations to access sensitive information in their respective com-
panies. Current research findings demonstrate the state of (in-)security of modern
information technology at all levels. Examples include:
Cybersecurity research thus faces enormous challenges that require not only tech-
nological innovations, but also a rethinking of the development and operation of
secure, software-intensive cyber-physical systems.
The specific risks that result from this insecure information technology in products,
services, and infrastructures become clear when we look at the various areas of
application of information and communications technology and the sectors of in-
dustry affected.
Industry 4.0: In the world of production, a fourth industrial revolution is taking place
where production systems are being integrated and networked by means of infor-
mation and communications technology, both vertically with other business systems
as well as horizontally across entire value chains. This allows production processes
to be designed more flexibly and efficiently and to be adjusted in real time. The
openness and interconnectedness of components in these production processes har-
bors the risk of IT manipulation by attackers, e.g. by competitors or extortionist
hackers. If IT vulnerabilities are found in industry control systems this often pro-
duces a large number of attack points since products from a small number of man-
ufacturers are generally in use among a large number of users. A summary of the
challenges and potential solutions for cybersecurity in Industry 4.0 can be found in
[15].
Energy supply: The energy transition relies on decentralization and intelligent con-
sumption management, e.g. in the context of smart grids. To achieve this, devices
for energy use and components of energy supply must be linked via information
270 Claudia Eckert • Michael Waidner
Fig. 16.3 Digitization opens up opportunities for optimization and new industrial value
chains, but also harbors risks due to unauthorized access to sensitive systems and informa-
tion (Fraunhofer SIT).
technology. This produces entry points for cyberattacks that can lead to economic
risks to those affected or even to problems of supply.
Finances: The players in the financial world are already highly connected with one
another via information and communications technology today. These connections
form the nervous system for the economy and industry. Breakdowns and manipula-
tion could lead to huge economic losses. Today a bank robbery no longer requires
16 Safety and Security 271
the use of firearms – all an attacker needs is a computer to access the bank’s IT
systems. It is thus highly important that financial systems are secure and remain
available. This requires cybersecurity.
Healthcare: Doctors, hospitals, and insurers today rely heavily on information and
communications technology, contributing to an increase in efficiency in healthcare
and a reduction in costs. Security and data protection requirements are particularly
important for medical data.
Software: Cybersecurity is becoming ever more important for the software industry.
Practically every company utilizes application software in business processes that are
critical to their respective business success. This application software is characterized
by special functions required for the most diverse range of purposes. Nevertheless,
when application software is developed often only the functions relevant to the appli-
cation domains is considered. Since cybersecurity is only considered marginally in these
cases, and due to software’s increasing complexity, security loopholes inevitably ensue.
Cloud: For cloud services there are security requirements that go far beyond those
for “traditional” application software [23]. Making IT resources available via cloud
services has particular economic potential, especially for small and mid-size com-
panies for whom there are significant costs to running in-house IT departments.
Cloud computing offers companies the opportunity to reduce investment and oper-
ational costs while at the same time increase agility. The high availability require-
ments for cloud services, combined with their exposed location on the Internet,
provide challenges for updating and patching processes and necessitate high to
maximum levels of protection from attack, permanent monitoring of the threat
landscape, and elaborate attack and intrusion detection mechanisms.
272 Claudia Eckert • Michael Waidner
Today, innovation takes place almost exclusively with the aid of or by means of
information technology. The short innovation cycles in the IT industry lead to a
constant pressure to modernize, accompanied by an equally constant pressure to
reduce costs. Whether it is the Internet of Things, Industry 4.0, big data, blockchain,
artificial intelligence, or machine learning: every new trend in information technol-
ogy intensifies the interaction between information, services, and end devices. This
constantly opens up new threats with regard to security and data protection.
Security ad hoc,
primarily reactive Systematic security, proactive and Systematic security for large real
attack-tolerant systems, for example „industrie
4.0“ or „smart services“
Fig. 16.4 Paradigm development in cybersecurity: from reactive security, through “secu-
rity and privacy by design”, to “cybersecurity at large” (Fraunhofer SIT)
16 Safety and Security 273
Industry 4.0, values distant from IT nevertheless increasingly stand at the center of
security technology considerations now.
If cybersecurity is not taken into account across a product’s entire lifecycle, then
this leads to negative effects for providers: either products and services are not se-
cure enough, or cybersecurity becomes far more expensive than necessary. It is thus
important to consider cybersecurity right from the design stage and in the develop-
ment and integration of new products, services, and production processes.
“Security and privacy by design” is the term used to refer to the consideration of
cybersecurity and privacy protection throughout the entire lifecycle of IT products
and services [12]. Paying attention to security questions at the earliest possible stage
is especially important, since most disclosed weak points are based on errors in the
design and implementation of IT systems.
“Security at large” considers security not only during the design and implemen-
tation of products and services but also during the integration of IT components for
large and complex systems. This also encompasses taking account of the require-
ments of specific fields of application and technology where numerous components
are integrated into large, complex systems. These include, for example, business
software, cyber-physical systems, cloud computing, critical infrastructure, or Indus-
try 4.0, and in particular the Internet itself as the fundamental infrastructure in the
IT domain.
Mobile app: For mobile apps, Fraunhofer SIT and Fraunhofer AISEC have respec-
tively developed a semi-automated service (Appicaptor) and an automated tool
(AppRay) for security analysis. Both analysis tools analyze not only Android apps
but iOS ones, too. Appicaptor brings together several tools for analysis and test case
generation, ranking apps using a range of different non-static techniques. These
security tests facilitate fast and practical testing for known errors and implementa-
tion weaknesses, amongst others, by means of mass testing where apps are exam-
ined quickly and with a minimal failure rate for a specific class of error. Here, re-
searchers gathered characteristics of different error behaviors that correspond to
attacks or may be able to be used for attacks so that a practically useful catalog of
potential attacks (or error behavior supporting attacks) was produced, enabling to
test the resilience of Android and iOS apps to these attacks in various scenarios. This
analysis technique also enables testing in cases where the source code of the app is
unavailable. Appicaptor detects, amongst others, a range of problems that arise via
the incorporation of foreign code in apps and thus represents a significant contribu-
16 Safety and Security 275
Fig. 16.6 Without end-to-end encryption, emails can be intercepted en route and read like
a postcard. (Fraunhofer SIT).
tion to security at large. The AppRay analysis tool enables the flow of data and in-
formation in apps to be automatically revealed, along with infringements of data
protection guidelines, for example, or of other security rules that are individually
configurable.
Enterprise apps: The large codebases of enterprise apps make access for analysis
purposes difficult. Here, Fraunhofer SIT introduced the Harvester solution to solve
part of the problem of software security analysis in the context of security at large:
Harvester extracts relevant runtime values from an app completely automatically,
even when obfuscation and anti-analysis techniques such as encryption have been
used. The precise program code that calculates the relevant values is cut out from an
app and then executed in a secure environment/in isolation. Irrelevant program state-
ments are thus first removed and the amount of code for examination thus minimized.
The cut out code is then executed directly in a separate area (without e.g. waiting
for a holding time or restart). In this way, it becomes possible to solve a complex,
dynamic analysis problem. Harvester can be used at various points, both as a plugin
and also as an independent tool or in-house webservice. Harvester is aimed at di-
verse user groups such as developers, security experts within companies and secu-
rity authorities, app store operators, and antivirus providers.
security. One example of this is the naming system in the Internet (domain name
system, DNS), in particular the caching strategies used. The topology and architec-
ture of name servers usually utilizes temporarily stored caches. A range of serious
vulnerabilities and misconfigurations have been identified: large-scale experiments
and measurements show that these caches were generally run very unprofessional-
ly, thus leading to potential for attack and decreased performance [8]. Anyone in a
position to manipulate DNS can intercept email and telephone calls or conduct
practically undiscoverable phishing attacks and thus gain access to login data and
passwords, for example. Fraunhofer SIT is thus working on tools to allow Internet
infrastructure to be better secured and is developing recommendations for actions
for manufacturers and network operators in order to be able to respond to vulnera-
bilities, also at a short notice [8].
Hardware and embedded security: Hardware and software protection measures for
increasing the security of electronic and digital devices and components have
formed part of Fraunhofer AISEC’s offering for years now. The institute develops
personalized and tailored solutions for different sectors and products. This may be
for embedded systems in industrial machinery and plant construction for example,
for embedded hardware and software components in industrial control systems and
Fig. 16.7 Without analyses, the dangers involved in using an app cannot be assessed
(Fraunhofer SIT)
16 Safety and Security 277
in the automotive or avionics field, but also for IoT components in home automation
or healthcare. Embedded systems are mostly made up of an assembly of several
chips and are generally easily physically accessible. They are thus defenseless to
attackers with competencies in the fields of electronics, telecommunications, im-
plementations, and hardware attacks. In addition, attackers have the opportunity to
exploit internal interfaces such as debug interfaces or to gain direct access to an
integrated memory chip. It is thus imperative to strive for a high degree of hardware
security for these kinds of systems right from the start. Core areas of focus at the
AISEC are in developing secure system-on-chip solutions [24], protecting embed-
ded software from manipulation [25][33], but also in safeguarding secure digital
identities for embedded components.
Secure IoT and data sovereignty: Insecure configurations and insufficient monitor-
ing of IoT devices pose a high risk of potential attacks/manipulation, above all in
the field of industrial automation/Industry 4.0. If companies use the data from these
devices as the basis for decision-making within their business processes then this
can have fatal consequences. The Trusted Connector developed at Fraunhofer
AISEC protects sensitive business processes from threats that arise due to network-
ing. It ensures that only reliable data is incorporated into critical decision-making.
The pre-processing of this data by applications in the Trusted Connector facilitates
reliable processing chains within the company and beyond company borders. A
secure container-based execution environment facilitates strict isolation of running
applications. Data and software are thus protected from loss and unwanted modifi-
cation. Integrity verification for the data and installed applications combined with
a hardware-based security module (Trusted Platform Module – TPM) together en-
sure a high level of reliability [25]. This is supplemented with flexible monitoring
of access and dataflows facilitating fine-grained organization of dataflows both
within and outside of the company [26]. The AISEC Trusted Connector is also the
central security component in the Industrial Data Space (IDS) that is currently being
developed by Fraunhofer together with partners from industry [34]. The IDS aims
to create a reference architecture for a secure data space that enables companies
from various industries to manage their data assets confidently. This data space is
based on a decentralized architectural approach where data owners are not required
to surrender their data superiority or sovereignty. A central component of the archi-
tecture is the Industrial Data Space Trusted Connector which facilitates the super-
vised exchange of data between participants in the Industrial Data Space.
partners change, but also new software artifacts such as apps and software updates
are loaded and executed while operation is ongoing. Techniques are being devel-
oped at Fraunhofer AISEC to continuously evaluate the current security state of
IT-based systems such as cloud platforms [27]. Using advanced analysis techniques
based on machine learning, malicious code can be identified early, for example, so
that potential damage can be minimized [28]. Processes developed specially at
AISEC also enable security analysis to be carried out via encrypted communications
paths [31] such that systems can be supervised from afar, for example, without
losing the protection of the secure communications channel. Using the isolation and
supervision techniques developed at AISEC [30] as well as measures for continuous
integrity measurement [29], a system can continuously compare its system state to
specified rules and requirements to be observed, and identify and defend against
deviations early on that indicate potential preparatory steps for attacks.
A quick look at the instruction by the Federal Office for Information Security [38]
shows how extensive the security specific competency profile is. According to the docu-
ment, the typical weak points that make IT systems vulnerable can be divided into four
categories:
1. Software programming errors
2. Weak points in the software’s specification
3. Configuration errors
4. Users of IT systems as an insecurity factor
We need only remember that the source code in commonly used software products may
be several million lines long to understand how great the task of developing secure
software is. Security by design approaches only find their way into development praxis,
however, if the knowledge of, for example, security-oriented programming languages
forms part of vocational or practical training courses. At the same time, the application
of principles for secure software development also requires a response to organizatio-
nal questions. How should the interfaces between the different roles (developer, tester,
system integrator, etc.) be best designed, for example?
However, resisting threats is just one of the elements. It is equally important that com-
panies and authorities identify successful attacks in the first place. They need to be able
to assess the degree of damage and find ways to “get affected systems going” again.
This knowledge is spread across several individuals within the company and can only
be retrieved through all of the different staff working well together.
280 Claudia Eckert • Michael Waidner
But how can companies and authorities be in a position to carry on keeping up. How
can IT security risks be reduced and better overcome? Professional development is the
decisive key here. Thereby four conditions need to be met so that new knowledge of IT
security can quickly be brought into effect.
1. Only the latest research knowledge allows organizations to be a step ahead of po-
tential attackers and to use, for example, new development and testing processes for
secure software.
2. Not all actors have the same need for knowledge. Professional development content
must be tailored for specific topics, roles, and sectors and take account of differing
levels of prior knowledge.
3. Knowing comes from doing. Training in laboratories with current IT infrastructure
allows participants to experience actual threat scenarios and test out the applicability
of solutions.
4. Utilize limited time resources sensibly. Compact formats with practical issues at the
forefront can be meaningfully integrated into day-to-day professional life and com-
bined with appropriately designed learning media made available electronically.
[19] F. Fischer, K. Böttinger, H. Xiao, Y. Acar, M. Backes, S. Fahl, C. Stransky. „Stack Over-
flow Considered Harmful? The Impact of Copy & Paste on Android Application Secu-
rity“ , IEEE Symposium on Security and Privacy 2017.
[20] A. Zankl, J. Heyszl, G. Sigl, „Automated Detection of Instruction Cache Leaks in RSA
Software Implementations“, 15th International Conference on Smart Card Research and
Advanced Applications (CARDIS 2016)
[21] N. Jacob, J. Heyszl, A. Zankl, C. Rolfes, G. Sigl, „How to Break Secure Boot on FPGA
SoCs through Malicious Hardware“, Conference on Cryptographic Hardware and Em-
bedded Systems (CHES 2017)
[22] C. Teichmann, S. Renatus and J. Eichler. „Agile Threat Assessment and Mitigation: An
Approach for Method Selection and Tailoring“. International Journal of Secure Software
Engineering (IJSSE), 7 (1), 2016.
[23] Niels Fallenbeck and Claudia Eckert. „IT-Sicherheit und Cloud Computing“. In: In-
dustrie 4.0 in Produktion, Automatisierung und Logistik: Anwendung, Technologien,
Migration“, ed. by Thomas Bauernhansl, Michael ten Hompel, and Birgit Vogel-
Heuser. Springer Vieweg, 2014, pp. 397–431.
[24] N. Jacob, J. Wittmann, J. Heyszl, R. Hesselbarth, F. Wilde, M. Pehl, G. Sigl, K. Fisher:
„Securing FPGA SoC Configurations Independent of Their Manufacturers“, 30th IEEE
International System-on-Chip Conference (SOCC 2017)
[25] M. Huber, J. Horsch, M. Velten, M. Weiß and S. Wessel. „A Secure Architecture for
Operating System-Level Virtualization on Mobile Devices“. In: 11th International Con-
ference on Information Security and Cryptology Inscrypt 2015. 2015.
[26] J. Schütte and G. Brost. „A Data Usage Control System using Dynamic Taint Tracking“.
In: Proceedings of the International Conference on Advanced Information Network and
Applications (AINA), March 2016.
[27] P. Stephanow, K. Khajehmoogahi, „Towards continuous security certification of Soft-
wareasaService applications using web application testing“, 31th International Confe-
rence on Advanced Information Networking and Applications (AINA 2017)
[28] Kolosnjaji, Bojan, Apostolis Zarras, George Webster, and Claudia Eckert. Deep Lear-
ning for Classification of Malware System Call Sequences. In 29th Australasian Joint
Conference on Artificial Intelligence (AI), December 2016.
[29] Steffen Wagner and Claudia Eckert. „Policy-Based Implicit Attestation for Microkernel-
Based Virtualized Systems“. In: Information Security: 19th International Conference,
ISC 2016,Springer 2016, pp. 305–322.
[30] Lengyel, Tamas, Thomas Kittel, and Claudia Eckert. Virtual Machine Introspection with
Xen on ARM. In 2nd Workshop on Security in highly connected IT systems (SHCIS),
September 2015.
[31] Kilic, Fatih, Benedikt Geßele, and Hasan Ibne Akram. Security Testing over Encrypted
Channels on the ARM Platform. In Proceedings of the 12th International Conference on
Internet Monitoring and Protection (ICIMP 2017), 2017.
[32] Muntean, Paul, Vasantha Kommanapalli, Andreas Ibing, and Claudia Eckert. Automated
Generation of Buffer Overflows Quick Fixes using Symbolic Execution and SMT. In
International Conference on Computer Safety, Reliability & Security (SAFECOMP),
Delft, The Netherlands, September 2015. Springer LNCS.
[33] M. Huber, J. Horsch, J. Ali, S. Wessel, „Freeze & Crypt: Linux Kernel Support for Main
Memory Encryption“ ,14th International Conference on Security and Cryptography
(SECRYPT 2017).
16 Safety and Security 283
Summary
The more we become dependent on the functioning of complex technical
systems, the more important their resilience becomes: they need to maintain
the required system performance even when internal and external failures and
disruptions occur. This applies both to individual systems (e.g. cars, medical de-
vices, airplanes) as well as to infrastructure (traffic, supply systems, information
and communications systems). Designing these complex systems to be resilient
requires Resilience Engineering, that is, a process of maintaining critical func-
tions, ensuring a graceful degradation (in the case where the critical functionality
cannot be retained due to the severity of the disruption) and supporting the fast
recovery of complex systems. This necessitates generic capabilities as well as
adaptable and tailored technical solutions that protect the system in the case of
critical issues and unexpected or previously nonexistent events. Cascade effects
that occur in critical infrastructures during disruption, for example, may thus be
simulated and their effects proactively minimized.
17.1 Introduction
Technical systems that function safely and reliably are vital for our society. 250
years after the beginning of the Industrial Revolution, more than 70 years after the
advent of the first computer, and almost 30 years after the invention of the World
285
286 Stefan Hiermaier • Benjamin Scharte
Wide Web, there is hardly any field conceivable without technical systems. In es-
sence, these systems determine the day-to-day lives of everyone living in modern
industrial societies. This triumph is due to the fact that the systems make day-to-day
lives easier in a multitude of ways. In industry, they provide efficiency and quality
gains, resulting in better products and services. For recreational life, they make
activities possible that were previously either impossible or at least required signif-
icant expense, from long-distance travel to e-sports.
The increasing digitization of industry, work, and private life is an additional
and, in view of its far-reaching effects, even revolutionary step towards a world that
is completely dependent on technical systems. The various chapters of the present
book span themes such as additive manufacturing, digital factories, and individual-
ized mass production, from the question of the usefulness of artificial intelligence
for challenging situations, to e-government – digital citizen-oriented administration.
The authors highlight opportunities and possibilities arising from these various
developments in digitization. But they also shed light upon the specific challenges
related to the topics mentioned.
If we take a top-level view of the trends in different fields, namely from a systems
perspective, it becomes clear that there is a salient commonality: the success of these
developments depends on the ever increasing connectedness of previously separate
societal and technical fields. Irrespective of the various positive aspects that increas-
ing connectedness can entail, this also gives rise to challenges at the systems level.
Even individual separate systems are becoming increasingly complicated due to
their inherent intelligence and are well past the point of anyone but specialists un-
derstanding them. Even specialists are starting to reach their limits however. If
several complicated systems are connected together, then, with increasing frequen-
cy, more complex (technical) systems result.
It is important then, as we progress through digitization, to ensure that these
complex technical systems demonstrate the maximum possible fault tolerance both
during day-to-day operation, but also and above all in exceptional cases (in the case
of disruptions of any kind). Complexity in itself is neutral at first, neither good nor
bad. The same is true of interconnecting different systems. In a crisis, however,
complexity and connectedness may increase negative effects or give rise to out-
comes that were neither foreseen nor planned. For this reason, the traditional secu-
rity approach of classic scenario-based risk management is no longer sufficient. An
enhanced systemic concept is required with which to analyze, understand, and ulti-
mately, increase the fault tolerance of complex technical systems [13]. The present
article introduces the need for this kind of concept, the ideas behind it, and a con-
crete implementation of it in four sections. First, the challenges facing complex
technical systems are explained in greater detail. Next, the concept of resilience is
17 Fault-Tolerant Systems 287
The whole is greater than the sum of its parts. This is precisely the effect we observe
when dealing with complexity. Complicated systems consist of a variety of individ-
ual parts which are logically connected with one another. Providing an adequate
description of their behavior may be extremely difficult in some conditions. They
can be explained reductively, however. That is, the behavior of the system as a whole
can be deterministically identified by observing the causal relationships of the in-
dividual parts of the system to one another. Thus, the computing power available is
essentially the only factor in determining whether and how quickly system behavior
can be correctly predicted. A complex system, in contrast, cannot be explained on
the basis of its individual parts; it is simply more than the sum of its parts. Complex
systems are able to develop emergent properties, that is, properties that can only be
explained when observing the system holistically [8]. The boundaries between
complicated and complex systems are fluid and it is often difficult or even impos-
sible to decide whether an actual system can better be described properly using the
one or the other term. As previously mentioned, complexity makes it impossible to
understand systems using reductive tools, which represents a challenge to the fault
tolerance of these kinds of systems. Traditional risk analysis and risk management
make use of explicitly reductive principles ranging from clearly specified scenarios
through precisely defined probabilities to exact assessments of damages [15][20].
This is no longer sufficient for analyzing complex technical systems, which is why
there is an increasing reliance upon the concept of resilience.
A second challenge for fault-tolerant systems is susceptibility to cascade effects
resulting from the ever-increasing networking of different technical systems. The
term “cascade” usually refers to a sequence of levels or processes in the context of
specific events. A common albeit somewhat inaccurate example is the domino ef-
fect. When applied to disruptive events, cascades refer to scenarios where an initial,
primary disturbance event is followed by a series of secondary events, which can
themselves be viewed as a new disturbance event [17]. A dramatic example of cas-
cading effects is provided by the earthquake off the coast of the Tōhoku region of
Japan on March 11, 2011. Measuring 9.0 on the Richter scale, it remains the strong-
288 Stefan Hiermaier • Benjamin Scharte
Fig. 17.1 A typical example of serious cascade effects: the Tōhoku earthquake of March
11, 2011 (Fraunhofer EMI)
est quake in Japan since records began and one of the strongest earthquakes ever.
Fig. 17.1 shows the direct and indirect (cascading) effects caused by the earthquake.
It becomes clear that the majority of the damage was caused by a secondary event
caused by the earthquake itself: the tsunami. It was this tsunami that, in a further
cascade, led to the Fukushima nuclear catastrophe and, at least indirectly, forced the
German federal government at the time to change course towards a definitive phase-
out from atomic energy.
The developments illustrated above lead to increased vulnerability of complex
technical systems with respect to serious disruptive events. At the same time, the
number of these events is also rising. Climate change, for example, is leading to
increasingly extreme weather events. Terrorist attacks, too, are becoming more
frequent, using different methods to target completely different aims. Cyberattacks
on important computer networks and infrastructures are of particular relevance in
the context of digitization, of course, such as the WannaCry malware attack in May
2017 (see Fig. 17.2). Over 300,000 computers in more than 150 countries were af-
fected by this attack – the worst yet known – which was conducted with the aid of
ransomware, malicious software that effectively “kidnaps” affected computers,
demanding payment of specific sums for the release of the encrypted data. In Great
Britain, the software affected numerous hospitals, in some cases causing patients to
be turned away, ambulances to be diverted, and routine operations to be cancelled.
In Germany, advertising billboards and ticket machines belonging to the Deutsche
17 Fault-Tolerant Systems 289
Fig. 17.2 Screenshot of a computer infected with the WannaCry malware (Fraunhofer
EMI)
Bahn (German Railway) were affected, with video surveillance technology in train
stations also partially disrupted. And in France, the Renault automotive group had
to temporarily suspend production in a number of factories [7][21]. These impacts
on critical infrastructures and important areas of industry show how the networking
and coupling of different complex technical systems can increase vulnerability
thereof in the presence of disruptive events such as cyberattacks.
In view of the challenges and developments just described, it is extremely clear that
our society needs an adequate security concept to prevent the failure of critical
systems. Where systems do fail in spite of all efforts, further mechanisms need to
be in place that ensure the speediest possible recovery of the relevant functionalities.
In order to arm complex connected systems against disturbances that are external
290 Stefan Hiermaier • Benjamin Scharte
as well as internal, expected as well as unexpected, and which occur abruptly as well
as those which develop more slowly, a holistic view of system security is needed.
The discussion among the security research community, pertaining to how this kind
of systemic approach should look and how concrete solutions to increase the fault
tolerance of complex connected systems can thus be developed is centered around
the terms “resilience” and “resilience engineering”. “Resilience” in particular has
become a dominant term in security research in recent years.
A definition of resilience
In his book, Resilient Nation, from 2009, Charlie Edwards drew extensively on
classic disaster management cycles to provide a better understanding of the
far-reaching concept of resilience and give it concrete expression [4]. The resulting
resilience cycle may be slightly expanded to consist of the five phases of prepare,
prevent, protect, respond, and recover (see Fig. 17.4). The first step is to seriously
prepare for adverse events, especially in terms of early warning systems. Reducing
the underlying risk factors should then prevent the occurrence of the event itself as
far as possible. If it nevertheless does occur, it is important that physical and virtu-
al systems protect from and minimize the negative impacts function without error.
In addition, fast, well-organized and effective emergency assistance is required.
During this time, the system must be able to maintain its essential ability to function
as much as possible (respond). After the immediate period of damage has ended, it
is important that the system is in a position to recover and draw appropriate learning
from what has happened so that it is better equipped for future threats [22]. Based
292 Stefan Hiermaier • Benjamin Scharte
on this resilience cycle and the aforementioned aspects, resilience overall can be
defined as follows:
“Resilience is the ability to repel, prepare for, take into account, absorb, recov-
er from and adapt ever more successfully to actual or potential adverse events.
Those events are either catastrophes or processes of change with catastrophic
outcome which can have human, technical or natural causes.” [19]
technical systems exist in, plus their interactions with said environment. Notwith-
standing this, the engineering sciences are investigating how the resilience of com-
plex technical systems to disturbances can be developed and their fault tolerance
increased. In scientific circles, the term “Resilience Engineering” is currently gain-
ing prominence as a suitable term for describing the process of increasing resilience
with the aid of engineering solutions [2][23]. The term was coined in recent years
by researchers such as Erik Hollnagel and David Woods. According to them, the
focus of measures to increase security traditionally lay on protective concepts for
common threat scenarios. This is where Resilience Engineering, as understood by
Hollnagel and Woods, comes in. It is a question of including the possible and not
merely the probable in planning, implementation, and execution. New and unex-
pected threats in particular, since they may differ in extent from all of the scenarios
considered, provide systems with challenges that are able to be met with the aid of
Resilience Engineering [14][16][25]. Systems of all kinds need to demonstrate
sufficiently large and appropriate security margins [14]. This orientation around the
possible rather than the probable leads to a necessity for reorientation, namely away
from the damaged and towards the normal state. After all, everyday life shows that
things normally function as they should. It is unusual for something to go (serious-
ly) wrong. Even complex systems operate relatively smoothly under normal circum-
stances. Understanding the functioning of complex systems is the appropriate and
necessary requirement for identifying and minimizing potential faults, problems,
and risks for these systems [11]. This understanding of Resilience Engineering
teaches us several things about how complex systems can be designed to be resilient.
Nevertheless, Hollnagel, Woods, and their colleagues focused less on complex
technical systems and more on organizationally complex systems such as hospitals
or air traffic control. Thus, their ideas for Resilience Engineering need to be further
developed if they are to make a contribution to the fault tolerance of complex tech-
nical systems.
the different states of equilibrium that ensure a system’s survival. In the case that
critical functionality cannot be maintained due to the severity of the disruptive
event, “graceful degradation” must at least be ensured. That is, an abrupt collapse
of the entire functionality must be avoided, providing the system’s operators and
rescue services with sufficient time to make functional alternatives available. As
soon as the event is over, technical systems developed using Resilience Engineering
begin to recover from the effects. This fast recovery from damage does not only
incorporate bouncing back to the original state but also the implementation of learn-
ing drawn from the experience, and an adaptation to changed circumstances [23]. A
key component of Resilience Engineering is providing complex technical systems
with generic capabilities. This idea was borrowed from the concept of generic com-
petencies that allow people to successfully overcome adverse events, even unex-
pected or hitherto nonexistent events. For example, an experimental study was able
to demonstrate that exercising generic competencies, compared with strictly keep-
ing to precisely stipulated rules and procedures, can increase the success of a ship’s
crew when dealing with critical situations [1]. Unlike people, however, technical
systems do not possess the ability a priori to improvise or adapt flexibly to changing
situations. This requires Resilience Engineering to implement generic capabilities
as heuristics in technical systems. Examples of these kinds of heuristics are redun-
dancy, the availability of backups, foreseeability, complexity reduction, and other
properties [10][14]. It may for example specifically be a question of accelerating
research towards new methods for modeling and simulating complex systems that
are able to simulate and investigate the impacts of adverse events, in particular with
respect to cascade effects (see Ch. 17.4).
At the same time, however, Resilience Engineering also signifies the targeted
use of the latest innovative technologies for the design and utilization of complex
technical systems. These technologies need to be customized for specific systems
and specific tasks. The efficient use of customized technologies to optimize the
functionality of complex technical systems during normal operation is one option
for increasing the number of processes functioning seamlessly and thus pursuing
the kind of Resilience Engineering conceived by Hollnagel and Woods. Overall,
Resilience Engineering offers complex technical systems the opportunity to suc-
cessfully interact with both known problems (by means of customized technologies)
as well as with unexpected interruptions or even previously nonexistent crises
(thanks to generic capabilities). In summary, the concept can be defined as follows:
“Resilience Engineering means preserving critical functionality, ensuring
graceful degradation and enabling fast recovery of complex systems with the
help of engineered generic capabilities as well as customized technological
17 Fault-Tolerant Systems 295
In the previous sections, the terms “resilience” and “Resilience Engineering” were
defined, and we explained why, in view of existing developments and challenges,
the fault tolerance of complex technical systems can only be raised by making use
of these kinds of holistic concepts. Building on these ideas, the following section
introduces a specific application project that allows for the simulation and under-
standing of cascade effects in complex coupled network infrastructures. A tool for
designing and analyzing resilient technical systems should ideally possess a range
of capabilities. It must for example be able to model the physical components of the
system and their interactions, define target system performance, and compare actu-
al and target performance. The opportunity to feed load cases caused by specific
Fig. 17.6 Effects of a storm on the system of coupled network infrastructures in Helsinki
(Fraunhofer EMI)
disruptive events into the system as well as generic (that is, event-independent)
damage scenarios and to simulate their effects is also important. In doing so, it
should be possible to facilitate both identification of critical system components as
well as assessment of the fault tolerance of the system. Subsequently, measures to
increase resilience can be integrated and, with the aid of a new calculation, the re-
silience of the improved system evaluated and compared with that of the original
system. CaESAR1 is a software tool developed at Fraunhofer EMI for simulating
and analyzing coupled network infrastructures that demonstrate a large number of
these abilities.
CaESAR is designed to simulate cascading effects within and especially between
various coupled infrastructures. The first systems considered here are the energy
grid, water supply, and mobile telephony network. These networks are shown on an
overview dashboard as nodes and arcs on a georeferenced map. Fig. 17.5 shows an
example of the networks identified in the Finnish capital of Helsinki. CaESAR in-
cludes a “crisis editor” which is used to either implement specific disruptive events
17.5 Outlook
When the whole really is more than the sum of its parts – and in view of the various
complex systems which exist in our everyday lives, there is no doubt that this is the
case – we need a systematic view from above or from outside in order to understand
“the whole”. Due to the ongoing digitization of our society, more and more previ-
298 Stefan Hiermaier • Benjamin Scharte
ously separate fields are being connected with one another. In order to be able to
nevertheless guarantee the maximum possible fault tolerance of societally relevant
systems, a holistic security concept such as resilience is required. This may be im-
plemented with the help of a Resilience Engineering approach, giving rise to tools
such as the CaESAR software.
Nevertheless, engineering and technological implementation of resilience prin-
ciples is still in its relative infancy [2][23]. Many opportunities are currently avail-
able here to implement resilience right from the start during the development of new
technologies and in particular when they are widely used. One example would be
autonomous driving, where questions regarding the security and reliability of sys-
tems in many ways play a decisive role. Another would be the digital management
of a society’s critical infrastructures. Here, too, it is important to ensure that securi-
ty aspects are taken seriously and integrated into systems as a matter of course in
the face of increasing automation and networking. At the same time, however, any
resulting potential risks need to be weighed up carefully. Here, too, the concept of
resilience provides excellent opportunities with its holistic approach. Analysis of
the key themes here is provided by the article in this volume on data security as a
prerequisite for digitization.
In summary, we can see that in the area of the resilience of complex technical
systems, a range of open (research) questions remain that in future need to be more
deeply engaged with academically by both engineering and technology as well as
by the natural and social sciences. Intensive work is already being carried out by
Fraunhofer-Gesellschaft and Fraunhofer EMI in particular, on innovative solutions
to increase the fault tolerance of complex technical systems.
[1] Bergström J, Dahlström N, van Winsen R, Lützhöft M, Dekker S, Nyce J (2009): Rule-
and role retreat: An empirical study of procedures and resilience. In: Journal of Maritime
Studies 6:1, S. 75–90
[2] Bruno M (2015): A Foresight Review of Resilience Engineering. Designing for the Ex-
pected and Unexpected. Aconsultation document. Lloyd’s Register Foundation, London
[3] DUDEN (2017): Resilienz. URL: http://www.duden.de/rechtschreibung/Resilienz
[Stand: 08.06.2017]
[4] Edwards C (2009): Resilient Nation, London: Demos
[5] Flynn S (2011): A National Security Perpesctive on Resilience. In: Resilience: Interdis-
ciplinary Perspectives on Science and Humanitarianism, 2, S. i-ii
[6] Goerger S, Madni A, Eslinger O (2014): Engineered Resilient Systems: A DoD Perspec-
tive. In: Procedia Computer Science, 28, S. 865–872
17 Fault-Tolerant Systems 299
[7] heise.de (2017): WannaCry: Was wir bisher über die Ransomware-Attacke wissen .
Url: https://www.heise.de/newsticker/meldung/WannaCry-Was-wir-bisher-ueber-die-
Ransomware-Attacke-wissen-3713502.html [Stand: 08.06.2017]
[8] Holland J (2014): Complexity. A Very Short Introduction (Very Short Introductions).
Oxford University Press, Oxford
[9] Holling C (1973): Resilience and Stability of Ecological Systems. In: Annual Review of
Ecology and Systematics, 4, S. 1-23
[10] Hollnagel E, Fujita Y (2013): The Fukushima disaster – systemic failure as the lack of
resilience. In: Nuclear Engineering and Technology, 45:1, S. 13-20
[11] Hollnagel E (2011): Prologue: The Scope of Resilience Engineering. In: Hollnagel E, Pa-
riès J, Woods D, Wreathall J (Hrsg.): Resilience Engineering in Practice. A Guidebook,
Farnham, Surrey: Ashgate, S. xxix-xxxix
[12] Kaufmann S, Blum S (2012): Governing (In)Security: The Rise of Resilience. In: Gander
H, Perron W, Poscher R, Riescher G, Würtenberger T (Hrsg.): Resilienz in der offenen
Gesellschaft. Symposium des Centre for Security and Society, Baden-Baden: Nomos,
S. 235-257
[13] Linkov I, Kröger W, Renn O, Scharte B et al. (2014): Risking Resilience: Changing the
Resilience Paradigm. Commentary to Nature Climate Change 4:6, S. 407–409
[14] Madni A, Jackson S (2009): Towards a Conceptual Framework for Resilience Enginee-
ring. In: IEEE Systems Journal, 32:2, S. 181-191
[15] Narzisi G, Mincer J, Smith S, Mishra B (2007): Resilience in the Face of Disaster: Ac-
counting for Varying Disaster Magnitudes, Resource Topologies, and (Sub)Population
Distributions in the PLAN C Emergency Planning Tool
[16] Nemeth C (2008): Resilience Engineering: The Birth of a Notion. In: Hollnagel E, Ne-
meth C, Dekker S (Hrsg.): Resilience Engineering Perspectives. Volume 1: Remaining
Sensitive to the Possibility of Failure. Farnham, Surrey: Ashgate, S. 3-9
[17] Pescaroli G, Alexander D (2015): A definition of cascading disasters and cascading
effects: Going beyond the „toppling dominos“ metaphor. In: GRF Davos Planet@Risk,
3:1, Special Issue on the 5th IDRC Davos 2014, S. 58-67
[18] Plodinec M (2009): Definitions of Resilience: An Analysis, Community and Regional
Resilience Institute
[19] Scharte B, Hiller D, Leismann T, Thoma K (2014): Einleitung. In: Thoma K (Hrsg.):
Resilien Tech. Resilience by Design: Strategie für die technologischen Zukunftsthemen
(acatech STUDIE). München: Herbert Utz Verlag, S. 9-18
[20] Steen R, Aven T (2011): A risk perspective suitable for resilience engineering. In: Safety
Science 49, S. 292-297
[21] telegraph.co.uk (2017): NHS cyber attack: Everything you need to know about ‚biggest
ransomware‘ offensive in history. Url: http://www.telegraph.co.uk/news/2017/05/13/
nhs-cyber-attack-everything-need-know-biggest-ransomware-offensive/ [Stand:
08.06.2017]
[22] The National Academies (2012): Disaster Resilience. A National Imperative, Washing-
ton, D.C.
[23] Thoma K, Scharte B, Hiller D, Leismann T (2016): Resilience Engineering as Part of
Security Research: Definitions, Concepts and Science Approaches. In: European Journal
for Security Research, 1:1, S. 3-19
[24] Wildavsky A (1988): Searching for Safety (Studies in social philosophy & policy, no.
10), Piscataway: Transaction Publishers
300 Stefan Hiermaier • Benjamin Scharte
[25] Woods D (2003): Creating Foresight: How Resilience Engineering Can Transform
NASA’s Approach to Risky Decision Making. Testimony on The Future of NASA for
Committee on Commerce Science and Transportation
Blockchain
Reliable Transactions
18
Prof. Dr. Wolfgang Prinz · Prof. Dr. Thomas Rose ·
Thomas Osterland · Clemens Putschli
Fraunhofer Institute for Applied Information Technology FIT
Summary
Blockchain technology has major relevance for the digitization of services and
processes in many different areas of application beyond the financial industry
and independent of cryptocurrencies in particular. Whilst for the Internet of
Things the potential for automation associated with smart contracts is especially
significant, for applications from the supply chain fields or for proofs of origin
it is the irreversibility of the transactions conducted. This article describes the
functioning of this new technology and the most important resulting qualities.
The chapter provides a list of criteria for identifying digitization projects for
which blockchain technology is suitable.
Because of the extent of blockchain technologies and their applications, develo-
ping the basic technologies requires a multidisciplinary approach, as does deve-
loping applications, carrying out studies of cost-effectiveness, and designing new
governance models. The diverse competencies offered by the various Fraunhofer
institutes, put the Fraunhofer-Gesellschaft in a position to make a significant
contribution to the ongoing development and application of blockchain techno-
logy.
301
302 Wolfgang Prinz • Thomas Rose • Thomas Osterland • Clemens Putschli
18.1 Introduction
Trust and reliability are the critical key elements for the digitization of business
processes, whether they take place between sales portals and customers or as in-
ter-organizational processes between business partners working together within
supply chains. While reputation management methods have attempted to use trans-
actional analyses to support seller confidence in business-to-consumer and consum-
er-to-consumer relationships, today, in an Internet of Value, the question of confi-
dence in transactions that depict different kinds of values immediately arises. Data-
bases and process management have traditionally always pursued a centralized
approach here, beginning with a nominated authority and central process synchro-
nization. This centralization nevertheless entails a range of potential risks. These
include for example performance bottlenecks, fault tolerance, authenticity, or inter-
nal and external attacks on integrity.
In the case of cryptocurrencies on the other hand, the central clearing function
of banks is replaced by shared algorithms for ensuring correctness in the network.
The central innovation of cryptocurrencies is thus guaranteeing the correctness of
transactions within a network and also shared consensus finding between partners
in the network. Consensus on the correctness of transactions and business process-
es is not managed centrally, but it is developed through shared consensus finding
between the partners.
Since the publication of Satoshi Nakamoto’s white paper in 2008 [3] and the
creation of the first bitcoins in early 2009, both cryptocurrencies and blockchain
technology have received ever more attention in the last two years. The reasons for
this are the following features of the technology:
• Documents and investments can be uniformly encoded in a forgery-proof way
and the transfer between senders and recipients can be stored as a transaction in
the blockchain.
• The storage of the transactions is irreversible and transparent based on distribut-
ed consensus building and encryption.
• Transactions are verified within a peer-to-peer (P2P) network rather than by a
central authority.
• Smart contracts offer the possibility of describing and executing complex trans-
actions and assuring their boundary conditions. They enable both the automation
of simple processes in the Internet of Things and new governance models by
establishing alternative organizational forms.
after the Internet of Things. In what follows, the present chapter first describes the
functioning of blockchain before illustrating the applications. An in-depth account
of the technology and its application is provided in [8].
18.2 Functioning
blockchain. Transactions reside in blocks and all types of transactions are convert-
ed into a standardized format using hash functions. To do this, all the individual
transactions are encoded into hash values and then compressed hierarchically. This
hierarchical compression is known as a hash tree or Merkle tree, which allows a
block of transactions to be represented unambiguously. This encoding is secure
against attempts at manipulation since changing even one transaction would
change the hash value of the block and the hash tree would thus no longer be
consistent.
Blocks are linked to preexisting blocks via concatenation to produce a (block)
chain. For a block to be accepted as a new element into the existing concatenation,
a process of consensus building as described in section 18.3 must take place, result-
ing in a correct and irreversible concatenation of blocks to form a blockchain. To
ensure persistence, these chains are replicated in all nodes of the network; that is,
all nodes have the same basic knowledge.
Blockchains can thus be described in a simplified form as distributed databases
that are organized by the participants in the network. In contrast to centralized ap-
proaches, blockchains are far less prone to error. Nevertheless, these systems also
entail various challenges. At present, critical discussions focus in particular on the
high data redundancy, since holding multiple copies of the same data within the
network requires a very large amount of storage space.
who perform work on behalf of the community are permitted to also use the com-
munity’s resources. In the blockchain context, the users are the miners who perform
the difficult task of computing the proof-of-work, and the providers are all of the
nodes who carry out relatively straightforward checks to see whether the successful
miner has computed the proof-of-work properly. In the bitcoin blockchain, the
proof-of-work algorithm is based on the hashcash process presented by Adam Back
[1]. The goal of the algorithm is to find a number (nonce, number used only once)
that, when combined with the new block to be attached to the existing blockchain,
produces a hash value that fulfills a specific condition. One example condition is
that the value to be identified must consist of a specific number of leading zeros.
This number can only be identified by means of extensive trial and error as hash
functions are one-way functions.
In this particular proof-of-work process, the computing power of the nodes is a
significant factor in who solves the problem and identifies a suitable nonce value.
Since miners are rewarded with new bitcoins for identifying the nonce, a competi-
tion is created among miners where they invest in ever-increasing computing pow-
er. This would decrease the time required to find a valid nonce. However, this is
inconsistent with the bitcoin network’s rule that a new block should only be gener-
ated approximately every 10 minutes, which is related to the fact that successful
miners are rewarded for “freshly minted” bitcoins. If the intervals between new
blocks being created were shortened, then the total amount of money in circulation
would grow too quickly. For this reason, the difficulty level of the puzzle is in-
creased whenever the period of time is shortened by the addition of new processing
capacity. For miners operating the computing nodes, this means increased work with
decreased prospects of success.
Since the work involved consists primarily of the energy used, alongside the
investment in computing power, the proof-of-work approach does not make sense
for all blockchain applications. This is particularly the case for applications where
this kind of competition is unnecessary. Therefore, alternative proof-of-work pro-
cesses were developed that are either memory or network-based. In the case of
memory-based approaches, the puzzle is solved not by computing power but by a
corresponding number of memory accesses [1], whereas in the network-based ap-
proach, by contrast, it is solved only by communication with other network nodes
(e.g. to gather information from them that is required to solve the puzzle) [9]. The
proof-of-work process makes sense when access to the blockchain network is pub-
lic and not subject to any access restrictions. An alternative process, primarily rel-
evant for private blockchains (see Ch. 18.4), where the nodes involved are known
and subject to access restrictions, is the proof-of-stake process. Here, nodes that are
able to validate a new block are selected according to their shares in the cryptocur-
306 Wolfgang Prinz • Thomas Rose • Thomas Osterland • Clemens Putschli
rency [5] or via a random process [7]. The selection of the most suitable process is
dependent on the specific use case and the blockchain solution used. An additional
important aspect is scalability for transaction volumes, especially in the case of
applications in the Internet of Things. Current approaches are not able to compete
with databases with respect to their frequency of transactions. This aspect, as well
as the fact that in a blockchain all of the data stored is replicated in each node, mean
that blockchain solutions initially cannot be used for data storage in the same way
that databases can. They take on special tasks in combination with databases when
the focus is on managing information reliably, with common agreement on the
transactions to be recognized as valid. In addition, databases store the payload,
while a fingerprint of the data is filed in the relevant blockchain to guarantee integ-
rity.
pating users. For validation, authorized individuals normally carry out a simplified
consensus process (e.g. proof of stake) that may be far more efficient.
Blockchain implementations may additionally be differentiated by the extent to
which they are oriented as a platform around solving logical problems, or whether
they are rather designed for “traditional” cryptocurrencies. On some blockchain
implementations such as Ethereum or Hyperledger Burrow, Turing complete smart
contracts can be executed, for example, that is, smart contracts may be complex
programs instead of simply conditional transactions. Fig. 18.2 contrasts the various
classification properties and lists a number of important blockchain implementa-
tions for comparison.
18.5 Applications
The terms “smart contract” or “chaincode” refer to programs that are executed
within the blockchain network. Once smart contracts are saved and instantiated in
the blockchain, they are immutable, and the execution of the processes defined in
the program code is independent of external entities. A smart contract can become
active via an external event or a user interaction.
The immutability of instantiated smart contracts and the additional option of
modeling complex processes permit the reliable handling of transactions between
various entities. The cryptographically secured execution of smart contracts in the
blockchain means they are not only beneficial for carrying out a defined process but
they also simultaneously document the process itself.
308 Wolfgang Prinz • Thomas Rose • Thomas Osterland • Clemens Putschli
ation partners without a stable and secure basis for transactions and trust. In these
cases, a blockchain can guarantee networked integrity.
4. Transferring value and protecting rights: blockchains facilitate the transfer of
value and rights. Thus, all processes are relevant where original copies, proofs
of origin, or rights need to be conveyed or transferred.
In addition to these criteria, it is important that the focus should not lie on the use
of a cryptocurrency itself and that no processes that are subject to strict regulation
should be selected for an initial assessment of the technology and the development
of demonstrators.
Summary
While the digital transformation is well underway in numerous areas of society,
medicine still faces immense challenges. Nevertheless, the potential resulting
from the interaction of modern biotechnology and information technology is
huge. Initial signs of the transformation can be seen in numerous places – a
transformation that will further be accelerated by the integration of previously
separate medical data silos and the focused use of new technologies. In this
chapter, we describe the current state of integrated diagnostics and the mecha-
nisms of action behind the emerging field of digital healthcare. One of the areas
of focus is the recent revolution caused by artificial intelligence. At the same
time, we have seen the emancipation of patients who now have access to an
enormous breadth of medical knowledge via social networks, Internet search en-
gines, and healthcare guides and apps. Against this backdrop, we will discuss the
change in the doctor-patient relationship as well as the changing roles of doctors
and computers, and the resulting business models.
19.1 Introduction
The digital transformation that is currently the topic of discussion in every segment
of the market and technology has barely taken place in the field of healthcare. Si-
multaneously, we are observing a rapid increase in complexity right across the
medical specialties that for years now has been stretching the limits of what is fea-
311
312 Horst Hahn • Andreas Schreiber
sible for those working in them. The following commentary is based on the hypoth-
esis that digital transformation is the key to raising healthcare to the next level in
terms of success rates, security, and cost-efficiency, while simultaneously realizing
the potential of modern medicine. The distribution of responsibilities within the
medical disciplines is just as much under scrutiny as remuneration mechanisms for
enhanced healthcare services, applicable quality standards, medical training curric-
ula, and, last but not least, the role of empowered patients.
One of the preconditions for this transformation, the digitization of all relevant
data, is already at an advanced stage: most patient information collected in this
country is already available in digital form. An exception is clinical pathology where
tissue samples generally continue to be assessed manually under optical micro-
scopes. At the other end of the spectrum, process automation in laboratory medicine
has long been part of the status quo. What remains neglected, however, is the con-
nection of the individual sectors as well as the structured use of integrated informa-
tion.
The digital transformation is supported by the interplay of several apparently
independent technologies that have become remarkably powerful in recent years
(cf. Fig. 19.1). On the one hand these are developments outside of medical technol-
ogy: the sheer computing power and storage capacity of modern computers of all
sizes with constantly increasing network bandwidth and cloud computing, as well
as far-reaching connectivity via the Internet, mobile devices, and not least social
media and artificial intelligence. On the other hand these are achievements in bio-
technology, laboratory automation, microsensors and imaging, as well as, very of-
ten, the findings of basic medical research. In a matter of just ten years, it has been
possible to reduce the cost of sequencing an entire genome by a factor of six to just
a few hundred Euros. Eric Topol, in his book The Creative Destruction of Medicine
[26], describes these simultaneous developments as a “super convergence” from
which the new healthcare arises.
In what follows, we describe the mechanisms of action of this digital healthcare
at its various stages from prevention and early diagnosis through to clinical treat-
ment. Thus it is possible, even now, to see numerous cases of integrated diagnostics
with new business models. We then discuss the revolution in artificial intelligence
being observed across every field (including healthcare), and the changing roles of
doctors and patients and of the different medical specialties. From a higher-level
perspective we also discuss the health-economic potential of digital medicine and
the changing industry landscape, where the battle for sovereignty over data integra-
tion and data access is increasingly emerging. The last section provides a brief
outlook, touching on additional and related subjects that are treated elsewhere due
to the brevity of this present text.
In the first decades of the 21st century, digital transformation is already in full swing
and is exercising far-reaching influence on numerous areas of society. Smartphones
and ubiquitous mobile access to the Internet in particular have changed the way we
communicate, work, access information, teach and learn, and – very importantly
– how, what, and where we consume. In the process, the gradual digitization of
everyday life has increased the efficiency of many processes and democratized
access to information and knowledge with the aid of participative platforms. Most
of the time, this has strengthened the role of the consumer thanks to increased
transparency. At the same time, however, the “data power” of large corporations is
growing, along with the risk that data will be manipulated or utilized to the disad-
vantage of the user.
314 Horst Hahn • Andreas Schreiber
Digitization presents society and its stakeholders with key challenges, since it
accelerates institutional change and demands a high degree of flexibility from
everyone involved. An example with numerous similarities to medicine would be
the media landscape, where digitization gave rise to new distribution channels and
empowered consumers, forcing a new market order into being that redefined the
role of established media producers. The resulting new diversity of media and fast-
er publication cycles, however, also increase the chances of manipulation and may
make it more difficult to receive high-quality information.
The disruption caused by digitization within manufacturing is no less far reach-
ing and extends to the complete redesigning of industrial value chains. Frequently
referred to as “Industry 4.0”, the digital transformation of production aims to
achieve a high degree of process automation together with the autonomous sharing
of data between machines in order to guarantee seamless processes across manufac-
turing and supply chains.
While the products and services we consume in the digital age are increasingly
tailored, healthcare is still dominated by the one-size-fits-all principle, closely allied
to the principles of evidence-based medicine. Already today, the majority of medical
data is captured digitally and, alongside record-taking and patient management, also
selectively used to support clinical decision-making such as diagnosis and therapy
selection.
Nevertheless, the true potential of broad computer-assisted medicine is going to
waste since data is at times documented and saved incompletely and in an insuffi-
ciently standardized format, and a lack of suitable interfaces as well as outdated
legal contexts prevent centralized storage in the form of electronic patient records.
The latter is a necessary precondition both for future support in personalized patient
decision-making as well as for the preliminary assessment of large medically curat-
ed databases for training self-learning algorithms [11].
This is where the medicine of the future will come to our aid. Alongside advances
in genetics and molecular biology as well as in diagnostic and interventional med-
ical technology, digitization is a pivotal foundation for future “personalized medi-
cine”, also known as “precision medicine”. Here, intelligent software solutions
function as integrators of all the relevant patient information and fields of care and
are the keys to integrated predictive diagnostics and therapy planning (cf. Fig. 19.2).
In future, personalized medicine will be based on individual patient risk assess-
ment in view of family history, genetic predisposition as well as socioeconomic and
19 E-Health 315
Against this backdrop the motivation also arises to incorporate the systematic and
long-term recording of environmental factors such as fine particulate air pollution
or the geographic distribution of viral infections into personal healthcare provision
and therapy planning. Without computer assistance, these complex fields of data
and knowledge cannot be exploited by doctors, since intelligent analysis systems
are required that merge regular knowledge with statistical models and are thus able
to generate individualized patient risk assessments.
Several projects were commenced in recent years that seek to gather comprehen-
sive health data (including imaging) from right across the population and make it
accessible for disease risk investigation. The goal in each case is to analyze the
health profiles of subjects over the long term in order to identify and interpret early
316 Horst Hahn • Andreas Schreiber
indications and reasons for diverse ailments. The most recent of these is the German
National Cohort (GNC), with 200,000 planned participants, of whom around one in
seven even undergoes an extensive whole-body MRI (see Fig. 19.3). Other similar
initiatives are the UK Biobank with 500,000 participants and a US study conducted
as part of the Precision Medicine Initiative. Also of particular interest here is Project
Baseline, a joint endeavor of Stanford and Duke universities which, alongside an-
nual examinations, is also collecting data via passive sleep sensors and smartwatch-
es.
ies to detect tumor DNA circulating freely in the blood. Proteome and metabolome
are also becoming of increasing interest to medicine since they are credited with
significance for overall health and the development of various diseases.
In future, these and other analyses will take place unnoticed alongside tradition-
al blood tests. Researchers at Stanford University’s Gambhir Lab, for example, are
working on lavatory-based detectors that automatically test stool samples for path-
ogens and indications of diseases [8]. And in the near future, we will be able to
carry out a range of in vitro tests in our emerging smart homes, even though still
today these tests require costly laboratory infrastructure. It also remains to be seen
how long it will take for the analysis of behaviors, speech, gestures, and facial ex-
pressions to contribute to the early detection of affective and other psychotic con-
ditions. Duke University’s Autism & Beyond1 research group, for example, promis-
es to detect signs of autism in early childhood based on automated video diagnostics.
Medical imaging will continue to play a significant role in personalized early
detection. Although large-scale screening programs, e.g. for the early detection of
breast cancer, are being vigorously debated today, imaging remains a procedure
with high specificity. In future, we will need to further stratify the patient popula-
tions benefiting from imaging and in each case employ the best possible imaging
workup to significantly minimize false positive results and side effects. Magnetic
resonance imaging and ultrasound in particular appear to be increasingly used in
screening.
Whereas MRIs will become even faster and more cost-efficient in the coming
years, sonography is benefitting from hardware and software advances that allow
for high-resolution, spatially differentiated and quantifiable diagnosis. The fact that
neither modality emits ionizing radiation represents a key safety advantage over
computed tomography and traditional projection radiography, despite successes in
radiation dosage reduction.
The personalized health monitoring expects that individuals increasingly assume
personal responsibility for their health and develop an awareness of the key lifestyle
factors. App-based guides on mobile devices contribute towards this by developing
tailored recommendations for nutrition and sporting activities. Apps such as Cara
for gastrointestinal complaints, mySugr for diabetes, and M-sense for migraines are
all designed to assist personal health management.
In case of acute conditions, we expect an even closer linking of monitoring and care
with the traditional diagnostic processes, the effectiveness of which will also con-
tinue to increase in the coming years. Results from laboratory diagnostics and im-
aging are in most cases already available in digital form today, and there is a trend
towards connecting those data silos, often currently still separated, via appropriate
interfaces. In ultrasound diagnostics, too, where analysis is still primarily carried
out live at the examination venue, an increasingly standardized image acquisition
process and centralized storage will form the basis in future. In the USA, storage
and appraisal often take place separately in the division of labor of sonographers
and radiologists.
Clinical pathology, where tissue samples are examined and analyzed under the
microscope, still operates the conventional way. Alongside traditional histological
assessments, modern immunohistochemistry and molecular pathology processes
yield more specific but also complex patterns. In future, slide scanning and consist-
ent digital characterization of tissue samples will play a significant role in this rich
information being processed in everyday clinical practice.
Virtual multiple staining, for example, allows several specific serial sections to
be combined so the overall tissue information contained can subsequently be ana-
lyzed automatically or semi-automatically (see Fig. 19.4). Examination under the
microscope does not even permit the visual precision correlation of two stains
Fig. 19.4 Left: virtual multiple staining using high-accuracy serial section image registra-
tion. Right: automatic analysis of lymphocytes in breast tissue images (Fraunhofer MEVIS)
19 E-Health 319
up
w-
llo
Fo n
io
Les king
a c
Tr
ast
Bre cer s g&
n tic
Ca nos nin f
lan rt o
ica
l Dia
g
Int P po
p
Su or
uro
log er Tum tion
Ne usion s ve Ab
la
rf
Pe nos
tic nt
ive
Dia
g ion PD tics
&
vas of al CO nos n
g tio
n-in ion s
No lorat amic Ra Dia rven
n y dio e
Int ning
log
Exp ody n
He
m
log Pla
-ba
sed
co y
MR rt
a
He nos
tics On
g
Dia ng
Lu ery
rg
Su ning
n
Pla
Di
ag
n y
Raost er
dioic rg
log Su
y
tive
tita l Ra
an ica
Qu rolog s dio gy
olo
u tic
Ne nos th
Dia
g
er
ap ath er
Liv ery &
y P rg
Su ning rt
n o
Pla Supp
al OR
gic
er a l sur
Liv ction uro tion
n Ne rven
Fu lysis e
Int ning
a
An Pla
n
al
od
ltim n
Mu iatio dic
s
d
Ra rapy pe
tho
e
Th ning Or
n
Pla l
ita y
Dig olog
th
Pa
Fig 19.6 Imaging plays a central role in nearly every clinical discipline and contributes
significantly to high-precision, integrated care. (Fraunhofer MEVIS)
320 Horst Hahn • Andreas Schreiber
(Fig. 19.4 ii). With subsequent generations of technology, the persistent deficits in
image quality and time required are expected to give way to an immense increase
in objectivity and productivity, particularly in the case of recording quantitative
parameters. We expect, as with the digitization of radiology, that pathology will
entirely adopt digital processes within just a few device generations and experience
a standardization of methods and reporting.
Nevertheless, the true catalyst of diagnosis is information technology that for
the first time allows us to consider diagnosis based on integrated data right across
disciplines. By “integrated diagnostics” we mean using software to bring together
all of the diagnostically-relevant information – from the laboratory, radiology,
pathology, or the individual health and case files – to allow statistical comparison
of the biomarker profile and, finally, intelligent differential diagnosis and deci-
sion-making support. The “digital twin” that results is not to be understood literal-
ly but is based on precisely this information integration and permits the predictive
modeling of potential courses of disease and the probabilistic prioritization of
therapy options.
Imaging data here plays an important role in phenotyping, planning of interven-
tions, and in detailed therapeutic monitoring (cf. Fig. 19.7). Innovative approaches
also permit the direct utilization of 3D planning data in the operating room. The
increasing dissemination of robotics in surgery and interventional radiology will
also further boost the significance of intraoperative imaging, because mechatronic
assistance systems achieve their full potential in particular in combination with
precise real-time navigation and corresponding simulation models (cf. Fig. 19.5 and
19.6)
The transformation towards highly efficient digital medicine will only occur when
it is possible to analyze and interpret the exponentially growing data volumes effi-
ciently. Artificial intelligence (AI) and machine learning methods – which have
been revolutionized in just these last few years – thus play a crucial role. In educat-
ed specialist circles and beyond, deep learning is on everyone’s lips. Dismissed as
a hype by some, there is nevertheless extensive agreement that a level of maturity
has been achieved that allows the most complex practical problems to be effective-
ly solved.
19 E-Health 321
AlphaGo’s victory in early 2016 [20] over the world’s best Go players is just
one example of how the tide has turned in a short space of time. Just a few years
ago, the common belief was that computers would require several decades before
they could play Go at the level of a grand master – if ever, in view of the enor-
mous diversity of variations. The theoretical total number of possible board po-
sitions is a figure with more than 170 digits. Even if we deduct the unrealistic
positions (a large proportion) from this, we are still left with a figure that makes
the approximate number of atoms in the universe (around 1081) appear vanish-
ingly small.
Public AI consciousness was boosted in 2011 when IBM’s Watson computer beat
the best players ever at the time, Ken Jennings and Brad Rutter [10] by an immense
distance, in the quiz show Jeopardy! Technically, Watson is just a huge database
with a search engine that not only recognizes logical but also meaningful connec-
tions in the data, and a so-called natural language processor that is able to under-
stand our spoken or written language. At the time, the Watson database comprised
around 200 million pages of text and tables – the system built on it could thus answer
the most commonly posed questions reliably even without an Internet connection.
The understanding of language alone has undergone a revolution due to deep learn-
ing, and today Alexa, Siri, Baidu et al. along with the latest translation machines
understand human language nearly as well as humans [17].
More relevant, perhaps, for medical applications are the successes of so-called
“convolutional neural networks” (CNNs), a special variant of deep neural networks
where the individual weight factors of multilayered linear convolution operations
are learnt from the training data. As a result, the network adapts itself to those visual
features that have the most significance for a given problem. In this context, the
network’s “depth” refers to the number of layers.
The breakthrough took place in 2012 in the context of the annual ImageNet
competition2, based on the eponymous image database of more than 14 million
photographs corresponding to over 21,000 terms. The goal of the competition is to
identify the most suitable term for a random image selection. Up until 2011, the
accuracy of automated computer systems had largely plateaued at an error rate of
over 25%, far inferior to the approx. 5% of the best human experts. Once Alex
Krizhevsky had used CNNs in his AlexNet to achieve a huge leap in accuracy in
2012, all of the top ten places were taken over by deep learning/CNN approaches
already by 2013. The next leap took place in 2015 when a team at Microsoft in-
creased the network depth from 22 to 152 layers, thus breaking the barrier of human
competence for the first time with an accuracy level of 3.7%.
The ideas behind deep learning as an extension of artificial neural networks
(ANNs) have been around for several decades now, and their successful application
in medicine was described more than 20 years ago [18, 28]. However, ANN research
grew quiet in the late 1990s, unable to keep pace with the expectations awakened,
in the face of simpler and more readily understood classification processes. The
breakthrough of the recent years took place due to the exponential growth in com-
puting power and in particular due to the use of graphics processors for solving the
kinds of complex numerical optimization problems arising from deep learning and
CNNs. Now, the first places in comparative medical image analysis competitions
are thus almost completely occupied by the various CNN variants.3
In what could almost be described as a kind of gold rush, the newly discovered
methods are being targeted at practically every problem within medical data analy-
sis, and corresponding startups are springing up all over the place [4]. If the niche
field had not caught their attention before, it was probably the $1 billion takeover
by IBM of Merge Healthcare in summer 2015 that finally moved AI right up the
CEO agenda of major medical technology groups worldwide.
At this point in time, integrated diagnosis shows most promise in radiomics appli-
cations. It combines phenotyping based on a large number of image-based quanti-
tative parameters with the results of genome sequencing, laboratory findings, and
(in future) even multisensory data from wearables and other sources. The goal is not
only the simple detection of patterns that humans, too, would be able to recognize
in the data, but the prediction of clinically relevant parameters such as drug therapy
response [14]. Overall, radiomics is the vehicle for providing machine learning and
integrated diagnostics with concrete demonstrable significance for solving complex
clinical problems.
Thus, in the case of cancer therapy, for example, the combination of radiologi-
cally identified tumor heterogeneity (cf. Fig. 19.7 and 19.8) and specific laborato-
ry parameters or results from molecular diagnostics could be decisive in discontin-
3 COMIC – Consortium for Open Medical Image Computing, Grand Challenges in Biomed-
ical Image Analysis: https://grand-challenge.org/
19 E-Health 323
Fig. 19.7 Image-based phenotyping using computed tomography scans of six different
lung tumors (Fraunhofer MEVIS, all rights reserved. CT data: S. Schönberg & T. Henzler,
Mannheim)
uing a therapy or starting a new one [2]. Ultimately, software needs to be able to
prioritize the huge volume of patient health information in specific individual
cases and make it useable for clinical decision-making. Patient selection and pre-
cise choice of therapy are also especially essential for highly specific immune
therapy, a form of treatment which has already demonstrated impressive results in
recent years [3].
The AI techniques described above will be a great help in comparing individ-
ual data with population-specific databases. And more and more examples of di-
agnostic or prognostic tasks are coming to light where computers and people are
meanwhile at least on a par with one another. One of these is the visual assessment
of skin cancer based on high-resolution photographs. At the beginning of 2017,
for example, researchers at Stanford University were able to demonstrate how a
CNN trained on 129,450 images achieved outcomes similar to 21 certified derma-
tologists when classifying the most frequently occurring varieties of skin cancer
[9].
324 Horst Hahn • Andreas Schreiber
Fig. 19.8 Deep learning segments the liver and liver lesions. Left: downstream classifiers
sort out false positive findings (dotted line), CT data: LiTS Challenge. Right: deep learning
distinguishes tumors (striped) from cysts (Fraunhofer MEVIS, all rights reserved. CT data:
R. Brüning, Hamburg).
19 E-Health 325
man gut feeling, also often likely learnt through experience but equally hard to
put into words.
And herein lies a key difference between people and computers, since trained
CNNs can be comprehensively described and statistically validated, a feature that
plays a significant role in authorization for use as a medical product. For the wider
adoption of AI-based assistants it will also be essential that their inner workings are
made comprehensible as much as possible and that remaining errors are analyzed
thoroughly.
Time and again, we are likely to see that people and machines have greatly dif-
fering sources of error as well as strengths, in keeping with their very different
learning and behavioral mechanisms. The correct design of user interfaces will thus
also be key to the successful optimization of human/computer teams.
With the spread of specialist digital systems, large databases, and integrated diag-
nostic solutions, the distribution of roles between the medical specialties will change
as well. We can assume that the extensive separation of working processes in radi-
ology, pathology, and other diagnostic disciplines will be combined into an equally
integrated cross-disciplinary working process.
The decisions made by tumor boards organized in numerous locations today,
based on the findings of individual specialties, will in future be made right from the
start by interdisciplinary diagnostic teams with extensive computer assistance, in a
highly efficient and personalized manner. In the ideal world, all relevant information
would flow together to these teams, which would represent a kind of control center
for the medical decision-making process. As a result, seamless cooperation with the
relevant neurologists, surgeons, cardiologists, oncologists, radiotherapists, gynecol-
ogists, urologists, etc. will take on an even greater significance than at present.
Connectivity will thus not only become a key competitive factor with respect to
data, but also with respect to the players involved.
326 Horst Hahn • Andreas Schreiber
A key motivator for the transformation of medicine lies in the necessity of reducing
the costs of healthcare provision. Industrialized nations use between 9% and 12%
of their gross domestic product for covering healthcare costs. With healthcare
spending equivalent to 17.8% of the GDP, the USA are particularly high in compar-
ison with other nations [5].
Fig. 19.9 Healthcare spending in the Federal Republic of Germany (Fraunhofer ME-
VIS, all rights reserved. Based on data from the Federal Statistical Office)
328 Horst Hahn • Andreas Schreiber
[15, 16]. Close to half of these costs are for medicines, in particular systemic chem-
otherapies and so-called “targeted therapies” that are focused more specifically on
the disease in question and are available in increasing numbers. Typical annual
therapy costs amount to between € 20,000 and € 100,000, and sometimes signifi-
cantly more.
In many cases these expensive therapies remain unsuccessful or merely contrib-
ute to a minor delay in the progress of the disease before being discontinued. With
the exception of specific types of cancer for which a therapy is possible, only around
a quarter of drug-based cancer therapy treatments are seen to be effective (cf. [21]).
Unsuccessful cancer therapies thus represent a considerable cost factor in the cur-
rent healthcare system, at an annual cost of far more than $ 50 billion without pro-
viding a successful cure. What is worse for the patients is the fact that most chem-
otherapies have significant side effects that are only justified in the case of the in-
tervention being successful.
Advances in specific immune therapy, in genomics as well as in early in vitro
and in vivo therapy monitoring promise to provide a remedy via far more precise
recommendations for the right combination of therapies, improved patient selection,
and verification of therapy response [3].
In view of the distribution of tasks across the healthcare value chain (cf. Fig. 19.9),
we see that by far the largest proportion of expenditure is on therapy, with an addi-
tional large proportion being allocated to care, and only a fraction remaining for
prevention, early detection, and diagnosis. Since late diagnosis not only reduces
chances of recovery but also increases the associated costs of therapy and care,
early diagnostic detection offers significant potential savings. We still tend to diag-
nose too late and thus spend too much on therapy.
Digital patient records will also contribute towards cost efficiency in general
patient management and the administration of medical services. Today’s data silos
and breakdowns in patient information sharing lead to a grave loss of information,
in turn giving rise to redundant procedures in the form of additional medical con-
sultations or diagnostic processes, for example.
19 E-Health 329
The vast majority of stakeholders in the healthcare system long for the digital trans-
formation to take place. Due to extensive regulatory authorization requirements,
tedious healthcare political processes surrounding design processes, and strong
institutional integration, the healthcare system has thus far tended towards long
technology lifecycles and incremental innovation. Disruptive innovation in medi-
cine thus requires an alignment to the respective predominant national contexts.
Disruption and cooperation need not stand in contradiction to one another. This is
demonstrated by innovative healthcare solutions such as Oscar, an American health-
care insurance company reinventing the insurer as a mobile-first healthcare assis-
tant, with a corresponding network of partner clinics already established. The con-
siderable financial investments of American providers and insurers in innovative
startups are also a witness thereof.
The vision of digitized medicine described above requires parallel innovations
across the medical value chain. Simultaneously, the healthcare market is exhibiting
characteristics of a network industry, tending, due to the effects of competition,
towards the development of oligopolies with a small number of dominant market
players. In the era of digitized medicine, comprehensive medically-curated patient
data is of significant strategic value for the continuous optimization and validation
of self-learning algorithms. In fact, we can assume that in the near future a standards
battle about integrative authority in medicine and the associated access to patient
data will arise. In order to safeguard the interoperability of the system and protect
users from dependence on selected providers, it is desirable that standards be devel-
oped for sharing data in the same way as is already the case for clinic data (FHIR,
HL7) and medical imaging data (DICOM).
The business magazine The Economist predicts that the transformation of medicine
will lead to the development of a new competitive landscape, with established
healthcare providers and pharmaceutical and medical technology giants on the one
hand and technology insurgents on the other [24]. The latter include large technol-
ogy companies such as Google/Alphabet, Apple, SAP, and Microsoft, but also a
range of venture capital-financed startups, which are developing solutions for sec-
330 Horst Hahn • Andreas Schreiber
tions of digital medicine. Annual venture capital investments of more than $4 billion
in the USA over the last three years demonstrate the hopes that investors are pinning
on the healthcare transformation [23]. No less impressive is Google’s latest funding
commitment of $500 million for the above-mentioned Project Baseline.
The decisive advantage of the technology insurgents with respect to established
medical technology players, alongside their agility and abundant financial resourc-
es may, above all, be that in the innovation process they will neither be influenced
by their own legacy nor by the danger of cannibalizing their existing business
through disruption. Since AI-based software solutions will be a key differentiating
feature in future competition, these firms have good prospects in the healthcare
market. In the medium term, we expect a variety of partnerships to develop between
providers, established players, and the new market participants, which will take on
various forms according to national circumstances.
19.7 Outlook
advances will actually be achieved. Many healthcare provider have recognized the
strategic and commercial value of their data and knowledge and are thus seeking
the right route towards realizing this value. Therefore, prospective partnerships and
business models need to also protect the legitimate interests of hospitals and patients
so that a continuous exchange and build-up of data and knowledge can take place
within the network. A well-known counterexample is the discontinued partnership
between the MD Anderson Cancer Center and IBM Watson Health [12].
332 Horst Hahn • Andreas Schreiber
[1] AHA – American Heart Association (2017) Alexa can tell you the steps for CPR, warn- ing
signs of heart attack and stroke. Blog. Zugriff im Juli 2017: http://news.heart.org/ alexa-
can-tell-you-the-steps-for-cpr-warning-signs-of-heart-attack-and-stroke/
[2] Aerts HJ, Velazquez ER, Leijenaar RT et al (2014) Decoding tumour phenotype by
noninvasive imaging using a quantitative radiomics approach. Nat Commun 5:4006.
doi:10.1038/ncomms5006
[3] ASCO (2017) Clinical Cancer Advances 2017, American Society of Clinical Oncology.
Zugriff im Juli 2017: https://www.asco.org/research-progress/reports-studies/clinical-
cancer-advances
[4] CB Insights (2017) From Virtual Nurses To Drug Discovery: 106 Artificial Intelligence
Startups In Healthcare. Zugriff im Juli 2017: https://www.cbinsights.com/blog/artificial-
intelligence-startups-healthcare/
[5] CMS – Centers for Medicare and Medicaid Services (2017) NHE Fact Sheet. Zugriff im
Juli 2017: https://www.cms.gov/research-statistics-data-and-systems/statistics-trends-
and-reports/nationalhealthexpenddata/nhe-fact-sheet.html
[6] Cooper DN, Ball EV, Stenson PD et al (2017) HGMD – The Human Gene Mutation
Database at the Institute of Medical Genetics in Cardiff. Zugriff im Juli 2017: http://
www.hgmd.cf.ac.uk/
[7] Destatis – Statistisches Bundesamt (2017) Gesundheitsausgaben der Bundesrepublik
Deutschland. Zugriff im Juli 2017: https://www.destatis.de/DE/ZahlenFakten/Gesell-
schaftStaat/Gesundheit/Gesundheitsausgaben/Gesundheitsausgaben.html
[8] Dusheck J (2016) Diagnose this – A health-care revolution in the making. Stanford Me-
dicine Journal, Fall 2016. Zugriff im Juli 2017: https://stanmed.stanford.edu/2016fall/
the-future-of-health-care-diagnostics.html
[9] Esteva A, Kuprel B, Novoa RA et al (2017) Dermatologist-level classification of skin
cancer with deep neural networks. Nature 542(7639):115–118. doi:10.1038/nature21056
[10] Ferrucci D, Levas A, Bagchi S et al (2013) Watson: Beyond Jeopardy! Artificial Intel-
ligence 199:93–105. doi:10.1016/j.artint.2012.06.009
[11] Harz M (2017) Cancer, Computers, and Complexity: Decision Making for the Patient.
European Review 25(1):96–106. doi:10.1017/S106279871600048X
[12] Herper M (2017) MD Anderson Benches IBM Watson In Setback For Artificial Intel-
ligence In Medicine. Forbes. Zugriff im Juli 2017: https://www.forbes.com/sites/matthe-
wherper/2017/02/19/md-anderson-benches-ibm-watson-in-setback-for-artificial-intelli-
gence-in-medicine
[13] Knight W (2017) The Dark Secret at the Heart of AI. MIT Technology Review. Zu-
griff im Juli 2017: https://www.technologyreview.com/s/604087/the-dark-secret-at-the-
heart-of-ai/
[14] Lambin P, Rios-Velazquez E, Leijenaar R et al (2012) Radiomics: extracting more infor-
mation from medical images using advanced feature analysis. Eur J Cancer 48(4):441–6.
doi:10.1016/j.ejca.2011.11.036
[15] Mariotto AB, Yabroff KR, Shao Y et al (2011) Projections of the cost of cancer care in the
United States: 2010-2020. J Natl Cancer Inst 103(2):117–28. doi:10.1093/jnci/djq495
[16] NIH – National Institutes of Health (2011) Cancer costs projected to reach at least $158
billion in 2020. News Releases. Zugriff im Juli 2017: https://www.nih.gov/news-events/
news-releases/cancer-costs-projected-reach-least-158-billion-2020
[17] Ryan KJ (2016) Who’s Smartest: Alexa, Siri, and or Google Now? Inc. Zugriff im Juli
2017: https://www.inc.com/kevin-j-ryan/internet-trends-7-most-accurate-word-recog-
nition-platforms.html
[18] Sahiner B, Chan HP, Petrick N et al (1996) Classification of mass and normal breast
tissue: a convolution neural network classifier with spatial domain and texture images.
IEEE Trans Med Imaging 15(5):598–610. doi:10.1109/42.538937
[19] Schmutzler R, Huster S, Wasem J, Dabrock P (2015) Risikoprädiktion: Vom Umgang mit
dem Krankheitsrisiko. Dtsch Arztebl 112(20): A-910–3
[20] Silver D, Huang A, Maddison CJ et al (2016) Mastering the game of Go with deep neural
networks and tree search. Nature 529:484–489. doi:10.1038/nature16961
334 Horst Hahn • Andreas Schreiber
Summary
A successful energy transition is inconceivable without extensive digitization. In
view of the complexity of the task of digitizing the energy sector and all of the
associated systems, previous efforts at defining essential components of digitiza-
tion (such as concretely usable reference architectures and research into the res-
ilience of the future energy system) currently still appear insufficient and unco-
ordinated. These components include smart management approaches capable of
integrating market mechanisms with traditional management technologies, and
comprehensive security concepts (including effective data utilization control)
that need to go far beyond the BSI security profile for smart meters. The digitiza-
tion of the energy system needs to be conceived and operated as a transformation
process designed and planned for the long term with reliable milestones.
Digitization facilitates the smart networking of people, machines, and resources, the
ongoing automation and autonomization of processes, the personalization of servic-
es and products, as well as the flexibilization and fragmentation, but also the inte-
gration of business models across the entire value chain [8]. In the context of this
definition, digitization is increasingly understood as a process of transformation,
which gives rise to the opportunity to scrutinize processes and procedures funda-
mentally and align them to revised or often even completely new business models.
335
336 Peter Liggesmeyer • Dieter Rombach • Frank Bomarius
In order to achieve this, standard system architectures in most cases need to either
be fundamentally revised or completely recreated so that conventional, modernized,
and new products and services can be offered via new network topologies and com-
munications. The term “smart ecosystem” was coined for these kinds of new busi-
1
ness models developed during the course of digitization. The fundamental charac-
teristic of the smart ecosystem is the interplay between individual business models
within a larger overall system oriented towards economic goals. The information
technology basis for this is formed by standards and open IT platforms with low
transaction costs, high reliability, and security for technically implementing the
business models. Internet technologies provide the universal infrastructures for
connecting business partners with one another as well as linking the digital rep-
resentations of things, devices, installations, goods, and services (so-called “digital
twins”) with the physical objects accessible via the Internet (Internet of Things,
IoT).
This transformation has already led, in numerous domains, to business models
previously linked to physical objects and personally provided services being super-
imposed or even substituted by a dematerialized data economy (hybrid value crea-
tion) in the course of digital transformation. Digitally provided services are thus
increasingly in the foreground:
• Amazon delivers without being a producer, Uber provides transport without
having its own fleet, Airbnb lets rooms without owning accommodation.
• Digitization avoids unprofitable downtime of assets through predictive mainte-
nance.
• Digitization replaces investments in facilities with service rental (logistics, heat-
ing, lighting, etc.)
• Digitization facilitates new contract models, e.g., on-time arrival contracts for
train journeys (Siemens)
• Digitization adds value to traditional products (John Deere FarmSight)
Key developer and operator competencies within a data economy are cloud and IoT
technologies, (big) data analytics, machine learning, and deep learning.
The act of fundamentally revising business processes frequently comes up
against the limits of applicable legal frameworks. Data protection and the protection
of privacy take on new meaning during the process of intensive digital networking
and need to be revised in order to guarantee sufficient protection on the one hand,
1 A smart ecosystem integrates information systems that support business objectives and
embedded systems that serve technical objectives so that they operate as one and can pursue
shared overarching (business) objectives.
20 Smart Energy 337
and to permit new and economically sustainable business models on the other hand.
Alongside economic reasons, the (current) lack of an internationally harmonized
legal framework may lead to the relocation of firms to countries with more accom-
modating legislation.
In highly regulated fields such as healthcare, food products, transport, and the
energy sector, legal regulations and compulsory standards such as the EU’s Gener-
al Data Protection Regulation should be taken into account in good time and, where
2
necessary, should be revised in order to not delay the transformation process unnec-
essarily or even smother it completely.
The digital transformation in the energy sector is a process that will unfold relative-
ly slowly over time, and will most likely require several decades. This is due to the
following factors:
• The long-term nature of investment decisions regarding extremely expensive
network infrastructure and power plants requires decision-making that provides
economic security, which is particularly difficult against the backdrop of the
far-reaching structural change provoked by the energy transition. It must never-
theless be ensured that this aspect is not misused as a reason for unnecessary
delays in digitization, which is needed so urgently.
• The weight of regulation in those areas of energy supply that operate as a natural
monopoly limits innovation since the regulatory adjustments necessary for op-
erating new equipment and business models generally lag behind the innovation
itself.
• The transformation process affects several sectors of the energy industry that
have thus far operated separately (electricity, gas, heating) and impacts associ-
ated sectors (transport, home automation, industrial automation). Due to the
dominance of the renewable energy sources of wind and sun and their volatile
feed-in, the electricity sector represents the controlling variable here to which
the other sectors must adjust.
ily invoiced according to the physical unit of work (in multiples of watt-hours, Wh),
that is, according to a volume tariff, or sometimes according to the maximum supply
availability (in multiples of Watts, W). The quality of the energy supply from cen-
trally managed, high-capacity power plants was usually very high but was not an
explicitly stated component of pricing. In other words, the costs of security of
supply and power quality3 have thus far been priced into the volume tariff.
The development of renewable sources of energy and the politically formulated
climate goals both result in the dismantling of power plants (nuclear phase-out and
planned phase-out of coal-fired power plants for decarbonization purposes) within
the electrical energy grid. Centrally provided security of supply, network services4,
and power quality will thus be lost. In the future, they must be provided and invoiced
by the renewable energy feeder operating in a decentralized manner, and will thus
gain the significance of tradable products.
Additional new products could, for example, include fixed levels of service even
during critical conditions within the energy network, similar to the Service Level
Agreements (SLAs) in the information and communications (ICT) sector. For in-
dustrial customers, scattered examples of such products already exist; in the course
of digitization, they may become standard, ensuring for example that basic, low-en-
ergy functions are maintained and dispensable devices are selectively turned off
during periods of extreme electrical energy undersupply, where until now complete
subnetworks had to be shut down. Developing this thought further means that in-
stead of paying for used, fed-in, or transported energy, flexible markets will devel-
op in the future where changes in the feed-in, requirements, or consumption profiles
can be traded, managed, and invoiced in cash value as flexibly as possible and
calculated by digital systems. Flexibility is above all a product motivated by the
technical necessities of operating a system with highly fluctuating renewable ener-
gy feed-in levels. Essentially, it focuses on the costs of operating under unbalanced
network conditions and only prices the actual amount of energy flow indirectly,
particularly since the marginal costs of renewable energy plants tend towards zero.
In the future then, energy quantities or power ratings will rather play the role of
unchangeable physical limitations for the digital markets. Even today, the actual
3 In simple terms, power quality refers to the maintenance of the target voltage (230V) and
the target frequency (50Hz). The actual voltage may thus deviate by 10 percent and the
frequency by 0.2 Hz. Greater deviation may entail damage in connected systems (both
producers and consumers), and power supplies are thus generally subjected to emergency
shutdown.
4 Network services refer to technically necessary functions, e.g., reactive power generation,
black-start capabilities (restarting after a blackout), and provision of short-circuit power
(maintaining the flow of electricity in the case of a short circuit such that fuses can trigger).
20 Smart Energy 339
Ambitious climate goals can only be achieved via massive decarbonization. The
German approach of largely replacing fossil fuels while simultaneously phasing out
nuclear energy requires a massive expansion of renewable energy installations for
years to come. Wind and sun are available in Germany, but water power only to a
340 Peter Liggesmeyer • Dieter Rombach • Frank Bomarius
limited degree. Biogases are only able to supply limited quantities of energy, espe-
cially due to the emerging competition with the food production industry. Thus,
supply based on sun (PV and solar heat) and wind is extremely dependent on the
weather and the time of day – in other words, we have volatile or fluctuating feed-in.
Fluctuation – which may mean both a dramatic temporary oversupply of electrical
energy and a massive undersupply (volatility) that may last for days – has to be
handled appropriately, without reducing the usual security of supply. A portfolio of
measures and goals for solving the problem is under discussion:
• Efficiency measures for reducing the energy requirements of devices and instal-
lations:
While measures to increase efficiency are basically to be supported since they
reduce energy costs and make it easier to achieve climate goals, they make very
little contribution to solving the problem of fluctuating provision of electrical
energy. Energy efficiency can also be increased by means of non-digital techni-
cal improvements, for example by reducing heat losses, pressure losses, or fric-
tion. In, numerous modern devices, however, smart digital controls are respon-
sible for increased energy efficiency, for example in the case of smart drive
control systems in pumps. The design of these control systems has thus far been
oriented towards optimization of their operation in the face of energy costs that
are neither time- nor load-dependent. It is only by opening the systems’ commu-
nications to signals from the supply network (e.g., coded as variable tariffs) that
these control systems can also become useful to the network. This is termed
“demand-side management”.
• Sector coupling in order to make the flexibility potential of other energy systems
(gas, heating, cooling) usable for the electricity sector:
20 Smart Energy 341
Fig. 20.1 Sector coupling of heating, electricity, and gas (Fraunhofer IESE)
I n sector coupling, flexibility potentials are harnessed in a way that goes beyond
the possibilities of DSM: either alternative energy sources are momentarily used
to cover energy demand in the face of a shortfall of electrical energy (e.g., by
using gas instead of electricity), or the storage capacity of another form of ener-
gy (heating, cooling, etc.) is used to adjust the load profile on the electricity side,
which is generally cheaper than using a battery (see Fig. 20.1). Sector coupling
is considered to have the greatest potential – in terms of quantity and cost – for
providing the necessary flexibility for the electricity grid of the future. It can be
used in both small end-customer installations, such as in the kW domain as bi-
valent heating using a choice of either gas or electricity, and in the MW domain,
e.g., for energy suppliers providing additional electric heating for heat storage in
district heating networks. If there are significant excess supplies of electricity,
variants of sector coupling with lower levels of effectiveness may also be eco-
nomically feasible, such as power to gas (electrolysis, methanation) or power to
liquid (production of liquid fuels) [5].
• Market mechanisms:
Flexibility requests arising from volatile feed-ins or imbalances in the system and
flexibility offerings resulting from demand-side management and sector coupling
may in the future be traded on new electronic markets in order to achieve a bal-
anced power economy within the electricity network, primarily through the use of
market mechanisms. Market mechanisms are favored politically, and the quest for
new business models that identify and test out these mechanisms is under way. A
number of research projects to this end are investigating the potential interaction
of technical, market-related, and adapted regulatory conditions (e.g., [9][15][16]).
With the exception of energy efficiency, the four approaches described above can
thus be implemented exclusively by using digitization: Devices and installations
342 Peter Liggesmeyer • Dieter Rombach • Frank Bomarius
Housing Production
Device Farm
Block Plant
Power
House Hospital Quarters
Plant
Power
kW / kWh MW /MWh
(Generation or Consumption)
Electricity
Gas Heat
Electricity Electricity
Fig. 20.3 Schematic diagram of a hierarchical cellular structure with sector coupling
(Fraunhofer IESE)
undesirable grid conditions. Nevertheless, in the age of the Internet and automated
attacks, this advantage may be lost if the architecture of the cell control systems does
not contain suitable lines of defense, or if the ICT in the cells is identically imple-
mented and thus any weak points disclosed can be attacked on a broad scale with
limited effort. Diversity5 of software implementation instead of digital monocultures
in control rooms, operating systems, and control algorithms makes broad-scale at-
tacks more difficult. Diversity is thus a desirable – if costly – system property, since
economies of scale are lost and any solutions found are more difficult to transfer.
Digitization makes it possible to utilize flexibility as a “currency” within the
energy system of the future. With flexibility as a product, new roles and actors will
be defined in the course of digital transformation, and new business models will
arise. The sector is currently beginning to think about which new personalized
products could be conceivable and profitable. A few examples are:
• Software plus consultancy services for planning and designing cell infrastructure
(for existing stock and new plans).
• Software for continuously monitoring cells vis-à-vis a range of indicators, e.g.,
climate protection contributions, energy flows, material flows, logistics, secu-
rity status. Services around the generation, analysis, and distribution of this data.
• Various control rooms for actors at the cell level (aggregators, contractors, data
resellers) with corresponding data aggregation and processing functions.
In recent years, the information and communications sector has clearly demonstrat-
ed that large ICT systems can be built and operated to be highly reliable, powerful,
and profitable. Leading examples are the Internet or the global mobile telephony
systems. In this respect, the digitization in the course of the energy transition – in-
volving several hundred million devices and installations across Germany in the
long term, or billions worldwide – is no more demanding in terms of numbers and
ICT performance than global mobile telephony, electronic cash, or IPTV.
Effective mechanisms for data security and data usage control are essentially
well-known. The issue now is to consistently build these into the resulting systems
right from the start instead of having to upgrade them later.
With significant participation from Fraunhofer, current German flagship projects
are demonstrating how industry standards have been systematically created for
years in the area of industry-relevant embedded systems in order to advance com-
prehensive digitization. Examples include AUTOSAR [9], BaSys 4.0 [10], or the
Industrial Data Space (IDS) [11]. Flagship projects for the energy systems sector
with comparably high demands are just starting: see projects within the programs
SINTEG [15] and KOPERNIKUS [16]. The criticality for society and industry of
the ICT systems of the future energy infrastructure will be key to their design. Based
on the high availability and power quality that is customary in Germany, the future
energy system, which will be several orders of magnitude more complex6, should
deliver at least the same availability and power quality.
Every system accessible via the Internet nowadays is exposed to a stream of
increasingly sophisticated cyber-attacks, which is growing by far more than a factor
of ten every year [18]. The energy system of the future, too, will be based on Inter-
net technology and will be a valuable target for attacks from the Internet. Despite
all precautionary safeguards and redundancy, breakdowns due to energy installation
6 In Germany, more than 560,000 transformers and more than 1.5 million renewable energy
installations were already part of the electricity grid as of 2015 [17], but there is hardly any
communication among them yet. A hypothetical area-wide smart meter roll-out would add
around 40 million networkable meters, not counting controllable consumers and batteries.
346 Peter Liggesmeyer • Dieter Rombach • Frank Bomarius
7 The system no longer works, but transitions to a safe state in any case.
8 In the automotive sector, this operational state is called fail-operational.
20 Smart Energy 347
The system itself needs to capture its condition, connected installations, as well
as their parameters and typical profiles, map them to models, and actively use
the results for control purposes (model learning).
• Self-optimization:
A cell’s many and diverse optimization goals may change dynamically. Climate
protection goals, for example, may vary in their importance according to the
time of day or year. The system needs to be able to adapt itself to optimization
goals, which may in part be contradictory and vary according to different
timescales.
The foundation for fulfilling all of these requirements is the massive collection,
processing, and secure sharing of data. To do this, an Energy Data Space (EDS)
must be created as an adaptation of the Industrial Data Space (IDS) [12].
A fundamental ICT reference architecture needs to be defined that, in particu-
lar, specifies and permits the implementation of all of the key requirements with
respect to security (with the goal of being safe to fail), sets standards, but does not
stand in the way of a multitude of implementations (goal: diversity).
Whereas in other sectors overarching ICT standards are being created in a tar-
geted manner, e.g., via AUTOSAR [9] for automotive engineering or BaSys 4.0 [10]
and IDS [11] for the field of embedded systems and Industry 4.0, the ICT landscape
in the energy field thus far still resembles more of a patchwork rug.
As explained in the previous section, the future energy system will not be con-
figured statically and will need to assert itself in the face of changing and at times
unexpected influences and attacks.
Resilience is the ability to adapt to previously unknown and unexpected
changes and simultaneously continue to provide system functions.
Modern definitions of the term “resilience” refer to its close relationship to the
concepts of security, forecasting, and sustainability [2][6]. To this extent, resilience
is not a schematic response to negative influences but also incorporates the ability
to self-adapt and to learn.
Traditional risk analyses are reaching their limits due to the complexity of the
energy system. They are also only suitable to a limited degree for identifying new
and unexpected events. In the future, criteria will therefore be required for oper-
ationalization and quantification of the resilience of the energy system during
operation. In some cases, however, the criteria, methods, and indicators for meas-
uring resilience first need to be developed. Monitoring technologies need to be
combined with a systemic approach in order to identify the energy system’s po-
tential vulnerabilities already during its transformation (i.e., without interruption)
[3]. Functioning examples of how system security and functionality can be mon-
itored and ensured during operation – even in the face of changes to the system
configuration – are well known from other sectors (Industry 4.0 or commercial
vehicles).
In the energy system of the future, a previously unknown interaction between
physical and virtual elements will develop. As a result, suitable new strategies for
redundancy also need to be worked out. The tried-and-tested “n-1” rule for redun-
dancy design will thus be insufficient to compensate for all of the diverse poten-
tially erroneous interactions between ICT systems (which are corruptible), poten-
tially maliciously influenced markets, regulated subsystems, and unchangeable
physical constraints. The property of being “safe to fail” always extends across all
of the system’s physical and virtual components – resilience is a property arising
from complex interactions that is specifically organized in each cell based on the
individual configuration of the installations within a cell. Suitable early warning
systems and the implementation of elastic system reactions are pressing research
questions.
Closely related to the issue of resilience are issues surrounding the forecasting
of system behavior, the immediate observability of the system state, and the reliable
transparency of system actions. How this kind of monitoring system should be con-
structed and how complex system states and complex interrelated processes can be
represented in a way that is clear and understandable for the user is yet another
important research question.
20 Smart Energy 349
Fig. 20.4 Present contractual relationships between energy sector players (Bretschneider,
Fraunhofer IOSB-AST)
Finally, set requirements and the services provided must be documented com-
prehensively and in a way that is subject to neither falsification nor dispute, and
must be made accessible to invoicing. Until now, this has been regulated via a
complex, static web of contracts (cf. Fig. 20.4) and corresponding reporting chan-
nels stipulated by the BNetzA (Federal Network Agency).
Consistent establishment of market mechanisms throughout the advancing en-
ergy transition will lead to very small volumes of energy (resp. flexibilities) being
traded and the contract partners here (mainly “prosumers”) not being able to first
conclude bilateral master contracts. The contractual agreements that are necessary
to protect the large volume of brief relationships between the numerous actors
within cells require contracts concluded by machine (“smart contracts”) that are
founded on framework agreements concluded by humans. Here, too, there is a need
for research in order to conclude legally protected agreements in real time between
machines that must then be translated into system actions, monitored during exe-
cution, traceably documented, and correctly invoiced. At the moment, the block-
chain approach is being propagated in this context. Its suitability remains to be
verified.
350 Peter Liggesmeyer • Dieter Rombach • Frank Bomarius
The energy transition is a process that will require several decades due to its tech-
nical and societal complexity. During the course of the transformation process, old
and new technologies will need to not only coexist but function in an integrated
manner over a long period of time. The authors are convinced that now is the time
to focus more intensively on the digitization of the energy transition. Only when the
energy transition is understood as a complex and systemic transformation process
can digitization actively support and successfully shape the necessary changes at
the technical and societal level and help to press ahead with the transformation
process. Very detailed support for this assessment is provided by the Münchner
Kreis in its 50 recommendations [1].
In the past, there were many important innovations with respect to renewable
energy technology. The accompanying digital networking and the resulting system-
ic challenges and opportunities have been neglected for a long time. Although it
appears necessary and has often been discussed, the cellular approach has thus far
been the focus of far too little research in the context of the energy system. The
specification and implementation of resilience – ultimately a critical system char-
acteristic affecting all of its components – also remain largely unexplored.
Last but not least, the success of the technical transformation process is, to a very
large extent, reliant on long-term social acceptance and support [19], which must
be continually verified and actively designed.
[6] Ahern, J.: From fail-safe to safe-to-fail: Sustainability and resilience in the new urban
world, Landscape and Urban Planning 100, 2011, S. 341–343.
[7] H.-J. Appelrath, H. Kagermann, C Mayer, Future Energy Grid – Migrationspfade ins
Internet der Energie, acatech-Studie, ISBN: 978-3-642-27863-1,2012.
[8] Bundesnetzagentur: Digitale Transformation in den Netzsektoren – Aktuelle Entwick-
lungen und regulatorische Herausforderungen, Mai 2017.
[9] Flex4Energy FLEXIBILITÄTS¬MANAGEMENT FÜR DIE ENERGIE-VERSOR-
GUNG DER ZUKUNFT, https://www.flex4energy.de
[10] AUTOSAR (AUTomotive Open SystemARchitecture) https://www.autosar.org/
[11] Basissystem Industrie 4.0 (BaSys 4.0). http://www.basys40.de/
[12] White Paper Industrial Data Space https://www.google.de/url?sa=t&rct=j&q=&e src
=s&source=web&cd=2&cad=rja&uact=8&ved=0ahUKEwiHp6X9qr7UAhU Fl
xoKHeBcAFUQFggvMAE&url=https%3A%2F%2Fptop.only.wip.la%3A443%2Fhttps%2Fwww.fraunhofer.de%2Fc- on
tent%2Fdam%2Fzv%2Fde%2FForschungsfelder%2Findustrial-data-space%2FIndu
strial-Data-Space_whitepaper.pdf&usg=AFQjCNGHaxIr7RxDWl6OPIaUZlwGaN0R
ZA&sig2=_mec3o8JzqMQDxLeQTkldw
[13] Dossier der Expertengruppe Intelligente Energienetze: Dezentralisierung der Energie-
netzführung mittels IKT unterstützen, Juni 2017
[14] Dossier der Expertengruppe Intelligente Energienetze: Branchenübergreifende IKT-
Standards einführen, Juni 2017
[15] https://www.bmwi.de/Redaktion/DE/Artikel/Energie/sinteg.html
[16] https://www.kopernikus-projekte.de/start
[17] https://www.bundesnetzagentur.de/DE/Sachgebiete/ElektrizitaetundGas/Unterneh-
men_Institutionen/ErneuerbareEnergien/ZahlenDatenInformationen/zahlenunddaten-
node.html
[18] https://www.tagesschau.de/inland/cyberangriffe-107.html
[19] BDEW und rheingold institut: Digitalisierung aus Kundensicht, 22. 3. 2017, https://
www.bdew.de/digitalisierung-aus-kundensicht
Advanced Software Engineering
Developing and testing model-based software
21
securely and efficiently
Summary
Software rules them all! In every industry now, software plays a dominant role
in technical and business innovations, in improving functional safety, and also
for increasing convenience. Nevertheless, software is not always designed, (re)
developed, and/or secured with the necessary professionalism, and there are
unnecessary interruptions in the development, maintenance, and operating chains
that adversely affect reliable, secure, powerful, and trustworthy systems. Current
surveys such as the annual World Quality Report put it bluntly, directly corre-
lated with the now well-known failures of large-scale, important and/or safety-
critical infrastructures caused by software. It is thus high time that software
development be left to the experts and that space be created for the use of current
methods and technologies. The present article sheds light on current and future
software engineering approaches that can also and especially be found in the
Fraunhofer portfolio.
21.1 Introduction
Let us start with the technological changes brought about by the digital transforma-
tion which, in the eyes of many people, represent revolutionary innovations for our
society. Buildings, cars, trains, factories, and most of the objects in our everyday
lives are either already, or will soon be, connected with our future gigabit society
via the ubiquitous availability of the digital infrastructure [1]. This will change in-
formation sharing, communication, and interaction in every field of life and work,
353
354 Ina Schieferdecker • Tom Ritter
ware. These include networked and non-networked control systems such as control
units in automobiles and airplanes, systems for connected and autonomous driving,
and systems of systems such as the extension of the automobile into the backbone
infrastructure of the OEMs. But also systems (of systems) in telecommunications
networks, IT, industrial automation, and medical technologies are understood by
this term.
Software-based systems today are often distributed and connected, are subject
to real-time demands (soft or hard), are openly integrated into the environment via
their interfaces, interact with other software-based systems, and use learning or
autonomous functionalities to master complexity. Independently of whether we are
now in a fourth revolution or in the second wave of the third revolution with digiti-
zation, the ongoing convergence of technologies and the integration of systems and
processes is brought about and supported via software. New developments such as
those in augmented reality, fabbing, robotics, data analysis, and artificial intelli-
gence, too, place increasing demands on the reliability and security of software-based
systems.
Let us examine things in greater depth. According to the IEEE Standard for Config-
uration Management in Systems and Software Engineering (IEEE 828-2012 [1]),
software is defined as “computer programs, procedures and possibly associated doc-
umentation and data pertaining to the operation of a computer system”. It includes
programmed algorithms, data capturing or representing status and/or context, and a
wide range of descriptive, explanatory, and also specifying documents (see Fig. 21.1).
A look at current market indicators reveals the omnipresence of software: ac-
cording to a 2016 Gartner study, global IT expenditures of $3.5 billion were expect-
ed in 2017. Software is thus the fastest-growing area, at $357 billion or 6% [4].
Bitkom, as well, supports this view [5]: according to its own survey of 503 compa-
nies with 20 or more staff, every third company in Germany is developing its own
software. Among large organizations with 500 or more staff, the proportion rises as
high as to 64%. According to this survey, already every fourth company in Germa-
ny employs software developers, and an additional 15% say they want to hire addi-
tional software specialists for digital transformation.
Nevertheless, 50 years after the software crisis was explicitly addressed in 1968,
and after numerous approaches and new methods in both software and quality en-
gineering, the development and operation of software-based connected systems is
still not smooth [8]. The term “software engineering” was initially introduced by
356 Ina Schieferdecker • Tom Ritter
F. L. Bauer as a provocation: “the whole trouble comes from the fact that there is so
much tinkering with software. It is not made in a clean fabrication process, which it
should be. What we need, is software engineering.” The authors Fitzgerald and Stol
identify various gaps in the development, maintenance, and distribution of soft-
ware-based systems that can be closed via methods of continuous development,
testing, and rollout.
Studies on breakdowns and challenges in the Internet of Things (IoT) complete
our view here: according to self-reports by German companies, four in five of them
have an “availability and data security gap” in IT services [9]. Servers in Germany,
for example, stand idle for an average of 45 minutes during an outage. The estimat-
ed direct costs of these kinds of IT failures rose by 26% in 2016 to $21.8 million,
versus $16 million in 2015. And these figures do not include the impacts that cannot
be precisely quantified such as reduced customer confidence or reputational damage
to the brand.
The top two challenges connected to IoT are security in particular IT security
and data protection, as well as functional safety and interoperability of the soft-
ware-defined protocol and service stacks [10].
In keeping with this, the latest 2016–17 edition of the World Quality Report [3]
shows that there is a change in the primary goals of those responsible for quality
assurance and testing that is accompanying the ongoing pervasion of the physical
world by the digital world with the Internet of Things. The change picks up the in-
creasing risk of breakdown and the criticality of software-based connected systems
from the perspective of business and security. Thus, increasing attention is given to
quality and security by design, and the value of quality assurance and testing is
being raised in spite of, or indeed due to, the increasing utilization of agile and
DevOps methods. Thus, with the complexity of software-based connected systems,
expenditures for the realization, utilization, and management of (increasingly vir-
tualized) test environments are also increasing. Even though extensive cost savings
are equally possible in this area through automation, the necessity of making qual-
ity assurance and testing even more effective at all levels remains.
bility, and portability [12]. Against the backdrop of current developments and rev-
elations, aspects of ethics as well as of guarantees and liability must also supplement
the dimensions of software quality.
For a long time, software was considered to be free from all of the imponderables
inherent to other technical products, and in this way was seen as the ideal technical
product [11]. A key backdrop to this is the fact that algorithms, programming concepts
and languages, and thus any computability is traced to the Turing computability (the
Turing thesis). According to Church’s thesis, computability here incorporates precise-
ly those functions which can be calculated intuitively by us. Thus, while non-com-
putable problems such as the halting problem elude algorithmics (and thus software),
for each intuitively computable function there is an algorithm with limited computing
complexity that can be realized via software. Here, the balance between function,
algorithm, and software is the responsibility of various phases and methods of soft-
ware engineering such as specification, design, and implementation as well as verifi-
cation and validation. If alongside this understanding of intuitive computability, soft-
ware now sounds like a product that is simple to produce, this is by no means the case.
What began with the software crisis still holds true today. Herbert Weber reiterated
this in 1992, “the so-called software crisis has thus far not yet produced the necessary
level of suffering required to overcome it” [13]. Also Jochen Ludewig in 2013 for-
mulated it as, “the requirements of software engineering have thus not yet been met”
[11]. The particular characteristics of software are also part of the reason for this.
First and foremost, software is immaterial, such that all of the practical values
for material products do not apply or are only transferable in a limited sense. Thus,
software is not manufactured but “only” developed. Software can be copied practi-
cally without cost, with the original and the copy being completely identical and
impossible to distinguish. This leads, among other things, to nearly unlimited pos-
sibilities for reusing software in new and typically unforeseen contexts.
On the one hand, using software does not wear it out. On the other hand, the
utilization context and execution environment of software are constantly evolving
such that untarnished software does in fact age technologically and indeed logical-
ly and thus must be continually redeveloped and modernized. This leads to mainte-
nance cycles for software that, instead of restoring the product to its original state,
generate new and sometimes ill-fitting, i.e. erroneous, conditions.
Software errors thus do not arise from wear and tear to the product but are built
into it. Or errors develop in tandem with the software’s unplanned use outside of its
technical boundary conditions. This is one way that secure software can be operat-
ed insecurely, for example.
In addition, the days of rather manageable software in closed, static, and local
use contexts for mainly input and output functionalities are long gone. Software is
358 Ina Schieferdecker • Tom Ritter
In what follows, we illustrate selected model-based methods and tools for the effi-
cient development of reliable and secure software that are the result of current R&D
studies at Fraunhofer FOKUS.
Models have a long tradition in software development. They originally served
the specification of software and its formal verification of correctness. In the mean-
time they are commonly used as abstract, technology-independent bearers of infor-
mation for all aspects of software development and quality assurance [15]. They
thus serve to mediate information between software tools and to provide abstrac-
tions for capturing complex relationships. One example of this, in the context of risk
analysis and assessment, or of systematic measurement and visualization of soft-
ware characteristics, and also of software test automation, is via model-based testing
or test automation. As Whittle, Hutchinson, and Rouncefield argue in [16], the
particular added value of model-driven software development (Model-Driven En-
gineering, MDE) is the specification of the architectures, interfaces, and compo-
nents of software. Architecture is also used by FOKUS as the foundation for docu-
21 Advanced Software Engineering 359
mentation, functionality, interoperability, and security in the methods and tools in-
troduced in what follows.
Process automation
Modern software development processes often use teams at various sites for indi-
vidual components and to integrate commercial third-party components or open
source software. Here, a wide range of software tools are used, with various indi-
viduals participating in different roles, whether actively or passively. The central
problems here are the lack of consistency of the artifacts created in the development
process, the shortage of automation, and the lack of interoperability between the
tools.
ModelBus® is an open-source framework for tool integration in software and
system development and closes the gap between proprietary data formats and soft-
ware tool programming interfaces [17]. It automates the execution of tedious and
error-prone development and quality assurance tasks such as consistency assurance
across the entire development process. To do this, the framework uses service or-
chestrations of tools in keeping with SOA (service-oriented architecture) and ESB
(enterprise service bus) principles.
The software tools of a process landscape are connected to the bus by the provi-
sion of ModelBus® adaptors. Adaptors are available for connecting IBM Rational
Doors, Eclipse and Papyrus, Sparx Enterprise Architect, Microsoft Office, IBM
Rational Software Architect, IBM Rational Rhapsody, or MathWorks Matlab Simu-
link.
Safety-critical software-based systems are subject to careful risk analysis and eval-
uation according to the ISO Risk Management Standard [18] in order to capture and
minimize risks. For complex systems, however, this risk management may be very
time-consuming and difficult. While the subjective assessment of experienced ex-
perts may be an acceptable method of risk analysis on a small scale, with increasing
size and complexity other approaches such as risk-based testing [21] need to be
chosen.
An additional opportunity for more objective analysis is provided by the use of
security tests in line with ISO/IEC/IEEE “Software and systems engineering – Soft-
ware testing” (ISO 29119-1, [19]). A further option is to first have experts carry out
a high-level assessment of the risks based on experience and literature. In order to
make this initial risk assessment more accurate, security tests can be employed at
360 Ina Schieferdecker • Tom Ritter
precisely the point where the first high-level risk picture shows the greatest vulner-
abilities. The objective test results may then be used to enhance, refine, or correct
the risk picture thus far. However, this method first becomes economically applica-
ble with appropriate tool support.
RACOMAT is a risk-management tool developed by Fraunhofer FOKUS, which
in particular combines risk assessment with security tests [20]. Here, security testing
can be directly incorporated into event simulations that RACOMAT uses to calcu-
late risks. RACOMAT facilitates extensive automation of risk modelling through to
security testing. Existing databases such as those of known threat scenarios are used
by RACOMAT to ensure a high degree of reuse and avoid errors.
At the same time, RACOMAT supports component-based compositional risk
assessment. Easy-to-understand risk graphs are used to model and visualize an
image of the risk situation. Common techniques such as fault tree analysis (FTA),
event tree analysis (ETA), and conducting security risk analysis (CORAS) may be
used in combination for the risk analysis in order to be able to benefit from the
various strengths of the individual techniques. Starting with an overall budget for
risk assessment, RACOMAT calculates how much expenditure is reasonable for
security testing in order to improve the quality of the risk picture by reducing vul-
Fig. 21.1 Risk analysis and security testing with RACOMAT (Fraunhofer FOKUS)
21 Advanced Software Engineering 361
nerabilities. The tool offers recommendations on how these means should be used.
To do this, RACOMAT identifies relevant tests and places them in order of priority.
Software-based systems are becoming ever more complex due to their increasing
functions and their high security, availability, stability, and usability requirements.
In order for this not to lead to losses of quality and so that structural problems can
be identified early on, quality assurance must commence right at the beginning of
the development process. A model-driven development process where models are
key to the quality of the software-based system is well suited to this. Up to now,
however, quality criteria for this were neither defined nor established. In future,
model characteristics and their quality requirements need to be identified, and ad-
ditionally mechanisms found with which their properties and quality can be deter-
mined.
Metrino is a tool that checks and ensures the quality of models [22]. It may be
used in combination with ModelBus® but can also be employed independently. With
the aid of Metrino, metrics for domain-specific models can be generated, indepen-
dently defined, and managed. The metrics produced can be used for all models that
accord with the meta-model used as the basis for development. Metrino thus ana-
lyzes and verifies properties such as the complexity, size, and description of soft-
ware artifacts. In addition, the tool offers various possibilities for checking the
computational results of the metrics and representing them graphically – for ex-
Fig. 21.2 Model-based software measurement and visualization with Metrino (Fraunho-
fer FOKUS)
362 Ina Schieferdecker • Tom Ritter
ample in a table or spider chart. Since Metrino saves the results from several evalu-
ations, results from different time periods can also be analyzed and compared with
one another. This is the only way that optimal quality of the final complex soft-
ware-based system can be guaranteed.
Metrino is based on the Structured Metrics Metamodel (SMM) developed by the
Object Management Group (OMG) and can be used both for models in the Unified
Modeling Language (UML) as well as for domain-specific modelling languages
(DSLs). On top of that, Metrino’s field of application includes specialized, tool-spe-
cific languages and dialects.
Whether for designing embedded systems or for software in general, Metrino
can be used in the widest variety of different domains. The tool can manage metrics
and apply them equally to (model) artifacts or also to the complete development
chain, including traceability and test coverage.
The quality of products is the decisive factor of being accepted on the market. In
markets with security-related products in particular, such as medicine, transporta-
tion, or automation sectors, for example, quality assurance is thus accorded high
priority. In these sectors, quality is equally decisive for product authorization. Qual-
ity assurance budgets, however, are limited. It is thus important to managers and
engineers that the available resources are utilized efficiently. Often, manual testing
methods are still being used, even if only a comparatively small number of tests can
be prepared and conducted in this way, and they are additionally highly prone to
error. The efficiency of manual testing methods is thus limited, and rising costs are
unavoidable. Model-based test generation and execution offers a valuable alterna-
tive: the use of models from which test cases can be automatically derived offers
enormous potential for increasing test quality at lower costs. In addition, case stud-
ies and practical uses have shown that when model-based testing techniques are
introduced necessary investment costs in technology and training pay off quickly
[24].
Fokus!MBT thus offers an integrated test modelling environment that leads the
user through the Fokus!MBT methodology and thus simplifies the creation and use
of the underlying test model [25]. A test model contains test-relevant, structural,
behavior- and method-specific information that conserves the tester’s knowledge in
a machine processable fashion. In this way, it can be adapted or evaluated at any
time, say for the generation of additional test-specific artifacts. Additional benefits
of the test model are the visualization and documentation of the test specification.
21 Advanced Software Engineering 363
Fokus!MBT uses the UML Testing Profile (UTP), specified by the Object Manage-
ment Group and developed with significant contributions from FOKUS, as its mod-
eling notation. UTP is a test-specific extension of the Unified Modeling Language
(UML) common in industry. This allows testers to use the same language concepts
as the system architects and requirements engineers, thus preventing communica-
tion issues and encouraging mutual understanding.
Fokus!MBT is based on the flexible Eclipse RCP platform, the Eclipse Modeling
Framework (EMF), and Eclipse Papyrus. As a UTP-based modeling environment,
it has all of the UML diagrams available as well as additional test-specific diagrams.
Alongside the diagrams, Fokus!MBT relies on a proprietary editor framework for
describing and editing the test model. The graphical editor user interfaces can be
specifically optimized for the needs or abilities of the user in question. In doing so,
if necessary, it is possible to completely abstract from UML/UTP, allowing special-
ists unfamiliar with IT to quickly produce a model-based test specification. This is
also supported by the provision of context-specific actions that lead the user through
the Fokus!MBT methodology. In this way, methodically incorrect actions or actions,
which are inappropriate for the context in question are not even enabled. Based upon
this foundation, Fokus!MBT integrates automated modeling rules that guarantee ad-
herence to guidelines, in particular modelling or naming conventions, both after and
during working on the test model. These preventative quality assurance mechanisms
distinguish Fokus!MBT from other UML tools, accelerate model generation, and
minimize costly review sessions.
The fundamental goal of all test activities is validating the system to be tested
vis-à-vis its requirements. Consistent and uninterrupted traceability, in particular
between requirements and test cases, is indispensable here. Fokus!MBT goes one
step further and also incorporates the test execution results into the requirements
traceability within the test model. In this way, a traceability network is created be-
tween requirement, test case, test script, and test execution result, thus making the
status of the relevant requirements or the test progress immediately assessable. The
visualization of the test execution results additionally facilitates the analysis, pro-
cessing and assessment of the test case execution process. The test model thus
contains all of the relevant information to estimate the quality of the system tested,
and support management in their decision-making related to the system’s release.
Analytical methods and dynamic testing approaches in particular are a central and
often also exclusive instrument for verifying the quality of entire systems. Software
tests thereby require all of the typical elements of software development, because
tests themselves are software-based systems and must thus be developed, built,
tested, validated, and executed in exactly the same way. In addition to that, test
systems possess the ability to control, stimulate, observe and to assess the system
being tested. Although standard development and programming techniques are
generally also applicable for tests, specific solutions for the development of a test
system are important in order to be able to take its unique features into account. This
approach expedited the development and standardization of specialized test speci-
fication and test implementation languages.
One of the original reasons for the development of Tree and Tabular Combined
Notation (TTCN) was the precise conformity definition for telecommunications
component protocols according to their specification. Test specifications were uti-
lized to define test procedures objectively and assess, compare, and certify the
equipment on a regular basis. Thus, the automated execution became exceptionally
important for TTCN, too.
Over the years, the significance of TTCN grew and various pilot projects demon-
strated a successful applicability beyond telecommunications. With the conver-
gence of telecommunications and information technology sectors, the direct appli-
cability of TTCN became obvious to developers from other sectors. These trends,
along with the characteristics of more recent IT and telecommunications technolo-
21 Advanced Software Engineering 365
gies also placed new requirements on TTCN: the result is TTCN-3 (Testing and Test
Control Notation Version 3, [27]).
TTCN-3 is a standardized and modern test specification and test implementation
language developed by the European Telecommunication Standards Institute
(ETSI). Fraunhofer FOKUS played a key role in TTCN-3’s development and is
responsible for various elements of the language definition, including Part 1 (con-
cepts and core languages), Part 5 (runtime interfaces), and Part 6 (test control inter-
faces), as well as TTCN-3 tools and test solutions [28][29]. With the aid of TTCN-
3, tests can be developed textually or graphically, and execution can be automated.
In contrast to many (test) modeling languages, TTCN-3 comprises not only a lan-
guage for test specification but also an architecture and execution interfaces for
TTCN-3-based test systems. Currently, FOKUS uses TTCN-3 for developing the
Eclipse IoT-testware for testing and securing IoT components and solutions, for
example [30].
366 Ina Schieferdecker • Tom Ritter
It is not possible to introduce all of our methods, technologies, and tools here. Our
publications (see also [31]) contain further information on
• Security-oriented architectures,
• Testing and certifying functional security,
• Model-based re-engineering,
• Model-based documentation,
• Model-based porting to the cloud, or
• Model-based fuzz tests.
It is not enough to simply develop new methods, technologies, and tools. These also
need to be distributed and supported during their introduction to projects and pilots.
Fraunhofer FOKUS has thus for a long time been involved in professional de-
velopment. The institute has initiated and/or played a key role in developing the
following professional development schemes in cooperation with the ASQF (Arbe-
itskreis Software-Qualität und -Fortbildung – “Software Quality and Training
Working Group”) [32], GTB (German Testing Board) [33], and the ISTQB (Inter-
national Testing Qualifications Board [34]):
• GTB Certified TTCN-3 Tester
• ISTQB Certified Tester Foundation Level – Model-Based Testing
• ISTQB Certified Tester Advanced Level – Test Automation Engineer
• ASQF/GTB Quality Engineering for the IoT
Further, Fraunhofer FOKUS, together with HTW Berlin and Brandenburg Univer-
sity of Applied Sciences, is also forming a consortium, the Cybersecurity Training
Lab [35], with training modules on
• Secure software engineering
• Security testing
• Quality management & product certification
• Secure e-government solutions
• Secure public safety solutions
This and other offerings, such as on semantic business rule engines or open govern-
ment data, are also available via the FOKUS-Akademie [36].
21 Advanced Software Engineering 367
21.11 Outlook
Software development and quality assurance are both subject to the competing re-
quirements of increasing complexity, the demand for high-quality, secure, and reli-
able software, and the simultaneous economic pressure for short development cy-
cles and fast product introduction times.
Model-based methods, technologies, and tools address the resulting challenges,
and in particular support modern agile development and validation approaches.
Continuous development, integration, testing and commissioning benefit from
model-based approaches to a particular degree. This is because they form a strong
foundation for automation and can also support future technology developments due
to their independence from specific software technologies.
Additional progress in model-based development is to be expected or indeed
forms part of current research. Whereas actual integration and test execution are
already conducted in a nearly entirely automated fashion, the analysis and correc-
tion of errors remains a largely manual task, one that is time-consuming and itself
subject to error and can thus lead to immense delays and costs. Self-repairing soft-
ware would be an additional step towards greater automation, borrowing from the
diverse software components in open source software using pattern recognition and
analysis through deep learning methods, and repair and assessment using evolution-
ary software engineering approaches. In this way, software could become not only
self-learning but also self-repairing.
[19] ISO/IEC/IEEE: Software and systems engineering – Software testing – Part 1: Concepts
and definitions. ISO/IEC/IEEE 29119-1:2013, https://www.iso.org/standard/45142.
html, besucht am 22.7.2017.
[20] Johannes Viehmann und Frank Werner. „Risk assessment and security testing of large
scale networked systems with RACOMAT.“ International Workshop on Risk Assess-
ment and Risk-driven Testing. Springer, 2015.
[21] Michael Felderer, Marc-Florian Wendland und Ina Schieferdecker. „Risk-based testing.“
International Symposium On Leveraging Applications of Formal Methods, Verification
and Validation. Springer, Berlin, Heidelberg, 2014.
[22] Christian Hein, et al. „Generation of formal model metrics for MOF-based domain spe-
cific languages.“ Electronic Communications of the EASST 24 (2010).
[23] Marc-Florian Wendland, et al. „Model-based testing in legacy software modernization:
An experience report.“ Proceedings of the 2013 International Workshop on Joining
AcadeMiA and Industry Contributions to testing Automation. ACM, 2013.
[24] Ina Schieferdecker. „Model-based testing.“ IEEE software 29.1 (2012): 14.
[25] Marc-Florian Wendland, Andreas Hoffmann, and Ina Schieferdecker. 2013. Fokus!MBT:
a multi-paradigmatic test modeling environment. In Proceedings of the workshop on
ACadeMics Tooling with Eclipse (ACME ‚13), Davide Di Ruscio, Dimitris S. Kolovos,
Louis Rose, and Samir Al-Hilank (Eds.). ACM, New York, NY, USA, Article 3, 10 pages.
DOI: https://doi.org/10.1145/2491279.2491282
[26] ETSI: TTCN-3 – Testing and Test Control Notation, Standard Series ES 201 873-1 ff.
[27] Jens Grabowski, et al. „An introduction to the testing and test control notation (TTCN-
3).“ Computer Networks 42.3 (2003): 375-403.
[28] Ina Schieferdecker und Theofanis Vassiliou-Gioles. „Realizing distributed TTCN-3 test
systems with TCI.“ Testing of Communicating Systems (2003): 609-609.
[29] Juergen Grossmann, Diana Serbanescu und Ina Schieferdecker. „Testing embedded
real time systems with TTCN-3.“ Software Testing Verification and Validation, 2009.
ICST‘09. International Conference on. IEEE, 2009.
[30] Ina Schieferdecker, et al. IoT-Testware – an Eclipse Project, Keynote, Proc. of the 2017
IEEE International Conference on Software Quality, Reliability & Security, 2017.
[31] FOKUS: System Quality Center, https://www.fokus.fraunhofer.de/sqc, besucht am
22.7.2017.
[32] ASQF: Arbeitskreis Software-Qualität und Fortbildung (ASQF), http://www.asqf.de/,
besucht am 25.7.2017.
Automated Driving
Computers take the wheel
22
Prof. Dr. Uwe Clausen
Fraunhofer Institute for Material Flow and Logistics IML
Prof. Dr. Matthias Klingner
Fraunhofer Institute for Transportation and Infrastructure
Systems IVI
Summary
Digital networking and autonomous driving functions mark a new and fascina-
ting chapter in the success history of automobile manufacture, which stretches
back well over a century already. With powerful environment recognition, highly
accurate localization and low-latency communication technology, vehicle and
traffic safety will thus increase dramatically. Precise, fully automated vehicle
positioning of autonomously driven electric cars creates the conditions for
introducing innovative high-current charging technologies located underground.
If autonomous or highly automated vehicles share information with intelligent
traffic controls in future, then this may lead to a significantly more efficient uti-
lization of existing traffic infrastructures and marked reductions in traffic-related
pollutant emissions. These are three examples that underscore the enormous
significance of electric mobility together with autonomous driving functions
for the development of a truly sustainable mobility. In this process, the range
of scientific-technical challenges in need of a solution is extraordinarily broad.
Numerous Fraunhofer institutes are involved in this key development process
for our national economy, contributing not merely expert competencies at the
highest scientific-technical level, but also practical experience in the industrial
implementation of high technologies. In what follows, we take a look at some
of the current topics of research. These include autonomous driving functions
in complex traffic situations, cooperative driving maneuvers in vehicle swarms,
low-latency communication, digital maps and precise localization. Also, security
of functions and security against manipulation for driverless vehicles, digital
371
372 Uwe Clausen • Matthias Klingner
22.1 Introduction
The initial foundations for highly automated driving functions were already de-
veloped more than 20 years ago under the European PROMETHEUS project
(PROgraMme for a European Traffic of Highest Efficiency and Unprecedented
Safety, 1986– 1994). With €700 m. in funding, it was Europe’s largest ever re-
search project to date and involved not only the vehicle manufacturers but also
nearly all of the main European supply firms and academic institutions. The goal
was to improve the efficiency, environmental sustainability, and security of road
traffic. Fraunhofer institutes such as the then IITB and today’s IOSB in Karlsruhe
have been able to successfully continue this research right up to the present day.
Many of the findings from these first years have now found broad application in
modern vehicle technology. These include, for example, proper vehicle operation,
friction monitoring and vehicle dynamics, lane keeping support, visibility range
monitoring, driver status monitoring or system developments in collision avoid-
ance, co-operative driving, autonomous intelligent cruise control, and automatic
emergency calls.
Then as now, the migration to ever higher degrees of automation is closely linked
to the development of driver assistance systems. Emergency braking assistants,
multi collision brakes, pre-crash security systems, and lane-keeping and lane-
change assistants all offer ever more comprehensive automated driving functions
for vehicle occupant protection. While these driver assistance systems primarily
actively intervene in vehicle control in security-critical driving situations, autono-
mous cruise control (ACC) relieves the driver in monotonous traffic flows on ex-
pressways and highways. Based on ACC and lane-keeping systems together with a
high-availability car-to-car communication, the first forms of platooning are cur-
rently developing. Here, vehicles are guided semi-autonomously as a group over
longer expressway distances at high speed and with minimal safety distances be-
tween them. It is commonly expected that within the coming years platooning will
22 Automated Driving 373
In the automobile sector, too, automated expressway driving and automatic maneu-
vering and parking represent challenges that have been largely overcome today.
With the addition of active assistive functions for longitudinal and lateral guidance
– for example, in the current Mercedes E-Class – vehicles are already shifting into
the gray area between semi- and high automation and are thus defining the state of
the art for market-ready and legally compliant standard features.
The function-specific form of the most highly developed approaches for auto-
mated driving functions was first demonstrated during the DARPA Grand Challenge
hosted by the Defense Advanced Research Projects Agency (DARPA) of the US
Defense Department. This was a competition for unmanned land vehicles, last held
in 2007, that significantly advanced the development of vehicles capable of driving
completely autonomously.
Since then, the IT corporation Google (now Alphabet) has become one of the
technological leaders in the field of autonomous vehicles. As of November 2015,
Google’s driverless cars reached the mark of 3.5 million test kilometers with a new
record proportion of 75% of these being fully autonomous. Google currently op-
erates around 53 vehicles capable of driving completely automatically using a
combination of (D)GPS, laser scanners, cameras, and highly accurate maps. Even
if the safety of the autonomous driving functions implemented still requires a
certain amount of verification, Tesla – with sales of 500,000 highly automated
passenger vehicles per year – is also planning to set new standards in the field of
highly automated electric vehicles from 2018. Established vehicle manufacturers
such as BMW, Audi, and Nissan have thus far primarily demonstrated autonomous
driving maneuvers within limited areas such as expressways and parking garages.
In 2013, Mercedes-Benz showed what can be implemented today under ideal
conditions, with close-to-production sensors and actuators, with its legendary
autonomous cross-country trip from Mannheim to Pforzheim. Swedish vehicle
manufacturer Volvo, itself active in the field of autonomous driving, is currently
implementing a cloud-based road information system in Sweden and Norway. The
rapid sharing of highly accurate digital map content, road status information, and
374 Uwe Clausen • Matthias Klingner
current traffic data is a key requirement for ensuring adequate safety for future
autonomous driving.
Alongside automobile manufacturers, large tier-one suppliers such as ZF and
Continental are also active in the field of automation and are increasingly beginning
to present their own solutions. The first markets for prototype driverless shuttle
vehicles are already beginning to open up.
The demonstrations by OEMs, suppliers, and shuttle manufacturers described
above show that completely automatic driving will be possible even with production
vehicles in the medium term. Here, less complex environments such as expressways
are the initial focus, especially when considering production vehicle development,
and the safety driver serves as the supervisory entity. This concept is only suitable
to a limited extent for comprehensively guaranteeing the safety of highly automat-
ed vehicles. The fact that the spontaneous transfer of the driving function to the
driver in genuinely critical situations can lead to driving mistakes has been under-
scored by the first, partly tragic accidents. Centrally operated roadside vehicle
monitoring with externally synchronized vehicle movement in situations of extreme
danger represents one alternative to the safety driver, but has thus far received little
to no consideration from the OEMs.
The Google and Daimler demonstrations show that vehicles can be safely local-
ized with decimeter accuracy in urban environments, using a fusion of odometry,
laser sensors, and DGPS. The development of roadside safety systems is thus en-
tirely possible. One such roadside monitoring system – based on lidar systems – is
currently being implemented at the University of Reno’s Living Lab in the state of
Nevada. The Fraunhofer IVI is involved there with corresponding projects. In ad-
dition, current developments make clear that the necessary products for networking
completely autonomous vehicles with the infrastructure will be widely available in
the short to medium term, and thus that the use of these communication structures
may also serve to ensure traffic safety.
The growing demand for automated driving functions has been emphasized
recently by a number of user studies, conducted by, among others, the Fraunhofer
IAO. Surveys by Bitkom, Bosch, Continental, and GfK reveal a high level of in-
terest (33% to 56%) and stress the future significance of the factors of safety and
convenience as the most important criteria for new purchases. By 2035, market
penetration for automated vehicles of between 20% [1] and 40% [2] is considered
a possibility. Just during these next few years, highly developed driver safety
systems and convenience-oriented assistance systems will dominate the rapidly
growing future car IT market. In their Connected Car 2014 study, Pricewaterhouse-
Coopers estimates an annual growth in global revenues in the two categories men-
tioned above of around €80 billion by 2020 [3]. The proportion of value added by
22 Automated Driving 375
electronics in cars currently lies at between 35% and 40% of the sale price and
rising [4].
The spending of German automotive firms for patents and licenses in the elec-
trics/electronics fields has risen significantly since 2005, currently lying above €2
billion [5]. For in-vehicle connectivity services, McKinsey predicts a fivefold in-
crease in potential sales volumes by 2020 [6]. The combination of these factors
makes the automobile market increasingly attractive for companies from the IT
sector. In an American Internet study, for example, Google and Intel are already
highlighted as the most influential players in the automated driving field [7]. While
US firms dominate the IT field, a leading position in the development and introduc-
tion of highly advanced automation systems is accorded to German vehicle manu-
facturers and suppliers. The Connected Car 2014 study mentioned above identifies
three German groups (VW, Daimler, and BMW) as the leading automobile manu-
facturers for innovation in driver assistance systems. American studies, too, reach
similar conclusions, both with respect to Audi, BMW, and Daimler [7], the three
premium German manufacturers, as well as the German suppliers Bosch and Con-
tinental [3].
Autonomous driving at VDA standard level 5 (fully driverless) will only be approv-
able in the public transport space when it is supervised from the roadside by pow-
erful car-to-infrastructure and car-to-car networking, and can be influenced by ex-
ternal safety systems in hazardous situations. Supervised and cooperative driving
in so-called automation zones will make its mark in the urban space. This will first
take place in the field of inner-city logistics and local public transportation in the
form of autonomous shuttle lines, e-taxis, and delivery services. In the medium
term, highly automated or fully autonomous vehicles in cooperatively functioning
vehicle convoys will enter the stage. In the context of comprehensive automation
of inner-city transport networks, these will make a significant contribution to the
easing of traffic flows as well as to the more efficient capacity utilization of existing
transportation infrastructures.
Currently, various technologies for cooperative driving in automation zones are
being developed within Fraunhofer institutes, in particular, institutes in the Fraun-
hofer ICT Group. Alongside the linear guidance of vehicles following one another
(platooning), cooperative driving also includes the multi-lane guidance of hetero-
geneous swarms of vehicles. Technologies under development are:
376 Uwe Clausen • Matthias Klingner
As well as established open source robotics frameworks such as ROS, plus YARP,
Orocos, CARMEN, Orca, and MOOS, in-house simulation and development tools
are also used across institutes here such as OCTANE from Fraunhofer IOSB. Con-
trol of the vehicle during autonomous driving is generally transferred to various
trajectory controls by implemented finite state machines in keeping with the situa-
tion identified. These trajectory controls are responsible for the individual driving
tasks such as straight-line driving, lane changes, left- or right-hand turns, etc. Even
though impressive outcomes have already been achieved with these situation-de-
pendent trajectory controls, the functional security is not sufficient to model fully
automatic driving on public roads. In contrast to the DARPA Challenge’s open
terrain, real road traffic offers far more complex and sometimes ambiguous situa-
tions to overcome, destabilizing the situationally discrete selection processes for
allocating the trajectory controls. In [8][9], Fraunhofer IOSB researchers intro-
duced, among other things, a significantly more powerful path planning process that
reduces insecurities significantly and is now also being pursued by other authors
[10]. For each point in time, the process [10][11] calculates a multicriteria optimized
trajectory that incorporates all of the decision-making criteria of a human driver
such as collision avoidance, rule conformity, convenience, and journey time.
In probabilistic modeling, not only the fallibility of sensor signals is taken into
account but also uncertainties, for example, with respect to the behavior of traffic
22 Automated Driving 377
of up to 80% have been possible, for example [18]. In addition to traditional object
detection, it has now been possible to demonstrate processes for fine-grained pix-
el-level object differentiation [19].
As part of the EU’s AdaptIVe research project, 28 different partners – including
Europe’s ten largest automobile manufacturers, suppliers, research institutes, and
universities – are investigating various use cases for autonomous driving on ex-
pressways, within car parks, and in complex urban areas. In 2014, the aFAS project
commenced, with eight partners from research and industry developing an autono-
mous security vehicle for expressway roadwork sites. In the IMAGinE project, re-
search is being carried out into the development of new assistance systems for the
cooperative behavior context. Within the European SafeAdapt research project, new
electric and electronic architecture concepts for increased autonomous electric ve-
hicle reliability are being developed, evaluated, and validated under Fraunhofer’s
leadership. The C-ITS (Intelligent Transport System) project is drafting a Eu-
rope-wide infrastructure and coordinating international cooperation, particularly in
the field of communications infrastructures. In 2017, the AUTOPILOT project be-
gan, viewing the vehicle as sensor nodes and incorporating it into an IoT structure
where the vehicle forms a specific info node.
Tests are ongoing in five different European regions.
As demonstrated by swarms of fish, birds, and insects and the herd instinct of some
mammals, synchronized movements mean that an incredible number of moving
objects can be concentrated in an extremely limited space. Well-known examples
of spontaneous synchronization outside of the engineering sphere include rhythmic
applause or audience waves. Swarm movements in the animal kingdom or the syn-
chronous firing of the synapses in the nervous system; in the engineering sphere,
examples include the synchronous running of power generators in a supply system.
In the highly automated transport systems of the future, synchronization will
• Increase the traffic density of existing infrastructures,
• Accelerate traffic flows via coordinated prioritization and synchronized signal-
ing systems,
• Extensively optimize transportation chains in public transport via dynamic, syn-
chronized connection conditions and thus
• Significantly reduce traffic-related emissions of CO2 and pollutants.
22 Automated Driving 379
Over the course of evolution, the ability of living organisms to respond reflexively
to the broadest range of hazardous situations has proved vital to their survival. In
the split seconds before an unavoidable collision, pre-crash safety systems (PCSSs)
even today protect vehicle occupants and those involved in the accident by tighten-
ing the seatbelts, releasing various airbags, or raising the bonnet. Fully autonomous
380 Uwe Clausen • Matthias Klingner
vehicles will in future be able to largely avoid crash situations or, in mixed traffic
situations outside of supervised automation zones, at least significantly alleviate
them. This is supported by the following:
• Early identification of hazardous situations,
• Car-to-car communication and coordinated trajectory selection,
• Extremely maneuverable chassis design, and
• Highly dynamic and autonomously controlled driving maneuvers.
The safety philosophy governing the development of autonomous vehicles is the
assumption that it will remain impossible even in future to identify every hazardous
situation sufficiently early, at least for standard evasive or braking maneuvers. For
these kinds of dangerous moments, distributed electric drivetrains, multi-axle steer-
ing control processes, ABS, ESP and torque vectoring, as well as familiar vehicle
stabilization techniques using combined longitudinal/latitudinal control offer addi-
tional degrees of freedom. In extreme driving situations, autonomously controlled
driving maneuvers will mean that even residual risks can be safely overcome.
Coordinating autonomous driving functions in hazardous situations requires
particularly powerful car-to-x communications (LTE-V, 5G) with very short laten-
cies and maximum transmission rates. Fraunhofer institutes such as IIS and HHI are
significantly involved in developing and defining standards for these communica-
tions technologies.
Which car-to-x communications technologies will eventually prevail over the com-
ing years cannot yet be predicted with certainty. It is possible to see potential develop-
ment both in technologies already introduced, such as WLAN 11p, as well as in the en-
hanced 4G standard LTE-V (Vehicular) and LTE Advanced Pro with basic functions
from 3GPP Release 14 through to the emerging 5G mobile telephony standard. For
narrowband applications, an IoT service was introduced by the Deutsche Telekom in
2017 that facilitates high signal strengths and ranges based on the LTE-Cat-NB1
specification. Fraunhofer IML is involved in the development related to innovative
applications in logistics and transportation.
Currently, the communications technology equipment for autonomous pilot ve-
hicles is designed to be highly open and migratable. This is so that the latest research
findings and technology standards for functionally secure, low-latency, robust, and
IT secure IoT communication can be quickly and flexibly integrated. At the mo-
ment, WLAN 11p represents the only car-to-x communications technology with
unrestricted availability. The focus of developments is currently concentrated on
application-specific functions such as WLAN 11p-based “cooperative perception”
or fast video data transmission. From 2018, a standardized mobile telephony variant
for car-to-x communication will be available in LTE-V. The future developments
22 Automated Driving 381
For the most demanding level of automation –fully driverless vehicles – safety re-
quirements that lie far beyond current technical achievements must be fulfilled.
Autonomous vehicles need to comply with the highest security integrity level (SIL
4) and thus guarantee an error rate below 10-4 ... 10-5. In order to at least correspond
to the average safety level of a human driver, on the other hand, more than 300,000
km of accident-free driving on expressways and in highly congested urban condi-
tions must have been demonstrated without a single intervention from the safety
driver. Isolated maximums for driverless journeys without safety intervention today
lie between 2,000 km and 3,000 km for Google/Waymo cars, which, however, were
mostly completed on highways. Analysis of the various testing projects in California
shows that automobile manufacturers are currently working at an average level of
one intervention every 5 km. Based on findings from the Digitaler Knoten (“digital
nodes”), Road Condition Cloud, and iFUSE projects, two strategies stand out for
increasing the security of autonomous driving. Many contributions follow the tra-
ditional approach to increasing the security of autonomous driving via on-board
system improvements:
• Highly integrated multisensory systems for environment recognition,
• More powerful signal and image processing,
• Learning processes for situation recognition and improving responsiveness,
• Reliable vehicle electric system infrastructure with micro-integrated electronic
systems,
• Cooperative vehicle guidance, and
• Development of robust and low-latency car-to-x communication.
In addition to this, roadside safety systems based on high-performance car-to-infra-
structure communications and stationary environment sensors plus object recogni-
tion, tracking, and behavior modeling are being developed. Research here focuses
on
382 Uwe Clausen • Matthias Klingner
The roadside safety system thus transfers a defined portion of the responsibility for
the safety to an intelligent infrastructure and can be understood as a kind of “virtu-
al safety driver” in supervised automation zones.
Here, stationary environment sensors, situation prediction, and pathway plan-
ning are enhanced with information from the vehicles on recognized situations and
intended trajectory. In moments of danger, the safety system coordinates the re-
sponse of the supervised vehicles and resolves collision conflicts in advance or in-
tervenes via emergency and evasive maneuvers. With the advancing development
of the perceptive and predictive abilities of the vehicle guidance systems and of ever
more powerful car-to-x networking, goal conflicts in external emergency interven-
tions are increasingly avoided. Autonomous driving functions for defined scenarios
outside of automation zones may be approved after a minimum number of test
kilometers without external intervention. This minimum number specified is based
on accident statistics from the GIDAS database. Whether roadside safety and coor-
dination systems will remain on busy highways and intersections in future is some-
thing further development will show.
22 Automated Driving 383
“Driverless vehicles” demand the highest standards in functional safety from the
safety-related components implemented. Primary research goals at Fraunhofer in-
stitutes such as the ESK in Munich and IIS in Erlangen are focused on the modular,
multifunctional prototyping of e/e (electric/electronic) and software architectures
with fault-tolerant designs, of error-tolerant sensor and actuator equipment through
to hot plug and play mechanisms. But they also primarily concentrate on the func-
tional safety of connected e/e components and the autonomous driving functions
built upon them. Over the last 10 to 12 years, the number of connected e/e systems
in a vehicle has doubled to around 100 in premium cars and 80 for the mid-range.
The level of connection will further increase by leaps and bounds in autonomous
electric vehicles.
It is thus anticipated that modular e/e and software architectures will be key
outcomes of Fraunhofer’s research. They will facilitate flexible and safe interaction
of the e/e systems and of the individual interfaces and data and energy flows within
the e/e systems, and variable linking of vehicle and environment data.
Sensor data can be provided by the utilization of a greater number of sensors,
different physical sensor principles, and suitable sensor fusion with sufficient failure
safety. A particularly high degree of sensor integration can be achieved through the
application of extremely thin and flexible sensor systems, such as those developed
at Fraunhofer EMFT, for example. Data, supply, and control paths within these
systems may be selectively interconnected in the case of failure. For safety-related
actuator systems, this integration scenario is significantly more complex in auton-
omous vehicles.
Reconfiguration and backup level management mechanisms are being devel-
oped and integrated into the software architecture and networking concept of auton-
omous vehicles [23] for maintaining data and energy flows in the case of the failure
of individual components.
Fault tolerance and tamper proofing require demanding technical solutions even
in conventional vehicle engineering. But the challenges of equipping autonomous
vehicles with the ability to quickly identify and independently solve every hazard-
ous situation inside and outside of the vehicle are significantly more complex. The
particular agility of the reactions of electric vehicles and the ability of autonomous
driving functions to stabilize vehicles even in extreme driving situations are ad-
dressed in a more comprehensive research scenario by Fraunhofer institutes in
Dresden, Karlsruhe, Kaiserslautern and Darmstadt. Also, the extent of connected
vehicles’ detection horizon (far exceeding the optical field of vision in convention-
al vehicles) forms part of this research. From today’s standpoint, a completely safe
vehicle in an accident-free traffic environment still seems visionary. But autono-
mous driving nevertheless provides all of the technical conditions for actually
reaching this goal in the not-too-distant future.
Car-to-x communication with the potential for fast ad hoc networking as well as
making extensive vehicle data available (e.g. positional, behavioral, and drivetrain
data; operating strategies; energy requirements; ambient data; environmental infor-
mation; and traffic, car-to-x, multimedia system and interaction and monitoring
data) clearly risk direct manipulation of vehicle controls and the misuse of person-
al or vehicle-specific data. Primary research goals at various Fraunhofer institutes
within the ICT Group are thus focused on
• Fault-tolerant communication systems with corresponding coding, encryption,
and data fusion,
• Security technologies with end-to-end encryption for intervehicle communica-
tions, and
• Signing and verifying sensor data.
Developing independent data rooms for industry, science, and society is also equal-
ly the subject of an extensive research and development project involving numerous
Fraunhofer ICT institutes. The Industrial Data Space Association, a corporate plat-
22 Automated Driving 385
form founded in 2016 can now point to more than 70 member companies, including
listed companies. The Fraunhofer data space reference model architecture [24] is
superbly suited to ensuring data-sovereign, Internet-based interoperability as well
as secure data sharing, also within intelligent traffic systems. The structure of a
Germany-wide mobility data space is currently being drafted in partnership with the
BMWI (Federal Ministry of Transport and Digital Infrastructure) and the German
Federal Highway Research Institute (Bundesanstalt für Straßenwesen, BASt). This
mobility data space is also intended to form the framework architecture for real-time
supply of data to autonomous vehicles. In this way, digital mapping materials, driv-
ing time forecasts, traffic signs, etc. can be provided to autonomous driving func-
tions in a location-referenced manner. Also, sensor information and floating car data
for identifying the traffic situation can be passed back retrospectively.
tors – which guarantee a sufficient level of functional safety even under conditions
of extreme stress [26].
With the vehicle itself taking over the driving function, the construction of future
automobiles will change fundamentally. The first design studies and prototypes
demonstrate that the removal of engine compartments, steering wheels, and displays
in sensor-driven vehicles with electric drivetrains opens up a vast creative scope.
This can be used to increase flexibility, energy efficiency, and vehicle security both
for passenger as well as goods transport.
The modularization strategy was already pursued in conventional automobile
engineering, which meant concentrating all of the driving and control functions in
a small number of base modules and completing these base modules with use
case-specific body modules. The same can be implemented significantly more con-
sistently for autonomous electric vehicles that have no driver’s cockpit, which leads
to a flexibilization of production, logistics, service, and utilization. This process has
not even begun to be achieved in traditional automobile engineering. Even the in-
dependent exchange of body modules is possible. One example of this would be the
design studies (Fig. 22.3 to 22.6) for bipart eV, an experimental electric vehicle with
autonomous driving functions to VDA Level 5 (“driverless”). All of the necessary
components for autonomous electric driving are concentrated in the base module
(Fig. 22.3). The body modules, so-called CAPs, may be offered by any manufacturer
for very specific application scenarios. Economies of scale through module reuse
and decoupling of the lifecycles of individual vehicle modules thus increase the
efficiency and sustainability of vehicle construction enormously.
The extent to which autonomous driving functions reliably rule out individual
crash load cases and thus contribute to a simplified bodywork design is a complex
issue linking the mechanical structural strength of the vehicle construction with the
safety of autonomous driving.
Automated guided vehicles (AGVs) have been used within intralogistics since the
1960s. The control for early systems was based on the optical recognition of strips
on the floor and later on inductive track guidance. Modern AGVs, such as those
developed at Fraunhofer IML, on the other hand, move around freely in space and
use hybrid odometry- and radio transmitter-based localization systems for position-
ing. This latest generation of transportation robots finds optimal routes to the desti-
nation of the goods independently and without track guidance.
Built-in sensors enable people and other vehicles to be recognized and obstacles
avoided. At present, a Smart Transport Robot (STR) for BMW production logistics
has been developed, which for the first time uses automobile industry components
for an AGV. The STR’s energy supply, for example, is provided by recycled BMW
i3 batteries. Additional standard parts from automobile production allow the inex-
pensive production of the suitcase-sized transport robot that, with a weight of just
135 kg, is capable of lifting, transporting, and depositing loads of up to 550 kg. The
highly flexible system can thus be used in production logistics to transport dollies
loaded with car parts, for example. In future, 3D camera systems will facilitate even
more precise navigation and also further lower the costs of safety sensors compared
with conventional AGVs.
Outside of production logistics, too, a dynamic change towards automated driv-
ing is taking place. This applies on the one hand to the automation of serial trucks
– currently primarily in delimited areas (factory/yard logistics) – and on the other
hand to completely new concepts for very small vehicles for use in the public
sphere. Starship Technologies’ delivery robot, which operates under electric power
at little more than a walking pace and is primarily used for last-mile package deliv-
ery on pavements, is one such example.
Although fully automated truck journeys still require many years of testing and
additional development, automation in transportation logistics nevertheless holds
great potential. Alongside digitization and new drivetrain and vehicle concepts, it is
one of the key drivers of change, and application scenarios are being researched at
Fraunhofer IML, for example.
Simple, robust tools such as plowshares, harrows, scythes, and rakes have for cen-
turies dominated agriculture in the world’s various regions. Heavy-duty tractors,
complex tillage units, and fully automated sowers, fertilizer spreaders and harvest-
ers shape our image of modern, high-productivity agricultural technology today.
From a historical perspective, the development of these technologies has taken place
over an extremely short timeframe and has not been without inconsistency. With
increasing mechanization, greater operating widths, and rising automation, there
has been a dramatic rise not only in productivity but also in machine weight. Trac-
tors with a tractive performance of more than 350 kilowatts are able to generate
turning moments of a few thousand newton meters and thus achieve tillage working
widths of up to 30 m. The weight of these machines, however, stretches to 20 t and
390 Uwe Clausen • Matthias Klingner
more. Moving this kind of heavy machinery over open terrain leads to high fuel
requirements and to extreme compression of the soil with deep and lasting damage.
The development outlined above illustrates that sustainable agriculture that is
efficient with respect to the global food situation must in future break completely
new ground. This is so that highly automated agricultural machines that are as
completely electrified as possible can be used in a manner that protects the soil.
So-called Feldschwarm® units, which are visionary at the moment, could over the
coming years prove to be a genuinely marketable migration path for future agricul-
tural machine technology. The units are equipped with zero-emission high-efficien-
cy drivetrain systems and work the fields as a swarm. The required pulling and work
force in swarms like these is distributed electrically across the swarm vehicles’
wheels and tools. Feldschwarm® units are able to move around autonomously and
handle flexible working widths and variously staggered work processes as a group.
Feldschwarm® technologies in lightweight design, like those currently being devel-
oped within the BMBF-sponsored Feldschwarm research project, protect the soil.
Due to electrification, precise navigation, comprehensive sensor setups, and au-
Fig. 22.7 Autonomous Feldschwarm® (“field swarm”) units for tillage (© TU Dresden
Chair of Agricultural Systems Technology) (Fraunhofer IVI)
22 Automated Driving 391
tonomous driving functions, they also achieve significantly greater degrees of au-
tomation and energy efficiency than the drawn technology that is thus far still
largely mechanically or hydraulically driven. Feldschwarm® units are thus destined
to provide key technologies for the global development trend towards precision
farming or computer aided farming. Fraunhofer, with its IVI and IWU institutes, is
playing a key role in these developments.
Due to the mostly external control over rail vehicle movements provided by track
guidance, switch towers, and dispatchers, the automation of rail travel is far easier
to solve technologically than that of road travel, and in certain cases, it is already
well advanced. Driverless rail traffic between stations has been a reality in closed
systems such as parts of the Nuremberg subway [27] or the Dortmund suspension
railway for years. New systems (such as in Copenhagen, or lines 13 and 14 on the
Paris metro [28]) are increasingly designed to be driverless. Existing subway sys-
tems are increasingly being converted to driverless operation [29]. One example,
alongside the Nuremberg subway, is line 1 of the Paris metro [28].
In contrast to road automation, key control functions are still retained in central
control rooms (switch towers). For regional travel, the hope is that automatic and
pre-planned control will both provide improved utilization of capacity as well as
contribute to energy-optimized travel. In railroad-based goods transportation in
particular, however, significant costs are incurred wherever trains are joined or
separated again, during switching, and where loading points are serviced. Here,
significant personnel and material costs are incurred for limited ton kilometers.
However, these functions are indispensable for bundling trains and servicing cus-
tomers. Automation approaches are being discussed both for efficiency reasons as
well as for improving workplace safety. The longevity of rolling stock and railroad
infrastructure requires tailored migration concepts with, for example, a semi-auto-
mation of switching processes within factory premises and maneuvering facilities
where protecting against the mistakes of other traffic participants or people is fea-
sible. In doing so, it is entirely possible for automated guidance to be combined with
conventional switching and, where relevant, even supported with sensors from the
automotive sector [30], for example via radar sensors for space detection during
switching. With increasing experience from semi-automated operations and addi-
tional sensor improvements, the utilization of driverless operated vehicles over re-
gional journeys is being considered as a next step.
392 Uwe Clausen • Matthias Klingner
More than 70% of the earth’s surface is water. Ocean-going ships often travel for
many weeks, requiring crews with specific nautical competencies. As recently as
2014, the production of autonomously guided ships was viewed as highly unlikely
by 96% of German ship owners. Just two years later, 25% were convinced that
unmanned shipping was possible (PwC, ship owners study, issue 2014 and 2016).
Demanding operating environments on the high seas, the restrictiveness of data
transmission, and constant monitoring of the technical systems on board, alongside
unanswered legal questions made autonomous shipping unlikely so far. But indus-
trial companies, classification societies, and research institutes (not least Fraunhofer
CML in Hamburg) have in the meantime made promising progress. Due to improve-
ments in and the acceleration of data transmission, continuing digitization, and the
development of specific solutions for autonomous shipping, the future realization
of this vision appears conceivable. More recent studies have focused on ships that
are intended for autonomous use on fixed routes rather than for global use. This is
because present conditions make constant digital supervision and control more
possible. The areas of application selected are suitable for low-maintenance electric
or hybrid propulsion. In Norway, for example, an electric container feeder ship for
distances of up to 30 nautical miles is currently being developed that is due to op-
erate autonomously from 2020 following a transition phase.
Fig. 22.9 Visualization of an automated crow’s nest – object recognition via camera sys-
tems (Fraunhofer CML)
Fraunhofer CML has been working intensively with solutions for unmanned and
autonomously guided commercial ships in recent years: first under the EU-spon-
sored MUNIN research project [31], and then in partnership with South Korea’s
Daewoo Shipbuilding and Marine Engineering DSME. Three ship guidance simu-
lators are being utilized here, e.g., in the ongoing development of systems for inde-
pendently conducting evasive maneuvers. Autonomously guided ships must also
operate in accordance with international collision avoidance regulations in order to
avoid dangerous situations and collisions. Their steering must thus respond at least
as reliably as would be expected from a human helmsman, in keeping with the
regulations. The solution developed at Fraunhofer CML is able to combine data
from different sensor sources such as radar, the automatic identification system
(AIS), or day- and night-vision cameras and produce a picture of the traffic situa-
tion. Should the situation require it, an evasive maneuver based on the internation-
al collision prevention rules is computed and executed. The simulation environment
is supplemented by a Shore Control Center (SCC). The SCC permits the monitoring
and control of a fleet of autonomously navigating ships from the shore. Generally,
these fleets operate on the high seas without relying on external support. Should,
however, automated on-board systems become overwhelmed by a situation, the
Shore Control Center is able to intervene instantly. In this way, by increasing effi-
ciency and safety on board, the development of assistance systems for commercial
shipping is already making important contributions to the transition towards auton-
omous shipping.
394 Uwe Clausen • Matthias Klingner
Vehicles also navigate underneath the water’s surface in order to produce under-
water maps, investigate marine sediments, and carry out inspections and measure-
ments. Areas of application, alongside marine research, include preparations for
laying deep sea cables and pipelines and their inspection [32].
Nevertheless, neither radio waves nor optical or acoustic signals permit contin-
uous communication under water. DEDAVE, a flexible deep-sea underwater vehicle
by Fraunhofer IOSB, is able to dive to depths of 6,000 m, to conduct missions
lasting up to 20 hours, and is equipped with numerous sensors. Due to the patented
fast-switching mechanism for batteries and data stores, maintenance and setup times
are shorter, reducing the duration and costs of a mission significantly. Current re-
search studies are investigating the operation of a swarm of 12 intelligent deep-sea
robots [33].
[1] IHS Automotive (2014): Emerging technologies. Autonomous cars – Not if, but when
[2] Victoria Transport Institute (2013): Autonomous vehicle implementation predictions. Im-
plications for transport planning. http://orfe.princeton.edu/~alaink/SmartDrivingCars/
Reports&Speaches_External/Litman_AutonomousVehicleImplementationPredictions.
pdf [letzter Zugriff: 03.07.2017]
[3] Strategy& (2014): Connected Car Studie 2014
[4] O. Wyman Consulting (2012): Car-IT Trends, Chancen und Herausforderungen für
IT- Zulieferer
[5] Zentrum für europäische Wirtschaftsforschung ZEW, Niedersächsisches Institut für
Wirtschaftsforschung (NIW) (2009): Die Bedeutung der Automobilindustrie für die
deutsche Volkswirtschaft im europäischen Kontext. Endbericht an das BMWi
[6] McKinsey (2014): Connected car, automotive value chain unbound. Consumer survey
[7] Appinions (2014): Autonomous cars. An industry influence study
[8] J.R. Ziehn (2012): Energy-based collision avoidance for autonomous vehicles. Master-
arbeit, Leibniz Universität Hannover
[9] M. Ruf, J.R. Ziehn, B. Rosenhahn, J. Beyerer, D. Willersinn, H. Gotzig (2014): Situation
Prediction and Reaction Control (SPARC). In: B. Färber (Hrsg.): 9. Workshop Fahreras-
sistenzsysteme (FAS), Darmstadt, Uni-DAS e.V., S. 55-66
[10] J. Ziegler, P. Bender, T. Dang, and C. Stiller (2014): Trajectory planning for Bertha. A
local, continuous method. In: Proceedings of the 2014 IEEE Intelligent Vehicles Sym-
posium (IV), S. 450-457
[11] J.R. Ziehn, M. Ruf, B. Rosenhahn, D. Willersinn, J. Beyerer, H. Gotzig (2015): Cor-
respondence between variational methods and hidden Markov models. In: Proceedings
of the IEEE Intelligent Vehicles Symposium (IV), S. 380-385
22 Automated Driving 395
[12] M. Ruf, J.R. Ziehn, D. Willersinn, B. Rosenhahn, J. Beyerer, H. Gotzig (2015): Global
trajectory optimization on multilane roads. In: Proceedings of the IEEE 18th International
Conference on Intelligent Transportation Systems (ITSC), S. 1908-1914
[13] J.R. Ziehn, M. Ruf, D. Willersinn, B. Rosenhahn, J. Beyerer, H. Gotzig (2016): A trac-
table interaction model for trajectory planning in automated driving. In: Proceedings of
the IEEE 19th International Conference on Intelligent Transportation Systems (ITSC),
S. 1410-1417
[14] N. Evestedt, E. Ward, J. Folkesson, D. Axehill (2016): Interaction aware trajectory plan-
ning for merge scenarios in congested traffic situations. In: Proceedings of the IEEE 19th
International Conference on Intelligent Transportation Systems (ITSC), S. 465-472
[15] T. Emter und J. Petereit (2014): Integrated multi-sensor fusion for mapping and local-
ization in outdoor environments for mobile robots. In: Proc. SPIE 9121: Multisensor,
Multisource Information Fusion: Architectures, Algorithms, and Applications
[16] T. Emter,A. Saltoğlu, J. Petereit (2010): Multi-sensor fusion for localization of a mobile ro-
bot in outdoor environments. In: ISR/ROBOTIK 2010, Berlin, VDE Verlag, S. 662- 667
[17] C. Frese, A. Fetzner, C. Frey (2014): Multi-sensor obstacle tracking for safe human-robot
interaction. In: ISR/ROBOTIK 2010, Berlin, VDE Verlag, S. 784-791
[18] C. Herrmann, T. Müller, D. Willersinn, J. Beyerer (2016): Real-time person detection in
low-resolution thermal infrared imagery with MSER and CNNs. In: Proc. SPIE 9987,
Electro-Optical and Infrared Systems: Technology and Applications
[19] J. Uhrig, M. Cordts, U. Franke, T. Brox (2016): Pixel-level encoding and depth layering
for instance-level semantic segmentation. 38th German Conference on Pattern Recogni-
tion (GCPR), Hannover, 12.-15. September
[20] C. Frese und J. Beyerer (2011): Kollisionsvermeidung durch kooperative Fahrmanöver.
In: Automobiltechnische Zeitschrift ATZ Elektronik, Jg. 6, Nr. 5, S. 70-75
[21] C. Frese (2012): Planung kooperativer Fahrmanöver für kognitive Automobile. Disserta-
tion, Karlsruher Schriften zur Anthropomatik, Bd. 10, KIT Scientific Publishing
[22] M. Düring, K. Lemmer (2016): Cooperative maneuver planning for cooperative driving.
In: IEEE Intelligent Transportation Systems Magazine, Bd. 8, Nr. 3, S. 8-22
[23] J. Boudaden, F. Wenninger, A. Klumpp, I. Eisele, C. Kutter (2017): Smart HVAC sen-
sors for smart energy. International Conference and Exhibition on Integration Issues of
Miniaturized Systems (SSI), Cork, 8.-9. März
[24] https://www.fraunhofer.de/en/research/lighthouse-projects-fraunhofer-initiatives/indus-
trial-data-space.html [letzter Zugriff: 03.07.2017]
[25] http://www.edda-bus.de/ [letzter Zugriff: 03.07.2017]
[26] F.-P. Schiefelbein, F. Ansorge (2016): Innovative Systemintegration für elektrische Steck-
verbinder und Anschlusstechnologien. In: Tagungsband GMM-Fb. 84: Elektroni- sche
Baugruppen und Leiterplatten (EBL), Berlin, VDE Verlag; F.Ansorge (2016): Stei- gerung
der Systemzuverlässigkeit durch intelligente Schnittstellen und Steckverbinder im Bord-
netz. Fachtagung Effizienzsteigerung in der Bordnetzfertigung durch Automa- tisierung,
schlanke Organisation und Industrie-4.0-Ansätze, Nürnberg, 5. Oktober
[27] G. Brux (2005): Projekt RUBIN. Automatisierung der Nürnberger U-Bahn. Der Eisen-
bahningenieur, Nr. 11, S. 52-56
[28] http://www.ingenieur.de/Branchen/Verkehr-Logistik-Transport/Fahrerlos-Paris-Die-
Metro-14-um-sechs-Kilometer-verlaengert [letzter Zugriff: 03.07.2017]
[29] A. Schwarte, M. Arpaci (2013): Refurbishment of metro and commuter railways with
CBTC to realize driverless systems. In: Signal und Draht, Jg. 105, Nr. 7/8, S. 42-47
396 Uwe Clausen • Matthias Klingner