Business Information Systems Compress
Business Information Systems Compress
Business
LNBIP 320
Information Systems
21st International Conference, BIS 2018
Berlin, Germany, July 18–20, 2018
Proceedings
123
Lecture Notes
in Business Information Processing 320
Series Editors
Wil M. P. van der Aalst
RWTH Aachen University, Aachen, Germany
John Mylopoulos
University of Trento, Trento, Italy
Michael Rosemann
Queensland University of Technology, Brisbane, QLD, Australia
Michael J. Shaw
University of Illinois, Urbana-Champaign, IL, USA
Clemens Szyperski
Microsoft Research, Redmond, WA, USA
More information about this series at http://www.springer.com/series/7911
Witold Abramowicz Adrian Paschke (Eds.)
•
Business
Information Systems
21st International Conference, BIS 2018
Berlin, Germany, July 18–20, 2018
Proceedings
123
Editors
Witold Abramowicz Adrian Paschke
Poznan University of Economics Fraunhofer FOKUS
and Business Berlin
Poznan Germany
Poland
This Springer imprint is published by the registered company Springer International Publishing AG
part of Springer Nature
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Preface
BIS 2018 was organized by the Fraunhofer Institute for Open Communication Systems
and Poznań University of Economics and Business, Department of Information Systems.
Program Committee
Witold Abramowicz Poznań University of Economics and Business, Poland
(Co-chair)
Adrian Paschke (Co-chair) Fraunhofer FOKUS and Freie Universität Berlin,
Germany
Frederik Ahlemann University of Duisburg-Essen, Germany
Rainer Alt Leipzig University, Germany
Dimitris Apostolou University of Piraeus, Greece
Timothy Arndt Cleveland State University, USA
Sören Auer TIB Leibniz Information Center Science
and Technology and University of Hannover,
Germany
Eduard Babkin INSA Rouen; State University, Higher School
of Economics (Nizhny Novgorod), Russia
Morad Benyoucef University of Ottawa, Canada
Tiziana Catarci Università di Roma la Sapienza, Italy
François Charoy Université de Lorraine, LORIA, Inria, France
Rafael Corchuelo University of Seville, Spain
Christophe Debryne University College Odisee, Belgium
Josep Domingo-Ferrer Universitat Rovira i Virgili, Spain
Suzanne Embury The University of Manchester, UK
Vadim Ermolayev Zaporozhye National University, Ukraine
Werner Esswein Technische Universität Dresden, Germany
Anna Fensel Semantic Technology Institute (STI) Innsbruck,
University of Innsbruck, Austria
Agata Filipowska Poznań University of Economics and Business, Poland
Adrian Florea Lucian Blaga University of Sibiu, Romania
Johann-Christoph Freytag Humboldt Universität zu Berlin, Germany
Naoki Fukuta Shizuoka University, Japan
Jaap Gordijn Vrije Universiteit Amsterdam, The Netherlands
Volker Gruhn Universität Duisburg-Essen, Germany
Francesco Guerra UniMo, Italy
Hele-Mai Haav Institute of Cybernetics at Tallinn University
of Technology, Estonia
Martin Hepp Universität der Bundeswehr München, Germany
VIII Organization
Organizing Committee
Adrian Paschke (Co-chair) Fraunhofer FOKUS and Freie Universität Berlin,
Germany
Bartosz Perkowski Poznań University of Economics and Business, Poland
(Co-chair)
Barbara Gołębiewska Poznań University of Economics and Business, Poland
Marko Harasic Fraunhofer FOKUS and Freie Universität Berlin,
Germany
Włodzimierz Lewoniewski Poznań University of Economics and Business, Poland
Milena Stróżyna Poznań University of Economics and Business, Poland
Additional Reviewers
Process Management
Mining Hybrid Business Process Models: A Quest for Better Precision . . . . . 190
Dennis M. M. Schunselaar, Tijs Slaats, Fabrizio M. Maggi,
Hajo A. Reijers, and Wil M. P. van der Aalst
Smart Infrastructures
1 Introduction
Globalization, shorter product life cycles, and rapidly changing customer needs
lead to increasing competitive pressure in the manufacturing industry [1]. In
addition to the high product quality and variety, flexibility and short delivery
times are also important success factors [2]. Thus, efficient and continuously
improved production processes are key prerequisites for a manufacturing com-
pany to become and remain successful in the market [3]. In other business sectors,
especially e-commerce and internet based services, big data computing and ana-
lytics are successfully used for data driven process optimization [4]. This fact
puts forth the potential of data driven optimization as a means to boost busi-
ness processes. Manufacturing companies can also exploit this potential and use
data driven optimization in order to meet the ever increasing demands.
In most cases, companies use the bottom-up approach, where business-relevant
knowledge is searched in all the available data, for example, by using data min-
ing techniques [5]. This approach is characterized by insufficient focus on specific
business objectives and strategies of the respective company, as well as relatively
Company goal
top down
Knowledge
bottom up Information
Data
Gain substantiated
understanding of the
Goals interdependencies
Identify improvement potential
Derive a course of action
towards improved processes
Data analysis
as well as the used equipment and infrastructure. By the means of data driven
optimizations, the automotive manufacturer seeks to gain deep, numerically sup-
ported understanding of the interdependencies within the selected business pro-
cesses and identify improving potentials, as well as deduce an adequate course
of action to exploit this potential. Using data analytics (cf. Fig. 2), the man-
ufacturer aims at reaching optimization goals, e.g., a course of action towards
improved processes.
At the beginning of this project, the responsible employees of the car man-
ufacturer are confronted with the task of gaining and keeping a perspective of
the large amounts of inhomogeneous, apparently incoherent data. Furthermore,
several factors make it difficult to decide, which data sets should be considered
significant for the analysis. First, the examined business processes show a lack
of transparency due to their high complexity and the numerous interdepartmen-
tal interfaces. Secondly, the documentation of the processes and the meta data
might show some incoherency. Lastly, the conjunction of data sources is not
always conducive, which can undermine data with a potential for business rele-
vant knowledge. These conditions render the data preprocessing and integration
a cumbersome task that can also affect the motivation.
Interdependencies between:
processing lead times
Knowledge first pass yield
productivity
Data
result entry, DB-record, XML, logs, …
3.2 Requirements
The hybrid approach, we aim for in this paper, minimizes the risks and com-
bines the advantages of the approaches described above. We define the following
requirements for our approach:
– (R3) Optimal cost-benefit ratio: The hybrid approach aims to achieve an opti-
mal cost-benefit ratio out of data driven optimization: it avoids investments
with a long payback period. Instead, it relies on incremental investments with
many “low-hanging fruits”.
– (R4) Promotion of feasibility: The approach must promote the feasibility of
data driven optimization within the business- and operating-guidelines of the
company: data driven optimization is not conducted for its own sake, but
rather to bring a practical benefit for the company. Therefore, they should be
conducted pragmatically and with minimum distortion of the core business.
Case specific
insight
Data treasure
Case specific
insight
Use case Data treasure
Execute use
case Case specific
Business insight Holistic
Use case
goals Insights
Derive, Execute use Data treasure
prioritize use case Integrate and
cases analyze data
treasures
Case specific
insight
Use case
Execute use Data treasure
case
– Which are the sources of the underlying data and which format does the raw
data have?
– Which requirements must the data processing meet, e.g., real-time or incre-
mental processing?
– How long is the period of time that is considered in the analysis and how
frequently will the analysis be conducted?
– What practical benefit for the company comes with answering the query?
After converting a query into a use case and based on the answers to the
questions above, a potential analysis is to be conducted. At this point, we rec-
ommend to look into the following features to assess the priority of a given use
case: (i) acuteness, defining to what extent the use case attends to urgent issues
of the company, (ii) feasibility, describing how much effort goes into providing
and processing the needed data, and (iii) relevance, examining the bearing of
the benefit. The potential analysis helps the company identify result-oriented,
data-based use cases in an efficient, structured and repeatable manner.
For the highly-prioritized use cases, the required data sources are made avail-
able, access authorization is managed, and data security measures are taken.
Afterwards, the data is processed, e.g., through validation, cleaning, and aggre-
gation, in order to prepare it for the subsequent analysis. In the analysis step,
statistical evaluation is used in order to answer the query with the help of the
data. The results of the analysis are then made comprehensible by means of
appropriate visualization. The latter is then evaluated by the domain experts
and used as support to conclude a course of actions. The sequence of the first
phase of the hybrid approach is depicted in Fig. 5.
The first phase of the hybrid approach are conducted in a cyclical manner.
This means that the output of a successfully executed use case, i.e., the gained
in-sights, may influence the input of the next use case, for example, through
the adaptation of a defined business goal or the definition of new ones. The
Data Driven Optimization 11
Fig. 5. The first phase of the hybrid approach (Steps in red color are conducted by the
business department, steps in grey color by the IT department) (Color figure online)
execution of the first phase of the hybrid approach calls for the collaboration
of the respective business departments and the IT department. In Fig. 5, the
steps marked in red color are to be conducted by the business department, while
the steps marked in grey color are the tasks of the IT department. To reach
maximum benefit, it is recommended to comply with the allocation of tasks, so
that each department can concentrate on its core expertise.
The executed use cases conduce to identify the parts of the data jungle, which
contain information with business value. As already mentioned, this data shall
be referred to as data treasure. Once a data treasure has been identified, it is
made available in a central data storage, for example, a data lake, (cf. Fig. 6).
This way, the central data storage, i.e. the data lake, will only contain data with
confirmed usefulness, and will expand with every conducted use case. The data
treasures of a specific use case, which show mutual correlation, are then assigned
to one cluster and should be considered as a coherent entity.
In the style of the bottom-up approach, the established entities are then exam-
ined in order to find correlations with each other or with further parameters from
different yet related use cases. By doing so, the analysis is carried out not only
within the boundaries of single use cases, but rather on a holistic level. Since
12 R. Ghabri et al.
Data storage
this step is likely to be sophisticated and costly, it should be ensured that the
efforts are well-invested. For that reason, the risk of trailing away with irrelevant
or pseudo-correlations needs to be minimized. We recommend this step of the
hybrid approach to adhere specific frame conditions in order to maintain the effi-
ciency. For instance, integrating and analyzing the data treasures can be carried
out for a specific period of time, a specific car model, or a specific manufacturing
technology.
5 Discussion
not call for a primary large investment. It rather favors gradual investments with
perceptible impact. Furthermore, the company is able to avoid committing to
a costly, sophisticated IT-solutions before thoroughly investigating the specific
circumstances. For these reasons the hybrid approach is in line with the require-
ment (R3) of achieving an optimal cost-benefit ratio. In terms of the requirement
of promoting the feasibility (R4), the hybrid approach is characterized by the
sensible, practical usage of data analysis in production environments. Due to
the sequence of its phases, the hybrid approach provides the company with the
opportunity to readjust its course of action in the manner of a loop control-
system. Moreover, the design of the hybrid approach allows to achieve quick-
wins, which accounts for a sense of achievement among the involved employees
and results in a higher motivation.
References
1. Khan, A., Turowski, K.: A survey of current challenges in manufacturing indus-
try and preparation for industry 4.0. In: Abraham, A., Kovalev, S., Tarassov, V.,
Snášel, V. (eds.) Proceedings of the First International Scientific Conference “Intel-
ligent Information Technologies for Industry” (IITI 2016). AISC, vol. 450, pp. 15–
26. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-33609-1 2
2. Jacob, F., Strube, G.: Why go global? The multinational imperative. In: Abele, E.,
Meyer, T., Näher, U., Strube, G., Sykes, R. (eds.) Global Production, pp. 2–33.
Springer, Heidelberg (2008). https://doi.org/10.1007/978-3-540-71653-2 1
3. Slack, N., Chambers, S., Johnston, R.: Operations Management. Pearson Educa-
tion, New York (2010)
4. Bryant, R., Katz, R.H., Lazowska, E.D.: Big-data computing: creating revolution-
ary breakthroughs in commerce, science and society (2008)
5. Raval, K.M.: Data mining techniques. Int. J. Adv. Res. Comput. Sci. Softw. Eng.
2(10) (2012)
14 R. Ghabri et al.
6. Couldry, N., Powell, A.: Big data from the bottom up. Big Data Soc. 1(2),
2053951714539277 (2014)
7. Lovelace, R.: The data revolution: big data, open data, data infrastructures and
their consequences, by rob kitchin. 2014. Thousand Oaks, California: Sage Publi-
cations. 222+XVII. ISBN: 978-1446287484. J. Reg. Sci. 56(4), 722–723 (2016)
8. McAfee, A., Brynjolfsson, E., Davenport, T.H., Patil, D., Barton, D.: Big data: the
management revolution. Harvard Bus. Rev. 90(10), 61–67 (2012)
9. Gartner: Data lake. Gartner IT Glossary (2017)
10. Held, J.: Will data lakes turn into data swamps or data reservoirs? (2014)
11. Han, J., Kamber, M., Pei, J.: Mining frequent patterns, associations, and correla-
tions. In: Data Mining: Concepts and Techniques, 2nd edn., pp. 227–283. Morgan
Kaufmann Publishers, San Francisco (2006)
12. Fayyad, U., Piatetsky-Shapiro, G., Smyth, P.: The KDD process for extracting
useful knowledge from volumes of data. Commun. ACM 39(11), 27–34 (1996)
13. Hirmer, P., Behringer, M.: FlexMash 2.0 - flexible modeling and execution of data
Mashups. In: Daniel, F., Gaedke, M. (eds.) RMC 2016. CCIS, vol. 696, pp. 10–29.
Springer, Cham (2017). https://doi.org/10.1007/978-3-319-53174-8 2
14. Hirmer, P., Wieland, M., Schwarz, H., Mitschang, B., Breitenbücher, U., Sáez,
S.G., Leymann, F.: Situation recognition and handling based on executing situation
templates and situation-aware workflows. Computing 99, 163–181 (2017)
15. Wieland, M., Hirmer, P., Steimle, F., Gröger, C., Mitschang, B., Rehder, E., Lucke,
D., Rahman, O.A., Bauernhansl, T.: Towards a rule-based manufacturing inte-
gration assistant. In: Westkämper, E., Bauernhansl, T. (eds.) Proceedings of the
49th CIRP Conference on Manufacturing Systems (CIRP-CMS 2016), Stuttgart,
Germany, 25–27 May 2016, Procedia CIRP, vol. 57, pp. 213–218. Elsevier, January
2017
16. Hermann, M., Pentek, T., Otto, B.: Design principles for industrie 4.0 scenarios.
In: 2016 49th Hawaii International Conference on System Sciences (HICSS), pp.
3928–3937, January 2016
17. Tönne, A.: Big Data is no longer equivalent to Hadoop in the industry. In: Proceed-
ings of 17. Datenbanksysteme für Business, Technologie und Web (BTW) (2017)
18. Hagerty, J.: 2017 planning guide for data and analytics (2016)
Human Perception of Enriched Topic
Models
1 Introduction
Topic modeling, similarly to text classification, is an established and thoroughly
researched field in computer science. Traditionally, both techniques are based on
a bag-of-words (BOW) document representation, where one feature corresponds
to one word (its stem or lemma), i.e. the word order doesn’t count, only the
frequencies. As Gabrilovich and Markovitch [3] describe the state of the art
in text classification in their 2005 paper, “after a decade of improvements, the
performance of the best document categorization systems became more or less
similar, and it appears as though a plateau has been reached [...]”. For this reason,
researchers started working on developing different approaches. Considering the
limitations of BOW model, the most natural idea was to enhance the method of
document representation.
Scott and Matwin [11] did one of the first efforts of feature vector engineer-
ing for the purpose of text classification by using WordNet, a lexical database
for english, and converting documents to feature vectors based on this new rep-
resentation. Recent papers, such as Garla and Brandt [4] and Zong et al. [14],
c Springer International Publishing AG, part of Springer Nature 2018
W. Abramowicz and A. Paschke (Eds.): BIS 2018, LNBIP 320, pp. 15–29, 2018.
https://doi.org/10.1007/978-3-319-93931-5_2
16 W. Lukasiewicz et al.
employ semantic information during the feature engineering step and apply
machine learning techniques to learn text categories.
These attempts inspired us to perform feature engineering in the context
of topic modeling. We want to incorporate semantic information in order to
extend the traditional bag-of-words approach into a novel bag-of-features app-
roach when preparing feature vectors. We plan to consider not only words but
also disambiguated Named Entities linked to DBpedia resources and several
related entities.
The underlying idea and motivation for our work is based on the fact that
topic modeling algorithms draw their information based on the frequencies
and co-occurrences of tokens in single documents and across the whole corpus.
Because of that, we formulated a hypothesis that, in thematically related docu-
ments, the entities and/or their types, hypernyms or categories of corresponding
Wikipedia articles should also be overlapping and thus summed frequencies of
these terms should be more meaningful and lift up their relevance in discovered
topics.
For example, consider a text snippet from a Spiegel Online1 article that a
human would assign a label “politics”: “Barack Obama is only passing through
Germany on his trip to Europe later this week and does not plan to hold substan-
tial talks with Angela Merkel. The White House views the chancellor as difficult
and Germany is increasingly being left out of the loop”. The word politics itself
has a zero frequency. But if we perform Named Entity Recognition and Dis-
ambiguation, the entities Barack Obama and Angela Merkel will be considered
politicians thanks to the enrichment we perform.
In this work we present an approach of mining topic models enriched with
background knowledge. We focus on the feature engineering aspect of topic mod-
eling and leave the underlying generative statistical model intact. We assess
the quality of this approach based on the evaluation strategy which consists of
inspecting the internal coherence of topics and the topic-document assignments
in terms of human understanding.
2 Related Work
In contrary to pure word-based LDA algorithm and its variations (such as differ-
ent sampling techniques or online learning proposed by Hoffman et al. [5] which
enables streaming-wise model mining and is thus much less resource-hungry) or
applications (Gibbs-sampling-based LDA for gene prediction [10]), topic mod-
eling approaches using enriched feature vectors have not been subject to much
research so far.
One of the first methods that contributes to topic modeling using entities
instead of words as features has been published by Newman et al. [8]. The
authors propose five new models that modify LDA in order to learn pure entity-
topic models. They evaluate them with regard to entity prediction and not their
information-theoretic value, e.g. by measuring perplexity.
1
http://www.spiegel.de/international/.
Human Perception of Enriched Topic Models 17
Computed log-likelihood can be used to compare topic models – the higher, the
better the model.
For LDA, to take the size of the vocabulary into account, we define perplexity
as the exponential of negative log-likelihood divided by the number of tokens
(note that the quality of the model increases while perplexity decreases):
−LL(D)
perplexity(D) = exp
#tokens
Wallach et al. [13] published an overview of evaluation methods for topic models.
They address certain challenges, such as the difficulty of estimating p(d|Φ, α) and
propose using sampling to overcome it.
18 W. Lukasiewicz et al.
shuffle and present to a human who is expected to select the intruder. In order to
quantitatively measure how well the topics match human understanding, Chang
et al. introduced model precision, i.e. to which extent the intruders selected by
humans correspond to the “real” intruders. Model precision is defined as the
fraction of correctly selected intruders (Eq. 1, where ws is the word selected by
the evaluation subject and w is the real intruding word), and thus ranges from
0 (worst) to 1 (best).
MP = 1(ws = w)/S (1)
The second task measures how understandable the topics are in terms of
assigning them to a given text. To prepare a topic intrusion question the authors
draw a random article and consider its topical coverage. They take three most
relevant topics and one irrelevant topic, each represented by its eight top words,
shuffle them and present to an evaluation subject to select the intruder. The
results for this task are evaluated using topic log odds. This measure, introduced
by Chang et al. measures how good were the guesses of humans. In topic intru-
sion task, every answer (a topic) has certain probability of generating a given
document. Topic log odds sums and normalizes differences of logs of probabil-
ities of real intruder belonging to the document and the intruder selected by
the evaluation subject. Intuitively, this way of evaluating makes a lot of sense,
since it doesn’t binarily count right/wrong answers, but works as a kind of error
function.
T LO = ( log θ̂d,∗ − log θ̂d,s )/S (2)
Simplifying the notation of Chang et al., Eq. 2 is the definition of topic log
odds, where θ̂d,∗ is the probability of the intruder topic belonging to the docu-
ment and θ̂d,s is the probability of the topic selected by the subject belonging to
the document. Because latter is greater or equal to the former, the topic model
in terms of T LO is better when T LO is higher (closer to 0).
3 Approach
Our approach differs in several ways from the state of the art methods using
entities in mining topic models. First of all, we do not modify the underlying
probabilistic generative model of LDA and can therefore apply our method on
any variation and implementation of the algorithm. Second of all, we mine topics
that contain named entities linked to a knowledge base and might be used for
knowledge acquisition purposes, e.g. taxonomy extraction or knowledge based
population. Moreover we only employ one KB – DBpedia and DBpedia Spotlight
as the NERD tool and focus on finding the best topic models in this setup.
Lastly, we combine two evaluation techniques – we concentrate on achieving low
perplexity, as well as measure human perception and interpretability of mined
models.
20 W. Lukasiewicz et al.
4 Evaluation
We evaluated our approach on three datasets: BBC2 , The New York Times
Annotated Corpus3 (NYT) and DBpedia Abstracts.
The BBC dataset is a collection of 2225 selected news articles from BBC news
website4 which correspond to stories from five domains (business, entertainment,
politics, sport and tech), published in 2004 and 2005.
NYT is a collection of over 1.8 million selected New York Times articles that
span for 20 years, from 1987 to 2007. Over 650 thousand of them have been
manually summarized by the library scientists, 1.5 million have been manually
tagged with regard to mentions of people, organizations, locations and topic
descriptors and over 250 thousand of them have been algorithmically tagged
and manually verified. Additionally Java utilities for parsing and processing the
corpus which is provided in XML format are included in the download. For
these reasons it is one of the most widely used datasets for approaches from the
domain of natural language processing. We didn’t take the full NYT dataset but
reduced it to over 46000 articles which have been pre-categorized into at least
one of following ten taxonomy classes: science, technology, arts, sports, business,
health, movies, education, u.s., world.
We already introduced the small and medium-sized datasets we used for the
evaluation. We chose one more dataset that can be categorized as big. Namely
it is the corpus of DBpedia abstracts – first paragraphs of Wikipedia arti-
cles extracted as free text. Abstracts are connected to the DBpedia resources
using the property abstract from the DBPedia ontology (http://dbpedia.org/
ontology). The dataset Long Abstracts is available to download from the DBpe-
dia downloads website5 .
After cleaning the documents and annotating the datasets (some articles
contain characters which cause Spotlight to fail) we end up with the numbers of
articles per data source displayed in Table 1.
Dataset Size
BBC 2224
NYT 46406
Abstracts 4641642
Table 2. Perplexity of topic models mined on the BBC dataset using different feature
types and numbers of clusters.
4.1 Results
We sent out the link to the questionnaire among fellow computer science/math-
ematics students, i.e. people with a bigger than average technical sense and
understanding of algorithmic and mathematical concepts. This fact might have
been reflected in final results. We closed the survey after receiving 600 answers.
It corresponds to 300 answers for each task type and thus 5 answers for each
single task. Our study had 10 participants, a similar number compared to 8 in
experiments by Chang et al.
We started the evaluation by assessing the model precision and topic log odds
overall, as well as for each dataset separately. Out of 300 entity intrusion answers,
260 were correct, which accounts for the model precision of 0.87. We achieved
the overall topic log odds of −4.23. The value itself is not very expressive, but we
see that the BBC dataset performs best. All topic log odds and model precision
values are presented in Table 5.
Human Perception of Enriched Topic Models 23
Table 3. Perplexity of topic models mined on the New York Times Annotated Corpus
dataset using different feature types and numbers of clusters.
Table 4. Perplexity of topic models mined on the DBpedia Abstracts dataset using
different feature types and numbers of clusters.
Table 5. Quality of estimated models (BBC 30, NYT 125 and Abstracts 1000) in
terms of model precision and topic log odds.
Table 6. Quality of estimated models in terms of model precision and topic log odds
for every feature combination.
Next, we calculated and evaluated the model precision separately for every
feature combination. The values can be found in Table 6 left. We were very happy
to notice that the standard, bag-of-words based model, has been outperformed
by seven enriched models, five of which being pure resource topic models. Also
the fact that both remaining models based on words occupy first two places
shows the meaningfulness of enriching topic models.
As a next step we took a look at topic log odds. Similarly to the case of
entity intrusion, also in topic intrusion the baseline (bag-of-words based model)
is outperformed. In this case only by five models out of nine but all of them
are pure resource topic models. The values can be found in Table 6 right. These
are very good outcomes that confirm our initial assumptions that incorporating
background knowledge into feature vectors brings additional context and makes
the topic model easier to understand.
Figure 1 depicts histograms of both used measures. Alongside the accumu-
lated values for mp and tlo, we presented example topics and documents account-
ing for certain values. We can see that the topic about technological companies
is easily separable from the entity London Stock Exchange, while University, as
an intruder in a topic about fish species is harder to spot and “infiltrates” the
topic better. On the right side, in the topic log odds histogram, we see that
the document about rugby and the protest for a bad referee is easy to decom-
pose into the topical mixture, while a DBpedia Abstract document about a 2007
Americas Cup sailing regatta is a very specific piece of text.
Apart from getting measurable, quantitative results from the intrusion-based
evaluation we also told the survey participants we would be happy to receive
feedback in a free form if they had some remarks about the tasks they saw. As
expected in such cases, only a few (three) decided to make an extra effort and use
this option. However, all three stated harmonically that often right answers to
Human Perception of Enriched Topic Models 25
Fig. 1. Histograms of the model precisions of all 30 topics (left) and topic log odds of
all 60 documents (right). On both histograms we presented several examples of topics
and documents for the corresponding values, mp and tlo, respectively.
the topic intrusion tasks were guessed not by the descriptiveness of the presented
text itself but rather by searching for an outlier within the group of four topics
with no regard to the text.
To illustrate this phenomenon on an example, let’s first take the following
topic intrusion task, generated from the model mined on NYT Annotated Corpus
dataset:
Of the 35 members in the Millrose Hall of Fame, all but one are renowned ath-
letes. The only exception is Fred Schmertz, the meet director for 41 years.
Now, Schmertz will be joined by his son, Howard, 81, who succeeded him and
directed the Millrose Games, the world’s most celebrated indoor track and field
meet, for 29 years, until 2003.
The 100th Millrose Games, to be held Friday night at Madison Square Garden,
will be the 73rd for Howard Schmertz, now the meet director emeritus. He will
be honored tomorrow night as this year’s only Hall of Fame inductee.
FRANK LITSKY
SPORTS BRIEFING: TRACK AND FIELD
The generated answers can be seen below. The intruder is marked.
– dbo:City, dbo:Settlement, dbo:AdministrativeRegion, dbo:Person, dbo:Office
Holder, dbo:Company, dbo:Town, dbo:EthnicGroup
– dbo:City, dbo:Settlement, dbo:AdministrativeRegion, dbo:Person, dbo:
University, dbo:Disease, dbo:Town, dbo:Magazine
– dbo:Person, dbo:Film, dbo:Settlement, dbo:City, dbo:Magazine, dbo:
Company, dbo:TelevisionShow, dbo:Newspaper
– dbo:Country, dbo:PoliticalParty, dbo:MilitaryUnit, dbo:
Settlement, dbo:Person, dbo:OfficeHolder, dbo:Weapon, dbo:
MilitaryConflict
Now let us explain the thought process that decided about selecting the right
answer. These are our observations confirmed by the survey participants. First,
26 W. Lukasiewicz et al.
we see that two first topics are very similar which makes it very unlikely that one
of them is an intruder. Several of third topic’s top words are present either in the
top words of the first (dbo:City, dbo:Company) or the second (dbo:Settlement,
dbo:City, dbo:Magazine) topic, plus the ontology class Newspaper seems related
to class Magazine in human understanding. Following this reasoning we conclude
that the last topic is the intruder as it contains entities related to politics/mil-
itary strategy. Unfortunately, this strategy is the only way to choose a correct
answer for this task because all four answers seem unrelated to the presented
document.
While evaluating and trying to understand the results, we came across
another topic intrusion task from the NYC dataset which is a perfect exam-
ple of how counter-intuitive and uninterpretable the topics can be:
BASEBALL
American League TEXAS RANGERS–Acquired RHP Brandon McCarthy and
OF David Paisano from the Chicago White Sox for RHP John Danks, RHP
Nick Masset and RHP Jacob Rasner.
National League MILWAUKEE BREWERS–Agreed to terms with RHP Jeff Sup-
pan on a four-year contract.
FOOTBALL
National Football League MINNESOTA VIKINGS–Agreed to terms with
DT Kevin Williams on a seven-year contract extension. PITTSBURGH
STEELERS–Signed LB Richard Seigler from the practice squad. Released WR
Walter Young.
HOCKEY
National Hockey League (...)
The possible answers can be seen below. Number of votes per answer can be
seen in brackets. The intruder is marked.
To our giant surprise, we noticed that the intruder is the last answer – one of two
topics that we would connect with sports at all. This example clearly shows what
a difficult task it is to interpret associations between documents and enriched
topics.
has not been subject to much open research so far. In our experimental setup
we showed that our approach outperforms the established bag-of-words based
models, i.e. the enrichment step we perform and injecting linked data while
generating the feature vectors makes the topic models more understandable for
humans. This is a very meaningful result since interpreting and labeling topics
is an interesting and popular research area. Having topics containing resources
linked to knowledge graphs would allow for completely new possibilities in this
domain.
Automatic topic labeling would not be the only research and application area
where we think enriched topic models could be used. We imagine scenarios where
we apply enriched topic modeling, attach huge text corpora to knowledge bases
and thus have systems to automatically “classify” unseen documents and to find
their places as mixtures of subsets of the knowledge graph.
Assessing the quality of hierarchical topic models was out of scope for this
work. However, if we define a measure for goodness of topical hierarchies and
manage to estimate enriched models that satisfy certain condition (e.g. perform
similarly well as their bag-of-words counterparts), we could attach mined topical
hierarchies to knowledge bases. This way we could conduct a fully automatic
taxonomy mining from arbitrary document collections and thus offer support
for experts in every domain.
Naturally, the evaluation didn’t bring only positive outcomes. We encoun-
tered some issues that hinder unlocking the full potential of our approach. First
of all, a vital obstacle while considering enriched models and pure resource mod-
els is the length of documents. This problem plays an even bigger role in our
approach than in pure word-based models. When texts are too short and thus
the NER systems fail to deliver a reasonable amount of resources, the infer-
encer is unable to predict the coverage proper in terms of human understanding
because the feature vector is far too sparse. To overcome this issue, we could
try to mine a model using Wikipedia dump, or at least a representative and
reasonable subset of it, to have a universal, multi-domain topic model. Running
a NER system, such as DBpedia Spotlight, on Wikipedia articles (and maybe
experimenting with tuning its parameters, such as support and confidence) could
result in much denser feature vectors of higher quality.
Second, we find that the fact that topic models mined using DBpedia
Abstracts perform worse might, apart from the average document length, be
influenced by the number of topics in the model. In this work we focused on the
aspects of feature engineering and feature selection, i.e. defining new features
from background knowledge and investigating which subset of them would per-
form best for a fixed number of topics. More patience when selecting k’s and
tailoring them better for every dataset should bring improvement in the quality
of topic models.
Another important thing to note is that we didn’t touch the topic modeling
algorithm itself. We only changed ways a document is represented as feature
vector. Digging deeper into LDA in order to differentiate between words and
28 W. Lukasiewicz et al.
References
1. Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent dirichlet allocation. J. Mach. Learn.
Res. 3, 993–1022 (2003)
2. Chang, J., Gerrish, S., Wang, C., Boyd-Graber, J.L., Blei, D.M.: Reading tea
leaves: how humans interpret topic models. In: Advances in Neural Information
Processing Systems, pp. 288–296 (2009)
3. Gabrilovich, E., Markovitch, S.: Feature generation for text categorization using
world knowledge. IJCAI 5, 1048–1053 (2005)
4. Garla, V.N., Brandt, C.: Ontology-guided feature engineering for clinical text clas-
sification. J. Biomed. Inf. 45(5), 992–998 (2012)
6
https://www.mturk.com/.
Human Perception of Enriched Topic Models 29
5. Hoffman, M., Blei, D.M., Bach, F.: Online learning for latent Dirichlet allocation.
In: Advances in Neural Information Processing Systems, vol. 23, pp. 856–864 (2010)
6. Hu, Z., Luo, G., Sachan, M., Xing, E., Nie, Z.: Grounding topic models with
knowledge bases (2016)
7. Mimno, D., Wallach, H.M., Talley, E., Leenders, M., McCallum, A.: Optimizing
semantic coherence in topic models. In: Proceedings of the Conference on Empirical
Methods in Natural Language Processing, pp. 262–272. Association for Computa-
tional Linguistics (2011)
8. Newman, D., Chemudugunta, C., Smyth, P.: Statistical entity-topic models. In:
Proceedings of the 12th ACM SIGKDD International Conference on Knowledge
Discovery and Data Mining, pp. 680–686. ACM (2006)
9. Newman, D., Lau, J.H., Grieser, K., Baldwin, T.: Automatic evaluation of topic
coherence. In: Human Language Technologies: The 2010 Annual Conference of the
North American Chapter of the Association for Computational Linguistics, pp.
100–108. Association for Computational Linguistics (2010)
10. Pinoli, P., Chicco, D., Masseroli, M.: Latent Dirichlet allocation based on Gibbs
sampling for gene function prediction. In: 2014 IEEE Conference on Computational
Intelligence in Bioinformatics and Computational Biology, pp. 1–8. IEEE (2014)
11. Scott, S., Matwin, S.: Feature engineering for text classification. ICML 99, 379–388
(1999)
12. Todor, A., Lukasiewicz, W., Athan, T., Paschke, A.: Enriching topic models with
DBpedia. In: Debruyne, C., et al. (eds.) OTM 2016. LNCS, vol. 10033, pp. 735–751.
Springer, Cham (2016). https://doi.org/10.1007/978-3-319-48472-3 46
13. Wallach, H.M., Murray, I., Salakhutdinov, R., Mimno, D.: Evaluation methods
for topic models. In: Proceedings of the 26th Annual International Conference on
Machine Learning, pp. 1105–1112. ACM (2009)
14. Zong, W., Feng, W., Chu, L.-K., Sculli, D.: A discriminative and semantic feature
selection method for text categorization. Int. J. Prod. Econ. 165, 215–222 (2015)
Predictive Quality: Towards a New
Understanding of Quality Assurance
Using Machine Learning Tools
1 Introduction
Product defects are dreaded by manufacturers. Not only do they come with sig-
nificant costs for replacement products, shipping and customer care staff, a high
rate of defects will also damage a manufacturer’s reputation. The longer it takes
to detect a defect in the development and production process, the more expen-
sive undoing the resulting damage becomes (Fig. 1, orange line). At the same
time, quality management research [2] shows that most product failures can be
traced back to early decisions in the product design process where far-reaching
decisions are made (Fig. 1, yellow line). Unfortunately, negative consequences
do not show up until the product is actually produced (e.g., in quality checks)
or worse, until it is returned by the customer (Fig. 1, green line). To minimize
expenses, insights about failure rates have to be fed back from the quality assur-
ance and return processes into the design process. However, such a preventive
failure avoidance is not common [2]. Manufacturers try to minimize defect rates
in the production process but a causal connection to design flaws is not drawn,
partly because those are rarely obvious.
The original contribution of this paper is a reconsideration of what the ideal
properties of modern quality assurance should be and how those can be achieved
c Springer International Publishing AG, part of Springer Nature 2018
W. Abramowicz and A. Paschke (Eds.): BIS 2018, LNBIP 320, pp. 30–42, 2018.
https://doi.org/10.1007/978-3-319-93931-5_3
A New Understanding of Quality Assurance 31
Fig. 1. Defect causes and detection, according to DGQ Beratung GmbH [1]. (Color
figure online)
using machine learning tools. We do not only propose a concept but a complete
system architecture for Preventive Quality Assurance (PreQA). It comprises two
components, one for automated data analysis and another for real-time assis-
tance of the product development process. The analysis component (Sect. 4.2)
leverages data collected in quality assurance and the return process to iden-
tify correlations between product features and defect probabilities automatically
using supervised learning. The assistance component (Sect. 4.3) can augment
software such as design tools or ERP systems and inform their users in case an
increased risk for product flaws is detected, feeding back analysis results (Fig. 1,
blue arrows). Additionally, it offers help in resolving problems by suggesting
alternatives to achieve a balance between freedom of choice and product quality.
After reviewing previous work (Sect. 2), we introduce our concept of Pre-
ventive Quality Assurance, distinguishing it from the current understanding of
preventive and predictive quality (Sect. 3), and go on to propose an architecture
for its realization (Sect. 4). By a case study in clothing industry (Sect. 5) we
demonstrate the potential of PreQA even in creative industries with their vast
freedom of decision. We conclude with an outlook towards future work on the
topic (Sect. 6).
2 Previous Work
which failure modes are identified and rated regarding their severity, probability
of occurrence and probability of detection. The product of these ratings is called
Risk Priority Number (RPN) and used to order the failure modes in terms of
urgency. Another example is Fault Tree Analysis (FTA) [6] where the condi-
tions for a particular fault to occur are mapped to variables in a logic formula.
The value of the expression then indicates whether a particular configuration
of conditions may cause the fault or not. Both FMEA and FTA are based on
experience and subjective judgment of the people involved which makes them
time consuming and unreliable. We instead perform a fully autonomous analysis
which is unbiased and does not require the intervention of human experts.
Statistical process control (SPC) [7] is a semi-automatic way to realize quality
control using statistics. After manual identification of quality-relevant quantities,
those are monitored, usually in an automated fashion, using tools such as con-
trol charts or histograms. This way, the variation of a manufacturing processes
can be tracked and its violation of defined bounds can be noticed easily. Going
further, the reactions following steps undertaken to decrease variability become
immediately visible. As a whole, the tools of SPC can therefore aid an iterative
quality refinement process but still require a lot of human interaction, unlike the
preventive quality assurance system proposed in this paper.
Due to its manual nature, preventive quality management is not yet common.
In particular not in creative industries such as the clothing industry. According
to a recent study [2], 74% of textile manufacturers do not use any kind of quality
management methods when developing new products. The intent of PreQA is to
change this by automation and simplification, offering a way to achieve quality
assurance even in trades with a diverse and creative development process.
So far, little approaches exist to support product design by software-based
assistance systems. Schulte and Pape [8] propose an assistance tool embedded
into the CAD system used by a designer supporting her in complying with a
formal definition of the design process. While such an approach might increase
overall efficiency, it does not offer recommendations how to prevent product
failures based on actual data. Dienst et al. [9] present a prototypical system which
analyzes information collected throughout the life cycle of a product in order
to improve future generations of the same product, e.g., by replacing machine
parts known to be error-prone by higher quality parts of the same type. As such,
the system can only be used for iterative improvement of a specific product,
while PreQA offers predictions for entirely new product designs which may differ
drastically from any existing product for which data has been collected.
In some industrial branches, an iterative quality improvement can be achieved
using control systems. At the core of a (closed-loop) control system, a controller
tries to adjust the input to a pre-defined system such that its output approaches
an ideal value. The challenge lies in simultaneously accounting for the effects
that the system has on the controller itself. In case the controlled system is a
manufacturing device and the output value is directly related to quality of the
produced item, e.g., in wafer production [10], this yields a production environ-
ment which autonomously and rapidly stabilizes the quality of its output at a
A New Understanding of Quality Assurance 33
high level. The scope of such approaches is limited to selected areas and cannot
easily be transferred, for example to clothing industry (Sect. 5) where the input
to the manufacturing process is high-dimensional and the delay between design
decisions (control actions) and feedback (returns) can be a year or even longer.
Fig. 2. A sample generic hierarchy of possible defects that might arise in a product.
Analysis Flow. The upper part of Fig. 3 shows the analysis flow. In a pre-
liminary step, structured input is pre-processed (Fig. 3a) and transformed to
a representation (Fig. 3b) which allows sampling of vectors (p̄, d) ¯ as defined
above. Unstructured data associated with individual samples of structured data
is mined for details on the type and location of the respective defect (Fig. 3c).
If the identified defect d is more specific than the known d from the structured
data, the latter is overridden (Fig. 3d).
The training of new meta classifiers may be initiated by two events. When
initializing the system, we may train a base set of meta classifiers, e.g., for all
feature selections {i} with i ∈ {1, · · · , nfeat } containing just one feature (Fig. 3e).
Second, the assistance system may schedule new meta classifiers, too, if a good
classifier for a particular input product is missing (Fig. 3f). When a new meta
classifier mI is requested we train the corresponding sub-classifiers (Fig. 3g). To
train cI,d , we need relevant training and test data and, for this, select samples
¯ with p̄ ∈ SJ , I ⊆ J and d¯ ∈ {d ≤ d} ∪˙ {none}, i.e., samples where the
(p̄, d)
product information covers all features in I and the defect is at least as specific.
After mI has been trained, we rate its quality (Fig. 3h), e.g., by the average of
the area under the receiver operating characteristic curve (ROC AUC) metrics
[12] for all sub-classifiers, and store mI together with its rating r (mI ) (Fig. 3i).
To provide a justifying evidence for warnings issued by the assistance
(Sect. 4.1), we use a k-nearest neighbor (knn) query in the set of defective sam-
ples selecting the k most similar existing products for which a defect previously
occurred. Fast knn searches require appropriate data structures like k-d trees
[13]. As each search is specific to one sub-classifier—only neighbors regarding
36 O. Nalbach et al.
the specific feature subset and with a relevant defect are considered—we trig-
ger the construction of these data structures after training a new meta classifier
(Fig. 3j).
which we expect to be the most reliable one (Fig. 3k). When mI uses only a
subset of features (J = I), we schedule training of mI in the background (Fig. 3f)
to enable better predictions in the future. Then, we apply all sub-classifiers in
mI to p to estimate the probability of individual defects (Fig. 3l) and perform
a thresholding (Fig. 3m) which filters out unlikely defects to not bury users of
the system in warnings. If no defect passes the thresholding the response to the
query is empty. Otherwise pairs of defects and their estimated probabilities are
returned (Fig. 3n) together with the k most similar products determined by the
knn searches associated with the relevant sub-classifiers (Fig. 3o+p).
sizing, and the defect that occurred. The specialized defect catalog is defined
manually as the ERP system does not define a hierarchy itself. In total, we deal
with a dataset containing 120k product variants and information about 150k
return transactions.
Fig. 4. Comparison of classification methods regarding time (a) and memory con-
sumption (b) and of MLP layouts (c+d). Different methods (a+b) were compared on
a typical feature subset, while MLP layouts were compared on the full feature set to
estimate an upper bound on the necessary complexity.
to existing mixes. Instead, we parse the strings and add real-valued features in
the range [0, 1] encoding the contribution of each material (Fig. 5b → c). In the
last step, we label-encode remaining nominal features (Fig. 5c → d) using the
LabelEncoder class provided by sci-kit learn [14].
Training Details. We train our MLP classifiers using the Adam algorithm [17],
applying a log-loss. Training is automatically terminated once the loss does not
change by at least 5 · 10−4 for two consecutive iterations. The sub-classifiers are
trained in parallel worker threads. As an example, training a typical meta clas-
sifier including 15 sub-classifiers on 170-element inputs (after one-hot encoding)
requires about 50 s in four parallel threads on an Intel Xeon Gold 5122 server
CPU.
Fig. 6. Analysis results comparing design alternatives. Left: two knitted garments of
different materials. A higher percentage of polyester reduces the risk of knitting defects
at the cost of increased pilling. Right: the same shirt produced in different countries
shows varying risks of stains and weaving flaws. As expected, there is no risk of defect
zippers as they are not common in shirts. Note that probabilities are overestimated
because we typically down-sample the significantly larger class of flawless items by a
factor of about 100 due to performance reasons.
A New Understanding of Quality Assurance 41
References
1. FMEA - Fehlermöglichkeits- und Einflussanalyse. Whitepaper, Deutsche
Gesellschaft für Qualität (2012)
2. Morlock, S.: Qualitätsgesicherte Produktentwicklung in der Bekleidungsin-
dustrie (2013). https://www.hohenstein.de/de/inline/pressrelease 45248.xhtml?
excludeId=45248
3. Harding, J.A., Shahbaz, M., Srinivas, S., Kusiak, A.: Data mining in manufactur-
ing: a review. Manuf. Sci. Eng. 128, 969–976 (2006)
4. Veronesi, L., Kifonti, R., Cattaneo, G.M.: Data-driven innovation in the manufac-
turing industry. Eur. Data Market Study (2014)
5. Stamatis, D.H.: Failure Mode and Effect Analysis: FMEA from Theory to Execu-
tion. ASQ Quality Press, Milwaukee (2003)
6. Ericson, C.A.: Fault tree analysis. In: Hazard Analysis Techniques for System
Safety, pp. 183–221 (2005)
7. Oakland, J.S.: Statistical Process Control. Routledge, New York (2007)
8. Schulte, R., Pape, D.: Advanced digital mock-up based on the usage of assistances.
Des. Principles Pract. 5, 195–206 (2011)
9. Dienst, S., Fathi, M., Abramovici, M., Lindner, A.: Development of a knowledge-
based feedback assistance system of product use information for product improve-
ment. Int. J. Prod. Dev. 19, 191–210 (2014)
10. van der Meulen, S.H., Tousain, R.L., Bosgra, O.H.: Fixed structure feedforward
controller design exploiting iterative trials: application to a wafer stage and a desk-
top printer. Dyn. Syst. Meas. Control 130(5), 051006 (2008)
11. Quality management systems - fundamentals and vocabulary. Standard, Interna-
tional Organization for Standardization (2015)
12. Bradley, A.P.: The use of the area under the ROC curve in the evaluation of
machine learning algorithms. Pattern Recogn. 30(7), 1145–1159 (1997)
42 O. Nalbach et al.
13. Bentley, J.L.: Multidimensional binary search trees used for associative searching.
Commun. ACM 18(9), 509–517 (1975)
14. scikit-learn. http://scikit-learn.org
15. Python data analysis library (pandas). https://pandas.pydata.org
16. Dal Pozzolo, A., Caelen, O., Johnson, R.A., Bontempi, G.: Calibrating probability
with undersampling for unbalanced classification. In: 2015 IEEE Symposium Series
on Computational Intelligence, pp. 159–166 (2015)
17. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint
arXiv:1412.6980 (2014)
18. Javascript object notation (JSON). https://www.json.org
Business and Enterprise Modelling
Application of Inductive Reference Modeling
Approaches to Enterprise Architecture Models
1 Introduction
Enterprises need to be aware of the relations among their strategy, processes, appli-
cations and infrastructures to be able to rapidly react on changing demands in the
market and within their organization. The Enterprise Architecture (EA) research
domain contributes to this purpose by providing methods and tools to establish a more
holistic perspective on enterprises [1, 2]. This includes to systematically capture and
develop an EA using modeling languages like ArchiMate [3]. Such models represent
different architectural layers of an enterprise, such as business, application and tech-
nology architecture. Since EA projects are highly time- and resource-consuming,
organizations would benefit from reference models for EAM. A Reference Enterprise
Architecture (R-EA) can be defined as a generic EA for a class of enterprises that is
used as foundation in the design and realization of the concrete EA [4].
Many methods exist to develop reference models—both deductive and inductive
approaches [5]. Especially, the latter one gains increasing attention by the research
community in recent years [6]. Unfortunately, inductive approaches focus primarily on
business process model structures [7] and research lacks investigating their applica-
bility towards EA structure like ArchiMate [8]. Nevertheless, we claim that the benefits
inductive methods provide for reference process development will also qualify for R-
EA development once these methods can be adapted.
In order to fill this research gap, we propose a method for inductive derivation of a
R-EA using approaches of existing inductive methods. We therefore use design science
research (Sect. 2) in the domain of inductive reference modeling (Sect. 3). We perform
a literature review on current inductive methods and assess them regarding their
applicability to EA modeling against prior defined requirements (Sect. 4). We further
refine one suitable approach to the EA domain (Sect. 5) and demonstrate the new
method by dint of a use case (Sect. 6).
2 Research Design
In general, reference models are information models developed for a certain problem in
a certain application domain. The purpose of their development is to be reused in a
concrete application case in this domain. The reuse of a reference model is intended to
increase both efficiency and effectivity of an enterprise’s information systems and their
change management [12]. From a user-oriented perspective, Thomas understands a
reference model as a model used to support the construction of another model [13].
From the perspective of reusability, other authors such as vom Brocke argue that
reference models are characterized by the concepts of universality and recommendation
Application of Inductive Reference Modeling Approaches to EA Models 47
[12]. The life cycle of such reference models can be distinguished between the phase of
construction and the phase of application [5, 14]. By presenting insights in reference
enterprise architecture development we contribute to the first phase, i.e. the construc-
tion of reference models.
For constructing reference models, research discusses two generic strategies. While
the deductive reference modeling derives reference models from generally accepted
knowledge, the inductive approach abstracts from individual models to agree on a
common understanding within the reference model [15]. Most established reference
models have been developed based on deductive approaches [16]. However, inductive
reference modeling provides potential because more and more relevant data in terms of
logs and concrete information models of organizations are available. Further, induc-
tively developed reference models tend to have a higher degree of detail, are more
mature and seem to be more accepted when it comes to reference model application [6].
Fettke provides a seven phase method for inductive reference model construction
[17]. The core of the approach is the derivation of a reference model from presorted
individual information models. Although Fettke names clustering as a technique to
elicit the reference model, it is not made explicit how it is conducted. In recent years,
reference modeling related research addresses this topic and approaches were devel-
oped. One work applies abstraction techniques from business process mining to the
reference model domain [7]. Other approaches utilize for example natural language
processing techniques [18], graph theory [19] or clustering methods [20], but always
focus on business process model structures.
are related to EA stakeholders and group distinct interests on the EA model; archi-
tecture principles that guide the EA model; an EA model, which relates different
elements of the EA with each other; views that represent a projection of the EA model
regarding a certain concern; and, a framework that provides a meta model and mod-
eling guidelines. Further, EA models result in more complex model structures. For
example, event-driven process chains (EPCs) define six elements and three relation
types while ArchiMate’s core framework defines 36 elements and 12 relation types. In
summary, business process models can be interpreted as a partial model of EA models,
since they can be integrated into an EA as Lankhorst et al. demand [2].
REQ1: The Approach Should Be Suitable to Other Modeling Languages. The
authors of the approach have to mention the applicability to other modeling languages.
This might be other process modeling languages or any modeling language. We dis-
miss the approach if the authors state that it is solely applicable to the utilized model
structure. Otherwise, we concluded this by studying the approach in detail.
REQ2: The Approach Does Not Solely Rely on the Control Flow of the Model
Structure. Most inductive approaches focus on the control flow characteristic of
business process languages like ECP to derive a reference model. Next to such dynamic
relationships, EA models use many other relationship types like the assignment
(structural), serving (dependency) or specialization relationship. Thus, we dismiss
approaches that are not applicable to static model structures.
REQ3: The Approach Has to Be Adjustable Regarding Concepts of Viewpoint
and Concerns of ArchiMate. EA models focus on a more aggregated level of detail
in contrast to business process models, which represent a much higher level of detail
[2]. ArchiMate uses different viewpoints to address different concerns of the model’s
stakeholders [22]. The approach may use mechanisms to vary the decision what ele-
ments to integrate into the reference model depending on a viewpoint’s concern. For
example, more criteria than a frequency threshold could be used. This also might
support the identification of best practices among the individual models.
We consider any approach that fulfills all three requirements suitable for the
application of R-EA development. However, we do not directly dismiss approaches
fulfilling REQ1 and REQ2 but not REQ3, because it may still be applicable to sub-
models of the ArchiMate models.
title, abstracts and keywords and we used a NEAR-operator in order to gather result
like “reference process models”. Furthermore, we included the term “mining” since
related work often use it instead of “inductive” [23]. The query was performed on the
databases SCOPUS (296 results), AISeL (20 results) and IEEExplore (94 results) from
2007 until today. This period was chosen since prior work in [6] already conducted a
similar search, which we wanted to extend with the most recent approaches. The
authors only identified inductive approaches after 2007 in their work. Then, we
excluded irrelevant articles by reading all titles and abstracts. This resulted in 16
results, from which we eliminated three duplicates and then reviewed the 13 full
documents. We included every article in English and German language that explicitly
addressed inductive reference modeling and propose an approach for reference model
derivation. We excluded work that offered techniques, e.g. from the process mining
domain, but lacked in describing its application to reference model development.
Additionally, we used related literature reviews to include relevant articles that we did
not capture yet [6, 7, 23, 24]. In the end, we identified 21 approaches for inductive
reference modeling. We analyzed them against REQ1, REQ2 and REQ3. The results
are presented in Table 1.
Table 1. Identified approaches and their fulfilment of requirements. (“+” fulfils REQ; “−” does
not fulfil REQ; “o” may be applicable with major adjustments)
REQ1
REQ2
REQ3
REQ1
REQ2
REQ3
APPROACH DECISION APPROACH DECISION
remaining approaches were dismissed due to major adjustments one would have to
make using them for EA model structures. For example, some use graph theory to
derive reference models [32], while others use clustering algorithms [30] or transfor-
mation matrices for their approach [40]. Although it is possible to represent ArchiMate
models in graphs or matrices, the approaches’ effort highly increases with the com-
plexity of different relationships ArchiMate defines. We exclude them from the scope
of this work, but discuss their potential for future research at the end of this article.
Based on the defined requirements in Sect. 4.1 we chose the approach by Scholta
[27] for applying it to ArchiMate models. Scholta’s is the only approach that proposes a
method for inductively deriving a best practice reference model and therefore defines a
number of characteristics that also can be applied to ArchiMate models, which fulfils
REQ 3. Still, in a prior research paper we already successfully applied the approach by
Ardalani et al. [16] to ArchiMate models by manually calculating the minimal graph-
edit distances of AchiMate model parts. The same may be done for the approach by
Martens et al. [36]. The remainder of this paper depicts our application of Scholta’s
approach and demonstrates the results by dint of a use case.
users to solve this problem in the best way. Moreover, it has to be clarified whether the
available source models offer the required information. The ranking of element groups
may use multiple criteria. For different viewpoints, different criteria may be applied.
Step 5: Assemble Reference Model. After all reference models are developed, the
reference viewpoints are integrated into the final reference model. There may exist
duplicates, since some elements or relations are defined in various reference
viewpoints.
deduced during the interview design—each interviewee verified it. These functions
were namely “Building an Organizational Framework”, “Corporate Hazard Analysis”,
“Implementation of a Prevention System” and “Monitoring”.
In Step 2, we identified common elements and relations in the viewpoint. After a
CSV-Export we defined a threshold of 80% and identified 41 common business
functions, and five common business roles. We argue that using this threshold still
enables to capture commonalities, while we can ignore outliers that may occur due to
the interview as a model elicitation method. We further assumed that elements with the
same semantic were labeled with the same syntax because we used the same model
element library when modeling the source models. During conducting Step 3, we
decided to group the source viewpoint’s models by dint of the four generic business
functions identified during Step 1 (see Fig. 1). Group G3 splits into two groups, namely
“G3.1 Measures for Internal Hazards” and “G3.2 Measures for External Hazards”. The
intention was to gather not only the common business function related to the four
management functions, but also to accompany them with identified best practices from
the source viewpoints. To do this evaluation in Step 4 we found it most reasonable to
use a criterion that identifies the group, which contains most aspects mentioned by the
Federal Financial Supervisory Authority in [42]. In the case that multiple groups from
different source viewpoints met this criterion, we integrated the group that contained
the most business functions. During Step 5 we integrated the groups from the prior
steps and assembled the reference viewpoint.
Figure 1 depicts an extract of the developed reference viewpoint for implementing
a prevention system for other criminal acts on a strategic level. The complete view has
96 business function elements, which are assigned to eleven distinct business roles via
In this article we address the research gap that the reference modeling discipline lacks
an inductive approach tailored for EA models. By applying a DSR process we therefore
define requirements that EA models have towards an inductive derivation of reference
enterprise architecture (R-EA) based on multiple individual EA models. We apply
these to existing inductive methods that we identified during a systematic literature
review, and which almost merely address business process structures. We identify the
approach by Sholta [27] to be the most appropriate for our purposes and refine it in
order to demonstrate its successful application in a use case for R-EA development.
Domain experts validated the resulting R-EA in a first workshop session. Nevertheless,
there are several open issues observed for both (i) the developed artefact in particular
and (ii) inductive approaches for R-EA development in general we identify for future
research.
First, the current approach is conducted manually using ArchiMate modeling tools
and spreadsheets. Once the individual EA models meet the input requirements of Step
0, the following steps may be automatized—including the grouping of Step 3. Second,
the best practice elicitation of Step 4 needs to be investigated in more detail. Although
the experts in the validation phase considered our chosen criteria as plausible, we see
flaws in it. Since the individual models were gathered before, may have been certain
information we could have inquired, if we would have known what criteria we would
use in Step 4 of the approach. Further, the concept of best practice is very vague and
should be investigated in much more detail for appropriate criteria selection. Last,
when automatizing the approach, it may be interesting to move from a viewpoint-based
approach to a model library based approach. This would allow deriving the reference
model at once and preventing the effortful model integration.
From a general perspective towards the inductive development of a R-EA, some of
the ideas and algorithms of the identified approaches from the literature review should
still be considered for further adaptation. First, EA models could be transformed into
graph structures in order to automate the merging step; as done in [30, 39]. Second,
clustering methods could be applied in order to identify commonalities beyond single
EA elements but also sub-models; as done in [18, 39]. Third and last, natural language
process could help to analyze syntactical and semantical similarities of EA elements,
that are not labeled similarly, but describe the same phenomenon [30, 32].
Application of Inductive Reference Modeling Approaches to EA Models 55
References
1. Ahlemann, F., Stettiner, E., Messerschmidt, M.: Strategic enterprise architecture manage-
ment. In: Ahlemann, F., Stettiner, E., Messerschmidt, M., Legner, C. (eds.) Challenges, Best
Practices, and Future Developments. Springer, Heidelberg (2012). https://doi.org/10.1007/
978-3-642-24223-6
2. Lankhorst, M.: Enterprise architecture at work. In: Modelling, Communication and Analysis.
Springer, Heidelberg (2017). https://doi.org/10.1007/978-3-642-29651-2
3. ArchiMate 3.0 Specification: Open Group Standard (2016)
4. ten Harmsen van der Beek, W., Trienekens, J., Grefen, P.: The application of enterprise
reference architecture in the financial industry. In: Aier, S., Ekstedt, M., Matthes, F., Proper,
E., Sanz, Jorge L. (eds.) PRET/TEAR -2012. LNBIP, vol. 131, pp. 93–110. Springer,
Heidelberg (2012). https://doi.org/10.1007/978-3-642-34163-2_6
5. Fettke, P., Loos, P.: Referenzmodellierungsforschung. Wirtschaftsinf 46, 331–340 (2004)
6. Rehse, J.-R., Hake, P., Fettke, P., Loos, P.: Inductive reference model development: recent
results and current challenges. In: Mayr, H.C., Pinzger, M. (eds.) INFORMATIK 2016, 26–
30 September 2016, Klagenfurt, Austria, vol. P-259. GI, Bonn (2016)
7. Rehse, J.-R., Fettke, P., Loos, P.: Eine Untersuchung der Potentiale automatisierter
Abstraktionsansätze für Geschäftsprozessmodelle im Hinblick auf die induktive Entwick-
lung von Referenzprozessmodellen. In: Alt, R., Franczyk, B. (eds.) WI2013, Leipzig,
Germany (2013)
8. Timm, F., Sandkuhl, K., Fellmann, M.: Towards a method for developing reference
enterprise architecture. In: Leimeister, J.M., Brenner, W. (eds.) WI2017, St. Gallen (2017)
9. Peffers, K., Tuunanen, T., Rothenberger, M.A., Chatterjee, S.: A design science research
methodology for information systems research. J. Manag. Inf. Syst. 24, 45–77 (2007)
10. Kitchenham, B., Pretorius, R., Budgen, D., Pearl Brereton, O., Turner, M., Niazi, M.,
Linkman, S.: Systematic literature reviews in software engineering – a tertiary study. Inf.
Softw. Technol. 52, 792–805 (2010)
11. Webster, J., Watson, R.T.: Analyzing the past to prepare for the future: writing a literature
review. MIS Q. 26, 13–23 (2002)
12. Vom Brocke, J.: Referenzmodellierung. Gestaltung und Verteilung von Konstruktion-
sprozessen. Logos, Berlin (2003)
13. Thomas, O.: Understanding the term reference model in information systems research:
history, literature analysis and explanation. In: Bussler, C.J., Haller, A. (eds.) BPM 2005.
LNCS, vol. 3812, pp. 484–496. Springer, Heidelberg (2006). https://doi.org/10.1007/
11678564_45
14. Thomas, O.: Management von Referenzmodellen. Entwurf und Realisierung eines
Informationssystems zur Entwicklung und Anwendung von Referenzmodellen. Logos-
Verl., Berlin (2006)
15. Becker, J., Schütte, R.: Referenz-Informationsmodelle für den Handel: Begriff, Nutzen und
Empfehlungen für die Gestaltung und unternehmensspezifische Adaption von Referenz-
modellen. In: Krallmann, H. (ed.) Wirtschaftsinformatik 1997, Physica-Verlag HD (1997)
16. Ardalani, P., Houy, C., Fettke, P., Loos, P.: Towards a minimal cost of change approach for
inductive reference model development. In: ECIS (2013)
17. Fettke, P.: Eine Methode zur induktiven Entwicklung von Referenzmodellen. In: Tagungs-
band Multikonferenz Wirtschaftsinformatik 2014, MKWI 2014, pp. 1034–1047 (2014)
56 F. Timm et al.
18. Ling, J., Zhang, L.: Generating hierarchical reference process model using fragments
clustering. In: Asia-Pacific Software Engineering Conference, vol. 2016, May 2016
19. Li, C., Reichert, M., Wombacher, A.: Discovering reference models by mining process
variants using a heuristic approach. In: Dayal, U., Eder, J., Koehler, J., Reijers, H.A. (eds.)
BPM 2009. LNCS, vol. 5701, pp. 344–362. Springer, Heidelberg (2009). https://doi.org/10.
1007/978-3-642-03848-8_23
20. Fettke, P.: Integration von Prozessmodellen im Großen: Konzept, Methode und experi-
mentelle Anwendungen. In: Thomas, O., Teuteberg, F. (eds.) Proceedings der 12.
Internationalen Tagung Wirtschaftsinformatik (WI 2015), pp. 453–467. Osnabrück (2015)
21. TOGAF Version 9. Van Haren Publishing, Zaltbommel (2010)
22. Greefhorst, D., Proper, E.: Architecture Principles: The Cornerstones of Enterprise
Architecture. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-20279-7
23. Rehse, J.R., Fettke, P.: Towards situational reference model mining - main idea, procedure
model & case study. In: Leimeister, J.M., Brenner, W. (eds.) Proceedings der 13.
Internationalen Tagung Wirtschaftsinformatik (WI2017), pp. 271–285. St. Gallen (2017)
24. Schoknecht, A., Thaler, T., Fettke, P., Oberweis, A., Laue, R.: Similarity of business process
models—a state-of-the-art analysis. ACM Comput. Surv. 50, 1–33 (2017)
25. Rehse, J.R., Fettke, P., Loos, P.: An execution-semantic approach to inductive reference
model development. In: 24th ECIS (2016)
26. Li, C., Reichert, M., Wombacher, A.: Discovering reference process models by mining
process variants. In: Proceedings of the IEEE International Conference on Web Services
(2008)
27. Scholta, H.: Semi-automatic inductive derivation of reference process models that represent
best practices in public administrations. In: 24th ECIS 2016 (2016)
28. Leng, J., Jiang, P.: Granular computing–based development of service process reference
models in social manufacturing contexts. Concurrent Eng. 25, 95–107 (2016)
29. Li, C., Reichert, M., Wombacher, A.: Mining business process variants. challenges,
scenarios, algorithms. Data Knowl. Eng. 70, 409–434 (2011)
30. La Rosa, M., Dumas, M., Uba, R., Dijkman, R.: Business process model merging: an
approach to business process consolidation. ACM Trans. Softw. Eng. Methodol. 22, 9
(2013)
31. Rehse, J.-R., Fettke, P.: Mining reference process models from large instance data. In:
Dumas, M., Fantinato, M. (eds.) BPM 2016. LNBIP, vol. 281, pp. 11–22. Springer, Cham
(2017). https://doi.org/10.1007/978-3-319-58457-7_1
32. Yahya, B.N., Bae, H., Bae, J.: Generating valid reference process model using genetic
algorithm. Int. J. Innov. Comput. Inf. Control 8(2), 1463–1477 (2012)
33. Gottschalk, F., van der Aalst, Wil M.P., Jansen-Vullers, M.H.: Mining reference process
models and their configurations. In: Meersman, R., Tari, Z., Herrero, P. (eds.) OTM 2008.
LNCS, vol. 5333, pp. 263–272. Springer, Heidelberg (2008). https://doi.org/10.1007/978-3-
540-88875-8_47
34. Rehse, J.R., Fettke, P., Loos, P.: A graph-theoretic method for the inductive development of
reference process models. Softw. Syst. Model. 16, 833–873 (2017)
35. Aier, S., Fichter, M., Fischer, C.: Referenzprozesse empirisch bestimmen. Von Vielfalt zu
Standards. Wirtschaftsinformatik & Management: WuM: die Praktikerzeitschrift für
Wirtschaftsinformatiker 3, 14–22 (2011)
36. Martens, A., Fettke, P., Loos, P.: A genetic algorithm for the inductive derivation of
reference models using minimal graph-edit distance applied to real-world business process
data. Tagungsband MKWI 2014, 1613–1626 (2014)
Application of Inductive Reference Modeling Approaches to EA Models 57
37. Li, C., Reichert, M., Wombacher, A.: The MinAdept clustering approach for discovering
reference process models out of process variants. Int. J. Coop. Info. Syst. 19, 159–203
(2010)
38. Sonntag, A., Fettke, P., Loos, P.: Inductive reference modelling based on simulated social
collaboration. In: Leimeister, J.M., Brenner, W. (eds.) WI2017. St. Gallen (2017)
39. Yahya, B.N., Bae, H.: Generating reference business process model using heuristic approach
based on activity proximity. In: Smart Innovation, Systems and Technologies, vol. 10 (2011)
40. Martens, A., Fettke, P., Loos, P.: Inductive development of reference process models based
on factor analysis. In: Thomas, O., Teuteberg, F. (eds.) Proceedings der 12. Internationalen
Tagung Wirtschaftsinformatik (WI 2015), pp. 438–452. Osnabrück (2015)
41. Gesetz über das Kreditwesen. KWG (2017)
42. Bundesanstalt für Finanzdienstleistungsaufsicht (ed.): Auslegungs- und Anwendungshin-
weise zu § 25c KWG (“sonstige strafbare Handlungen”) (2011)
Towards a Typology of Approaches
for Sustainability-Oriented Business
Model Evaluation
Abstract. While numerous studies related to business models are dealing with
understanding and representing businesses, limited research focus on the eval-
uation aspect. However, due to the highly dynamic environment, changing
availability of resources and booming digitalization, the evaluation of such
models and alternatives is an essential task. Analysing, evaluating and inno-
vating businesses is a multidimensional issue, which poses challenges that need
to be facilitated. Thus, in this study, we aim to explore approaches and criteria
for sustainability-oriented business model evaluation. Based on an extensive
literature review, we obtained 50 articles and determined more than 40 different
evaluation methods.
1 Introduction
The rapid deterioration of the natural environment as well as concerns over wealth
disparity and corporate social responsibility present fundamental issues for our entire
society [1]. Thus, ‘sustainability’ has increasingly gained importance in business
research and practice [2, 3] and, although companies often pay attention to be eco-
nomically competitive and effective, there is also an emerging stream that respects
ecological and social issues, for example by trying to achieve cleaner production
processes [4]. To foster this, approaches that support changes in behaviour and practice
are required.
Here, methods and tools from business model research can contribute, which are
usually applied to visualize, innovate and evaluate business models [5]. Although the
evaluation of business models including their configuration options is emphasized by
various authors (e.g., [4–6]), current research mainly addresses the representation [8–
10], the categorization (e.g., via taxonomies) and the understanding or the relevance of
such models [11–13]. However, evaluation is a fundamental aspect to facilitate fast
decisions, which are essential today because of the increasingly dynamic and uncertain
environment (e.g., through changing availability of resources, growing demands,
incremental pollution or booming digitalization). Evaluation approaches should support
such fast and high quality decision-making and the adjustment of business strategies,
for example to manage risks and uncertainty [14], analyse consequences of options
[6, 15], or measure the performance of a business [16] also in respect sustainability.
Sustainability is usually divided into three dimensions, for example Triple Bottom
Line (economic, ecological and social) [17]. According to the Brundtland report, it is
the “development that meets the needs of the present without compromising the ability
of future generations” [1]. Nonetheless, prior studies regarding business model eval-
uation tend to deal with economic-oriented assessment [18, 19] in term of aspects such
as efficacy, robustness, profitability and feasibility [6, 11, 20–22]. Hence, we currently
lack of approaches that respect ecological and social issues (e.g., emissions or impact
on the society). Moreover, existing approaches are varying and are often focused on
separated aspects because of different disciplinary backgrounds of the authors like
management, accounting or IT [23]. Thus, we need an overview that structures existing
approaches regarding multidimensional assessment strategies. This gap is problematic
because it inhibits the management to make informed decision (e.g., selection of
resources), manage their business model effects, improve their strategy in respect
sustainable criteria and allocate future efforts in the right way. From an IS/tool designer
perspective, it hinders to develop supporting software that facilitates these tasks.
Accordingly, this paper is guided by the following question: What are current
approaches and criteria that allow for sustainability-oriented business model
evaluation?
In order to answer this, we carry out an extensive literature review and develop a
typology of current methods, tools and criteria. In doing so, we aim to contribute to
both practice (e.g., providing tools for managing sustainability-oriented effects) and
research (e.g., design new tools and methods as well as derive advanced theories that,
for example attempt to explain how specific methods affect the performance of business
model evaluation). Following our research methodology (Sect. 3), we conduct a lit-
erature review, code the results and classify them in form of a typology. Next, we
present our main findings, which contribute to the sustainability-oriented assessment,
comparison, benchmarking of different business model aspects (Sect. 4). Finally, we
discuss results, provide implications and conclude with our paper (Sect. 5).
2 Background
3 Research Methodology
The primary goal of our study is to identify strategies for evaluating multidimensional
sustainability in business models. Therefore, we conduct a literature review [35].
Because the researcher has to document literature findings, the selection of keywords
and the evaluation of the results [36], we follow vom Brocke et al. [36] who allow for
Towards a Typology of Approaches 61
methodological rigorousness. The authors define five steps: definition of review scope
(Sect. 1), conceptualization of topic (Sect. 2), literature search (Sect. 4.1), literature
analysis and synthesis (Sects. 4.2 and 4.3), and research agenda (Sect. 5).
[55] ● – – ● – – ● – ● – ● – ● ●
[56] ● – – ● – ● – – ● – – – – ●
(continued)
63
64
Table 1. (continued)
#Ref. Sustainability Level Assessment Type of contribution
Economic Ecological Social Instance General Qualitative Quantitative Mixed Framework Construct Model Method Instantiation Measure
[57] ● – – – ● – – ● ● – – – – ●
[4] ● ● ● – ● – – ● ● – – ● – ●
[58] ● – – ● – – – ● – – – ● – –
[59] ● – – – ● – – ● – – – ● – ●
[60] ● – – – ● – – ● – – ● ● – –
[61] ● – – – ● – – ● ● – – – – ●
T. Schoormann et al.
[62] ● – – – ● ● – – ● – – – – –
[63] ● – – ● – ● – – – – – ● – –
[64] ● ● – ● – ● – – – – – – – ●
[65] ● ● – ● – – – ● – – – – – –
[66] ● – – ● – – – ● – – ● ● – ●
[67] ● – ● ● ● ● – – ● – ● – ● ●
[30] ● – – ● – ● – – – – ● – – –
[68] ● – – ● – ● – – ● – – – – ●
[15] ● – – ● – ● – – – – – ● – ●
[69] ● – – ● – – ● – ● – – – – –
[70] ● – – ● – – ● – ● – – ● – ●
[14] ● – – ● – – – ● ● – – ● – –
[71] ● – – ● – – ● – – – – ● – ●
[72] ● – – ● – – ● – – – – ● – ●
[73] ● – – ● – – – ● – – – ● – ●
[74] ● – – ● – – – ● – – ● – – ●
[75] ● ● ● – ● – ● – – – ● – – ●
[76] ● – – ● – – ● – – – ● ● – ●
[77] ● – ● ● – – ● – – – – ● – ●
Towards a Typology of Approaches 65
Frequency Sustainability
Type/Description/References
Ratio No. Econ. Ecol. Social
(I) Benchmark-, Comparison- and Trade Off-oriented Evaluation 64% 5
Compare/Identify Alternatives [6, 14, 15, 29, 30, 39, 40, 42–45, 47–49, 51, 61, 65, 66, 68, 70, 71, 74,
77], Benchmark [16], Qualitative Comparative Analysis [50], Trade Off Analysis [15, 65],
Collect/Value Impacts [18, 43, 44, 47, 66]
(II) Economic-/Financial-oriented Evaluation and Metrics 36% 14 - -
Return on Investment [38, 55], Life Cycle Costing [71], {Net Present Value [55, 71], Breakeven [48],
Market Share [48, 58, 74], Valuation [48], Scalability [74], Channel Recurring [74], Transaction Sunk
Cost [74], Profitability [52], Price-based Technical [58], Value-based Fundamental [58],
Technology Competitiveness [52], Competitive Advantage [74]} Analysis
(III) Mathematical-oriented Evaluation Methods 32% 4
Analytical Hierarchy Process [29, 38, 51, 53, 54, 69, 72, 76, 77], Fuzzy Comprehensive Evaluation
Method [51, 69, 72, 76], DEA-model [75], Sensitivity Analysis [38, 77]
(IV) Survey- and Questionnaire-oriented Evaluation 28% 4
Questionnaire [16, 38, 45, 47, 52, 54, 59, 65], Delphi Technique [70, 76], Group Decision [6, 18, 69],
Consensual Assessment Technique [45]
(V) Simulation-based Evaluation Modelling Techniques/Tools 25% 4 - -
Simulation [52, 54, 61, 66], System Dynamics [60], Stress Test [6],
Scenario Planning [6, 30, 39, 45–47, 52, 70]
(VI) Strategy-oriented Evaluation Tools 20% 6
Balanced Scorecard [19, 69], PESTEL [6], PEST [39, 52], Five Forces Framework [52],
SWOT Analysis [57, 58, 62], SUST-BMA [19]
(VII) Business Model Ontology-oriented Evaluation 14% 3 - -
Business Model Canvas [6, 19, 62, 63, 67], 4C Net Framework [73], Ballon Framework [65]
(VIII) Decision Structuring-oriented Evaluation 14% 3 -
Decision Matrix [51, 61], Event-Decision Tree [31], Morphological Box/Taxonomy [29, 30, 47, 51]
(IX) Pattern- and Key Question-based evaluation 8% 2 - -
Pattern-based Analysis [46, 52], Central/Key Questions [59, 62]
(X) Value Proposition-oriented Evaluation Tools 6% 2
Value Mapping Tool [18], Value Proposition Traceability Diagram [43, 44]
measures, and the SWOT Analysis that takes strengths, weaknesses, opportunities, and
threats into account. In addition, strategic frameworks facilitate to analyze factors
effecting an organization like PEST and PESTEL (macro-environment factors: politi-
cal, economic, social, technological; for PESTEL additionally: environmental and
legal) as well as the Five Forces Framework. Next to those strategy-related tools, there
are also tools for analyzing the value proposition (X) and providing innovative forms of
value.
Acknowledgments. This research was conducted in the scope of the research project “Smart-
Hybrid—Process Engineering” (ZW 6-85003451), which is partly funded by the European
Regional Development Fund and the State of Lower Saxony (NBank).
References
1. Brundtland, G.H.: World Commission on Environment and Development (WCED, 1987):
Our Common Future. Oxford University Press, Oxford (1987)
2. Melville, N.P.: Information systems innovation for environmental sustainability. MIS Q. 34,
1–21 (2010)
3. Seidel, S., Recker, J.C., vom Brocke, J.: Sensemaking and sustainable practicing: functional
affordances of information systems in green transformations. MIS Q. 37, 1275–1299 (2013)
Towards a Typology of Approaches 67
4. Lüdeke-Freund, F., Freudenreich, B., Schaltegger, S., Saviuc, I., Stock, M.: Sustainability-
oriented business model assessment—a conceptual foundation. In: Carayannis, E.G.,
Sindakis, S. (eds.) Analytics, Innovation, and Excellence-Driven Enterprise Sustainability,
pp. 169–206. Palgrave Macmillan US, New York (2017)
5. Veit, D., Clemons, E., Benlian, A., Buxmann, P., Hess, T., Kundisch, D., Leimeister, J.M.,
Loos, P., Spann, M.: Business models: an information systems research agenda. Bus. Inf.
Syst. Eng. 6, 45–53 (2014)
6. Haaker, T., Bouwman, H., Janssen, W., de Reuver, M.: Business model stress testing: a
practical approach to test the robustness of a business model. Futures 89, 14–25 (2017)
7. Schaltegger, S., Hansen, E.G., Lüdeke-Freund, F.: Business models for sustainability:
origins, present research, and future avenues. Organ. Environ. 29, 3–10 (2016)
8. Osterwalder, A., Pigneur, Y.: Business Model Generation: A Handbook for Visionaries,
Game Changers, and Challengers. Wiley, Hoboken (2010)
9. Schoormann, T., Behrens, D., Kolek, E., Knackstedt, R.: Sustainability in business models–a
literature-review-based design-science-oriented research agenda. In: Proceedings of the
European Conference on Information Systems (ECIS), Istanbul, Turkey (2016)
10. Täuscher, K., Abdelkafi, N.: Visual tools for business model innovation: recommendations
from a cognitive perspective. Creat. Innov. Manag. 26, 160–174 (2017)
11. Afuah, A., Tucci, C.L.: Internet Business Models and Strategies. McGraw-Hill, New York
(2001)
12. Kamoun, F.: Rethinking the business model with RFID. Commun. Assoc. Inf. Syst. 22, 35
(2008)
13. Gordijn, J., Akkermans, H.: Designing and evaluating e-business models. IEEE Intell. Syst.
16, 11–17 (2001)
14. Tesch, J., Brillinger, A.: The evaluation aspect of the digital business model innovation:
aliterature review on tools and methodologies. In: Proceedings of the European Conference
on Information Systems, pp. 2250–2268 (2017)
15. Si, Y., Chen, T., Zheng, W.: An approach to evaluating business models - a case study of
Taobao. In: 2011 IEEE 18th International Conference on Industrial Engineering and
Engineering Management, pp. 31–35 (2011)
16. Díaz-Díaz, R., Muñoz, L., Pérez-González, D.: The business model evaluation tool for smart
cities: application to SmartSantander use cases. Energies 10, 262 (2017)
17. Isaksson, R., Garvare, R.: Measuring sustainable development using process models. Manag.
Audit. J. 18, 649–656 (2003)
18. Bocken, N.M.P., Rana, P., Short, S.W.: Value mapping for sustainable business thinking.
J. Ind. Prod. Eng. 32, 67–81 (2015)
19. Lüdeke-Freund, F., Freudenreich, B., Schaltegger, S., Saviuc, I., Stock, M.: Sustainability-
oriented business model assessment—a conceptual foundation. In: Analytics, Innovation,
and Excellence-Driven Enterprise Sustainability, pp. 169–206. Palgrave Macmillan,
New York (2017)
20. Hamel, G.: Leading the Revolution. Harvard Business School Press, Boston (2000)
21. Wirtz, B.W.: Business Model Management, Design-Instrumente-Erfolgsfaktoren von
Geschaftsmodellen (2011)
22. Zott, C., Amit, R.: Measuring the Performance Implications of Business Model Design:
Evidence from Emerging Growth Public Firms. Insead, Fontainebleau (2002)
23. Alexa, M.: Business model evaluation–a conceptual approach. Rev. Econ. Bus. Stud. 14,
245–260 (2014)
24. Magretta, J.: Why business models matter (2002)
25. Osterwalder, A., Pigneur, Y., Tucci, C.L.: Clarifying business models: origins, present, and
future of the concept. Commun. Assoc. Inf. Syst. 16, 1 (2005)
68 T. Schoormann et al.
26. Lüdeke-Freund, F., Dembek, K.: Sustainable business model research and practice:
emerging field or passing fancy? J. Clean. Prod. 168, 1668–1678 (2017)
27. Stubbs, W., Cocklin, C.: Conceptualizing a “sustainability business model”. Organ. Environ.
21, 103–127 (2008)
28. Upward, A., Jones, P.: An ontology for strongly sustainable business models: defining an
enterprise framework compatible with natural and social science. Organ. Environ. 29, 97–
123 (2016)
29. Jun, M., Di Muro, A.: Holistic methodology to analyze EV business models. Int. J. Innov.
Manag. Technol. 4, 259 (2013)
30. Seim, S., Scheller, F., Götz, M., Kondziella, H., Bruckner, T.: Assessment of PV-based
business models in urban energy systems with respect to political and economic targets: a
model-based scenario analysis. In: 10 Internationale Energiewirtschaftstagung IEWT 2017
Wien (2017)
31. Copani, G., Rosa, P.: DEMAT: sustainability assessment of new flexibility-oriented business
models in the machine tools industry. Int. J. Comput. Integr. Manuf. 28, 408–417 (2015)
32. Weill, P., Vitale, M.: Place to Space: Migrating to eBusiness Models. Harvard Business
Press, Boston (2001)
33. Hansen, E.G., Grosse-Dunker, F., Reichwald, R.: Sustainability innovation cube—a
framework to evaluate sustainability-oriented innovations. Int. J. Innov. Manag. 13, 683–
713 (2009)
34. Ness, B., Urbel-Piirsalu, E., Anderberg, S., Olsson, L.: Categorising tools for sustainability
assessment. Ecol. Econ. 60, 498–508 (2007)
35. Webster, J., Watson, R.T.: Analyzing the past to prepare for the future: writing a literature
review. MIS Q., xiii–xxiii (2002)
36. vom Brocke, J., Simons, A., Niehaves, B., Riemer, K., Plattfaut, R., Cleven, A.:
Reconstructing the giant: on the importance of rigour in documenting the literature search
process. In: Proceedings of the European Conference on Information Systems, Verona, Italy,
pp. 2206–2217 (2009)
37. Hevner, A., March, S., Park, J., Ram, S.: Design science in information systems research.
MIS Q. 28, 75–105 (2004)
38. Ali, A.: An MCDM approach towards m-payment business models evaluation. Int. J. Anal.
Hierarchy Process. 7, 273–294 (2015)
39. Bilgeri, D., Brandt, V., Lang, M., Tesch, J., Weinberger, M.: The IoT business model
builder. A White Paper of the Bosch IoT Lab in collaboration with Bosch Software
Innovations GmbH (2015)
40. Brea-Solís, H., Casadesus-Masanell, R., Grifell-Tatjé, E.: Business model evaluation:
quantifying Walmart’s sources of advantage. Strateg. Entrep. J. 9, 12–33 (2015)
41. Cagliano, A.C., De Marco, A., Mangano, G., Zenezini, G.: Assessing City Logistics
projects: a business-oriented approach. In: Proceedings of the Summer School “Francesco
Turco”, pp. 204–208. Italian Association of Industrial Operations Professors, Napoli (2016)
42. Clauss, T.: Measuring business model innovation: conceptualization, scale development, and
proof of performance. RD Manag. 47, 385–403 (2017)
43. Costa, C., Cunha, P.: The social dimension of business models: an Actor-Network Theory
perspective. In: Proceedings of the American Conference on Information Systems (AMCIS)
(2015)
44. Costa, C.C., Cunha, P.R.: More than a gut feeling: Ensuring your inter-organizational
business model works. In: Proceedings of the Bled eConference, p. 31 (2015)
45. Ebel, P., Bretschneider, U., Leimeister, J.M.: Leveraging virtual business model innovation:
a framework for designing business model development tools: leveraging virtual business
model innovation. Inf. Syst. J. 26, 519–550 (2016)
Towards a Typology of Approaches 69
46. Echterfeld, J., Amshoff, B., Gausemeier, J.: How to use business model patterns for
exploiting disruptive technologies. In: 24th International Conference on Management of
Technology, International Association for Management of Technology, pp. 2294–2313
(2015)
47. Ghezzi, A., Georgiades, M., Reichl, P., Le-Sauze, N., Cairano-Gilfedder, C.D., Managiaracina,
R.: Generating innovative interconnection business models for the future internet. info 15, 43–68
(2013)
48. Grustam, A.S., Vrijhoef, H.J.M., Koymans, R., Hukal, P., Severens, J.L.: Assessment of a
Business-to-Consumer (B2C) model for Telemonitoring patients with Chronic Heart Failure
(CHF). BMC Med. Inf. Decis. Mak. 17, 145 (2017)
49. Hart, S.L., Milstein, M.B.: Creating sustainable value. Acad. Manag. Exec. 17, 56–67 (2003)
50. Hvass, K.: A Boolean Approach to Airline Business Model Innovation. Copenhagen:
CIBEM Working Paper Series 13 (2012)
51. Im, K., Cho, H.: A systematic approach for developing a new business model using
morphological analysis and integrated fuzzy approach. Expert Syst. Appl. 40, 4463–4477
(2013)
52. Ishida, F., Sakuma, H., Abe, H., Fazekas, B.: Remodeling method for business models of R
D outputs. In: 2006 Technology Management for the Global Future - PICMET 2006
Conference, pp. 708–714 (2006)
53. Kampker, A., Heimes, H.H., Ordung, M., Lienemann, C., Hollah, A., Sarovic, N.:
Evaluation of a remanufacturing for lithium ion batteries from electric cars. World Acad. Sci.
Eng. Technol. Int. J. Mech. Mechatron. Eng. 3 (2016)
54. Kayaoglu, N.: A Generic Approach for Dynamic Business Model Evaluation (2013)
55. Laplume, A., Anzalone, G.C., Pearce, J.M.: Open-source, self-replicating 3-D printer factory
for small-business manufacturing. Int. J. Adv. Manuf. Technol. 85, 633–642 (2016)
56. Lee, Y.H., Shim, H.D., Park, N.I.: A new convergence business model framework for
context-aware computing service market. Commun. Converg. Rev. 2, 1–12 (2010)
57. Leitão, A., Cunha, P., Valente, F., Marques, P.: Roadmap for business models definition in
manufacturing companies. Procedia CIRP 7, 383–388 (2013)
58. Martínez, A.B., Galván, R.S., Alam, S.: Financial analysis of retail business organization: a
case of Wal-Mart Stores, Inc. Nile J Bus. Econ. 3, 67–89 (2017)
59. Mateu, J.M., March-Chorda, I.: Searching for better business models assessment methods.
Manag. Decis. 54, 2433–2446 (2016)
60. Moellers, T., Bansemir, B., Pretzl, M., Gassmann, O.: Design and evaluation of a system
dynamics based business model evaluation method. In: Maedche, A., vom Brocke, J.,
Hevner, A. (eds.) DESRIST 2017. LNCS, vol. 10243, pp. 125–144. Springer, Cham (2017).
https://doi.org/10.1007/978-3-319-59144-5_8
61. O Riordan, N., O’Reilly, P., Duane, A., Andreev, P.: Business model innovation: a temporal
perspective. In: Proceedings of the Australasian Conference on Information Systems,
Auckland New Zealand, 8–10 December 2014 (2014)
62. Osterwalder, A.: How to describe and improve your business model to compete better. Trobe
Univ. Acesso Em. 26, 2015 (2007)
63. Prado, A.M., Calderon, D., Zúñiga, R.: Providing low-cost and high-quality medications to
rural communities in developing countries: the case of Accion Medica Cristiana in
Nicaragua. J. Bus. Res. 69, 3910–3922 (2016)
64. Qin, Q., Liang, F., Li, L., Wei, Y.-M.: Selection of energy performance contracting business
models: a behavioral decision-making approach. Renew. Sustain. Energy Rev. 72, 422–433
(2017)
70 T. Schoormann et al.
65. Raju, A., Lindmark, S., Yaron, O., De Poorter, E., Tytgat, L., Delaere, S., Ballon, P.:
Business model assessment of Green wireless sensor ecosystems. In: 11th Conference of
Telecommunication, Media and Internet Techno-Economics (CTTE-2012), pp. 1–8. Ghent
University, Department of Information Technology (2012)
66. Sánchez, P., Ricart, J.E.: Business model innovation and sources of value creation in low-
income markets. Eur. Manag. Rev. 7, 138–154 (2010)
67. Schüle, S., Schubert, M., Hoyer, C., Dressel, K.-M.: Development of an assessment tool to
evaluate and improve SME business models. J. Bus. Models 4, 5 (2016)
68. Sharma, S., Gutiérrez, J.A.: An evaluation framework for viable business models for m-
commerce in the information technology sector. Electron. Mark. 20, 33–52 (2010)
69. Su, C.H., Hung, Y.H., Tzeng, G.H.: Fuzzy multiple attribute decision making theory with
the balanced scorecard application in mobile industry. In: 2011 IEEE International
Conference on Fuzzy Systems (FUZZ-IEEE 2011), pp. 1479–1484 (2011)
70. Suikki, R.M., Goman, A.M.J., Haapasalo, H.J.O.: A framework for creating business models –
a challenge in convergence of high clock speed industry. Int. J. Bus. Environ. 1, 211–233
(2006)
71. Van Ostaeyen, J., Duflou, J.: Assessing the potential of business model innovation for
investment goods through Life Cycle Costing. In: Proceedings of the CIRP IPS2
Conference, pp. 527–534. Linköping University Electronic Press (2012)
72. Wang, X., Brooks, S., Sarker, S.: A review of green IS research and directions for future
studies. Commun. Assoc. Inf. Syst. 37, 395–429 (2015)
73. Wirtz, B.W., Lihotzky, N.: Customer retention management in the B2C electronic business.
Long Range Plann. 36, 517–532 (2003)
74. Young, L.W., Johnston, R.B.: A framework for evaluating the impact of the internet on
business-to-business e-commerce value delivery options. J. Syst. Inf. Technol. 7, 129–152
(2003)
75. Zeng, D.-B., Lu, H.-X.: Evaluation Study of Internet Commercial Sustainable Development
under the New Commercial Civilization (2014)
76. Zhang, G., Jiang, Y., Chen, X., Wang, Z.: The research on business model evaluation based
on internet free services. In: 2012 Proceedings of Computer Science and Information
Management (ICIIM 2012), pp. 97–102 (2012)
77. Zografos, K.G., Androutsopoulos, K.N., Sihvola, T.: A methodological approach for
developing and assessing business models for flexible transport systems. Transportation 35,
777–795 (2008)
78. Bailey, K.D.: Typologies and Taxonomies: An Introduction to Classification Techniques.
Sage, Thousand Oaks (1994)
Towards Agility in IT Governance Frameworks
1
University of the West of Scotland, Paisley Campus, Paisley PA1 2BE, Scotland, UK
[email protected]
2
Reutlingen University, Alteburgstrasse 150, 72762 Reutlingen, Germany
[email protected]
Abstract. Digital transformation has changed corporate reality and, with that,
firms’ IT environments and IT governance (ITG). As such, the perspective of ITG
has shifted from the design of a relatively stable, closed and controllable system
of a self-sufficient enterprise to a relatively fluid, open, agile and transformational
system of networked co-adaptive entities. Related to this paradigm shift in ITG,
this paper aims to clarify how the concept of an effective ITG framework has
changed in terms of the demand for agility in organizations. Thus, this study
conducted 33 qualitative interviews with executives and senior managers from
the banking industry in Germany, Switzerland and Austria. Analysis of the inter‐
views focused on the formation of categories and the assignment of individual
text parts (codings) to these categories to allow for a quantitative evaluation of
the codings per category. Regarding traditional and agile ITG dimensions, 22
traditional and 25 agile dimensions were identified. Moreover, agile strategies
within the agile ITG construct and ten ITG patterns were identified from the
interview data. The data show relevant perspectives on the implementation of
traditional and new ITG dimensions and highlight ambidextrous aspects in ITG
frameworks.
1 Introduction
Digital transformation has fundamentally changed corporate reality. The role of informa‐
tion technology (IT) in business is now pervasive. Whereas the dominant logic of the
Industrial Age was linear and product-oriented, the Digital Age is nonlinear and service-
oriented [1]. This broad paradigm shift in logic is also reflected in changing IT environ‐
ments and, consequently, in firms’ IT governance (ITG). The perspective has shifted from
the relatively stable, closed and controllable system of a self-sufficient enterprise to the
relatively fluid, open, agile and transformational system of networked co-adaptive entities
[2]. Thus, the demand for agility has become more important to enterprises. Existing
studies have recognised efficiency and stability as core concepts in ITG design [3]. This
makes sense as the “old” world was characterized by a stable, placid environment, in which
neither the core technology nor the markets in which companies were operating changed
drastically over time. Organizations could afford to use “command-and-control” mecha‐
nisms to govern IT [3]. In this context, many proposed methodologies, reference guides,
sets of best practices (e.g. COBIT), and frameworks such as the IT Infrastructure Library
(ITIL) have emerged and developed in recent years. However, the adoption of such ITG
models, or so-called conventional or traditional frameworks, does not necessary yield the
desired return in the context of digital transformation [4]. By contrast, agile processes and
methods have evolved in the past years, especially in the area of software development [5,
6]. Independent of the business area, these agile frameworks can “add value” to business
organizations, through a process in which the principles of communication and collabora‐
tion are essential [7]. Thus, adopting the agile principles, values and best practices to the
context of ITG can bring even more meaningful results to organizational management.
Their benefits can lead to an increase in the speed of decision making, the insurance of
business processes, organizational competitiveness and other aspects [4]. Thus, agile ITG
has become highly relevant to keep up with competitors in today’s dynamic world. To
date, the broad scientific community has not analysed the impetus of agile strategies on ITG
[4, 8]. However, research interest in this topic is growing [5]. As such, this study aims to
investigate the impact of agility on effective ITG for today’s dynamic world, which yields
the following research question (RQ): How is the concept of an effective ITG framework
changing in response to the demand for agility in organizations?
To answer this RQ, we analysed several traditional and agile aspects of governance
dimensions gleaned from 33 qualitative interviews with executives and senior managers
from the banking industry in Germany, Switzerland and Austria. In doing so, we were
able to elicit major patterns for ITG dimensions in the digital world. Strategies used to
implement an agile ITG can be grouped under the same general dimensions as in the
conventional ITG literature, which enables us to compare traditional and agile aspects
and to derive patterns for the effectiveness of an agile ITG framework. Such patterns
help explain “real-world” problems because they capture and allow for reuse of expe‐
riences of best practices in a specific professional domain [9]. Thus, the contribution of
this paper is threefold. First, the study uncovers agile ITG dimensions that complement
the traditional elements of ITG; second, it provides an overview of agile strategies used
within the governance construct; third, it elicits patterns for effective ITG dimensions.
2 Theoretical Background
Since late 1990s, many researchers and practitioners have investigated ITG from
different perspectives [10]. As such, several definitions have been proposed. While some
definitions such as those from Peterson [3] and Weill and Ross [11] focus on the decision-
making process within the ITG framework and do not address any role aspects, other
definitions such as those from van Grembergen and de Haes [12] and the IT Governance
Institute [13] adequately address objectives, objects and subjects of ITG decisions [14].
We argue that an ITG definition should include both structure and process aspects.
Therefore, enhancing the definition of the IT Governance Institute with Weill and Ross’s
characterization should help cover the most relevant dimensional concepts of current
ITG research, leading to the following definition:
Towards Agility in IT Governance Frameworks 73
ITG is the responsibility of executives and the board of directors and consists of the leadership
and organizational structures and processes that ensure that the organization’s IT sustains and
extends firm strategies and objectives. ITG represents the framework for decision rights and
accountabilities to encourage desirable behaviour in the use of IT.
With this definition, implementing ITG effectively requires a set of ITG instruments to
gain congruence with the firm’s mission, strategy, values, norms and culture [15], which
in turn leads to desirable IT behaviours and governance outcomes [11]. However,
implementing ITG is a complex issue because it is contingent on a variety of sometimes
conflicting internal and external factors [16–18]. Effective ITG doesn’t happen by acci‐
dent and top-performing firms should consequently carefully design governance [17].
The effectiveness of a firm’s ITG can be assessed by evaluating how well it enables IT
to deliver on four objectives: cost-effectiveness, asset utilization, business growth and
business flexibility [17]. Consequently, in the literature several studies have argued that
organizations should use ITG [15, 19, 20]. Therefore, consistent with the previous
definition and the literature, implementing an effective ITG requires a framework based
on three major dimensions [21].
Structure. The framework needs to answer the following questions: Who makes the
decisions? Which organizational units will be created? Who will take part in these
organizational units? What responsibilities will they assume [22]? Examples of tradi‐
tional structures are IT steering committees, IT project steering committees and IT
strategy committees as well as structures that enable CIOs to report to CEOs [15].
Process. The process aspect targets the following questions: How are IT investment
decisions made? What are the decision-making processes for proposing, reviewing,
approving and prioritizing investments? Conventional processes, for example, contain
portfolio management, IT budget control and reporting, project governance methodol‐
ogies or information systems planning [23].
The manifesto and all its values and principles represent the philosophy behind agile
strategies and ideally should be present in all practices proposed by various agile
methods [48].
Therefore, the core objective of the current study is to try to obtain traditional as well
as agile strategies put in place in firms, to identify how the concept of an effective
framework is changing with respect to the demand for agility in organizations. Then, to
elicit ITG patterns, the traditional dimensions and agile strategies must be set in context
to some control variables. This is important because determining the right ITG dimen‐
sions is a complex endeavour, and it should be recognised that what strategically works
for one company does not necessarily work for another [19]. Therefore, to elicit some
ITG patterns of success, this study uses control variables such as firm size, IT strategy
and regional differences in context, as also highlighted in the approaches of Weill [20]
and Sambamurthy and Zmud [16].
3 Research Method
Furthermore, all respondents gave permission to have the interview digitally recorded.
Thus, all 33 semi-structured interviews were conducted and audio-recorded by two
members of the panel. Voice recording is essential because it saves the recorded informa‐
tion at hand, thus ensuring accurate information of the data [51]. Each interview session
took approximately 30 min. All the interviewees were at managerial levels and held exec‐
utive or senior managerial positions with decision rights to define digital transformation
projects.
In stage four, the audio records were fully transcribed to enable analysis of the data set.
The transcribed source texts were first analysed case by case, to set the basis for the devel‐
opment of a system of categories for a structured evaluation of the text material. In doing
so, the overall approach corresponds to the approach of a qualitative data analysis [52].
From a methodological point of view, the qualitative data analysis is based on the forma‐
tion of categories and the assignment of individual text parts to these categories. To
support the encoding of the data, analysis software MAX QDA was used. The transcribed
texts were first imported into MAX QDA, and then individual text parts (= codings) of the
transcribed interviews were assigned to the defined categories. When necessary, the main
categories were further differentiated into subcategories, so that the data could be organ‐
ized and interpreted in a category-based manner. In doing so, the quantitative evaluation
referred to the quantity of the codings per category. Consequently, the data per category
were quantitatively evaluated and qualitatively interpreted. The thematic codings allowed
us to highlight the links and concepts related to the organizational integration of gover‐
nance dimensions in the Digital Age. Many dimensions, especially the traditional ITG
dimensions, were confirmed by the literature mentioned previously. However, our anal‐
ysis led to several new elements of governance, mainly in the context of agility.
4 Data Analysis
Data analysis uncovered several traditional and agile ITG dimensions mentioned by the
respondents. Table 1 provides an overview of the bank characteristics and control varia‐
bles. We included firm size, degree of agility, duration of strategic focus on digital trans‐
formation, and regional differences as control variables in our questionnaire. Different
categories of banks are well distributed on all control variables, with a slight focus on
small and medium-sized banks (<500 employees). A differentiation between subsamples
and categories per control variable extends the scope of this paper but opens a fruitful
avenue for further analysis.
Table 2 lists all the information gathered from the 33 banks regarding different ITG
dimensions. Overall, 47 dimensions for ITG appear in the sample. The ITG dimensions
used by the firms are grouped into the categories “structures”, “processes” and “relational
mechanisms”. To group the dimensions into the three mentioned categories, the study of
Almeida et al. [23] served as a template. Many other studies, such as that of de Haes and
van Grembergen [53] and Symons [21], use this type of framework to structure ITG
dimensions. Moreover, we subdivided the dimensions into “traditional” and “agile” to
differentiate between these two subcategories. To identify traditional ITG dimension, we
used the framework of Almeida et al. [23].
78 S. Vejseli and A. Rossmann
Banks
Control Variables 11111111 112 222 222 222 3333
123456789
01234567 890 123 456 789 0123
Small (<500 employees)
Medium-sized
Firm size
(500–2000 employees)
Large (>2000 employees)
Very agile
Degree of agility
A little agile
in ITG
Not agile
Germany
Regional
Switzerland
differences
Austria
The interviewees reported on how the goals were inherent in their business strategy and
how they strived to manage digital transformation projects. Therefore, in exploring the
traditional ITG dimensions implemented by banks in Germany, Switzerland and Austria
we identified 22 traditional ITG dimensions (six structure elements, seven processes and
nine relational/communication mechanisms; see Table 2). Many of the elements are also
confirmed in the literature [23].
In terms of structure, the most frequently mentioned element was “CIO/COO on
executive committee”. Of the 33 interviewees, 26 noted that the CIO or COO is a full
member of the executive committee or has a direct reporting structure to the CEO. This
situation allows a firm to ensure that IT is a regular agenda item and reporting issue for
the board of directors. Furthermore, boards and committees (25 codings) are formed to
help align business and IT and determine business priorities in IT investments. To
prioritise and manage IT projects, project steering committees (18 banks) and IT organ‐
ization structures (17 banks) are primarily used. Surprisingly, only in few institutions
the CIO is a member of the executive board (3 codings).
Regarding traditional processes, portfolio management was highlighted as the most
essential element. Of the 33 interviewees, 32 mentioned that within portfolio manage‐
ment, in which business and IT are involved, IT investments and projects are prioritised.
In terms of controlling, monitoring and reporting projects, most banks use the processes
“IT budget control and reporting” (20 codings), “project governance” (20 codings) and
Towards Agility in IT Governance Frameworks 79
Banks
Codings
ITG Dimensions 11 1 1 1 1 1 1 1 1 2 2 2 2 2 2 2 2 2 2 3 3 33
123456789
01 2 3 4 5 6 7 8 9 0 1 2 3 4 5 6 7 8 9 0 1 23
1 CIO/COO on executive committee
26
2 Boards and committees 25
Traditional
15 Project governance
20
16 Strategic information systems planning
16
17 Project tracking
15
18 Benefit management
8
19 Demand management
6
Using agile practices (Scrum,
20 20
Processes
Devops, design thinking, lean start-ups)
21 Taking higher risk (trial and error)
17
22 Fast/agile decision-making processes
15
Using innovative key performance
23
10
indicators (KPIs)
Agile
“project tracking” (15 codings). Formal processes to define and update IT strategy are
implemented in almost half the institutions (process: strategic information systems
planning; 16 codings). In relatively few banks are the processes of benefit management
(8 codings) and demand management (6 codings) implemented.
Communication in the financial institutions is crucial. Of the 33 interviewees, 27
highlighted the importance of regular internal communication. Here, leadership aspects
play a key role. Seventeen interviewees indicated that management should give good
examples in communicating. Furthermore, 14 executives mentioned IT leadership as
important to articulate a vision for IT’s role in the company and to ensure that this vision
is clearly understood by managers throughout the organization.
To ensure that business and IT work together, more than 50% of the firms locate
business and IT close to each other (business/IT collocation; 17 codings). Only seven
banks use cross-functional business/IT training instead. Ten banks use informal meet‐
ings of executives and senior management announcements. Finally, seven interviewees
mentioned practicing knowledge management in their firms.
One of the most relevant dimensions, in terms of relational mechanisms, the respond‐
ents highlighted was external collaboration. Seventy-six percent of the executives agreed
that collaboration (25 codings) with external partners (e.g. start-ups, business partners,
outsourcing partners and research partners) had become increasingly important to
achieve their strategic goals. Furthermore, transformational leadership (18 codings), and
open communication and participation (18 codings) play key roles in empowering
employees and being transparent in communication. Moreover, executives from 12
banks regard the setup of lean communication structures (12 codings) and the usage of
social and digital media (12 codings), such as enterprise social networking, Twitter,
Facebook, Webex, and so on, in their communication initiatives as a way to engage
people. A few bank executives also mentioned mechanisms such as regular trainings
and teamwork (9 codings), cross-functional trainings (7 codings), specific innovation
rooms/meetings, and management dialogues and campaigns (6 codings) as improving
the communication and transparency within their companies.
The main objectives of this paper were to conduct a qualitative analysis to identify agile
aspects of the governance construct and to elicit major patterns of ITG dimensions in
the digital world. Uncovering the traditional and agile dimensions implemented in the
banking industry to master digital innovations allows us to identify how the concept of
an effective ITG framework has changed with regard to the demand for agility (RQ). As
indicated in the “Data analysis” section, the analysis identified both traditional and agile
ITG dimensions. Furthermore, it outlined agile characteristics of the uncovered agile
mechanisms and offered several key ITG patterns.
Weil and Ross [11] illustrated the importance of implementing an effective ITG
framework by means of the three major dimensions - structural, processual, relational -
of ITG in an organization. They showed that up to 40% higher returns could be generated
82 S. Vejseli and A. Rossmann
by using effective ITG processes. However, because of sweeping changes in the new
and fast-moving economy, leading to new requirements in and expectations of IT, the
task of aligning business and IT remains a prime challenge, even though ITG frameworks
are being implemented. Today’s firms need flexible, complementary, adaptive and
collaborative ITG dimensions to be put in place, if they are to prosper in a turbulent
environment, in which the challenge for firms is to sustain value realization from IT
instead of restraining its importance by emphasizing control. Therefore, what is required
in the Digital Age is the underlying capability to flexibly adapt to changing business
environments and requirements, to select and adopt promising new technologies, to
effectively anticipate future needs, and to ensure that the potential residing in IT is
proactively communicated within the firm, thus ensuring its effective exploitation [54].
Hence, the digital era demands a culture of speed and collaboration, if it is to differentiate
and deliver extraordinary customer service to drive business growth. In this era, the rise
of mobility and the velocity to deliver differentiated business processes is critical to
success, which calls for a more agile ITG dimensions [14].
With regard to the understanding of the agile strategies within the governance
construct, the respondents all believed that implementing agile strategies helps improve
their agility. Therefore, 25 agile dimensions were identified. The most frequently
employed agile mechanisms are the ITG dimension of the processes and the relational/
communication mechanisms. Moreover, our study also indicates that traditional ITG
dimensions are important for sustaining control. In this way, 22 traditional ITG dimen‐
sions were uncovered. In the context of the punctuated equilibrium theory, the concept
of an effective ITG framework changing in response to the demand for agility in organ‐
izations calls for more ambidextrous approaches.
The two systems—traditional and agile—should work together, with a constant flow
of information and activity between them [2]. In other words, to be effective an agile
ITG governance must work seamlessly and organically with traditional, as well as agile
ITG dimensions, so that the whole organization is both ensuring that tasks are completed
with efficiency and reliability, constantly and incrementally improving itself, and
handling today’s increasingly strategic challenges with speed and agility. Therefore, the
interaction between the traditional and the agile dimensions needs to be optimized in
managing strategies to impact positively on the agility of a company.
Finally, this study is not without limitations. First, the scope of the data collected is
restricted to firms from Germany, Switzerland and Austria. Considering other
geographic locations could provide additional insights. Second, the analysis is limited
to the banking industry. As such, understanding might be advanced from investigating
other sectors. Third, we mainly interviewed bank executives. Use of different group,
such as employees working in IT and business, could lead to further relevant insights.
In conclusion, our interview data provide relevant perspectives of traditional and
new ITG dimensions implemented within the banking industry of the German-speaking
part of Europe, which highlights the ambidextrous aspect within the governance
construct. As such, the current study helps stimulate further investigation into combining
agile capabilities with governance capabilities.
Towards Agility in IT Governance Frameworks 83
References
1. Collin, J., Halén, M., Helenius, M., Hiekkanen, K., Itälä, T., Korhonen, J.J.: IT Leadership
in Finnish Organizations and Digital Transformation. Alto University (2014)
2. Kotter, J.P.: Accelerate Building Strategic Agility for a Faster-Moving World. Harvard
Business Review Press, Boston (2014)
3. Peterson, R.R.: Integration strategies and tactics for information technology governance. In:
van Grembergen, W. (ed.) Strategies for Information Technology Governance, pp. 37–80.
IGI Global, Hershey (2004)
4. Luna, A.J.H., Costa, C.P., Moura, H.P.D., Novaes, M.A., do Nascimento, C.A.: Agile
governance in information and communication technologies, shifting paradigms. J. Inf. Syst.
Technol. Manag. 7(2), 311–334 (2010)
5. Cheng, T.-H., Jansen, S., Remmers, M.: Controlling and monitoring agile software
development in three Dutch product software companies. In: Proceedings of the ICSE
Workshop on Software Development Governance, pp. 29–35. IEEE, Washington, D.C.
(2009)
6. Qumer, A.: Defining an integrated agile governance for large agile software development
environments. In: Agile Processes in Software Engineering and Extreme Programming, pp.
157–160 (2007)
7. Fruhling, A., McDonald, P., Dunbar, C.: A case study: introducing eXtreme programming in
a US Government system development project. In: Proceedings of the 41st Annual Hawaii
International Conference on System Sciences (HICSS), Waikoloa, HI, USA, p. 464. IEEE
(2008)
8. Almeida, N.H.R., de Magalhaes, E.M.C., de Moura, H.P., de Teixeira, F.J.G.D.A., Cappelli,
C., Martins, L.M.F.: Evaluation of a maturity model for agile governance in ICT using focus
group. In: Proceedings of the Annual Conference on Brazilian Symposium on Information
Systems, 15–22. Brazilian Computer Society, Porto Alegre, Brazil (2015)
9. Schadewitz, N., Jachna, T.: Comparing inductive and deductive methodologies for design
patterns identification and articulation. In: International Design Research Conference IADSR
2007 Emerging Trends in Design Research, Hong Kong (2007)
10. Bhattacharjya, J., Chang, V.: Adoption and implementation of IT governance. cases from
Australian higher education. In: Strategic Information Systems: Concepts, Methodologies,
Tools, and Applications, pp. 1308–1326. IGI Global, Hershey (2010)
11. Weill, P., Ross, J.W.: IT governance. How top performers manage IT decision rights for
superior results. Harvard Business School Press, Boston, Mass (2004)
12. van Grembergen, W., de Haes, S.: Measuring and improving IT governance through the
balanced scorecard. Inf. Syst. Control J. 2(1), 35–42 (2005)
13. IT Governance Institute: Cobit 4.1: Framework, Control Objectives, Management
Guidelines, Maturity Models. Meadows (2007)
14. Vejseli, S., Rossmann, A.: The impact of IT Governance on firm performance: a literature
review. In: Proceedings of the 21st Pacific Asia Conference on Information Systems (PACIS).
AIS Electronic Library (2017)
15. de Haes, S., van Grembergen, W.: IT governance structures, processes and relational
mechanisms: achieving IT/Business alignment in a major Belgian financial group. In:
Proceedings of the 38th Annual Hawaii International Conference on System Sciences
(HICSS), p. 237b. Los Alamitos, CA (2005)
16. Sambamurthy, V., Zmud, R.W.: Arrangements for information technology governance. a
theory of multiple contingencies. MIS Q. 23(2), 261–290 (1999)
84 S. Vejseli and A. Rossmann
17. Weill, P., Ross, J.: A matrixed approach to designing IT governance. MIT Sloan Manag. Rev.
46(2), 26 (2005)
18. de Haes, S., van Grembergen, W.: IT governance and its mechanisms. Inf. Syst. Control J.
1, 27–33 (2004)
19. van Grembergen, W. (ed.): Strategies for Information Technology Governance. IGI Global,
Hershey (2004)
20. Weill, P.: Don’t just lead, govern. How top-performing firms govern IT. MIS Q. Executive
3(1), 1–17 (2004)
21. Symons, C.: IT governance framework. Forrester Research (2005)
22. Burtscher, C., Manwani, S., Remenyi, D.: Towards a conceptual map of IT governance: a
review of current academic and practitioner thinking. In: Proceedings of the UK Academy
for Information Systems Conference, p. 15. AISeL (2009)
23. Almeida, R., Pereira, R., Mira da Silva, M.: IT governance mechanisms: a literature review.
In: Falcão e Cunha, J., Snene, M., Nóvoa, H. (eds.) IESS 2013. LNBIP, vol. 143, pp. 186–
199. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-36356-6_14
24. de Haes, S., van Grembergen, W.: An exploratory study into IT governance implementations
and its impact on business/IT alignment. Information Systems Management (2009)
25. Gottschalk, P.: E-business Strategy, Sourcing and Governance. IGI Global, Hershey (2005)
26. Gersick, C.J.G.: Revolutionary change theories, a multilevel exploration of the punctuated
equilibrium paradigm. Acad. Manag. Rev. 16, 10–36 (1991)
27. Romanelli, E., Tushman, M.L.: Organizational transformation as punctuated equilibrium: an
empirical test. Acad. Manag. J. 37(5), 1141–1166 (1994)
28. Greiner, L.E.: Evolution and revolution as organizations grow. a company’s past has clues
for management that are critical to future success. Family Bus. Rev. 10(4), 397–409 (1997)
29. Tushman, M.L., O’Reilly, C.A.: Ambidextrous organizations: managing evolutionary and
revolutionary change. California Manag. Rev. 38(4), 8–30 (1996)
30. Dunphy, D.C., Stace, D.A.: Transformational and coercive strategies for planned
organizational change: beyond the O.D. model. Organ. Stud. 9(3), 317–334 (1988)
31. Higgs, M., Rowland, D.: All changes great and small: exploring approaches to change and
its leadership. J. Change Manag. 5(2), 121–151 (2005)
32. O’Reilly, C.A., Tushman, M.L.: Organizational ambidexterity. past, present, and future.
Acad. Manag. Perspect. 27(4), 324–338 (2013)
33. Markides, C.C.: Business model innovation: what can the ambidexterity literature teach US?
Acad. Manag. Perspect. 27(4), 313–323 (2013)
34. March, J.G.: Exploration and exploitation in organizational learning. Organ. Sci. 2(1), 140–
145 (1991)
35. Markides, C., Charitou, C.D.: Competing with dual business models: a contingency approach.
Acad. Manag. Executive 18(3), 22–36 (2004)
36. Kotter, J.: How the most innovative companies capitalize on today’s rapid-fire strategic
challenges-and still make their numbers. Harvard Bus. Rev. 90(11), 43–58 (2012)
37. van Roosmalen, M.W., Hoppenbrouwers, S.: Supporting corporate governance with
enterprise architecture and business rule management: a synthesis of stability and agility. In:
Proceedings of the International Workshop on Regulations Modelling and Deployment
(ReMoD 2008), pp. 13–24. CEUR-WS.org (2008)
38. Sloane, E., Beck, R., Metzger, S.: AGSOA - Agile Governance for Service Oriented
Architecture (SOA) systems: a methodology to deliver 21st century military net-centric
systems of systems. In: Proceedings of the 2nd Annual IEEE Systems Conference, pp. 1–4.
IEEE (2008)
Towards Agility in IT Governance Frameworks 85
39. Luftman, J.N., Lewis, P.R., Oldach, S.H.: Transforming the enterprise: the alignment of
business and information technology strategies. IBM Syst. J. 32(1), 4–16 (1993)
40. Sambamurthy, V., Bharadwaj, A., Grover, V.: Shaping agility through digital options:
reconceptualizing the role of information technology in contemporary firms. MIS Q. 27(2),
237–263 (2003)
41. Sherehiy, B., Karwowski, W., Layer, J.K.: A review of enterprise agility: concepts,
frameworks, and attributes. Int. J. Indus. Ergon. 37(5), 445–460 (2007)
42. Yang, C., Liu, H.M.: Boosting firm performance via enterprise agility and network structure.
Manag. Decis. 50(6), 1022–1044 (2012)
43. Goldman, S.L., Nagel, R.N., Preiss, K.: Agile Competitors and Virtual Organizations:
Strategies for Enriching the Customer. Van Nostrand Reinhold, New York (1995)
44. Luna, A.J.H.D.O., Kruchten, P., Moura, H.P.D.: GAME. Governance for Agile Management
of Enterprises: a management model for agile governance. In: Proceedings of the 8th
International Conference on Global Software Engineering workshops (ICGSEW), pp. 88–90.
IEEE, Piscataway (2013)
45. Luna, A.J.D.O., Kruchten, P., Pedrosa, M.L.D.E., Neto, H.R., de Moura, H.P.: State of the
art of agile governance: a systematic review. In: IJCSIT (2014)
46. Beck, K., Beedle, M., van Bennekum, A., Cockburn, A., Cunningham, W., Fowler, M.,
Grenning, J., Highsmith, J., Hunt, A., Jeffries, R., Kern, J.: The agile manifesto (2001). http://
agilemanifesto.org/. Accessed 15 Nov 2017
47. Madi, T., Dahalin, Z., Baharom, F.: Content analysis on agile values: a perception from
software practitioners. In: Proceedings of the 5th Malaysian Conference in Software
Engineering (MySEC), pp. 423–428. IEEE, Piscataway (2011)
48. Fernandes, J.M., Almeida, M.: Classification and comparison of agile methods. In: Abreu,
F.B.E. (ed.) Proceedings of the 17th International Conference on the Quality of Information
and Communications Technology (QUATIC), pp. 391–396. IEEE, Piscataway (2010)
49. Benbasat, I., Goldstein, D.K., Mead, M.: The case research strategy in studies of information
systems. MIS Q. 11, 369–386 (1987)
50. Saunders, M., Lewis, P., Thornhill, A.: Research Methods for Business Students: Financial
Times, 5th edn. Prentice Hall, Harlow (2009)
51. Saetang, S., Haider, A.: The impacts of IT governance implementation: a case study on
banking industry in Thailand. In: Proceedings of the Portland International Center for
Management of Engineering and Technology (PICMET), pp. 2619–2627. IEEE (2013)
52. Kuckartz, U., Dresing, T., Rädiker, S., Stefer, C.: Qualitative evaluation. Der Einstieg in die
Praxis, 2nd edn. VS Verlag für Sozialwissenschaften/ GWV Fachverlage GmbH Wiesbaden,
Wiesbaden (2008)
53. de Haes, S., van Grembergen, W.: Enterprise Governance of Information Technology,
Achieving Alignment and Value, Featuring COBIT 5, Management for Professionals, 2nd
edn. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-14547-1
54. Schlosser, F.: Mastering the social IT/business alignment challenge. In: Proceedings of the
18th Americas Conference on Information Systems 2012 (AMCIS), pp. 1843–1849. AIS
Electronic Library (2012)
Organizations in Transformation: Agility as
Consequence or Prerequisite of Digitization?
1
Chair of IT Management, Friedrich Alexander Universität Erlangen-Nürnberg,
Lange Gasse 20, 90403 Nuremberg, Germany
[email protected]
2
Chair of Information Systems, ESP IS in Manufacturing and Commerce,
Technische Universität Dresden, Helmholtzstr. 10, 01069 Dresden, Germany
Abstract. With the aim of meeting new technological and social requirements,
nearly all companies are facing the so-called digital transformation. A starting
point to respond to the changing demands is to improve the company’s ability to
operate flexibly and, thus, increase its agility. Thereby, with this paper, we
contribute to the ongoing discussion on changing corporate organizational struc‐
tures. We used a three-stage study design to provide insights on the question of
whether agility is a prerequisite or a consequence of the digital transformation.
In addition to a literature analysis, we present the results of focus group interviews
with managers of selected enterprises.
Nowadays enterprises are facing a complex environment with rapid and profound
changes. According to Bajer [1], one of the main drivers of these changes is the expo‐
nential technological development, which is often strongly related to buzzwords like
“digital transformation,” “digital age,” or “digital change.” These advanced technolog‐
ical opportunities result in new fundamental paradigm shifts and affect almost every
enterprise, regardless of size or industry sector [2, 3]. This also applies to the German
corporate landscape, which is predominantly characterized by small and medium-sized
enterprises (SMEs) [4]. However, SMEs are often still classically organized and opti‐
mized for controllability and control, which is in contradiction to the fundamental varia‐
tions and high velocity of digital changes [5]. To address this contradiction, our article
focusses on SMEs (according to the SME definition of the European Union [6]).
Thereby, traditional SMEs are usually characterized by shorter decision-making
processes and greater flexibility for customers and employees, as well as smaller budgets
than large companies [7]. According to Urbach and Ahlemann [5], the majority of these
traditional SMEs are also strongly hierarchically structured and have clear communi‐
cation structures oriented towards hierarchical relationships. The traditional SMEs are,
therefore, optimized for control, monitoring, and coordination. In the course of digital
2 Study Design
could be categorized (in accordance with the approach of [17]) in two categories: Six
papers were categorized under “Agility,” and six papers were categorized under “Digital
Organization.” In addition, to those two categories, we analyzed the papers for their
relevance regarding SMEs in combination with our research focus. Ultimately, we
identified six additional relevant papers. The list of all 18 papers is not included in this
paper but can be requested from the first author.
This section shows the results of the literature analysis, which is aimed at highlighting
the ongoing discussion on corporate organizational structures. Nevertheless, although
this discussion expounds upon the two directions of agility and digitization, among other
things, the connection between these two aspects has hardly been investigated, and that
there is hardly any literature with reference to classic SMEs that address agility and
digitization.
Since we are using in the following discussion the two central terms “agility” and
“digitization” extensively, a short definition shall be given beforehand:
“Agility” means the ability of a company to be prepared for and to respond to
changing capacity demands and changing functional requirements very quickly (if
possible in real time), as well as the possibilities of the used information technology to
extend the capacity if necessary. Thereby, agility and flexibility are not clearly differ‐
entiated [22]. Therefore, the assumption for our study is that agility leads to more flex‐
ibility.
The term “digitization” is defined according to Petry [13] as the (exponential) imple‐
mentation and usage of (information) technology in companies. Thereby, digitization
or digital transformation (as synonym) affects many sections and components of enter‐
prises, even across company borders. Linking departments, as well as the contact to
customers or suppliers, is important in the context of digitization. ICT triggers and
enables this “transformation” of the company towards a holistic network This means
digitization is not only defined by a restructuring or optimization of the enterprises’ ICT
landscapes, though this is the main aspect of digitization [2].
Organizations in Transformation 91
In contrast to the papers from Sect. 3.1, there is also a discussion with more focus on
the agility of companies. Gastaldi et al. [25] believe that the required IT reorganization
in the digital age can demand disruptive changes from companies. They stated that new
digital technologies cannot be implemented in a classical/Taylorist organization and
demanded more agile structures. Arnold [26] stated a similar opinion and described the
example of “Haufe-umantis,” which use different methods like the democratic vote of
the leaders and self-organized teams. To attain a higher agility, Arnold [26] and Jova‐
nović et al. [27] recommended experiments to determine how higher agility can be
achieved in combination with the digital transformation.
Lindner et al. [10] recommended adapting organizations to the vision of the New
Work (see [28]): Work when, how, and where you want. In this way, employees are
given maximum freedom in the execution of their work duties. Lindner et al. [10] noted
that the flexibilization of work is strongly dependent on mobile technology, but that the
necessary organizational restructuring in terms of more agility should be undertaken
before new technology is implemented as this facilitates the implementation of tech‐
nology. Furthermore, the potential that new technology offers cannot be achieved suffi‐
ciently without appropriate agile organizational structures [11, 12].
The majority of the sources identified in this section, thus, support the viewpoint that
increasing agility in the company is a precursor to successful digitization.
92 D. Lindner and C. Leyh
The discussion addressing corporate organizational structures has been long debated.
Some of the authors believe that digitization with flexible hardware and software can
increase agility. For them, it is also not possible to work in a more flexible or agile way
in terms of time and place without appropriate technology. However, other authors
believe the implementation of technology requires increased agility beforehand, and that
the required reorganization of IT in the digital age is only barely possible in enterprises
with traditional organizational structures. This, again, shows the topicality of the ques‐
tion of whether agility is a prerequisite or a consequence of digitization.
A similar discussion is taking place with regard to SMEs, but with a slightly different
context. For example, SMEs often have a more centralized and less developed IT system
landscape than large companies due to their lower budget, and it is often necessary to
expand this step by step. SMEs are already described as very agile in contrast to large
corporations; therefore, it is necessary to maintain or increase their agility, especially
when SMEs face rapid growth. However, even if the question of agility vs. digitization
has not been clearly answered yet in the literature, the SME focus still has not been
addressed in the literature to a large extent. Hence, due to the specific characteristics of
the SME and their importance for the economy, it is essential to discuss these aspects
with a special focus on SMEs. Therefore, in the following sections, we show and discuss
the results of our focus group interviews.
Organizations in Transformation 93
The aim in the first group discussion was to examine current agile or digital pilot projects
and to consider, in a mutual dialogue, whether agility is a prerequisite or consequence
of digitization. In our study, a “pilot project” is a company project where a company
implements and tries out new technology or working method with one single team. The
goal of a pilot project is to evaluate if the new technology/method increases the efficiency
of the team and leads to a benefit for the company.
In order to evaluate the theses, four of the participants (No. 1, 2, 3, and 4, Table 2)
presented their current pilot projects. This assisted in enabling the deduction of whether
they see agility as a prerequisite for an increased agility or as a consequence. In this
section, the examples are categorized. They address agility on the basis of employee
satisfaction as well as agility on the basis of the complexity of digitization.
conditions, which allow us to have this flexibility.” The participants of the group discus‐
sion agreed that the demands of digitization projects for employees necessitate more
flexibility in their work execution. According to the consensus of the participants,
rigorous processes and limited flexibility encourage negative dynamics, such as dissat‐
isfaction, burnout, and demotivation. They also agreed with the union’s executive board
member, that current employment agreements are not suitable for the required flexibility
requirements of agile work. To sum up, all participants agreed that the guidelines for an
agile organization are an important prerequisite for digitization.
After the first group discussion showed that increased agility is a basic precondition for
digitization, action recommendations for agility enhancement should be conceived to
derive a reference model in the second group discussion. The question for the group
discussion was, “How would you increase agility in a mature business like yours?” and
the participants were then allowed to freely respond. The group discussions took place
at intervals of almost three months.
According to the participants, the central field of tension concerns how agility should
be scaled in the company. The executive board member of the consulting company (No.
8, Table 2) stated: “Because we have been in the company for thirty years, [it means:]
I have always done it this way. On the other hand, we have things that are extremely
new, and finding the right balance is quite an exciting thing.” Therefore, the participants
recommended establishing different degrees of agility in the company. The executive
board member of the data center (No. 5, Table 2) emphasized the following statement:
“We have highly creative areas where cooperation must constantly reinvent itself, every
day, as new requirements appear. But on the other hand, the classic data center, where
it is really about that: one has to do some electrics that is repetitive.” In this context, as
a degree of agility the participants understand the need of clear processes and self-
organization of employees. For example, high agility means less control, less processes
96 D. Lindner and C. Leyh
restrictions and high self-organization of employees while low agility means strict
processes as well as a command and control leadership.
In addition, not every employee wants to work agilely, added the manager of the
pharmaceutical company (No. 10, Table 2), who quoted an employee: “I do my job!
Tell me what to do! And then I go home!” Thus, two key findings in the conception of
the recommended actions were foregrounded:
• Not every employee wants to work agilely.
• Not every area of business needs agility.
In this section, based on the recommendations for action, a reference model is formed
to increase agility in a classic and fictitious SME. As indicated previously, since a refer‐
ence model is not always fully applicable in practice nor can be presented in a simplified
scheme, a limitation has been subsequently constructed.
The model in Fig. 2 is a summary of the participants’ recommendations for action
in the context of agile SME transformation. It can be used as the basis for a first consid‐
eration to enhance agility. In order to display the demanded degree of agility, the partic‐
ipants recommended a division into three sectors differing in the proximity to the market.
Outside the circle is the market. Thus, the central services often do mostly internal tasks
and do not need to contact the customer or the market. Larger stable departments (e.g.,
addressing stable and long-term customers) with a mix of standard and individual
Organizations in Transformation 97
services require medium agility, and new pilot teams, in particular, require a high degree
of agility. A pilot team (startup) can be separated from a department. Startups or pilot
projects are projects that are very close to the market and often use agile methods – the
closer to the market, the greater the autonomy in the work execution or the need for
agility. With these different degrees of agility, the participants wanted to create a larger
flexibility for employees. As such, agility should also be used purposefully. It is beyond
question that agility may even be counterproductive in accounting. This can be justified
due to strict regulations and standards. Outside are the departments of the company,
which, as previously described, split from a certain size and form so-called startups. The
departments are supported by the central services through standardized services such as
HR, accounting, etc. As in the previous passage, in the absence of agility, these should
be divided into smaller units. However, this is not easy, and it is currently not possible
to specify exactly when a split or an integration should occur. The manager of the phar‐
maceutical company (No. 10, Table 2) stated: “You can do it well with the startups. But
then it is hard to manage to recognize the point when it no longer works as a startup and
you have to create other structures and mechanisms.” According to the participants, the
startups do not necessarily have full-time employees, but rather employees can be
assigned to a department and to one or more startups. The participants saw the basis for
a company as a mix of classical and agile methods, depending on the area.
Fig. 2. Reference model based on the recommendations of the focus group participants
The current discussion about business organization shows that, according to the Taylo‐
rist approach, a discussion about the humanization of work has occurred, which also
covers the topic of agile organization. In the face of the exponential digitization, this
debate is still tackling a general question: Is agility a prerequisite or consequence of
digitization? Experts have not reached an agreement here.
In the first group discussion, all participants preferred the thesis stating that an in-
creased agility is a prerequisite for digitization. Thus, the participants of the group
discussion are currently trying to agilely transform their companies and increase their
attractiveness for skilled workers, customers, and service providers through agile pilot
projects. In the second group discussion, the participants developed concrete recom‐
mendations for action in the form of a model dividing the company systematically and
in a self-organized manner, into small units. The classification is made according to the
degree of required agility. Areas of high agility are subject to loosened framework
conditions, greater autonomy in task completion and have direct customer contact. In
this way, employee and customer satisfaction can be enhanced. Based on these recom‐
mendations, an abstract reference model could be designed and considered as a recom‐
mendation to agilely convert a company regarding the digital transformation.
As limitation, our study is based on a small sample of participants and as already
stated, the reference model is only a simplified representation of reality and not always
fully applicable. However, although these results represent only the views of a small
group, it still broadens and continues the discussion on corporate organizational struc‐
tures in the digital age. Thus, the question that remains unanswered is:
How, taking into account the smaller budgets of SMEs, can a meaningful organizational and IT
architecture be designed to promisingly continue the increase of agility and flexibility of SMEs?
In our future research we will foster this unanswered question. Besides this question
there are many other interesting fields of digitization of SMEs, e.g., the role of leadership
in this kind of organizations. It would be interesting to understand how fast changing
and self-organized teams can be controlled and motivated. Therefore, we recommend
discussing current knowledge which is summarized under the term “digital leadership”.
Also, we are focusing on the aspect how classical work-models will change with the
increase of self-organization in companies (e.g., place and time flexible work). First
approaches can be found with the terms “newwork,” “work 4.0” or “smart working.” In
addition, we aim to further analyze our results by focusing on literature in strategic
management that has provided plenty of insights regarding the appropriate balance of
stability and flexibility of organizational structures, processes and work practices.
Organizations in Transformation 99
Appendix
References
1. Bajer, J.: Digital transformation needs the human touch. Strateg. HR Rev. 16, 91–92 (2017)
2. Mathrani, S., Mathrani, A., Viehland, D.: Using enterprise systems to realize digital business
strategies. J. Enterp. Inf. Manag. 26, 363–386 (2013)
3. Bley, K., Leyh, C., Schäffer, T.: Digitization of German enterprises in the production sector
– do they know how “digitized” they are? In: Proceedings of the 22nd Americas Conference
on Information Systems, AMCIS 2016 (2016)
100 D. Lindner and C. Leyh
4. Ludwig, T., Kotthaus, C., Stein, M., Durt, H., Kurz, C., Wenz, J., Doublet, T., Becker, M.,
Pipek, V., Wulf, V.: Arbeiten im mittelstand 4.0 – KMU im spannungsfeld des digitalen
wandels. HMD Prax. Wirtsch. 53, 71–86 (2016)
5. Urbach, N., Ahlemann, F.: Der Wissensarbeitsplatz der Zukunft: Trends, Herausforderungen
und Implikationen für das strategische IT-Management. HMD Prax. Wirtsch. 53, 16–28
(2016)
6. European Union: Commission Recommendation of 6th May 2003 concerning the definition
of micro, small and medium-sized enterprises. Off. J. Eur. Union 20th May 2003. (2003/361/
EC), pp. 36–41 (2003)
7. Lindner, D., Ludwig, T., Amberg, M.: Arbeit 4.0 – Konzepte für eine neue Arbeitsgestaltung
in KMU. HMD Praxis Der Wirtschaftsinformatik 6(1), 17 (2018)
8. Arbussa, A., Bikfalvi, A., Marquès, P.: Strategic agility-driven business model renewal: the
case of an SME. Manag. Decis. 55, 271–293 (2017)
9. Branicki, L.J., Sullivan-Taylor, B., Livschitz, S.R.: How entrepreneurial resilience generates
resilient SMEs. Int. J. Entrep. Behav. Res. (2017)
10. Lindner, D., Ott, M., Leyh, C.: Der digitale Arbeitsplatz – KMU zwischen Tradition und
Wandel. HMD Prax. Wirtsch. 54, 900–916 (2017)
11. Denning, S.: The next frontier for agile: strategic management. Strategy Leadersh. 45, 12–
18 (2017)
12. Beaumont, M., Thuriaux-Alemán, B., Prasad, P., Hatton, C.: Using agile approaches for
breakthrough product innovation. Strategy Leadersh. 45, 19–25 (2017)
13. Petry, T.: Digital Leadership: Erfolgreiches Führen in Zeiten der Digital Economy. Haufe
Gruppe, Freiburg München Stuttgart (2016)
14. Martensen, M., Ryschka, S., Blesik, T., Bick, M.: Collaboration in the consulting industry:
analyzing differences in the professional use of social software. Bus. Process Manag. J. 22,
693–711 (2016)
15. Schrauzer, S.: Computerisierung in der globalen softwareentwicklung – Eine
arbeitsmethodische betrachtung. HMD Prax. Wirtsch. 53, 42–54 (2016)
16. Fettke, P.: State-of-the-art des state-of-the-art: eine untersuchung der forschungsmethode
“Review” innerhalb der Wirtschaftsinformatik. Wirtschaftsinformatik 48, (2006)
17. Mayring, P.: Qualitative Inhaltsanalyse. In: Mey, G., Mruck, K. (eds.) Handbuch Qualitative
Forschung in der Psychologie, pp. 601–613. VS Verlag für Sozialwissenschaften, Wiesbaden
(2010)
18. Krueger, R.A., Casey, M.A.: Focus Groups: A Practical Guide for Applied Research. Sage,
Thousand Oaks (2015)
19. Greenbaum, T.L.: The Handbook for Focus Group Research. Sage, Thousand Oaks (1998)
20. Wilde, T., Hess, T.: Methodenspektrum der Wirtschaftsinformatik: Überblick und
Portfoliobildung. Institut für Wirtschaftsinformatik und Neue Medien der Ludwig-
Maximilians-Universität München (2006)
21. Schütte, R.: Grundsätze ordnungsmäßiger Referenzmodellierung. Gabler Verlag, Wiesbaden
(1998)
22. Termer, F.: Determinanten der IT-Agilität: Theoretische Konzeption, empirische Analyse
und Implikationen. Springer Fachmedien, Wiesbaden (2016)
23. Bair, J.H.: Personal computers and the office of the future. Telemat. Inform. 2, 113–117 (1985)
24. Noonan, M., Richter, G., Durham, L., Pierce, E.: Learning and the digital workplace: what?
so what? now what? Strateg. HR Rev. 16, 267–273 (2017)
25. Gastaldi, L., Corso, M., Raguseo, E., Neirotti, P., Paolucci, E., Martini, A.: Smart working:
rethinking work practices to leverage employees’ innovation potential. In: Proceedings of the
15th International CINet Conference (2014)
Organizations in Transformation 101
26. Arnold, H.: Digitalisierung der Unternehmensführung: Fallstudie und Plädoyer für mehr Mut
in einer flachen Führung und Organisation. Z. Führ. Organ. ZfO. 85, 330–336 (2016)
27. Jovanović, M., Mesquida, A.-L., Mas, A., Lalić, B.: Towards the development of a sequential
framework for agile adoption. In: Mas, A., Mesquida, A., O’Connor, Rory V., Rout, T.,
Dorling, A. (eds.) SPICE 2017. CCIS, vol. 770, pp. 30–42. Springer, Cham (2017). https://
doi.org/10.1007/978-3-319-67383-7_3
28. Hackl, B., Wagner, M., Attmer, L., Baumann, D.: New Work: Auf dem Weg zur neuen
Arbeitswelt. Springer Fachmedien, Wiesbaden (2017)
29. O’Connor, C., Kelly, S.: Facilitating knowledge management through filtered big data: SME
competitiveness in an agri-food sector. J. Knowl. Manag. 21, 156–179 (2017)
30. Essers, M.S., Vaneker, T.H.J.: Design of a decentralized modular architecture for flexible and
extensible production systems. Mechatronics 34, 160–169 (2016)
31. Ng, H.S., Kee, D.M.H.: Entrepreneurial SMEs surviving in the era of globalisation: critical
success factors. In: Sindakis, S., Theodorou, P. (eds.) Global Opportunities for
Entrepreneurial Growth: Coopetition and Knowledge Dynamics within and across Firms, pp.
75–90. Emerald Publishing Limited (2017)
32. Robertson, B.J.: Holacracy: The New Management System for a Rapidly Changing World.
Macmillan Publishers, New York (2015)
33. Gloger, B.: Scrum: Produkte zuverlässig und schnell entwickeln. Hanser, München (2016)
Information Security Management Systems -
A Maturity Model Based on ISO/IEC 27001
1
Instituto Superior Técnico, Universidade de Lisboa, Lisbon, Portugal
{diogo.proenca,jlb}@tecnico.ulisboa.pt
2
INESC-ID - Instituto de Engenharia de Sistemas e
Computadores Investigação e Desenvolvimento, Lisbon, Portugal
1 Introduction
In a growing and overly competitive world, only organizations that take advantage of
the benefits the best information can deliver for decision-making are able to profit and
thrive. Organizations should understand that information is such a valuable asset that it
must be protected and managed properly. Information security should be used as a way
to protect information against loss, exposure or destruction of its properties. [1] One of
the goals of information security is to ensure business continuity while minimizing the
impact of security incidents. In this sense, information is an asset that, like any other
important asset, is essential to an organization and therefore needs to be adequately
protected. This is especially important in the increasingly interconnected business
RQ1 - What are key requirements for an Information Security Management System
process according to the ISO/IEC 27001 relevant for the purpose of maturity
assessment?
RQ2 - How could a maturity Model specific to ISMS be designed which targets the
challenges of different organizations and industries?
To address these research questions, this paper is structured in six sections. First, the
key terms and concepts are explained in Sect. 2. This is Followed by Sect. 3, where the
research methodology is outlines. Section 4, details the findings from a literature review
in existing Information Security Management Maturity models and a comparison
between the existing maturity models for the Information Security Management domain.
Then Sect. 5, presents the ISMS Maturity model and the iterative development method
used. The evaluation of the ISMS Maturity Model is presented in Sect. 6 which evaluates
the mapping between the ISMS Maturity Model dimensions and the ISO/IEC 27001
requirements. This section also details the results of five assessments performed to five
104 D. Proença and J. Borbinha
different organization using the proposed maturity model. Finally, Sect. 7 details the
conclusions and the limitations of the ISMS maturity model.
2 Foundation
This section explains the key terns and concepts within this paper, such as, “maturity
models” and “information security management system” to ensure a common under‐
standing.
In 1986, the US Department of Defense needed a method to assess the capabilities
of the software companies with whom it worked, so Watts Humphrey, the SEI team and
Miter Corporation were tasked with this task. In 1991 was released the first version, the
CMM maturity model of capabilities. This model has achieved remarkable success and
has been revised and improved having evolved into CMMI, the currently integrated
capability maturity model integration version 1.3 [4].
Due to the success obtained, the principles used to develop the SEI maturity models
served as inspiration to other authors, both academics and practitioners, and there are
now hundreds of models applied to different domains [2]. Currently, the two major
references of maturity models are CMMI and ISO/IEC 15504, both of which are related
to Software Engineering processes.
In general, maturity can be defined as “an evolutionary progression in the demon‐
stration of a specific skill or in the achievement of an objective from an initial state to a
desired final state” [5]. In addition to the general definitions, there are many definitions
of maturity that are directly related to the domain to which this term refers. As this work
will develop a maturity model applied to a process of ISMS, it is also important to define
maturity applied to a process. Maturity can then be defined as the “degree to which an
organization executes processes that are explicitly and consistently documented,
managed, measurable, controlled, and continuously improved. Maturity can be meas‐
ured through appraisals” [4]. According to Loon [15], a maturity model is a sequence
of maturity levels for certain objects, usually people, organizations or processes. In these
models is represented the evolutionary path, anticipated, desired or typical, through
discrete levels. In addition to the above, these models provide the necessary criteria to
reach each of the model’s maturity levels. Thus, maturity models allow us to see at what
level of the evolutionary process certain objects meet. The maturity levels are organized
from an initial level of lower capacity to an advanced level corresponding to the
maximum capacity of the reality in question. In order to reach higher maturity levels, it
is necessary that there is a continuous progression of the capability of a given object.
ISO/IEC 27001 was based on the British standard BS7799 and ISO/IEC 17799. It
was prepared to provide the requirements to establish, implement, operate, monitor,
critically analyze, maintain and improve an ISMS [2]. An ISMS as defined by this
standard is “that part of the overall management system, based on a business risk
approach, to establish, implement, operate, monitor, review, maintain and improve
information security” [2]. This standard is used around the world by all types of organ‐
izations as the basis for the organization’s policy management and implementation of
information security. It is being used by small, medium and large organizations. In fact,
Information Security Management Systems 105
Fig. 1. PDCA model applied to ISMS processes and ISO/IEC 27001 mapping [2]
3 Research Methodology
In order to address the research questions of this paper, we selected the DSR paradigm
[17, 19]. DSR is described by “a designer answering questions relevant to human prob‐
lems via the creation of innovative artifacts, thereby contributing new knowledge to the
body of scientific evidence. The designed artifacts are both useful and fundamental in
understanding that problem” [19]. The major benefit of DSR is the fact that it addresses
real-world problems and simultaneously contributes to the body of knowledge [17].
However, the development of maturity models within the Information Security Manage‐
ment domain is not new but has been popular for quite some time [6]. Mettler, et al. [12]
count more than 100 models in the information systems domain, Poeppelbuss et al. [14]
counts even many more. One significant fault within this research area is the lack of
specific contributions regarding how to develop maturity models. Moreover, most
authors rarely describe their development process. Up to our knowledge there are only
a few development procedure models for maturity models. The models of Becker et al.
[16] and De Burin et al. [13] seem to be quite popular among the community based their
citation counts. We decided to apply the model of Becker et al. [16] to develop our
maturity model because it is based on DSR and therefore provides a methodological
foundation very suitable for application in our research approach. Furthermore, Becker
et al. provide a stringent and consistent development process according to the DSR
guidelines of Hevner et al. [17].
Becker et al. [16] argue that maturity models are artifacts that serve to solve the
problem of appreciating capacity and obtain improvement measures. According to [19]
design science allows you to create artifacts such as constructs, models, methods, and
106 D. Proença and J. Borbinha
instantiations that help improve problem-solving capabilities. Thus, the authors state
that design science research is appropriate for the development of maturity models.
In the same study [16], the author proposes a procedure for the development of
maturity models composed of eight steps. All steps should be documented. As depicted in
the procedure model in Fig. 2 the first steps focus on the problem identification (step 1). In
this step the research problem is identified and detailed, the practical relevance of the
problem is specified and the value of the artifact is justified. This step is followed by the
comparison with existing maturity models (step 2). This second step is based on the problem
identification of the first step and analysis of existing maturity model in the Information
Security Management domain, which leads to the identification of weaknesses in these
models. We conducted a literature analysis, which was based on an extensive online search
to find existing maturity models focused on the Information Security Management domain.
Thus, the analysis of the maturity models was performed according to their functionality,
as well as, their capability to address the ISO/IEC 27001 requirements.
Fig. 2. Procedure model of the research approach (adopted from Becker et al. [16])
The next step deals with the determination of the research strategy (step 3) outlined in
this section of the paper. This is followed by the iterative maturity model development (step
4). In this step, we used model adoption techniques, such as, configuration, instantiation,
aggregation, specialization and analogy [18] to incorporate the ISO/IEC 27001 in the
maturity model. This allowed us to create a rigorous maturity model regarding both the
structure and content. In the last step, evaluation (step 5), we combined the steps of Becker
et al. [16], conception of transfer and evaluation, implementation of transfer media, and
evaluation. All steps will be conducted, but to match the structure of this paper we made
this change.
Information Security Management Systems 107
4 Problem Analysis
In order to provide a consistent and precise problem definition, we gathered the ISMS
process requirements from ISO/IEC 27001. According to the ISO/IEC 27001, the activi‐
ties for ISMS Processes can be summarized as follows:
• A1: Establish the ISMS – “Establish ISMS policy, objectives, processes and proce‐
dures relevant to managing risk and improving information security to deliver results
in accordance with an organization’s overall policies and objectives.” [2];
• A2: Implement and operate the ISMS – “Implement and operate the ISMS policy,
controls, processes and procedures.” [2];
• A3: Monitor and Review the ISMS – “Assess and, where applicable, measure process
performance against ISMS policy, objectives and practical experience and report the
results to management for review.” [2];
• A4: Maintain and Improve the ISMS – “Take corrective and preventive actions, based
on the results of the internal ISMS audit and management review or other relevant
information, to achieve continual improvement of the ISMS.” [2].
These are the activities that the ISMS process must perform in order to be in line with
the recommendations of the ISO/IEC 27001. The activities are used as a reference base‐
line to assess the appropriateness of several existing Information Security Management
Maturity Models. Based on the results of the literature review we conducted within the
Information Security Management domain, we identified several papers dealing with
maturity models. We selected maturity models that used different methodological
approaches. Then, each maturity model was analyzed according to the degree to which
they cover and fit to the previously defined reference baseline. Each maturity model was
ranked for every requirement according to the degree of matching, using a Likert-scale,
from 1 (very low) to 5 (very high). After this analysis, we concluded that only six
maturity models scored an aggregate of at least 10 points according to the defined ISO/IEC
27001 activities baseline: (1) Open Information Security Management Maturity Model (O-
ISM3) [6]; (2) Systems Security Engineering – Capability Maturity Model (SSE-CMM)
[7]; (3) ISF Maturity Model Accelerator (ISF MM) [8]; (4) Control Objectives for Infor‐
mation and Related Technologies - Version 5 (COBIT 5) [9]; (5) Cyber Security Capa‐
bility Maturity Model (C2M2) [10], and (6) Building Security in Maturity Model
108 D. Proença and J. Borbinha
(BSIMM) [11]. Table 1 presents the assessment results of the above as the most signifi‐
cant identified maturity model in detail. Based on this set an average total score of 12 was
achieved (maximum score 20).
In accordance to the maturity model development approach of Becker et al. [16] a new
maturity model has to be developed, if no existing or the advancement of an existing
one is capable of addressing the identified problem. So, based on the findings of our
analysis there is no maturity model which satisfactorily fulfill the entire ISO/IEC 27001
activities baseline. Therefore, we will develop a new maturity model. The newly devel‐
oped maturity model presented in Table 2 adopts established structural elements,
domains and functions of the best practice in ISO/IEC 27001. As detailed within the
research methodology, we applied an iterative process for the maturity model develop‐
ment. In total we needed two iterations which can be detailed as follows:
First iteration: As a first step, we defined the characteristics and structure of the maturity
model. We started by proposing five maturity levels, Initial, Managed, Defined, Quan‐
titively Managed, and Optimizing. These maturity levels can be found in various estab‐
lished maturity models, such as, CMMI [4]. In this initial iteration, we focused in just a
part of the ISO/IEC 27001 ISMS process namely the Plan step. For each criterion of the
maturity model we modeled what was the manifestation of that criterion at the different
maturity levels.
Second Iteration: In the second iteration we completely overhauled the definition of the
maturity levels by proposing five new maturity levels, Initial, Planning, Implementation,
Monitoring, and Improvement. These maturity levels are based on the PDCA cycle used
within the ISO/IEC 27001 as depicted in Fig. 1. Table 3 details the activities on which
our maturity model is based, along with a mapping to the ISO/IEC 27001 ones they were
derived from. This made it easier for a user accustomed with the ISO/IEC 27001 to
understand the maturity model and make a connection between what was being asked
in each assessment criterion and the requirements specified in the ISO/IEC 27001, which
resulted in an easily understandable maturity model that is presented in Table 2. Finally,
this leads to the following maturity levels: (Level 1) Initial Stage; (Level 2) Planning
Stage; (Level 3) Implementation Stage; (Level 4) Monitoring Stage; (Level 5) Improve‐
ment Stage.
To improve from level X to level X + 1, the organization must comply with all the
criteria from level X, which makes this maturity model follow a “stages” approach. What
an organization can expect from progressing through the maturity levels is that their
ISMS process will become increasingly managed, defined and optimized.
Information Security Management Systems 109
The evaluation step is a main element of DSR. It is necessary to show the “utility, quality,
and efficacy of a design artifact” [19]. To be compliant with these requirements we
evaluated the ISMS Maturity Model by using a multi-perspective approach which
consists of three stages: (1) Evaluation of the mapping using the Wand and Weber
110 D. Proença and J. Borbinha
Table 3. Mapping of the ISMS maturity model and ISO/IEC 27001 requirements, and the
resulting evaluation using the Wand and Weber (W&W) ontological deficiencies.
PCDA ISMS maturity model activities ISO/IEC W&W
cycle 27001 ontological
requirements deficiencies
Plan Maturity level: planning
Define scope and boundaries of the ISMS 4.2.1 – (a) Complete
Define ISMS Policy 4.2.1 – (b) Complete
Define risk assessment approach 4.2.1 – (c) Complete
Risk identification 4.2.1 – (d) Complete
Risk analysis and evaluation 4.2.1 – (e) Complete
Risk treatment options 4.2.1 – (f) Complete
Risk treatment control objectives and controls 4.2.1 – (g) Complete
Obtain management approval for residual risks 4.2.1 – (h) Complete
Obtain management authorization to implement and operate the 4.2.1 – (i) Complete
ISMS
Prepare a statement of applicability 4.2.1 – (j) Complete
Do Maturity level: implementation
Formulate risk treatment plan 4.2.2 - (a) Complete
Implement risk treatment plan 4.2.2 - (b) Complete
Implement controls to meet control objectives 4.2.2 - (c) Complete
Define effectiveness measurement procedure of the selected 4.2.2 - (d) Complete
controls
Implement training and awareness programmes 4.2.2 - (e) Complete
Manage operation of the ISMS 4.2.2 - (f) Complete
Manage resources for the ISMS 4.2.2 - (g) Complete
Implement procedures and controls for detection and response 4.2.2 - (h) Complete
to security events.
Check Maturity level: monitoring
Execute monitoring and reviewing procedures and other 4.2.3 – (a) Complete
controls
Undertake regular reviews of the effectiveness of the ISMS 4.2.3 – (b) Complete
Measure the effectiveness of controls 4.2.3 – (c) Complete
Review risk assessments 4.2.3 – (d) Overload
Review residual risks 4.2.3 – (d) Overload
Review identified acceptable levels of risks 4.2.3 – (d) Overload
Conduct internal ISMS audits 4.2.3 – (e) Complete
Undertake management review of the ISMS 4.2.3 – (f) Complete
Update Security Plans 4.2.3 – (g) Complete
Record Actions and Events 4.2.3 – (h) Complete
Act Maturity level: improvement
Implement the identified improvements in the ISMS 4.2.4 – (a) Complete
Take appropriate corrective and preventive actions 4.2.4 – (b) Complete
Communication 4.2.4 – (c) Complete
Ensure that improvements achieve their objectives 4.2.4 – (d) Complete
Information Security Management Systems 111
The ontological evaluation of the mapping between the ISMS Maturity Model and
ISO/IEC 27001 chapters (see Fig. 1) and requirements is detailed in Table 3. A first
observation is that the mapping is complete, since every proposed activity can be mapped
to an ISO/IEC 27001 requirement. As for the other attributes, there is no redundancy
and excess. However, regarding overload, the ISO/IEC 27001 “4.2.3 - (d)” requirement
was overloaded as in our understanding it defines a requirement for three different
activities. As a result, we created three different assessment criteria for this requirement.
Finally, the ISMS Maturity Model covers all the requirements detailed in Sect. 4, which
means that the total score using the same scale is 20.
Following the first two evaluation steps, we assessed five real organizations by
following an assessment method, anonymized due to consent issues. Organization Alpha
is the public institute responsible for promoting and developing administrative modern‐
ization in its country. Its operation is in three axes: customer service, digital transfor‐
mation and simplification. Organization Beta is part of the business sector in its country
government that produces and supplies goods and services that require high security
standards, namely: coins, banknotes, and documents, such as, citizen’s card and
112 D. Proença and J. Borbinha
passports. Organization Gamma is a public higher education institution that has approx‐
imately 11.500 students being the largest school of engineering, science and technology
in its country. Organization Delta is a public institution for scientific and technological
research and development whose purpose is to contribute to the creation, development
and diffusion of research in fields related to civil engineering. Organization Omega is a
private organization which focus on software development and maintenance providing
services all over the globe with various offices in Europe.
For each of these five organizations we took the role of assessors, assessed the
organization collecting objective evidence for the assessment criteria defined in the
maturity model. Then, the results were analyzed which resulted in the assessment results
depicted in Table 4. In this table, “Y” stands for criterion satisfied, an empty cell stands
for criterion not satisfied, and the last columns shows the final maturity level for each
of the assessed organizations.
Criterion 2.1 2.2 2.3 2.4 2.5 2.6 2.7 2.8 2.9 2.10 3.1 3.2 3.3 3.4 3.5 3.6 3.7 3.8
ALPHA Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
Organization
BETA Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
GAMMA Y Y Y Y Y Y Y Y Y Y Y Y Y
DELTA Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
OMEGA Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y Y
Organization 4.1 4.2 4.3 4.4 4.5 4.6 4.7 4.8 4.9 4.10 5.1 5.2 5.3 5.4 Maturity Level
ALPHA Y Y Y Y Y Y Y Y Y Y Y Y Y Y 5
Organization
BETA Y Y Y 3
GAMMA Y Y 2
DELTA Y Y Y Y Y Y Y Y Y Y 4
OMEGA Y Y Y Y Y Y 3
In order to achieve a certain maturity level, the organization must comply with all
the criteria for that specific level and the levels below, which means that an organization
at maturity level 3 complies with all the criteria for maturity levels 2 and 3.
As can be perceived from Table 4, we were able to assess each of the assessment
criteria, which in turn allowed us to determine the ISMS maturity level for each of the
five organizations. From our analysis, the assessment results shown that the maturity
model correctly determined the maturity levels and these in fact correspond to our
perception of the maturity of the ISMS implemented in the organization. These results
were then used by the organizations to create improvement plans specially tailored to
their organizational context.
7 Conclusions
The aim of this paper is to detail the development of a maturity model for the ISMS
process based on the ISO/IEC 27001 standard. The latter can serve as a governance
instrument that could be used by the Information Security Management function to
analyze and evaluate the current strengths and weaknesses of the ISMS process.
Information Security Management Systems 113
However, the model is not restricted to analytical purposes only. It can also be used
to derive a roadmap towards an evolutionary improvement of the Information Security
Management function regarding its capabilities and its effectiveness and efficiency.
The first part of the paper elaborates the ISMS activities requirements which were
used as a reference baseline to investigate whether existing maturity models are capable
of holistically assessing an ISMS process (RQ1). The findings revealed that existing
maturity models cover the entire reference baseline insufficiently, since they only selec‐
tively address the activities. Hence, no existing maturity model is able to solve the
identified problem. Finally, we decided to design a new maturity model in consistency
to the defined research strategy.
In the second part of the paper, we described the development of a maturity model
for ISMS, including the model itself as well as its evaluation to address the second
research question (RQ2). The developed model is based on existing maturity model
structures and inherits concepts and methodologies of the ISO/IEC 27001. The
researchers took care during the development to provide a consumable research result.
Moreover, the ISMS maturity model benefited from the multi-perspective evaluation
approach by further advancements.
Naturally, the applied research approach comes along with certain limitations. This
paper presents the assessment results for five organizations using the ISMS Maturity
Model. However, in order to extend usefulness of the maturity model, as well as, provide
additional validation scenarios and further improve the research aspect, we suggest
evaluating (and refining) the ISMS maturity model within different industry sectors, this
would lead to a more generic ISMS maturity model and would enable cross-industry
benchmarking.
Acknowledgements. This work was supported by national funds through Fundação para a
Ciência e a Tecnologia (FCT) with reference UID/CEC/50021/2013.
References
1. Dubois, E., Heymans, P., Mayer, N., Matulevicius, R.: A systematic approach to define the
domain of information system security risk management. In: Nurcan, S., Salinesi, C.,
Souveyet, C., Ralyté, J. (eds.) Intentional Perspectives on Information Systems Engineering,
pp. 289–306. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-12544-7_16
2. ISO/IEC 27001:2013, Information technology - Security techniques - Information security
management systems – Requirements (2013)
3. Miller, H., Murphy, R.: Secure cyberspace: answering the call for intelligent action. IT
Professional (2009)
4. CMMI Product Team: CMMI for Development, Version 1.3, Carnegie Mellon Univ., no.
November, p. 482 (2010)
5. Mettler, T.: A design science research perspective on maturity models in information systems.
Institute of Information Management, University of St. Gallen, St. Gallen (2009)
6. The Open Group: Open Information Security Management Maturity Model (O-ISM3) (2011)
7. Carnegie-Mellon-University: Systems Security Engineering Capability Maturity Model
(SSE-CMM) - Model Description Document. Version 3.0 (2003)
114 D. Proença and J. Borbinha
8. ISF: Time to grow using maturity models to create and protect value, in Information Security
Forum ISF (2014)
9. IT Governance Institute: COBIT 5 – A business Framework for the Governance and
Management of Enterprise IT (2012)
10. Department of Energy, U.S. Department of Homeland Security, Cybersecurity Capability
Maturity Model (C2M2 v1.1) (2014)
11. McGraw, G., Migues, S., West, J.: Building Security in Maturity Model (BSIMM) Version
8 (2015)
12. Mettler, T., Rohner, P., Winter, R.: Towards a classification of maturity models in information
systems. In: D’Atri, A., De Marco, M., Braccini, A., Cabiddu, F. (eds.) Management of the
Interconnected World. Physica-Verlag, Heidelberg (2010). https://doi.org/10.1007/978-3-
7908-2404-9_39
13. De Bruin, T., Freeze, R., Kaulkarni, U., Rosemann, M.: Understanding the main phases of
developing a maturity assessment model. In: Proceedings of the Australasian Conference on
Information Systems (ACIS) (2005)
14. Poeppelbuss, J., Niehaves, B., Simons, A., Becker, J.: Maturity models in information systems
research: literature search and analysis. In: Communications of the Association for
Information Systems, vol. 29 (2011)
15. van Loon, H.: Process Assessment and Improvement: A Practical Guide. Springer, New York
(2015)
16. Becker, J., Knackstedt, R., Pöppelbuβ, J.: Developing maturity models for IT management:
a procedure model and its application. Bus. Inf. Syst. Eng. 3, 213–222 (2009)
17. Hevner, A., Ram, S., March, S., Park, J.: Design science in information systems research.
MISQ 28, 75–105 (2004)
18. Vom Brocke, J.: Design principles for reference modeling-reusing information models by
means of aggregation, specialization, instantiation, and analogy. In: Fettke, P., Loos, P. (eds.)
Reference Modeling for Business Systems Analysis. Idea Group Inc., Hershey (2007)
19. Hevner, A., Chatterjee, S.: Design Research in Information Systems: Theory and Practice.
Springer, Heidelberg (2010). https://doi.org/10.1007/978-1-4419-5653-8
20. Wand, Y., Weber, R.: On the ontological expressiveness of information systems analysis and
design grammars. Inf. Syst. J. 3(4), 217–237 (1993)
Repairing Outlier Behaviour
in Event Logs
1 Introduction
Process Mining bridges the gap between traditional data analysis techniques like
data mining and business process management analysis [1]. It aims to discover,
monitor, and enhance processes by extracting knowledge from event data, also
referred to as event logs, readily available in most modern information systems
[2]. In general, we identify three main branches of process mining being process
discovery, conformance checking, and process enhancement. In process discovery,
we try to discover a process model that accurately describes the underlying
c Springer International Publishing AG, part of Springer Nature 2018
W. Abramowicz and A. Paschke (Eds.): BIS 2018, LNBIP 320, pp. 115–131, 2018.
https://doi.org/10.1007/978-3-319-93931-5_9
116 M. Fani Sani et al.
process captured within the event data. In conformance checking we try to assess
to what degree a given process model (possibly the result of a process discovery
algorithm) and event data conform to one another. Finally, process enhancement
aims at improving the view on a process by improving its corresponding model
based on the related event data.
Most process mining algorithms, in any of these branches, work under the
assumption that behaviour related to the execution of the underlying process is
stored correctly within the event log. Moreover, completeness of the behaviour,
i.e., each instance of the process as stored in the event log is already finished, is
assumed as well. However, real event data often contains noise, i.e., behaviour
that is/should not be part of the process. Furthermore, it often contains data
related to infrequent behaviour, i.e., behaviour that is rather rare due to the
handling of exceptional cases. The presence of such behaviour, which we subse-
quently refer to as outlier, makes most process mining algorithms return complex,
incomprehensible or even inaccurate results. To reduce these negative effects,
process mining projects often comprise of a pre-processing phase in which one
tries to detect and remove traces that contain such undesired behaviour. This
cleaning phase is usually performed manually and is therefore rather costly and
time consuming.
Despite the negative impacts of the presence of noisy, incomplete and infre-
quent behaviour, little research has been done towards automated data cleans-
ing techniques that improve process mining results. Recently, research has been
performed aiming at filtering out traces that contain outlier behaviour from an
event log [3,4]. Even though both techniques show improvements in process min-
ing results, in particular w.r.t. process discovery, only a little fragment of outlier
behaviour within a trace of event data leads to ignoring the trace as a whole.
This potentially leads to a distortion of the general distribution of common
behaviour of the process, yielding potentially inaccurate or even wrong process
mining results.
Therefore, we propose a general purpose event log repairing technique that,
given event potentially containing outlier behaviour, identifies and modifies
such behaviour in order to obtain a more reliable input for all possible pro-
cess mining algorithms. In particular, we use a probabilistic method to detect
outlier behaviour according to the context of a process i.e., fragments of activ-
ity sequences that occur before and after the potential outlier behaviour. After
outlier identification, the corresponding behaviour is replaced with another frag-
ment of behaviour that is more probable to occur according to the context in
which the outlier behaviour occurs.
Using the ProM [5] based extension of RapidMiner, i.e., RapidProM [6], we
study the effectiveness of our approach, using both synthetic and real event data.
The obtained results show that our approach adequately identifies and repairs
outlier behaviour and as a consequence increases the overall quality of process
model discovery results. Additionally, we show that our method achieves better
results compared to one of the best performing existing filtering techniques in
the process mining domain.
Repairing Outlier Behaviour in Event Logs 117
2 Motivation
Real event logs often contain noise/anomalous behaviour. The presence of such
behaviour causes many problems for process mining algorithms. In particular,
most process discovery algorithms incorporate all behaviour in event logs as
much as possible. As a result, most outlier behaviour is incorporated as well
which decreases the overall accuracy of the discovered model. Therefore, it is
essential to accurately pre-process event data. In fact, according to the process
mining manifesto [7], cleaning event logs is one of the main challenges in the
process mining field.
It is possible to define outlier behaviour in a variety of ways. For example,
in [2], infrequent behaviour is considered as outlier behaviour. In turn, infrequent
behaviour relates to rare behaviour that is supposed to be part of the process, yet
severely hampers the feasibility of process mining results. In practice, behaviour
that is not part of the process, i.e. caused by logging exceptions, faulty execution
of the process, is infrequent by its sheer nature. So, in this paper we define both
infrequent behaviour that is part of the process and (infrequent) faulty execution
behaviour, i.e. noise, as outlier behaviour.
An example event log with some outlier behaviour is shown in Table 1. In
this event log there are 74 events belong to 11 traces. Except for the first variant
(unique behaviour of a process instance) each variant has just one corresponding
trace, note that this is customary in some application domains, e.g. medical treat-
ment process [8]. The first three traces contain no outlier behaviour. However,
the next seven variants have different types of outlier behaviour. For example,
in the fourth and fifth rows the activities “h” and “a” are missing. Some pro-
cess discovery algorithms like the Alpha miner [9] are sensitive to such outlier
behaviour and yield inferior process discovery results when applied directly to
such event logs. Other process discovery algorithms like the Inductive Miner
[10], have embedded filtering mechanisms to deal with some types of outliers.
The results of applying various process discovery algorithms on this event log
are shown in Fig. 1. If we apply filtering method of [4] on this event log, variants
1 − 5 are retained. A resulting process model using these traces combined with
the Inductive Miner is shown in Fig. 1e. As it is shown in this figure, all men-
tioned process discovery algorithms have problem to discover an accurate model
from the given event log. However, if we first repair the event log and then apply
the Alpha miner (or any other mentioned process discovery methods), we obtain
an accurate process model, i.e. as presented in Fig. 1f. It is because there is no
outlier behaviour in the repaired event log and the resulted process model is
more accurate and understandable.
118 M. Fani Sani et al.
Variant Frequency
a, b, c, d, e, f, h 2
a, b, d, c, e, f, h 1
a, b, c, d, e, g, h 1
a, b, d, c, e, g 1
b, d, c, e, g, h 1
a, b, c, d, e 1
a, b, c, d, g, h 1
a, b, b, c, d, e, f, h 1
a, b, c, d, g, h 1
a, b, d, c, a, e, g, f, h 1
(c) Result of the Inductive Miner with integrated filtering (IMi) on event log of Table 1
(e) Result of the Inductive Miner on the filtered event log by [4]
(f) Applying Alpha miner on the even log of Table 1 after repairing outlier behaviour in it.
3 Related Work
Some process mining algorithms are designed to/have been extended to be able
to handle outliers as well [11–14]. However, these filtering techniques are tailored
towards the internal working of the corresponding algorithms and cannot be used
for general purpose event log cleansing. Additionally, they typically focus on a
specific type of noise, e.g. incompleteness.
Most of filtering techniques present in commercial tools are based on just the
frequency of activities and variants. However, the presence of parallelism and
loops often hampers the applicability of such filtering techniques. There are also
some basic filtering plug-ins developed in ProM [5] based on activity frequencies
and users inputs.
Outlier detection for general temporal data is addressed in some research,
e.g. in [15] a survey on different methods of detecting outliers in sequential data
is presented. Also, there are some related techniques that specifically proposed
for process mining domain. In [16,17] the authors propose filtering techniques
that use additional information such as training event data or a reference pro-
cess model. In reality, providing a sufficiently complete set of training traces
or having a reference process model is impractical. Recently, some general pur-
pose filtering techniques are proposed in the process mining domain. In [3] by
constructing an Anomaly Free Automaton (AFA) based on the whole event log
and a given threshold, all non-fitting behaviour, w.r.t. the AFA, is removed
from the event log. In [4], we propose a filtering method that detects outliers
based on conditional probabilities of subsequences and their possible following
activities. In [18],an adjustable on-line filtering method is proposed that detect
outlier behaviour for streaming event logs that also works based on conditional
probabilities.
All aforementioned methods, after detecting outlier behaviour, try to remove
such behaviour. As motivated in Sect. 2, modifying outlier behaviour or repairing
it is more valuable than just removing it. There is some research [19,20] that
120 M. Fani Sani et al.
tries to repair process models based on event logs. Moreover, [21] uses process
model to repair event logs. However, as we aim to design a general purpose
repairing method, we assume there does not exist a process model as an input
of our proposed algorithm.
ple event log that presented in Table 1, f req(b, a, b, b, c, d, e, f, h) = 2 and
f req(b, d, a, b, d, c, e, g) = 1, etc.
Event logs describe sequences of executed business process activities, typ-
ically in context of some cases (or process instances), e.g., a customer or an
order-id. The execution of an activity in context of a case is referred to an event.
A sequence of events for a specific case is also referred to a trace. Thus, it is
possible that multiple traces describe the same sequence of activities, yet, since
events are unique, each trace itself contains different events. An example event
log, adopted from [2], is presented in Table 2.
Consider the events related to Case-id value 1. Sara registers a request, after
which Ali examines it thoroughly. William checks the ticket and checks resources.
Ava sends the request to manager and Fatima accept the request. Finally, Anna
Repairing Outlier Behaviour in Event Logs 121
emails the decision to the client. The example trace is written as a, b, c, d, e, f, h
(using short-hand activity names). In the context of this paper, we formally
define event logs as a multiset of sequences of activities (e.g., Table 1).
Also, the maximum length of covered subsequence (K) and context’s subse-
quences (CL ) respectively are specified by the user. Note that CL describes two
values, i.e. a maximal length for σ1 and σ2 .
The input of our proposed method is an event log L, a context frequency
threshold TC , a context’s subsequence lengths CL (l, r) and an upper bound for
length of covered subsequence K. At first, all traces are scanned to compute
covering probabilities of all contexts and their possible covered subsequences.
Next, for each trace and each subsequence in it (with length ≤ K), we check its
context frequency and covering probability (according to TC ). If the context is
significant and the covering probability is low, we replace the subsequence with
a better one according to that context. Otherwise, if a context is insignificant it
is not able to be used for repairing outlier behaviour.
A simple visual example of how the proposed method works is given in Fig. 2.
Fig. 2. A simple example of repairing an event log L with our proposed method. Sig-
nificant contexts and their probable subsequences are shown in the table for K = 1.
to the outlier subsequence we choose it and the trace is changed into a, c, b.
For the other outlier trace in this example, among the context (, b), it is more
probable that a occurs. So, by replacing with a, it changes to a, b, c.
To avoid having infinitive loops, after each replacement, the starting point of
the next scanned subsequence will be the first point of the right subsequence of
the previous context. For example, after replacement of a with and having
a, b, c, we will not check a again as a subsequence as we consider b as the
next subsequence.
5 Evaluation
1
Repair Log plugin svn.win.tue.nl/repos/prom/Packages/LogFiltering.
Repairing Outlier Behaviour in Event Logs 125
Table 3. Details of real event logs that are used in the experiment
Miner on cleaned event logs or R&IMi. Moreover, in most of cases, M &IM and
the Inductive Miner (here N &IMi ) are achieving their best possible process
model according to F-Measure, by sacrificing a lot in term of fitness. In Figs. 5
and 6, we see the result of different methods according to their conditional pre-
cision. According to these figures, the R&IM method helps the Inductive Miner
to discover process models with high precision without sacrificing the fitness.
However, we see that for event logs with lots of outlier behaviour the Inductive
Miner does not able to discover a process model with high fitness and precision
at the same time.
Note that for some event logs like the Sepsis− Cases neither M&IM nor
R&IM allow us to discover an outstanding process model, i.e., having high con-
ditional precision. The reason is for this is related to the fact that the discovered
process model of this event log is not precise and there is some behaviour that
possible to happen anywhere during execution of the process. We have shown
the best discovered process model for this event log to a business expert which
indicated that the process model is acceptable (but not perfect).
In Fig. 7, the best process model of each methods for the BP IC− 2017−
Offer− 05 are shown. It is obvious that Fig. 7a that is discovered on the repaired
event log is the best process model among them.
Note that, in the previous experiments, the original event log that is used
to compute F-Measures (our ground truth) potentially also contains noisy and
infrequent behaviour. Therefore, in another experiment, we create six artificial
process models with different behaviour. These process models consequently
describe different types of behaviour, i.e. one model just contains sequence con-
structs (Sequence), one contains sequence and choice constructs (Xor ), one con-
Fig. 4. Best F-measure of applying different methods on other real event logs.
Repairing Outlier Behaviour in Event Logs 127
Fig. 5. Best conditional precision of applying different methods on BPIC event logs.
Fig. 6. Best conditional precision of applying different methods on other real event
logs.
Fig. 7. The best discovered process models on BP IC− 2017− Offer− 05 using RL, MF
and NF methods.
128 M. Fani Sani et al.
Fig. 9. Best conditional precision of applying different methods on synthetic event logs.
tains sequence and parallel constructs (Parallel ), one contains sequence and loop
constructs (Loop), one contains all possible behavioural constructs (All ) and we
use a process model with locally structured subprocesses that has lots of vari-
ants (High Variants). Based on these reference process models, we generated six
event logs that each contains 5000 traces. Similar to the previous experiment,
5%, 10%, 20% and 50% of noise inserted to these original event logs. Note that,
an event log with 10% inserted noise not necessary has 90% original traces. Sim-
ilarly, we apply N &IMi, M &IM, R&IM and R&IMi to these event logs. Results
of this experiment are given in Figs. 8 and 9.
As shown in these figures, for all event logs if we insert some noisy behaviour
(even just 5%), the Inductive Miner has problems to discover a process model
with a high F-Measure. This means the embedded noise filtering method mech-
anism in this algorithm is not able to deal with all types of outlier. Furthermore,
Matrix Filter always works better than the embedded noise filtering method in
the Inductive Miner. Compared to these filtering methods, the proposed repaired
method performs better especially when there probability of noise is low. Only
for the Parallel event log, results of applying M &IM slightly outperforms our
proposed method. Like the previous experiment, the best results are achieved
when we apply the Inductive Miner with its filtering method to the repaired event
logs (R&IMi ). Except for the Parallel event logs, for all others this method finds
the underlying process models, i.e. almost equal to the reference process models.
For the High Variants event log that has local structured subprocesses, our pro-
posed repair method works perfectly and accurately detects and repairs outlier
behaviour. The reason for this is related to the fact that our method is able to
detect and repair outlier behaviour locally.
Repairing Outlier Behaviour in Event Logs 129
Table 4. The percentage of remaining traces in event logs for the best process model
Finally, note that filtering methods achieve the best results by removing a
lot of behaviour in event logs. The percentages of remaining traces in each event
log for the best model in M&IM are given in Table 4. In some cases the best
process model is discovered just with 5% of the traces. This means that we need
to remove a lot of behaviour from the event log. However, in the repair method
all the traces remain in the event log but they may be modified.
Note, for all methods we use a grid search on different parameters and show
the best results obtained. However, in reality like other state-of-the-art process
mining specific data cleansing methods, adjusting these thresholds is a challeng-
ing task for users.
6 Conclusion
Process mining provides insights into the actual execution of business processes,
by exploiting available event data. Unfortunately, most process mining algo-
rithms are designed to work under the assumption that the input data is outlier
free. However, real event logs contain outlier (noise and infrequent) behaviour
that typically leads to inaccurate/unusable process mining results. Detecting
such behaviour in event logs and correcting it helps to improve process mining
results, e.g. discovered process models.
To address this problem, we propose a method that takes an event log and
returns a repaired event log. It uses the occurrence frequency of a control-flow
oriented context pattern and the probabilities of different subsequences appear-
ing in middle of it to detect outlier behaviour. If such probability be lower than
a given threshold, the subsequence is substituted with a more probable one
according to the context.
To evaluate the proposed repairing method we developed a plug-in in the
ProM platform and also offer it through RapidProM. As presented, we have
applied this method on several real event logs, and compared it with other
state-of-the-art process mining specific data cleansing methods. Additionally,
we applied the proposed method on synthetic event logs. The results indicate
that the proposed repair approach is able to detect and modify outlier behaviour
and consequently is able to help process discovery algorithms to return models
130 M. Fani Sani et al.
References
1. van der Aalst, W.M.P.: Using process mining to bridge the gap between BI and
BPM. IEEE Comput. 44(12), 77–80 (2011)
2. van der Aalst, W.M.P.: Process Mining - Data Science in Action, 2nd edn. Springer,
Heidelberg (2016)
3. Conforti, R., La Rosa, M., ter Hofstede, A.H.M.: Filtering out infrequent behavior
from business process event logs. IEEE Trans. Knowl. Data Eng. 29(2), 300–314
(2017)
4. Sani, M.F., van Zelst, S.J., van der Aalst, W.M.P.: Improving process discovery
results by filtering outliers using conditional behavioural probabilities. In: Teniente,
E., Weidlich, M. (eds.) BPM 2017. LNBIP, vol. 308, pp. 216–229. Springer, Cham
(2018). https://doi.org/10.1007/978-3-319-74030-0 16
5. van der Aalst, W., van Dongen, B.F., Günther, C.W., Rozinat, A., Verbeek, E.,
Weijters, T.: ProM: the process mining toolkit. BPM (Demos) 489(31) (2009)
6. van der Aalst, W.M.P., Bolt, A., van Zelst, S.J.: RapidProM: mine your processes
and not just your data. CoRR abs/1703.03740 (2017)
7. van der Aalst, W., et al.: Process mining manifesto. In: Daniel, F., Barkaoui, K.,
Dustdar, S. (eds.) BPM 2011. LNBIP, vol. 99, pp. 169–194. Springer, Heidelberg
(2012). https://doi.org/10.1007/978-3-642-28108-2 19
8. Rebuge, Á., Ferreira, D.R.: Business process analysis in healthcare environments:
a methodology based on process mining. Inf. Syst. 37(2), 99–116 (2012)
9. van der Aalst, W.M.P., Weijters, T., Maruster, L.: Workflow mining: discovering
process models from event logs. IEEE Trans. Knowl. Data Eng. 16(9), 1128–1142
(2004)
10. Leemans, S.J.J., Fahland, D., van der Aalst, W.M.P.: Discovering block-structured
process models from event logs - a constructive approach. In: Colom, J.-M., Desel,
J. (eds.) PETRI NETS 2013. LNCS, vol. 7927, pp. 311–329. Springer, Heidelberg
(2013). https://doi.org/10.1007/978-3-642-38697-8 17
11. Leemans, S.J.J., Fahland, D., van der Aalst, W.M.P.: Discovering block-structured
process models from event logs containing infrequent behaviour. In: Lohmann, N.,
Song, M., Wohed, P. (eds.) BPM 2013. LNBIP, vol. 171, pp. 66–78. Springer, Cham
(2014). https://doi.org/10.1007/978-3-319-06257-0 6
12. van Zelst, S.J., van Dongen, B.F., van der Aalst, W.M.P., Verbeek, H.M.W.: Dis-
covering workflow nets using integer linear programming. Computing (2017)
13. Weijters, A.J.M.M., Ribeiro, J.T.S.: Flexible heuristics miner (FHM). In: CIDM
(2011)
Repairing Outlier Behaviour in Event Logs 131
14. Günther, C.W., van der Aalst, W.M.P.: Fuzzy mining – adaptive process simplifi-
cation based on multi-perspective metrics. In: Alonso, G., Dadam, P., Rosemann,
M. (eds.) BPM 2007. LNCS, vol. 4714, pp. 328–343. Springer, Heidelberg (2007).
https://doi.org/10.1007/978-3-540-75183-0 24
15. Chandola, V., Banerjee, A., Kumar, V.: Anomaly detection for discrete sequences:
a survey. IEEE Trans. Knowl. Data Eng. 24(5), 823–839 (2012)
16. Wang, J., Song, S., Lin, X., Zhu, X., Pei, J.: Cleaning structured event logs: a
graph repair approach. In: ICDE 2015, pp. 30–41 (2015)
17. Cheng, H.J., Kumar, A.: Process mining on noisy logs-can log sanitization help to
improve performance? Decis. Support Syst. 79, 138–149 (2015)
18. van Zelst, S.J., Fani Sani, M., Ostovar, A., Conforti, R., La Rosa, M.: Filtering
spurious events from event streams of business processes. In: Proceedings of the
CAISE (2018)
19. Fahland, D., van der Aalst, W.: Model repair-aligning process models to reality.
Inf. Syst. 47, 220–243 (2015)
20. Armas-Cervantes, A., van Beest, N., La Rosa, M., Dumas, M., Raboczi, S.: Incre-
mental and interactive business process model repair in Apromore. In: Proceedings
of the BPM Demos. CRC Press (2017)
21. Rogge-Solti, A., Mans, R.S., van der Aalst, W.M.P., Weske, M.: Improving doc-
umentation by repairing event logs. In: Grabis, J., Kirikova, M., Zdravkovic, J.,
Stirna, J. (eds.) PoEM 2013. LNBIP, vol. 165, pp. 129–144. Springer, Heidelberg
(2013). https://doi.org/10.1007/978-3-642-41641-5 10
22. Bolt, A., de Leoni, M., van der Aalst, W.M.P.: Scientific workflows for process min-
ing: building blocks, scenarios, and implementation. STTT 18(6), 607–628 (2016)
23. Weerdt, J.D., Backer, M.D., Vanthienen, J., Baesens, B.: A robust f-measure for
evaluating discovered process models. In: Proceedings of the CIDM, pp. 148–155
(2011)
ICT Project Management
Big Data Enabled Organizational
Transformation: The Effect of Inertia
in Adoption and Diffusion
Abstract. Big data and analytics have been credited with being a revolution
that will radically transform the way firms operate and conduct business.
Nevertheless, the process of adopting and diffusing big data analytics, as well as
actions taken in response to generated insight, necessitate organizational
transformation. Nevertheless, as with any form of organizational transformation,
there are multiple inhibiting factors that threaten successful change. The purpose
of this study is to examine the inertial forces that can hamper the value of big
data analytics throughout this process. We draw on a multiple case study
approach of 27 firms to examine this question. Our findings suggest that inertia
is present in different forms, including economic, political, socio-cognitive,
negative psychology, and socio-technical. The ways in which firms attempt to
mitigate these forces of inertia is elaborated on, and best practices are presented.
We conclude the paper by discussing the implications that these findings have
for both research and practice.
1 Introduction
While big data analytics have been in the spotlight of attention by researchers and
practitioners in the last few years, to date there has been limited attention on what
forces can potentially hinder the potential business value that these investments can
deliver. Much research has focused on the necessary investments that must be made to
derive business value [1], but the process from making the decision to adopt such
technologies, up to turning insight into action is seldom discussed, particularly with
respect to inertia. The underlying premise of big data dictates that such investments can
generate insight with the potential to transform the strategic direction of firms, and help
them outperform competition [2]. Nevertheless, this process entails organizational
transformation at multiple levels, and as with any case of organizational transformation,
is subject to path dependencies, routinization, and other hindering forces [3].
While big data literature has documented the importance that organizational
learning and a data-driven culture have on overall project success [4], there is to date a
very limited understanding on how these should be implemented and what factors may
inhibit successful deployment or even adoption. In this respect, there is not much
attention on the processes of big data adoption and implementation. Most studies to
date have attempted to provide a narrative on how big data can produce value [5], or
even empirically show an association between investments and performance measures
[1, 6]. Yet, in reality, managers and practitioners are faced with a number of hurdles
which need to be overcome, on individual, group, organizational, and industry levels.
The purpose of this study is therefore to attempt to understand how inertial forces in
these levels hinder the potential value of big data analytics. By doing so, it is possible
to isolate key success factors of implementation and help guide practitioners in
developing strategies for adoption and deployment [7].
Hence, this research is driven by the following research question which helps guide
our investigation:
How is inertia present in big data projects? At what stages do inertial forces appear and at
what levels?
Big data analytics is widely regarded as the next frontier of improving business per-
formance due to its high operational and strategic potential [8]. The literature defined
big data analytics as “a new generation of technologies and architectures, designed to
economically extract value from very large volumes of a wide variety of data, by
enabling high velocity capture, discovery and/or analysis” [9]. While most claims on
the value of big data analytics are anecdotal, the emerging literature has documented a
positive relationship between the decision to invest in firm-wide deployment of big data
analytics and performance [1]. Big data analytics enable firms to make sense of vast
amounts of data and reconfigure their strategies based on trends that are observed in
their competitive environment [10]. The importance of big data analytics is evident
from the increasing investments made from firms, and particularly those working in
complex and fast-paced environments [11]. Managers nowadays are relying ever more
on big data analytics to inform their decision-making and direct future strategic ini-
tiatives [12]. The value of investing in big data analytics is clearly reflected in a recent
article by Liu [13], who notes that big data analytics constitutes a major differentiator
between high-performing and low-performing firms, as it enables firms to be more
proactive and swift in identifying new business opportunities. Additionally, the study
Big Data Enabled Organizational Transformation 137
reports that big data analytics have the potential to decrease customer acquisition costs
by 47% and enhance revenues by about 8%. A report by MIT Sloan Management
Review shows that companies that are leaders in the adoption of big data analytics are
much more likely to produce new products and services compared to those that are
laggards [14]. Nevertheless, the value that firms realize from big data investments, is
highly contingent upon the idiosyncratic capabilities that they develop in deriving
meaningful insight [15, 16].
The first step however is to clearly define and understand how the different sources
of inertia have been examined in literature and at what level they are present. Negative
psychology inertia has been predominantly attributed to group and individual behavior,
and is based on threat perceptions of losing power or even their position. Uncertainty
about the role of individuals or groups in the face of novel technological deployments
thus causes negative psychological reactions which biases them towards the status quo
[25]. Socio-cognitive inertia emphasize mostly on malleability due to path dependen-
cies, habitualization, cognitive inertia and high complexity [26]. These forms of inertia
arise due to periods of sustained stability and routinization caused by a stable envi-
ronment. Socio-technical inertia on the other hand refers to the dependence on socio-
technical capabilities, which arise from the interaction of the social systems and
technical system and their joint optimization [27]. Economic inertia may be present in
the form of commitment to previously implemented IT solutions that do not pay off and
create sunk costs, or through transition expenses which cause organizations to not
adopt potentially better alternatives [19]. Finally, political inertia is caused by vested
interests and alliances which may favor that the organization remains committed to a
specific type of information technology so that partnerships are not broken.
While to date there has been no systematic study to examine the forms of inertia in
big data analytics implementations, several research studies have reported inhibiting
factors during adoption and diffusion. Mikalef, Framnes, Danielsen, Krogstie and
Olsen [28] mention that in some cases economic inertia caused a problem in the
adoption of big data analytics. The authors state that top managers in some cases were
reluctant to make investments in big data analytics, since their perceptions about the
cost of such investments in both technical and human resources greatly exceeded the
potential value. In addition, they mention that both socio-cognitive and socio-technical
issues rose at the group level, where people were reluctant to change their patterns of
work, and were also afraid of losing their jobs. Similar findings are reported by Janssen,
van der Voort and Wahyudi [15], where socio-cognitive inertia can be reduced by
implementing governance schemes [29], which dictate new forms of communication
and knowledge exchange. In their study, Vidgen, Shaw and Grant [4] note that inertial
forces impact the implementation of big data projects, and that the presence of the right
people that can form data analytics teams and implement processes is critical to
success.
4 Research Method
4.1 Design
Beginning from the theoretical background and the overview of existing literature on
big data-enabled organizational transformation and business value, the present work
aims to understand how the processes of deploying big data analytics within companies
is hindered by different forms of inertia as well as decision making barriers by top
managers in the process of decision making. We explain how inertia is presented at
different forms and stages throughout the deployment and routinization of big data
Big Data Enabled Organizational Transformation 139
5 Results
After transcribing the interviews and assigning them each a thematic tag as those
described in Table 2, we started aggregating finding and identifying common patterns.
More specifically, the inertial forces and how they are presented in big data projects are
summarized below.
142 P. Mikalef et al.
“The management was very skeptical about if we should go into big data. Our competitors were
doing that and it meant that we had to follow so that we are not left behind. We tried to
experiment in the beginning with some internal resources but then we understood that we have
to invest more. This was a hard decision to make and perhaps we delayed a bit on this
(adopting big data)”
Similar quotes were made by several other executives, showcasing that economic
inertia are a major inhibiting factor of big data adoption and deployment. A major
Big Data Enabled Organizational Transformation 143
inhibitor that leads to this is the unclear link between big data investments and business
value. On the other hand, competitive pressures seem to be driving mimetic behaviors
in companies mitigating the effects of economic inertia.
Similar examples were found in several companies working external partners such
as universities, public bodies, and other companies, as well as in firms that have highly
siloed business units. We found that in many cases, consulting firms were brought in to
resolve this issue and act as a mediating agent. They tried to create a common
understanding of the objectives of the big data projects, and bring representatives from
each business unit to the table so that cognitive structures can be in alignment.
144 P. Mikalef et al.
We saw that the way many IT managers handled this issue was by providing their
division with small challenges, and incrementally growing projects. Also, they
assigned a few hours a week where they had the freedom to experiment with big data
tools. This allowed them to try out these novel tools at their own pace, without the fear
of time-pressure to deliver results.
5.5 Socio-Technical
In terms of socio-technical inertia, it was observed that in many cases middle-level
managers exerted behaviors that stalled the implementation of big data projects. Their
primary fear was that decision-making would now reside in insight from analytics,
therefore replacing them. In many occasions this fear manifested itself as a distrust
towards the value of big data analytics, and general tendency to downplay the sig-
nificance of big data in operations. Despite the clear directive of top management to
diffuse big data analytics into operations, in many circumstances middle-level man-
agers would not invest time in clear implementation strategies leaving the IT depart-
ment with a bleak understanding of how they should proceed. The respondent from C.6
states the following.
“When we made the decision to start working with big data we were happy. Then after some
time we were not sure what we should do with it. We presented some examples to management
but they didn’t take them much into account. We then realized that they prefer to make decisions
based on their own knowledge, and didn’t trust the process we were following, or even, felt like
it may replace them. It was quite discouraging to work on something that is not applied”
Typically, these issues were resolved by a strong top management vision and
leadership. In addition, training seminars for middle-level managers on the value of big
data and their role were regarded as very beneficial in overall success.
Big Data Enabled Organizational Transformation 145
6 Discussion
In the current study we have examined how inertia during the implementation and
deployment phases of big data projects influence their success. Following the literature
which distinguishes between economic, political, socio-cognitive, negative psychol-
ogy, and socio-technical, we looked at how these forces of inertia are manifested in
contemporary organizations through 27 case studies. Our results show that value from
big data investments, and even actual implementation, can be hindered by multiple
factors and at multiple levels which need to be considered during the planning phase.
To the best of our knowledge this is one of the first attempts to isolate these inhibiting
forces and provide suggestions on which future research can build and managers can
develop strategies for adopting and diffusing their big data investments.
From a research perspective the major finding of this research is that even in the
presence of all necessary big data analytics resources, there are multiple ways in which
a business value can be hindered. This raises the question of how can these obstacles be
overcome. While there is a stream of research into the issues of information governance
these studies primarily focus on the issue of how to handle data and how to appropriate
decision making authority in relation to the data itself. There still seems to be an
absence of governance schemes that follow a holistic perspective and include man-
agement and organization of all resources, including human and intangible ones. In
addition, how firms should handle individual, group and industry-level dynamics is a
topic that is hardly touched upon.
From a managerial point of view, the outcomes of this study outline strategies that
can be followed to mitigate the effects of the different types of inertia. Our findings
indicate that inertia can be present at many phases of adoption and diffusion so action
need to be taken throughout projects. It is critical to consider the socio-technical
challenges that these technologies create for middle-level managers and clearly
understand how their decision-making is influenced or not by insight generated by big
data. In addition, it is important to develop strategies so that the whole organization
adopts a data-driven logic, and that a common understanding and language is estab-
lished. With regards to the IT department, educational seminars and incremental pro-
jects seem to be the way to limit negative psychology barriers. Also, providing a clear
sense of direction as to what kind of analytics are to be performed on what data is of
paramount importance. It is commonly observed that many companies delve into the
hype of big data without having a clear vision of what they want to achieve.
While this research helps to uncover forces of inertia and the levels at which they
present themselves, it does not come without limitations. First, we looked at companies
that have actually adopt big data, a more complete approach would be to look at what
conditions cause other firms to not opt for big data. Second, while we briefly touched
on the issue of middle-level managers not following insight generated from big data, it
is important to understand in more detail the decision-making processes that underlie
their reasoning. Also, the actions that are taken in response to these insights are seldom
put into question. This is a future area which should be examined since the value of big
data cannot be clearly documented in the absence of knowledge about strategic or
operational choices.
146 P. Mikalef et al.
Acknowledgments. This project has received funding from the European Union’s
Horizon 2020 research and innovation programme, under the Marie Sklodowska-Curie grant
agreement No. 704110.
References
1. Gupta, M., George, J.F.: Toward the development of a big data analytics capability. Inf.
Manag. 53, 1049–1064 (2016)
2. Prescott, M.: Big data and competitive advantage at Nielsen. Manag. Decis. 52, 573–601
(2014)
3. Sydow, J., Schreyögg, G., Koch, J.: Organizational path dependence: opening the black box.
Acad. Manag. Rev. 34, 689–709 (2009)
4. Vidgen, R., Shaw, S., Grant, D.B.: Management challenges in creating value from business
analytics. Eur. J. Oper. Res. 261, 626–639 (2017)
5. McAfee, A., Brynjolfsson, E., Davenport, T.H.: Big data: the management revolution.
Harvard Bus. Rev. 90, 60–68 (2012)
6. Wamba, S.F., Gunasekaran, A., Akter, S., Ren, S.J.-f., Dubey, R., Childe, S.J.: Big data
analytics and firm performance: effects of dynamic capabilities. J. Bus. Res. 70, 356–365
(2017)
7. Mikalef, P., Pappas, I.O., Giannakos, M.N., Krogstie, J., Lekakos, G.: Big data and strategy:
a research framework. In: MCIS, p. 50 (2016)
8. Brown, B., Chui, M., Manyika, J.: Are you ready for the era of ‘big data’. McKinsey Q. 4,
24–35 (2011)
9. Mikalef, P., Pappas, I.O., Krogstie, J., Giannakos, M.: Big data analytics capabilities: a
systematic literature review and research agenda. Inf. Syst. e-Bus. Manage., 1–32 (2017)
10. Chen, H., Chiang, R.H., Storey, V.C.: Business intelligence and analytics: from big data to
big impact. MIS Q. 36, 1165–1188 (2012)
11. Wang, G., Gunasekaran, A., Ngai, E.W., Papadopoulos, T.: Big data analytics in logistics
and supply chain management: certain investigations for research and applications. Int.
J. Prod. Econ. 176, 98–110 (2016)
12. Constantiou, I.D., Kallinikos, J.: New games, new rules: big data and the changing context
of strategy. J. Inf. Technol. 30, 44–57 (2015)
13. Liu, Y.: Big data and predictive business analytics. J. Bus. Forecasting 33, 40 (2014)
14. Ransbotham, S., Kiron, D.: Analytics as a source of business innovation. MIT Sloan Manag.
Rev. (2017)
15. Janssen, M., van der Voort, H., Wahyudi, A.: Factors influencing big data decision-making
quality. J. Bus. Res. 70, 338–345 (2017)
16. Mikalef, P., Pateli, A.: Information technology-enabled dynamic capabilities and their
indirect effect on competitive performance: findings from PLS-SEM and fsQCA. J. Bus. Res.
70, 1–16 (2017)
17. Karahanna, E., Straub, D.W., Chervany, N.L.: Information technology adoption across time:
a cross-sectional comparison of pre-adoption and post-adoption beliefs. MIS Q. 23(2), 183–
213 (1999)
18. Besson, P., Rowe, F.: Strategizing information systems-enabled organizational transforma-
tion: a transdisciplinary review and new directions. J. Strateg. Inf. Syst. 21, 103–124 (2012)
19. Haag, S.: Organizational inertia as barrier to firms’ IT adoption–multidimensional scale
development and validation (2014)
Big Data Enabled Organizational Transformation 147
20. Polites, G.L., Karahanna, E.: Shackled to the status quo: the inhibiting effects of incumbent
system habit, switching costs, and inertia on new system acceptance. MIS Q. 36(1), 21–42
(2012)
21. Kelly, D., Amburgey, T.L.: Organizational inertia and momentum: a dynamic model of
strategic change. Acad. Manag. J. 34, 591–612 (1991)
22. Hannan, M.T., Freeman, J.: Structural inertia and organizational change. Am. Sociol. Rev.
49, 149–164 (1984)
23. Stieglitz, N., Knudsen, T., Becker, M.C.: Adaptation and inertia in dynamic environments.
Strateg. Manag. J. 37, 1854–1864 (2016)
24. Barnett, W.P., Pontikes, E.G.: The Red Queen, success bias, and organizational inertia.
Manag. Sci. 54, 1237–1251 (2008)
25. Kim, H.-W., Kankanhalli, A.: Investigating user resistance to information systems
implementation: a status quo bias perspective. MIS Q. 33(3), 567–582 (2009)
26. Lyytinen, K., Newman, M.: Explaining information systems change: a punctuated socio-
technical change model. Eur. J. Inf. Syst. 17, 589–613 (2008)
27. Rowe, F., Besson, P., Hemon, A.: Socio-technical inertia, dynamic capabilities and
environmental uncertainty: senior management view and implications for organizational
transformation (2017)
28. Mikalef, P., Framnes, V.A., Danielsen, F., Krogstie, J., Olsen, D.H.: Big data analytics
capability: antecedents and business value. In: Pacific Asia Conference on Information
Systems (2017)
29. Mikalef, P., Krogstie, J., van de Wetering, R., Pappas, I., Giannakos, M.: Information
Governance in the big data era: aligning organizational capabilities. In: Proceedings of the
51st Hawaii International Conference on System Sciences (2018)
30. Benbasat, I., Goldstein, D.K., Mead, M.: The case research strategy in studies of information
systems. MIS Q. 11, 369–386 (1987)
31. Battistella, C., De Toni, A.F., De Zan, G., Pessot, E.: Cultivating business model agility
through focused capabilities: a multiple case study. J. Bus. Res. 73, 65–82 (2017)
32. Gregor, S.: The nature of theory in information systems. MIS Q. 30, 611–642 (2006)
33. Yin, R.K.: Case Study Research and Applications: Design and Methods. Sage Publications
(2017)
34. Boudreau, M.-C., Gefen, D., Straub, D.W.: Validation in information systems research: a
state-of-the-art assessment. MIS Q. 25(1), 1–16 (2001)
35. Myers, M.D., Newman, M.: The qualitative interview in IS research: examining the craft.
Inf. Organ. 17, 2–26 (2007)
Amalgamation of 3D Printing Technology
and the Digitalized Industry – Development
and Evaluation of an Open Innovation
Business Process Model
Abstract. Innovations and trends have a significant impact on industries and the
way that businesses are conducted. Therefore successful business models need to
be increasingly adaptable. We develop a business process model (BPM) focusing
primarily on manufacturing companies. This model not only illustrates how the
3D printing (3DP) technology could be implemented in production processes, but
also how the approach of open innovation (OI) could be adopted with the purpose
of increasing innovative capabilities. We first perform a literature review of the
relevant fields; in the second step we develop a BPM (using BPMN 2.0), which
is evaluated through expert interviews in a third step. It is demonstrated that the
convergence towards highly flexible and customer-oriented production processes
and the adoption of OI enhance the amalgamation of 3DP with the digitalized
industry.
1 Introduction
According to [1], in the future the manufacturing industry will be reshaped in large parts.
This means that advantages of volume production will gradually become insignificant,
making traditional production methods obsolete in the long run. 3D printing is
mentioned as one of the key trends significantly affecting the manufacturing industry.
The range of printable materials continues to expand, while prices continue to fall,
making production at low costs economically viable even for standardized parts in the
near future [2]. These developments increasingly require manufacturers to consider how
advanced information and communication technologies (ICT) might facilitate both
higher efficiency and adaptability of their operational processes to ultimately contribute
to their value generation [3].
This paper contributes to the ongoing restructuring of the manufacturing industry by
developing specific guidance for manufacturers with the purpose of enabling, through
2 Research Methodology
The underlying research process can be summarized in five steps as shown in Fig. 1:
These steps were complemented by the literature review process following [7]. After
the broad topic was selected an initial literature review confirmed the existence of a
research gap, which specified our research focus. The following search string was
developed and applied to the databases AISeL, EBSCO Host, science direct, Springer
and web of science:
(“open innovation” OR “open business model”) AND (“3D printing” OR “additive manufac‐
turing” OR “rapid prototyping”) AND (“digitaliz*” OR “industry”) AND (“process” OR
“crowdsourcing” OR “co-creation”).
In a second step the identified literature was analyzed and filtered for keywords
relevant to our research objectives, including business process model (BPM), crucial
success factor (CSF) or relevant actors and departments in keywords, titles and abstracts
of the literature sample. The third step was to assess the full texts of the extracted liter‐
ature. The data was considered adequate if it bridges the gap at least in parts and answers
important questions within this research. Further it helps to build the fundament in order
to establish the pursued business process model. Development of the BPM forms the
core of this contribution (Sect. 3). Visualization of the BPM was executed using BPMN
2.0. Additionally, hypotheses were developed in order to evaluate the developed BPM
(Sect. 4). For this purpose, seven industry experts were interviewed using a questionnaire
that was developed following [8]. The fourth step includes the discussion and interpre‐
tation of the findings, which is given in Sect. 5. Step five finally aims to summarize the
results, to enable an outlook for the future manufacturing industry and to give directions
for further research (Sect. 6).
150 D. Warnecke et al.
Our aim is to develop an innovative BPM focusing on the manufacturing industry and
to demonstrate how a future manufacturer could implement 3DP as well as the OI
approach in their business. To recognize and absorb the dynamics of the environment
through an innovative BPM it needs to be well defined [5]. As the use of information
systems is highly conceptual and design-oriented, it helps to identify, formalize and
visualize the key elements including their relations and operational sequences [27]. To
highlight the agile and automatable processes that constitute our BPM we are using
BPMN 2.0. The CSF obtained from the literature review function as a starting point for
development (see Table 1).
In the following example, three players are included, i.e. the company, a platform
and to begin with only one customer. The communicational path between this customer
and the platform exists during the entire process and is illustrated as information flows
between the pools. In a next step it is aspired to extend the BPM, e.g. to implement
multiple customers or partners in the innovation process more deeply.
In this initial version, there are several tasks in the BPM marked as subtasks, which
are actually more complex than displayed and would have required an extension of the
process. Company A is a manufacturing company and divided into five departments
[28]: Management, Technical Department (TD), Marketing Department (MD), Invest‐
ment Department (ID) and Production Department (ProdD), whereas TD is further
divided into Technical Service (TS) and Product Design (PD) and ProdD is further
divided into Production Planning (PP) and Production/3DP. The purpose of this classi‐
fication is to increase the transparency and structure of the example. The following
152 D. Warnecke et al.
Step 1 (Fig. 2)
Step 2 (Fig. 3)
OI-Platform: In this case, the idea is allocated to manufacturing company A and the
manufacturer is informed, so that collaboration with all participants can begin. Addi‐
tionally, the idea is categorized and passed on to the assessment.
Management: After the company is informed, all information is gathered and the coor‐
dination of the involved departments begins. Then the idea is internally classified. After
a clear coordination with the Technical Department (TD), the Management awaits the
blueprint.
Technical Service (TS): Now the processes in this subdivision begin. The TS checks
general technical requirements and provides necessary information for assessment of
the idea. Next, the TS coordinates the personnel including company experts.
Amalgamation of 3D Printing Technology and the Digitalized Industry 153
Step 3
OI-Platform: If the assessment concludes the idea is not feasible, reasons for this will
be specified and the idea will be added to the pool of “not possible ideas”. Initially
rejected ideas might still be considered at a later stage. If this happened in the BPM, the
token would move directly from the gateway “step 1, possible?” via the sequence flow
“No or not yet” to the second gateway “step 2, possible?”. This means, that if the idea
is not or not yet possible, all necessary information will be updated and coordinated with
all parties involved before the project ends. In turn, if the idea is deemed feasible, it will
be added to the pool of possible ideas and all necessary information will be updated
before the collaboration continues. To handle the possibility that insolvable problems
may also occur at a later stage, such as during the collaboration, a compensatory inter‐
mediate event is implemented in the BPM, which leads the idea to “not possible ideas”.
Ultimately, in a positive case a detailed concept is created (intermediate event) and the
OI platform/customer awaits the final product.
Step 4 (Fig. 4)
Product Design (PD): The processes start after the detailed concept is received. In order
to develop a design and a blueprint, design developers are provided. The process ends
with finalization of the blueprint.
Production Planning (PP): The realization processes start after the blueprint in the PD
is available and the coordination has taken place. Next, the request for the PP is checked
and aligned with own systems. Schedule time, capacity and other production require‐
ments are checked, so that a needs-based planning can continue. As can be seen in the
BPM, another compensatory intermediate event is implemented at this stage in case a
needs-based planning is not possible, for instance due to time restrictions or to avoid
obstruction of the company’s objectives. Such a compensatory event would lead to an
update of all information related to the order, inform the Management, and finally halt
the processes. Assuming that all requirements are fulfilled and Management gives their
154 D. Warnecke et al.
authorization, the PP continues by specifying the order, so that the production process
can be determined and the order transmitted appropriately to Production/3DP. For this
purpose, the next process step involves a complex gateway allowing an iterative iden‐
tification of the appropriate production type. The following step is an exclusive gateway.
It illustrates that the first five production types are to be produced by 3DP, whereas the
fifth type is a dummy position for another classification leading to 3DP and the sixth
type is classified as volume production. In this example, the product to be produced is
a highly customized product and passed on to the 3DP subdivision. Now, the processes
in the PP end.
Marketing Department (MD): Hence, this example does not include criteria for a
marketing plan.
Investment Department (ID): The processes begin upon coordination with the Manage‐
ment. As soon as all information required is collected, the evaluation, including a prof‐
itability, required budget and prioritization check, begins.
Amalgamation of 3D Printing Technology and the Digitalized Industry 155
Management: The Management now possesses all information required for a final
approval of the product.
Production/3DP: The processes for the final steps are initiated. As already determined,
3DP is required in this example; hence the process chain above the inclusive gateway
is activated. The order is now printed and then Management is informed that the product
has been completed.
Step 5 (Fig. 5)
OI-Platform: Once the product has been finalized, the step “Coordination with all
parties involved” ensures that all participants receive a progress update before the project
is declared successful.
The success of the developed concept depends on the CSF resulting from the literature
review (see Table 1). In order to test the five developed hypotheses, a survey was
conducted using a questionnaire following [8] (see Appendix B). The interviews were
carried out with seven experts working for different internationally active consulting
companies in the field of ICT. Divided into sections that aim the five hypotheses the
questionnaire consists of 16 questions using a seven step likert scale, which include
subquestions at certain points and an optional part for additional information and clar‐
ifications. The business process model was also provided as a DIN A2 sized poster and
explained briefly. The potential feasibility and practical relevance of the business
process model was addressed as final part of the interviews (see Sect. 5).
156 D. Warnecke et al.
5 Discussion of Findings
All participants substantiate the indispensability of 3DP for future production processes.
They state that spare parts, prototyping, complex parts and final products might be the
most attractive applications of 3DP. I4 states: “Trends such as 3DP offer great oppor‐
tunities for investing into companies’ future growth”, thus underpinning the significance
of such an implementation. On the other hand, I2 and I3 state that the significance of
3DP is essential mainly for larger organizations (average value is “slightly disagree”).
Nonetheless, all participants agree (average value is 2.25) that the overall benefits
outweigh the disadvantages of 3DP (Question 7). Accordingly, H1 can be confirmed.
Regarding H2, I1 indicates: “OI platforms generate more knowledge, they enable
more market participants to be included as co-creators and allow the faster detection of
changes in customer demands.” The average answer level (2,6) underpins this statement.
Additionally, H3 can be confirmed with an answer average score of 2,5. I3 specifies:
“System stability plays a crucial role when it comes to the compatibility between new
and existing systems […] Although it is important to consider security aspects, this might
prevent the open innovation process from fully unfolding.” Scores for the answers
concerning H4 bring up similar conditions (2,6). Given the fact that processes in organ‐
izations are becoming more and more digitalized and automated, future BPM need to
be increasingly coordinated with the digital world, state I1 and I5.
As an exception, H5 cannot be confirmed. I1 evaluated this aspect with “slightly
agree (1)” and added that the technical possibilities to collaborate digitally improve
continuously, making physical proximity and its effect on working results lose in signif‐
icance. However, I3, I6 and I7 evaluated this aspect with “strongly agree (3)”. I3 stated
that there are other qualities bound to physical proximity, such as sympathy and direct
eye or physical contact, which might influence the collaboration and consequently the
working results. Conclusively, a balance between automatized platform organizations
and personal interaction must be maintained.
I2 states that the idea of the BPM is plausible in theory but the culture of the company
has to be compatible in order to allow the concept to fully unfold its effect. I5 also
substantiated that the amalgamation of all mentioned aspects should be feasible, but
adds that the willingness of companies to adapt to such a concept largely depends on
the perceived additional value. Not only is the relevance of the generated knowledge
essential, but also whether this knowledge contributes to the company’s objectives.
Overall, the findings substantiate that the implementation of an OI platform and the
integration of 3DP will enable more flexible and customer-oriented production processes
with the ability to satisfy customer needs to a higher degree and enhance the aforemen‐
tioned amalgamation (see Table 2).
Amalgamation of 3D Printing Technology and the Digitalized Industry 157
6 Concluding Remark
This contribution considered how the open innovation approach might enhance the
amalgamation of the 3D printing technology with the digitalized industry. A BPM was
developed and modeled in BPMN 2.0. The result serves as an initial solution for
producers to cope with the increasing digitalization of the industry and changing market
conditions. A main finding is that future production processes as well as end-user
requirements are presumably becoming more complex, raising demand for more adjust‐
able production processes.
However, in order to allow the OI approach to fully uncover its potential, barriers
such as certain security aspects must be minimized. Further, it is important to have
adjustable business models, enabling the detection and incorporation of essential envi‐
ronmental dynamics. It was indicated that 3DP manufacturing companies may increase
their flexibility and enable a stronger customer-orientation by implementing an OI plat‐
form. It would be useful to execute a case study with a transforming company to itera‐
tively improve the BPM, including its business processes and practical relevance, and
allow insights regarding the quality of the generated knowledge. Moreover, the BPM
could be extended to a more interactive version, where multiple customers and partners
are enabled to cooperate within the innovation process, including several iteration steps.
Further research could also reveal more information regarding the negative effects
of collaborating through platforms rather than in physical proximity, specifically
158 D. Warnecke et al.
assessing the paradox mentioned by [29], i.e. exploring whether such strategies
contribute to a homogenization of companies, leading them to gradually lose unique
features and making it more difficult to achieve a competitive advantage.
References
1. D’Aveni, R.: The 3D printing revolution. Harvard Bus. Rev. 93(5), 40–48 (2015)
2. Khan, O., Mohr, S.: 3D printing and its disruptive impacts on supply chains of the future.
Logistics Transport Focus 18, 24–28 (2016)
3. Guzman, J.: Digitalization on the horizon. SMT Mag., 62–65 (2015)
4. Johnsson, H.: Production strategies for pre-engineering in house - building: exploring product
development platforms. Constr. Manag. Econ. 31(9), 941–958 (2013)
5. Pisano, P., Pironti, M., Rieple, A.: Identify innovative business models: can innovative
business models enable players to react to ongoing or unpredictable trends? Entrepreneurship
Res. J. 5(3), 181–199 (2015)
6. Baumeister, R., Leary, M.: Writing narrative literature reviews. Rev. General Psychol. 1(3),
311–320 (1997)
7. vom Brocke, J., Simons, A., Riemer, K., Niehaves, B., Plattfaut, R.: Standing on the shoulders
of giants: challenges and recommendations of literature search in information systems
research. Commun. Assoc. Inf. Syst. 37(1), 205–224 (2015)
8. Gläser, J., Laudel, G.: Experteninterviews und qualitative Inhaltsanalyse, 4th edn. VS Verlag,
Wiesbaden (2010)
9. Bertschek, I., Clement, R., Buhr, D., Hirsch-Kreinsen, H., Falck, O., Heimisch, A., et al.:
Industrie 4.0: Digitale Wirtschaft - Herausforderung und Chance für Unternehmen und
Arbeitswelt. ifo Schnelldienst, pp. 3–18 (2015)
10. Loupis, M.: Embedded systems development tools: a modus-oriented market overview. Bus.
Syst. Res. 5(1), 6–20 (2013)
11. Reischauer, G.: Industrie 4.0 organisieren. zfo Zeitschrift Führung und Organisation, pp. 271–
277 (2015)
12. Palmquist, D.: 5 trends driving the movement toward smart manufacturing. Supply Demand
Chain Executive 15, 20–21 (2014)
13. Berman, B.: 3-D printing: The new industrial revolution. Bus. Horiz. 55, 155–162 (2012)
14. Petrick, I., Simpson, T.: 3D Printing Disrupts Manufacturing. Research-Technology
Management (2013)
Amalgamation of 3D Printing Technology and the Digitalized Industry 159
1 Introduction
combine the traditional worlds of manual process modeling with data support,
thereby allowing a user to add domain knowledge during process discovery.
Conformance checking techniques use a process model and an event log to
calculate how well an event log fits the process model and vice versa. Tradi-
tionally, the conformance checking techniques are used to perform post-mortem
analysis of the execution of a process. That is, once a process model is made
available, conformance checking techniques determine the goodness-of-fit of the
event data and the process model. However, in a user-guided process discovery
setting, there is a need for active feedback regarding the goodness of a process
model after each user interaction (see Fig. 1). That is, after the user interactively
changes a process model by taking into account the domain knowledge, there is
a need to quantify how good or bad the change was according to the event data.
This task can indeed be performed by conformance analysis, and thereby confor-
mance analysis forms a natural fit with user-guided process discovery, and can
be used to provide feedback after each step in an interactive process discovery
setting.
In an interactive user-guided discovery
setting, it is required to have fast feedback
depending on the change made to the pro-
cess model. In order to address this problem,
we present a framework of calculating incre-
mental conformance depending on the change
made in the process model. Unlike other con-
formance analysis techniques, the framework
discussed in this paper exploits the underly-
ing structure used for interactive process dis-
covery in order to perform fast and approx-
imate conformance analysis. Moreover, we Fig. 1. Interactive process discov-
present an application of the framework that ery setting. During discovery, con-
formance analysis is required to be
shows that even though the conformance
computed fast enough for an unin-
results are approximated, they still contain terrupted experience for the user.
diagnostic information which could provide
useful feedback for the user for decision making in interactive process discovery.
The rest of the paper is structured as follows. In Sects. 2 and 3, we discuss the
related work from literature and the preliminaries resp. In Sect. 4 we discuss our
approach and in Sect. 5 we discuss the implementation and visualization details.
In Sect. 6 we evaluate our approach and in Sect. 7 we conclude and provide
indicators of future work.
2 Related Work
among the first one’s to devise conformance checking techniques in process min-
ing using the token based replay in Petri nets. In [7], the authors discuss con-
formance strategies for declarative models. The authors of [3,4] use alignment
based strategy for performing conformance checking. Some approaches look at
the conformance problem from various other dimensions such as natural lan-
guage processing, formal methods, real time setting etc. [5,6,17]. Most of these
approaches focus on the accuracy of the results, and hence do not emphasize
on the performance dimension and thereby are not advisable in an interactive
process discovery setting. In [18], the authors discuss strategies for incrementally
repairing a prior alignment based on a change in a process tree for improving
the performance of the ETM algorithm. However, as the authors noted, after a
certain number of repairs, the repaired alignment might be drastically different
from the so-called optimal alignment. Thereby, this could lead to highly inaccu-
rate results. Also, the class of models supported by this approach is limited to
block structured process models (process trees).
In order to improve the performance of conformance analysis, various divide
and conquer techniques were discussed in [1,13,16,19,20]. Typically, the cen-
tral idea behind these techniques is to decompose a process model into vari-
ous sub-models based on a specific decomposition strategy, and then to com-
pute alignments on the smaller model (and a corresponding smaller log) and
aggregate the information across all models. However, in certain scenarios, the
decomposed sub-models may still have large state spaces. Therefore, the confor-
mance calculation time in such scenarios may be similar, or even worse owing
to decomposition and aggregation time, compared to the complete model. In all
these divide-and-conquer techniques, there is more emphasis on the accuracy of
the results. However, in our case, we relax the accuracy of the results to ensure
short response times, similar to the technique discussed in [11]. In both [11]
and our approach, conformance is calculated and aggregated over all combina-
tions of sets of activities for a given cardinality. The main difference between
our approach and [11] is that we inherently exploit the incremental nature of
process modeling during interactive discovery. That is, in the case of [11], the
conformance for all the combinations of activities after each change in a process
model is recalculated. However, in our technique, we remember the prior con-
formance results, and typically recalculate only those conformance results which
may have changed depending on the change in the process model. Hence, the
proposed approach is much faster, robust and provides diagnostic information
useful during interactive process discovery.
3 Preliminaries
In this section, we discuss the relevant preliminaries. Events form the basic
building blocks of an event log. An event represents an occurrence of an activity.
Every event has a timestamp associated with it which indicates the time of
occurrence of that activity [2]. A trace is a sequence of events, and an event log
is a bag (or multiset) of traces.
166 P. M. Dixit et al.
Having discussed event logs, we now discuss process models. Process models
are graphical constructs used to represent the flow of activities. Typically, a
process model has a fixed start and a fixed end point, and a navigation from
start to end point of a process model results in a trace. Hence a process model
corresponds to a set of traces. Conformance analysis in process mining aims at
finding a fit between the traces from an event log and a process model.
We now define two concepts of projecting a log and a process model on a set
of activities. An event log can be projected on a set of activities by removing
all the events from the event log which are not a part of the projection set.
Consider an event log L = [a, b, c, d, a, e10 , a, b, c, d5 , b, d, e5 ]. Then the event
log projected on activities {a, c} is L ↓{a,c} = [a, c, a10 , a, c5 , 5 ].
A process model M can be projected on a set of activities A by making
all the activities from the process model which are not a part of projection
set as invisible (depicting no behavior) activities. These invisible activities can
be further removed (if possible) using language preserving reduction rules. The
reduction removes the unnecessary invisible activities from the net which do not
have any behavioral impact on the net. The projected and reduced model is
denoted by M ↓A . Due to space limitations, we refer the user to [8,14] for more
details about these reduction rules.
The first part of our approach is similar to the projected conformance checking
[11] approach. Here we exploit the fact that calculating conformance for a smaller
process model having a smaller state space and a smaller event log is often faster
than calculating conformance for a large complex process model. Let A be all
the activities present in a model M and let L be the corresponding event log.
For a user defined cardinality k, using A we calculate the set of all possible
activity combinations C, where ∀c∈C |c| = k. Let M ↓c and L ↓c denote the
process model M and event log L projected with the activities from activity
combination c (c ∈ C). Let Q(M ) define some function used to quantify the
quality of the model M w.r.t. the event log. Then the quality of the overall
model Q(M ) is the aggregation of the quality values of all the projected models:
Σ Q(M ↓c )
Q(M ) = c∈C .
|C|
Therefore, instead of calculating the conformance of one (complex) process
model with an event log, conformance of several projected (smaller) process
Incremental Conformance Analysis 167
models with projected event logs are calculated and then aggregated. Here we
exploit the fact that calculating conformance for one large process model is
often time consuming, and thereby distributing the conformance calculation over
several smaller projected process models improves efficiency. Similarly, based on
user’s preference, minimum (maximum) quality value can also be computed of
the overall model, using the minimum (maximum) value of each combination.
There is no need to calculate the conformance values for those activity com-
binations which exhibit the same projected behavior in Mi and Mi+1 (CS ). How-
ever, the activity combinations CD whose projected behavior is not the same,
there is a need to calculate the conformance values. Therefore, in an interactive
setting, we improve the projected conformance calculation times, by calculating
conformance for only the activity combinations from the set CD . It should be
noted that, if a new activity is added to the model interactively, then there will
be new activity combinations possible, which would all be a part of CD . The
amount of time needed for calculation of conformance for activity combinations
CS is saved by using an incremental way of calculating the conformance values.
We show the instantiation of the framework using a synthesis rules based inter-
active process modeling approach. In order to do so, we first introduce free choice
workflow (FC WF) nets which are a well-known class of Petri nets to represent
processes. A workflow (WF) net is a bi partite graph containing places (circles)
and transitions (rectangles), such that a transition can only be connected to a
place and vice versa. A WF net is navigated from left-to-right, such that the
first transition is always a ‘dummy’ (silent) transition called , and the last
transition is always a ‘dummy’ (silent) transition called ⊥. Transitions in a WF
net can be labeled with an activity, or they can be silent, i.e. not representing
any behavior from the event log. Moreover, a WF net is a FC WF net if for
any two places, the set of output transitions of both the places are exactly the
same, or completely disjoint (that is, there is no overlap at all). Figure 3a shows
an example of a FC WF net. The only possible complete trace of this FC WF
net is a, b, c. Note that the transitions and ⊥ are silent, and hence do not
form a part of this trace sequence. A short circuited version of a FC WF net can
be obtained by merging the places i and o, and renaming it as io. It should be
noted that we can easily obtain a FC WF net from a short circuited FC WF net
by splitting the place io into places i and o.
Three rules are used to interact and synthesize a short circuited FC WF net,
such that each rule leads to addition of a new place and/or a new transition in
the model. Using the synthesis rule based interactive modeling approach, at each
iteration we identify the set of combinations of activities (CS ) whose projected
behavior does not change. For all the other activity combinations (CD ), the
conformance is recalculated. We describe the incremental way of calculating the
change in the models depending on each type of rule. It should be noted that,
with the usage of synthesis rules, a process model can only grow.
Incremental Conformance Analysis 169
Figure 3d shows the addition of a new transition using the so-called linearly
dependent transition rule [8]. Addition of a new transition usually results in
the introduction of a choice or loop in the model. There exists a set of bags of
transitions Tset , which have the same effect on the short circuited FC WF net, as
the newly added transition. In Fig. 3d, this set corresponding to the newly added
transition td is {[tb ]}. We use this information to calculate the set of activity
combinations CS whose projected behavior does not change. No elements are
added to CS , if any bag from Tset contains or ⊥. This is for reasons similar
the one described in Subsect. 4.4. The second scenario is when none of the bags
from Tset contain or ⊥, i.e. ∀E∈Tset , ⊥ ∈ / E. Let TL be the set of all the labels
Fig. 5. Projected behavior for k = 2 corresponding to the new net after adding td
(Fig. 3d).
Incremental Conformance Analysis 171
Adding a new transition and a new place using the so-called abstraction rule [8]
results in a new sequence in the model. For example, Fig. 3e is derived by adding
a new transition (labeled e) and a new place (p6 ) to the model from Fig. 3d. If the
newly added transition is labeled with an activity which is not already present
in the model, then for any chosen cardinality, all the activity combinations from
Fig. 3d are behaviorally equivalent in Fig. 3e. That is, if the transition for newly
added activity e is made silent, then the net would be behaviorally equivalent
to the previous net. Hence all the activity combinations from Fig. 3d are added
to CS .
1
http://www.promtools.org/doku.php?id=nightly.
172 P. M. Dixit et al.
Fig. 6. (A) shows the conformance of the modeled activities according to the event log:
the density of blue color indicates the frequency, the red (darker) and green (lighter)
bars above the transitions indicate the distribution of fitting and non-fitting events.
The tabular view (B) shows the individual scores of the activity sets (k = 2 in this
case). The user can dig deeper to explore the relationship between any activity set by
choosing the desired visualization. (Color figure online)
6 Evaluation
We evaluate the approach presented in this paper by comparing it with state-
of-the-art conformance techniques. The goal of this evaluation is to show the
effectiveness of our approach in an interactive setting, measured in terms of
performance times and correctness of the result. We use two real-life event logs:
(i) the Sepsis event log2 containing the workflow data for roughly 1000 patients
suffering from Sepsis in a hospital. A normative process model is already available
for this event log at [12]. (ii) the BPIC 2011 event log3 containing hospital data
for cancer patients from a Dutch hospital - filtered to retain only top 40% of
the most frequent activities, resulting in a total of around 10000 events. We
use the inductive miner [10] to discover a process model from the filtered event
log. In order to replicate an interactive scenario, starting with a process model
without any activities, we interactively re-construct each process model using the
synthesis rules based approach. After each “interaction” (step), conformance is
recalculated using following techniques: the decomposed replay technique [20],
the recomposed replay technique [19], projected conformance checking [11] (with
k = 2), the technique presented in this paper (with k = 2) and the regular
alignments technique [4]. The fitness (i.e., the faithfulness of the model to the
log), precision (i.e., the extent to which behavior not seen in the log is allowed by
the model) and time taken for recalculation w.r.t. each technique are recorded.
2
https://data.4tu.nl/repository/uuid:915d2bfb-7e84-49ad-a286-dc35f063a460.
3
http://www.win.tue.nl/bpi/2011/challenge.
Incremental Conformance Analysis 173
The fitness and precision values are scored from 0 (worst) to 1 (best) in all the
techniques.
1,000 1,000
800 800
600 600
Time (ms)
Time (ms)
400 400
200 200
0 0
0 5 10 15 20 25 30 0 5 10 15 20 25 30 35 40
Step number Step number
This paper PCC Regular Decomposed Recomposed This paper PCC Regular Decomposed Recomposed
(a) Performance times for the filtered (b) Performance times for the Sepsis event
BPIC 2011 event log. log.
Fig. 7. The time(s) taken after each interaction (step) for conformance analysis.
changes in the process model wherein the projected behavior of all the activity
combinations was the same in the iterated model.
Fig. 8. The view of projected alignment on activities ‘CRP’ and ‘ER Registration’
from the Sepsis event log.
References
1. van der Aalst, W.M.P.: Decomposing petri nets for process mining: a generic app-
roach. Distrib. Parallel Databases 31(4), 471–507 (2013)
2. van der Aalst, W.M.P.: Process Mining - Data Science in Action, 2nd edn. Springer,
Heidelberg (2016). https://doi.org/10.1007/978-3-662-49851-4
3. Adriansyah, A., van Dongen, B.F., van der Aalst, W.M.P.: Conformance check-
ing using cost-based fitness analysis. In: 2011 15th IEEE International Enterprise
Distributed Object Computing Conference (EDOC), pp. 55–64. IEEE (2011)
4. Adriansyah, A., van Dongen, B.F., van der Aalst, W.M.P.: Towards robust con-
formance checking. In: zur Muehlen, M., Su, J. (eds.) BPM 2010. LNBIP, vol.
66, pp. 122–133. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-
20511-8 11
5. vanden Broucke, S.K.L.M., Munoz-Gama, J., Carmona, J., Baesens, B.,
Vanthienen, J.: Event-based real-time decomposed conformance analysis. In:
Meersman, R., Panetto, H., Dillon, T., Missikoff, M., Liu, L., Pastor, O.,
Cuzzocrea, A., Sellis, T. (eds.) OTM 2014. LNCS, vol. 8841, pp. 345–363. Springer,
Heidelberg (2014). https://doi.org/10.1007/978-3-662-45563-0 20
Incremental Conformance Analysis 175
1 Introduction
2 Theoretical Background
is
Business Process prerequisite
for
compliance process, IT components are also necessary for the proper execution of auto‐
matic compliance processes [23]. Third, both business processes and also IT components
are affected by numerous compliance requirements. For example, COSO [7] and the
Institute of Public Auditors in Germany [6] provide a lot of requirements for the proper
operation of IT components.
The insert change pattern inserts a new element into the business process at a defined
place. The delete pattern removes an existing element. The replace pattern replaces an
existing element with a new one. The update pattern modifies an attribute of an existing
element [30].
These change patterns can also be applied to our perspectives on enterprise archi‐
tecture. Therefore, we assume that IT components, compliance requirements, and
compliance processes are additional views of a business process that are similar to for
example, a data view. Due to space limitations, we will only apply the change patterns
“delete element” and “replace element”. Next, we discuss a running scenario to motivate
our research problem and to demonstrate the solution through our method.
3 Running Scenario
The left side of Fig. 3 shows a simplified purchase to pay process based on [31, 32],
including perspectives on compliance requirements, compliance processes, and IT
components. Some of the business activities are supported by IT components that are
modeled as triangles.
180 T. Seyffarth et al.
§ 238
German Replace ERP FI Delete ERP FI
Com.
Code Approve invoice
Logical access (CR)
Physical Internal Logical payment (CP)
access policy access § 238 German
Internal policy (CR)
Commercial Code (CR)
Inform purchase § 238 German
Hardware (IT)
requester about Commercial Code (CR)
received goods
Physical access (CR) Logical access (CR)
Create purchase
request § 238 German
Send Receive Approve Commercial Code (CR)
purchase goods and Pay invoice invoice payment
request invoice
Pay invoice (activity)
In the case of deleting “ERP FI” the compliance process “approve invoice payment”
cannot be executed. Thus, the compliance requirements “internal policy” and “§ 238 of
the German Commercial Code” are violated. Furthermore, the compliance requirement
“logical access” is obsolete. In the next section, we propose a method and its prototypical
implementation for analyzing the interactions between BPC and business process
change considering supporting IT components and compliance processes.
Figure 4 shows the three steps of our method for analyzing the interactions of BPC and
business process change. First, business process, compliance process, IT architecture,
and compliance requirements must be imported into the software prototype. Second, the
imported models must be transformed and integrated into a single graph. Third, the
interactions of BPC and business process change must be analyzed. In the following
sections, we explain each step in detail with respect to the running scenario introduced
in the previous section.
Fig. 4. Steps to analyze the interactions of BPC and business process change
Within our software prototype, we used the Java library “JGraphT” [16] to model
and visualize graphs. In the next step, the integrated graph is used as a basis for analyzing
the interactions between BPC and business process change.
4.3 Analyzing the Interactions Between BPC and Business Process Change
By taking the business process change patterns “replace element” and “delete element”
into consideration, their semantic interpretation to the business process model consid‐
ering IT components, compliance processes, and compliance requirements must be
defined. When replacing an element, the replaced elements are sometimes affected by
compliance requirements (see replace “ERP FI” within the running scenario). Thus, the
new element must take these respective compliance requirements into consideration to
ensure BPC. The removal of an element can also have an impact on BPC (see remove
“ERP FI” within the running scenario). First, it can have an impact on compliance
requirements and/or compliance processes that satisfy the impacted compliance require‐
ments. Second, due to a removal of an element, the respective compliance requirements
can become obsolete. In the following, we describe our algorithm to analyze the inter‐
actions between BPC and business process change. Due to space limitations, we will
refer to an IT component as the replaced or deleted element.
Replacing an Element
Figure 5 shows the algorithm to analyze the relations of compliance requirements and
compliance processes to the IT component that shall be replaced. To ensure BPC, the
replaced element must fulfill the related compliance requirements and be able to execute
a related compliance process. Thereby, we distinguish between direct and transitive
relations.
Second, all compliance requirements of the replaced element prerequisites must be taken
into consideration, as well. Within the running scenario, the compliance requirement
“Physical access” is direct related to the replaced element “ERP FI” because it is a
prerequisite for the IT component “hardware”, which is, in turn a prerequisite for “ERP
FI”.
A transitive relation between an updated IT component and a compliance require‐
ment or a compliance process also occurs in two different cases. First, the changed IT
component is a prerequisite of a business activity that is affected by at least one compli‐
ance requirement. Second, the updated IT component is a prerequisite of a compliance
process that helps to satisfy at least one compliance requirement. Within the running
scenario, this is the case at the business activity “pay invoice” and its prerequisite
compliance requirements “internal policy” and “§ 238 of the German Commercial
Code”.
Deleting an Element
Figure 6 shows the algorithm to analyze the impacts on BPC by removing an IT compo‐
nent. Here, we distinguish between two different impacts. First, the removal an IT
component can lead to a compliance process not being executable. This concerns the
compliance process “approve invoice payment”, since the IT component “ERP FI” is a
prerequisite for its execution. Consequently, the compliance requirements “internal
policy” and “§ 238 of the German Commercial Code” are violated.
5 Related Work
References
1. Schumm, D., Turetken, O., Kokash, N., Elgammal, A., Leymann, F., van den Heuvel, W.-J.:
Business process compliance through reusable units of compliant processes. In: Daniel, F.,
Facca, F.M. (eds.) ICWE 2010. LNCS, vol. 6385, pp. 325–337. Springer, Heidelberg (2010).
https://doi.org/10.1007/978-3-642-16985-4_29
2. Turetken, O., Elgammal, A., van den Heuvel, W.-J., Papazoglou, M.: Enforcing compliance
on business processes through the use of patterns. In: 19th ECIS 2011 (2011)
3. Schäfer, T., Fettke, P., Loos, P.: Towards an integration of GRC and BPM – requirements
changes for compliance management caused by externally induced complexity drivers. In:
Daniel, F., Barkaoui, K., Dustdar, S. (eds.) BPM 2011. LNBIP, vol. 100, pp. 344–355.
Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-28115-0_33
Business Process Compliance and Business Process Change 187
4. Sadiq, S., Governatori, G., Namiri, K.: Modeling control objectives for business process
compliance. In: Alonso, G., Dadam, P., Rosemann, M. (eds.) BPM 2007. LNCS, vol. 4714,
pp. 149–164. Springer, Heidelberg (2007). https://doi.org/10.1007/978-3-540-75183-0_12
5. Knackstedt, R., Eggert, M., Heddier, M., Chasin, F., Becker, J.: The relationship of is and
law - the perspective of and implications for IS research. In: ECIS 2013 Completed Research
(2013)
6. The Audit of Financial Statements in an Information Technology Environment. IDW AuS
330 (2002)
7. Committee of Sponsoring Organizations of the Treadway Commission (COSO): Internal
Control - Integrated Framework. Framework and Appendices (2012)
8. Rudzajs, P., Buksa, I.: Business process and regulations: approach to linkage and change
management. In: Grabis, J., Kirikova, M. (eds.) BIR 2011. LNBIP, vol. 90, pp. 96–109.
Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-24511-4_8
9. Fdhila, W., Indiono, C., Rinderle-Ma, S., Reichert, M.: Dealing with change in process
choreographies: design and implementation of propagation algorithms. Inf. Syst. 49, 1–24
(2015)
10. Rinderle, S., Reichert, M., Dadam, P.: Correctness criteria for dynamic changes in workflow
systems—a survey. Data Knowl. Eng. 50, 9–34 (2004)
11. Awad, A.: BPMN-Q: a language to query business processes. In: Proceedings of EMISA
2007, pp. 115–128 (2007)
12. Koetter, F., Kochanowski, M., Weisbecker, A., Fehling, C., Leymann, F.: Integrating
compliance requirements across business and IT. In: 2014 IEEE 18th International Enterprise
Distributed Object Computing Conference (2014)
13. Hevner, A.R., Gregor, S.: Positioning and presenting design science research for maximum
impact. MIS Q. 37 (2013)
14. Yu, P.S., Han, J., Faloutsos, C. (eds.): Link Mining: Models, Algorithms, and Applications.
Springer Science + Business Media LLC, New York (2010). https://doi.org/
10.1007/978-1-4419-6515-8
15. Camunda: Camunda BPMN model API. https://github.com/camunda/camunda-bpmn-model
16. Naveh, B.: JGraphT. http://jgrapht.org/
17. OMG (Hg): Business Process Model and Notation (BPMN). http://www.omg.org/spec/
BPMN/2.0/PDF/
18. Jonkers, H., Lankhorst, M., van Buuren, R., Hoppenbrouwers, S., Bonsangue, M., van der
Torre, L.: Concepts for modeling enterprise architectures. Int. J. Coop. Inf. Syst. 13, 257–287
(2004)
19. Juris (ed.): Gesetze im Internet. http://www.gesetze-im-internet.de
20. Seyffarth, T., Kühnel, S., Sackmann, S.: ConFlex: an ontology-based approach for the flexible
integration of controls into business processes. Multikonferenz Wirtschaftsinformatik
(MKWI) 2016, 1341–1352 (2016)
21. Kharbili, M., Medeiros, A.K.A.d., Stein, S., van der Aalst, W.M.P.: Business process
compliance checking: current state and future challenges. MobIS 141, 107–113 (2008)
22. Sackmann, S., Kittel, K.: Flexible workflows and compliance: a solvable contradiction?! In:
Vom Brocke, J., Schmiedel, T. (eds.) BPM - Driving Innovation in a Digital World, pp. 247–
258. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-14430-6_16
23. Seyffarth, T., Kühnel, S., Sackmann, S.: A taxonomy of compliance processes for business
process compliance. In: Carmona, J., Engels, G., Kumar, A. (eds.) BPM 2017. LNBIP, vol.
297, pp. 71–87. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-65015-9_5
24. IEEE: IEEE Recommended Practice for Architectural Description of Software Intensive
Systems, (IEEE Std 1 1471–2000). IEEE Computer Society, New York (2000)
188 T. Seyffarth et al.
25. TOGAF (ed.): Content Metamodel. Content Metamodel Vision and Concepts. http://
pubs.opengroup.org/architecture/togaf9-doc/arch/
26. Winter, R., Fischer, R.: Essential layers, artifacts, and dependencies of enterprise architecture.
In: 2006 10th IEEE International Enterprise Distributed Object Computing Conference
Workshops (EDOCW 2006), p. 30 (2006)
27. The Open Group (ed.): TOGAF 9.1. Content Meta Model. http://pubs.opengroup.org/
architecture/togaf9-doc/arch/chap34.html
28. Weber, B., Reichert, M., Rinderle-Ma, S.: Change patterns and change support features –
enhancing flexibility in process-aware information systems. Data Knowl. Eng. 66, 438–466
(2008)
29. Rinderle-Ma, S., Reichert, M., Weber, B.: On the formal semantics of change patterns in
process-aware information systems. In: Li, Q., Spaccapietra, S., Yu, E., Olivé, A. (eds.) ER
2008. LNCS, vol. 5231, pp. 279–293. Springer, Heidelberg (2008). https://doi.org/
10.1007/978-3-540-87877-3_21
30. Fdhila, W., Rinderle-Ma, S., Reichert, M.: Change propagation in collaborative processes
scenarios. In: 8th International Conference on Collaborative Computing: Networking,
Applications and Worksharing (CollaborateCom) (2012)
31. Namiri, K.: Model-Driven Management of Internal Controls for Business Process
Compliance. Karlsruhe (2008)
32. Frank, U., Heise, D., Ulrich, Kattenstroth, H., Ferguson, D.F., Hadar, E., Waschke, M.G.:
ITML: a domain-specific modeling language for supporting business driven IT management.
In: Proceedings of the 9th OOPSLA Workshop on Domain-Specific Modeling (2009)
33. Principles of Proper Accounting When Using Information Technology. IDW AcP FAIT 1
(2002)
34. Kirikova, M., Penicina, L., Gaidukovs, A.: Ontology based linkage between enterprise
architecture, processes, and time. Commun. Comput. Inf. Sci. 539, 382–391 (2015)
35. Vom Brocke, J., Simons, A., Niehaves, B., Riemer, K., Plattfaut, R., Cleven, A.:
Reconstructing the giant: on the importance of rigour in documenting the literature search
process. In: 17th European Conference on Information Systems, pp. 2206–2217 (2009)
36. Webster, J., Watson, R.T.: Analyzing the past to prepare for the future: writing a literature
review. MIS Q. 26, xiii–xxiii (2002)
37. Delfmann, P., Steinhorst, M., Dietrich, H.-A., Becker, J.: The generic model query language
GMQL – conceptual specification, implementation, and runtime evaluation. Inf. Syst. 47,
129–177 (2015)
38. Gacitua-Decar, V., Pahl, C.: Automatic business process pattern matching for enterprise
services design. In: 2009 World Conference on Services - II (2009)
39. Fellmann, M., Thomas, O., Busch, B.: A query-driven approach for checking the semantic
correctness of ontology-based process representations. In: Abramowicz, W. (ed.) BIS 2011.
LNBIP, vol. 87, pp. 62–73. Springer, Heidelberg (2011). https://doi.org/
10.1007/978-3-642-21863-7_6
40. Ghanavati, S., Amyot, D., Peyton, L.: Compliance analysis based on a goal-oriented
requirement language evaluation methodology. In: 2009 17th IEEE International
Requirements Engineering Conference (2009)
41. Fdhila, W., Rinderle-Ma, S., Knuplesch, D., Reichert, M.: Change and compliance in
collaborative processes. In: 2015 IEEE International Conference on Services Computing
(2015)
Business Process Compliance and Business Process Change 189
42. Knuplesch, D., Fdhila, W., Reichert, M., Rinderle-Ma, S.: Detecting the effects of changes
on the compliance of cross-organizational business processes. In: Johannesson, P., Lee, M.L.,
Liddle, Stephen W., Opdahl, Andreas L., López, Ó.P. (eds.) ER 2015. LNCS, vol. 9381, pp.
94–107. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-25264-3_7
43. Knackstedt, R., Braeuer, S., Heddier, M., Becker, J.: Integrating regulatory requirements into
information systems design and implementation. In: ICIS 2014 Proceedings (2014)
44. Elgammal, A., Turetken, O., van den Heuvel, W.-J., Papazoglou, M.: Root-cause analysis of
design-time compliance violations on the basis of property patterns. In: Maglio, Paul P.,
Weske, M., Yang, J., Fantinato, M. (eds.) ICSOC 2010. LNCS, vol. 6470, pp. 17–31. Springer,
Heidelberg (2010). https://doi.org/10.1007/978-3-642-17358-5_2
45. Halle, S.: Causality in message-based contract violations. a temporal logic “Whodunit”. In:
2011 IEEE 15th International Enterprise Distributed Object Computing Conference (2011)
46. Koetter, F., et al.: An universal approach for compliance management using compliance
descriptors. In: Helfert, M., Ferguson, D., Méndez Muñoz, V., Cardoso, J. (eds.) CLOSER
2016. CCIS, vol. 740, pp. 209–231. Springer, Cham (2017). https://doi.org/
10.1007/978-3-319-62594-2_11
Mining Hybrid Business Process Models:
A Quest for Better Precision
1 Introduction
In recent years, different comparative investigations have been conducted to bet-
ter understand the distinctive characteristics of imperative and declarative pro-
cess modeling languages and to support the choice of the most suitable paradigm
to be used in different scenarios [25,26]. While advantages and limitations of the
two paradigms are still a matter of investigation, both in academic research and
in industry [12,21], a trend has emerged to consider hybrid approaches combining
a mixture of imperative and declarative specifications. The motivations behind
this trend rely on the surmise that many real-life processes are characterized by
This work is supported in part by the Hybrid Business Process Management Tech-
nologies project funded by the Danish Council for Independent Research.
c Springer International Publishing AG, part of Springer Nature 2018
W. Abramowicz and A. Paschke (Eds.): BIS 2018, LNBIP 320, pp. 190–205, 2018.
https://doi.org/10.1007/978-3-319-93931-5_14
Mining Hybrid Business Process Models 191
a mixture of (i) less structured processes with a high level of variability, which
can usually be described in a compact way using declarative languages such as
Declare [24], SCIFF [23], or DCR Graphs [10], and (ii) more stable processes
with well-structured control flows, which are more appropriate for traditional
imperative languages such as Petri nets and BPMN.
As a result, there have been several recent efforts to fully develop and
formalize such hybrid modeling notations and methodologies. These hybrid
approaches can be categorized based on how tightly the imperative and declara-
tive paradigms are integrated, with each approach falling in one of three mayor
categories: (i) a complementary approach, where imperative and declarative
notations are used in separate models, without any semantic interaction, but
with the aim of representing different aspects of the same process (see, e.g., [11]).
(ii) A hierarchical approach, where both imperative and declarative notations
are used in the same overall model, but the model is separated into several sub-
processes and each sub-process uses a single uniform notation. To our knowledge
most existing proposals for hybrid approaches fall within this category (see,
e.g., [1,27,30]). (iii) A fully mixed approach, where imperative and declarative
constructs fully overlap in the same model. While this approach can be con-
sidered the most flexible, it is also the most complex since the modeler needs
to consider and understand in minute detail how the elements of the different
notations interact. An example of such an approach is the one proposed in [36].
One field where the hybrid approach may lead to significant advances is that
of automatic process discovery [5], which aims at generating useful process mod-
els from real-life event logs. Many imperative and declarative process discovery
algorithms (usually referred to as miners) exist, and it remains an open chal-
lenge to determine a suitable discovery approach given the characteristics of the
event logs [7]. When faced with very variable logs, imperative miners have a
tendency to create “spaghetti models”, or “flower models”. The former are gen-
erally unreadable, the latter allow any behavior and therefore contain very little
useful information about the process. Declarative miners, on the other hand, are
not well-suited to dealing with very structured logs: when faced with such a log
they will often generate an abundance of constraints, which makes the model
unreadable for human users. Hybrid process discovery approaches aim at provid-
ing a middle way between these two extremes, generating an imperative model
for the structured parts of the log and declarative constraints for the unstruc-
tured ones. Two primary approaches to hybrid discovery have been considered:
(i) a model-driven approach, where first an imperative or a declarative model is
created, from which “bad” parts are identified (e.g., flower sub-models in imper-
ative models, or over-constrained activities in declarative models) and secondly
these “bad” parts are re-mined with a more suitable mining algorithm. (ii) A
log-driven approach where an up-front analysis of the log is used to identify
structured and unstructured parts, which are then separated into sub-logs. Each
sub-log is mined with the most suitable discovery algorithm.
We can categorize the full spectrum of hybrid miners by orthogonally con-
sidering their overall mining technique and the type of models that they output.
192 D. M. M. Schunselaar et al.
The current paper falls within the Hierarchical/Model-driven section of this cate-
gorization. In particular, we propose a hybrid miner that builds on the Inductive
Miner [16] by Leemans et al., which, for a given input log, generates an impera-
tive process model in the form of a process tree. The Inductive miner guarantees
a fitness of 1. This guarantee sometimes comes at the cost of precision, e.g., when
no structure can be found in the log, the Inductive miner defaults to a flower
loop, which is able to capture all behavior and much more. Being hierarchical in
nature, process trees can be easily adopted into a hierarchical hybrid approach.
To do so, we take the process tree returned from the Inductive miner and iden-
tify nodes from the process tree to be replaced with Declare models. We have
used Declare since there are several well accepted mining algorithms for this
notation, whereas support to process discovery for other declarative notations is
either non-existent or in a very early stage.
In this paper, we primarily focus on the precision of the model. There are
two main reasons for this: (1) precision, contrary to other quality dimensions like
simplicity, has widely accepted metrics for measuring it, and (2) flower loops tend
to be very imprecise since they allow for any possible behavior, thus giving an
opportunity to have a more accurate understanding of the process captured in
the event log by replacing them with declarative models.
The approach has been implemented as a plug-in of the process mining tool
ProM and tested on both synthetic and real-life logs. The evaluation shows posi-
tive results for both synthetic and real-life logs, with up to 52.18% improvement
in the precision of the mined models for synthetic logs and up to 3471% increase
in precision for the real-life logs.
The rest of the paper is structured as follows: in Sect. 2, we discuss the
related work. In Sect. 3, we introduce the hybrid models we use. In Sect. 4, we
describe the proposed approach. In Sect. 5, we report on our evaluation. Section 6
concludes the paper and spells out directions for future work.
2 Related Work
In the literature, there are different families of approaches that have been
proposed for the discovery of imperative process models. The first family of
approaches extracts some footprint from the event log and uses this footprint to
directly construct a process model (see, e.g., [5,34]). The second family of process
discovery approaches first constructs an intermediate model and then converts
it into a more refined model that can express concurrency and other advanced
control-flow patterns (e.g., [4]). The third family of approaches tries to break
the discovery problem into smaller problems. For example, the Inductive miner
[16] aims at splitting the event log recursively into sub-logs, which are eventually
mined. Techniques originating from the field of computational intelligence form
the basis for the fourth family of process discovery approaches. An example of
this type of approaches is the genetic mining approach described in [22].
In the area of declarative process discovery, Maggi et al. [18,19] first proposed
an unsupervised algorithm for mining Declare models. They base the discovery
Mining Hybrid Business Process Models 193
of constraints on the replay of the log on specific automata, each accepting only
those traces that are compliant with one constraint. In [8,14,15], the authors
use inductive logic programming techniques for the supervised learning of con-
straints expressed using SCIFF [6]. The approach for the discovery of Declare
constraints presented in [13] is divided in two steps. The first step computes
statistic data describing the occurrences of activities and their interplay in the
log. The second step checks the validity of Declare constraints by querying the
statistic data structure built in the previous step. In [28], the authors present
a mining approach that works with RelationalXES, a relational database archi-
tecture for storing event log data. The relational event data is queried with
conventional SQL.
Recent research has put into evidence synergies between imperative and
declarative approaches [25,26]. Accordingly, hybrid process modeling notations
have been proposed. In particular, in [9], the authors provide a conservative
extension of BPMN for declarative process modeling, namely BPMN-D, and
show that Declare models can be transformed into readable BPMN-D models.
In [36], the authors propose to extend Colored Petri nets with the possibility
of linking transitions to Declare constraints directly. The notion of transition
enablement is extended to handle declarative links between transitions. A recent
implementation of this technique has been made available in CPN Tools 4.0 [35].
In [31], the authors extend the work in [36] by defining a semantics based on
mapping Declare constraints to R/I-nets and by proposing modeling guidelines
for the mixed paradigm. In [32], a discovery approach based on this paradigm
is proposed. Differently from these approaches devoted to obtain a fully mixed
language, in [30], a hybrid process model is hierarchical, where each of its sub-
processes may be specified in either a procedural or declarative fashion. In [20],
an approach for the discovery of hybrid process models based on this semantics
is presented. In contrast to the approach of the current paper, which is model-
driven, the approach in [20] is log-driven. In addition, our approach guarantees
that the resulting hybrid model has 100% fitness.
3 Hybrid Models
a b c d e
X1
init
2..2
s X2 0..1
Given this semantics, the language of the process tree in Fig. 1 is: {a, b,
X1, X2, o, c, d, e, a, b, X2, X1, o, c, d, e, a, b, X2, o, X1, c, d, e, a, b, X1, X2, c,
d, e, a, b, X2, X1, c, d, e}.
3.2 Declare
Declare [24] is a declarative process modeling language that encodes a set of
declarative templates into a formal logic. Encodings into several different logics
have been proposed, such as a linear temporal logic (LT L) and regular expres-
sions. When using Declare in the context of process mining, it is most sensible to
use one of their finite variants, as event logs contain only finite traces. Therefore,
we employ the encoding into the finite variant of LT L (LT Lf ) for the current
paper. Table 1 shows an overview of the core relations used in Declare models.
For the existence and absence templates, the language supports all countable
variations, e.g., 0..2, 0..3, 4..∗, etc.
The language of a Declare model M = (A, C) is the set of all traces that
satisfy every constraint in C. Formally:
a b c d e
s
0..1 0..1 0..1
o init q r t
2..2
s
p 0..1
For the first step of our approach, we use the Inductive miner [16] to generate
a process tree from the input log. The Inductive miner builds a directly-follows-
graph for the unique activities of the log and uses it to find structures and
construct a process tree. If the Inductive miner cannot find any structure, it will
return a flower model where any behavior is allowed. This makes the Inductive
miner ideal for our approach: not only is it one of the most popular mining
algorithms, but also it avoids finding structures when none exists and instead
simply returns a pocket-of-flexibility, which can then be mined declaratively. We
use the standard Inductive miner with a noise threshold of 0. In this way, the
generated model is guaranteed to have 100% fitness and we can make a fair
comparison between a pure imperative and a hybrid approach, where we only
consider improvements on the precision of the model. For our example log, the
Inductive miner creates the process tree shown in Fig. 3a.
Once we have mined the process tree, we determine which parts should be
replaced with a declarative sub-process. For every permutation of nodes, we
198 D. M. M. Schunselaar et al.
replace those nodes with Declare sub-processes, determine the precision of the
model, and return the most precise permutation. If a node is selected to be
replaced with a declarative sub-process and also a higher node in the tree is
selected, then we only substitute the higher node in the tree, i.e., the node and
all its descendants are replaced.
Note that, when we apply our approach to the example log and process tree
from Fig. 3a by replacing the loop sub-processes with declarative sub-processes,
this results in an increase of precision. This is because the loops allow for any
number of occurrences of s and p, q, r, t, while, in reality, the allowed number of
occurrences is bound. This bound can be found by the Declare miner.
5 Evaluation
Our approach has been implemented in the Hybrid package in ProM. The
main plug-in for mining is called “Mine Hybrid Model”. Figure 4 shows the
visualization of a hybrid model in the plug-in.
We evaluated our approach on both synthetic and real-life logs. We use the
synthetic examples to show that the algorithm is able to detect and construct
hybrid models when these are clearly present in the log. We use the real-life logs
to demonstrate how well the approach works on actual logs coming from indus-
trial settings, and to discover if any of these have a clear hybrid structure. We
have evaluated the output of the Hybrid miner (HM) in terms of precision. This
in order to show that a hybrid model indeed can help in making the unstructured
parts of the process tree resulting from the Inductive miner (IM) more precise.
To show this increase in precision, we introduce the relative increase in precision
(RIiP) as the percentage of increase in precision.
Mining Hybrid Business Process Models 199
5.1 Precision
We use a metric based on escaping edges [3] for the computation of precision.
Our main consideration for choosing this metric is that the precision is defined
based on a transition system, a formalism that both a process tree and a Declare
model can be transformed into. For every visited state in the transition system,
the outgoing transitions used by the log are compared to the total number of
possible outgoing transitions. Transitions that are never visited by the log, the
escaping edges, are counted against the precision of the model. The metric takes
into account the frequencies of states, i.e., a frequently visited state has a higher
impact on the precision than an infrequently visited state, and states that are
never visited by the log are not taken into consideration.
For this log, the Hybrid miner provides an improvement from 0.543 to 0.827
with respect to the Inductive miner. Compared to L1 , the hybrid model accounts
for the unstructured parts of the log being more pronounced by adding 2 addi-
tional constraints to the declarative sub-process. Note that, while there is an
absolute improvement that is slightly better (+0.284) than the one obtained for
L1 , the relative improvement RIiP is much higher (52.18%).
L3 : Abstract structure. For the third experiment, we used log L3 .
The log exhibits an abstract structure, i.e., while all the activities of the log
occur in an unstructured manner, they can be separated in two sets {a, b, c, d}
and {e, f, g, h}, where events of the second set always occur after the events of
the first set. As expected the Hybrid miner finds a hybrid model consisting of a
sequence flow and two declarative sub-processes, one containing 14 constraints,
the other 15. The hybrid model results in a precision improvement from 0.459
to 0.616.
L4 : Unstructured sequences. For the fourth experiment, we used log L4 .
This log contains four strict sequences, a, b, c, d, e, f, g and h, i, which
themselves occur in an unstructured manner. This example can be seen as being
in the inverse class as L3 , i.e., while all activities occur in a structured manner,
there is a very clear unstructured abstract behavior in the log. Since the best
way of representing this process would be by using a declarative model with
imperative sub-processes, mining the log with the Hybrid miner (which is devel-
oped to build imperative models with declarative sub-processes) results in a fully
declarative model. The declarative model provides a precision improvement from
0.743 to 0.897 with respect to the Inductive miner.
Mining Hybrid Business Process Models 201
5.4 Discussion
Table 2 shows the detailed results of all the experiments we ran. Overall the
Hybrid miner performed quite well on both synthetic and real-life examples.
However, in the case of the real-life logs this was mostly due to the fact that
the Hybrid miner was able to determine that a fully declarative model would
have the highest precision and provided such a model. Granted, this increase
in precision came at the cost of simplicity since more than 75 000 constraints
cannot be comprehended by an end-user. This is also one of the limitations of
1
https://data.4tu.nl/repository/.
202 D. M. M. Schunselaar et al.
our current approach: getting the right settings on the used miners to give a
precise and comprehensible result.
In future work, we would like to investigate why the Hybrid miner did not
generate any particularly interesting hybrid models from real-life logs. One par-
ticularly promising hypothesis is that these logs fit into the category represented
by L4 , which would be better represented by a hybrid model where the top-level
is declarative, and sub-processes are imperative. Another explanation could be
that the structure of the process tree returned by the Inductive miner hinders
further improvements. After all, we take the structure of the process tree as-is
without considering other structures that are language equivalent to the returned
process tree.
In addition, it should be noted that, while we are using one of the most widely
accepted precision metric, there is still an ongoing search for better metrics [33].
For example, for a log containing at most 2 occurrences of an activity, the current
metric gives only a small benefit to an absence3 constraint over a process tree
loop, even though, from a language inclusion perspective, the absence3 constraint
is infinitely more precise. Improvements to the way precision is calculated for
hybrid models may also lead to different results.
6 Conclusion
We presented a novel technique for mining hybrid models, which combine the
strengths of the imperative and declarative process modeling paradigms. We
implemented our technique as a ProM plug-in and evaluated the approach on
several synthetic and real-life logs. Compared to the Inductive miner, the miner
showed significant improvements in precision for both synthetic and real-life logs.
In the case of real-life logs, it mostly found that purely declarative models were
the most precise, whereas in the case of synthetic logs, proper hybrid models
were found. Precision improvements ranged up to 52.18% for synthetic logs and
up to 3471% for real-life logs.
The presented approach is easily extendible and customizable. For example,
we can already plug any other miner that provides a process tree as output
into our implementation and it is also possible to extend the general approach
to other types of imperative models and miners. It is also possible to apply
different metrics for precision and to use various declarative miners.
As future work, we would like to implement and experiment with different
variants of the approach (e.g., different miners and ways to partition behavior
in logs into declarative and imperative parts). Moreover, we also plan to use an
approach opposite to the one proposed here: we could first attempt to mine a
log declaratively, analyze the resulting model to detect parts which are overly
constrained, and replace these with imperative sub-processes.
Mining Hybrid Business Process Models 203
References
1. van der Aalst, W.M.P., Adams, M., ter Hofstede, A.H.M., Pesic, M., Schonenberg,
H.: Flexibility as a service. In: Chen, L., Liu, C., Liu, Q., Deng, K. (eds.) DASFAA
2009. LNCS, vol. 5667, pp. 319–333. Springer, Heidelberg (2009). https://doi.org/
10.1007/978-3-642-04205-8 27
2. van der Aalst, W.M.P.: Process Mining - Data Science in Action, 2nd edn. Springer,
Heidelberg (2016). https://doi.org/10.1007/978-3-662-49851-4
3. van der Aalst, W.M.P., Adriansyah, A., van Dongen, B.F.: Replaying history on
process models for conformance checking and performance analysis. Wiley Inter-
disc. Rew. Data Min. Knowl. Disc. 2(2), 182–192 (2012)
4. van der Aalst, W.M.P., Rubin, V.A., Verbeek, H.M.W., van Dongen, B.F., Kindler,
E., Günther, C.W.: Process mining: a two-step approach to balance between under-
fitting and overfitting. Softw. Syst. Model. 9(1), 87–111 (2010)
5. van der Aalst, W.M.P., Weijters, A.J.M.M., Maruster, L.: Workflow mining: dis-
covering process models from event logs. IEEE Trans. Knowl. Data Eng. 16(9),
1128–1142 (2004)
6. Alberti, M., Chesani, F., Gavanelli, M., Lamma, E., Mello, P., Torroni, P.: Ver-
ifiable agent interaction in abductive logic programming: the SCIFF framework.
ACM Trans. Comput. Log. 9(4), 29:1–29:43 (2008)
7. Back, C.O., Debois, S., Slaats, T.: Towards an entropy-based analysis of log vari-
ability. In: Teniente, E., Weidlich, M. (eds.) BPM 2017. LNBIP, vol. 308, pp. 53–70.
Springer, Cham (2018). https://doi.org/10.1007/978-3-319-74030-0 4
8. Chesani, F., Lamma, E., Mello, P., Montali, M., Riguzzi, F., Storari, S.: Exploiting
inductive logic programming techniques for declarative process mining. T. Petri
Nets Other Models Concurrency 2, 278–295 (2009)
9. De Giacomo, G., Dumas, M., Maggi, F.M., Montali, M.: Declarative process mod-
eling in BPMN. In: Zdravkovic, J., Kirikova, M., Johannesson, P. (eds.) CAiSE
2015. LNCS, vol. 9097, pp. 84–100. Springer, Cham (2015). https://doi.org/10.
1007/978-3-319-19069-3 6
10. Debois, S., Hildebrandt, T., Slaats, T.: Hierarchical declarative modelling with
refinement and sub-processes. In: Sadiq, S., Soffer, P., Völzer, H. (eds.) BPM 2014.
LNCS, vol. 8659, pp. 18–33. Springer, Cham (2014). https://doi.org/10.1007/978-
3-319-10172-9 2
11. Debois, S., Hildebrandt, T.T., Marquard, M., Slaats, T.: Hybrid process technolo-
gies in the financial sector. In: BPM (Industry track), pp. 107–119 (2015)
12. Debois, S., Hildebrandt, T.T., Slaats, T., Marquard, M.: A case for declarative
process modelling: Agile development of a grant application system. In: EDOC
Workshops, vol. 14, pp. 126–133 (2014)
13. Di Ciccio, C., Mecella, M.: A two-step fast algorithm for the automated discovery
of declarative workflows. In: CIDM, pp. 135–142. IEEE (2013)
14. Lamma, E., Mello, P., Montali, M., Riguzzi, F., Storari, S.: Inducing declarative
logic-based models from labeled traces. In: Alonso, G., Dadam, P., Rosemann,
M. (eds.) BPM 2007. LNCS, vol. 4714, pp. 344–359. Springer, Heidelberg (2007).
https://doi.org/10.1007/978-3-540-75183-0 25
15. Lamma, E., Mello, P., Riguzzi, F., Storari, S.: Applying inductive logic program-
ming to process mining. In: Blockeel, H., Ramon, J., Shavlik, J., Tadepalli, P. (eds.)
ILP 2007. LNCS (LNAI), vol. 4894, pp. 132–146. Springer, Heidelberg (2008).
https://doi.org/10.1007/978-3-540-78469-2 16
204 D. M. M. Schunselaar et al.
16. Leemans, S.J.J., Fahland, D., van der Aalst, W.M.P.: Discovering block-structured
process models from event logs - a constructive approach. In: Colom, J.-M., Desel,
J. (eds.) PETRI NETS 2013. LNCS, vol. 7927, pp. 311–329. Springer, Heidelberg
(2013). https://doi.org/10.1007/978-3-642-38697-8 17
17. Maggi, F.M.: Declarative process mining with the declare component of ProM. In:
BPM (Demos). CEUR Workshop Proceedings, vol. 1021. CEUR-WS.org (2013)
18. Maggi, F.M., Bose, R.P.J.C., van der Aalst, W.M.P.: Efficient discovery of under-
standable declarative process models from event logs. In: Ralyté, J., Franch, X.,
Brinkkemper, S., Wrycza, S. (eds.) CAiSE 2012. LNCS, vol. 7328, pp. 270–285.
Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-31095-9 18
19. Maggi, F.M., Mooij, A.J., van der Aalst, W.M.P.: User-guided discovery of declar-
ative process models. In: CIDM, pp. 192–199. IEEE (2011)
20. Maggi, F.M., Slaats, T., Reijers, H.A.: The automated discovery of hybrid pro-
cesses. In: Sadiq, S., Soffer, P., Völzer, H. (eds.) BPM 2014. LNCS, vol. 8659, pp.
392–399. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10172-9 27
21. Marquard, M., Shahzad, M., Slaats, T.: Web-based modelling and collaborative
simulation of declarative processes. In: Motahari-Nezhad, H.R., Recker, J., Wei-
dlich, M. (eds.) BPM 2015. LNCS, vol. 9253, pp. 209–225. Springer, Cham (2015).
https://doi.org/10.1007/978-3-319-23063-4 15
22. de Medeiros, A.K.A., Weijters, A.J.M.M., van der Aalst, W.M.P.: Genetic process
mining: an experimental evaluation. Data Min. Knowl. Disc. 14(2), 245–304 (2007)
23. Montali, M.: Specification and Verification of Declarative Open Interaction Models.
LNBIP, vol. 56. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-
14538-4
24. Pesic, M., Schonenberg, H., van der Aalst, W.M.P.: DECLARE: full support for
loosely-structured processes. In: EDOC, pp. 287–300 (2007)
25. Pichler, P., Weber, B., Zugal, S., Pinggera, J., Mendling, J., Reijers, H.A.: Impera-
tive versus declarative process modeling languages: an empirical investigation. In:
Daniel, F., Barkaoui, K., Dustdar, S. (eds.) BPM 2011. LNBIP, vol. 99, pp. 383–
394. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-28108-2 37
26. Reijers, H.A., Slaats, T., Stahl, C.: Declarative modeling–an academic dream or
the future for BPM? In: Daniel, F., Wang, J., Weber, B. (eds.) BPM 2013. LNCS,
vol. 8094, pp. 307–322. Springer, Heidelberg (2013). https://doi.org/10.1007/978-
3-642-40176-3 26
27. Sadiq, S., Sadiq, W., Orlowska, M.: Pockets of flexibility in workflow specification.
In: S.Kunii, H., Jajodia, S., Sølvberg, A. (eds.) ER 2001. LNCS, vol. 2224, pp.
513–526. Springer, Heidelberg (2001). https://doi.org/10.1007/3-540-45581-7 38
28. Schönig, S., Rogge-Solti, A., Cabanillas, C., Jablonski, S., Mendling, J.: Efficient
and customisable declarative process mining with SQL. In: Nurcan, S., Soffer, P.,
Bajec, M., Eder, J. (eds.) CAiSE 2016. LNCS, vol. 9694, pp. 290–305. Springer,
Cham (2016). https://doi.org/10.1007/978-3-319-39696-5 18
29. Schunselaar, D.M.M.: Configurable Process Trees: Elicitation, Analysis, and Enact-
ment. Ph.D. thesis, Eindhoven University of Technology (2016)
30. Slaats, T., Schunselaar, D.M.M., Maggi, F.M., Reijers, H.A.: The semantics of
hybrid process models. In: OTM CoopIS. pp. 531–551 (2016)
31. Smedt, J.D., Weerdt, J.D., Vanthienen, J., Poels, G.: Mixed-paradigm process
modeling with intertwined state spaces. Bus. IS Eng. 58(1), 19–29 (2016)
32. Smedt, J.D., Weerdt, J.D., Vanthienen, J.: Fusion miner: process discovery for
mixed-paradigm models. Decis. Support Syst. 77, 123–136 (2015)
33. Tax, N., Lu, X., Sidorova, N., Fahland, D., van der Aalst, W.M.P.: The imprecisions
of precision measures in process mining (2017). https://arxiv.org/abs/1705.03303
Mining Hybrid Business Process Models 205
34. Weijters, A.J.M.M., van der Aalst, W.M.P.: Rediscovering workflow models from
event-based data using Little Thumb. Integr. Comput.-Aided Eng. 10(2), 151–162
(2003)
35. Westergaard, M., Slaats, T.: CPN Tools 4: A process modeling tool combining
declarative and imperative paradigms. In: BPM (Demos) (2013)
36. Westergaard, M., Slaats, T.: Mixing paradigms for more comprehensible models.
In: Daniel, F., Wang, J., Weber, B. (eds.) BPM 2013. LNCS, vol. 8094, pp. 283–290.
Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-40176-3 24
Extending BPSim Based on Workflow
Resource Patterns
1 Introduction
In this section we discuss the level of support offered by BPSim 2.0 standard
to resources perspective required for resources representation and distribution.
Next, we discuss in detail our extended version of BPSim 2.0 meta-model. The
extended meta-model is introduced to refine the spotted limitations in BPSim
standard concerning the resources classification and distribution with respect to
the refined Workflow Resource Patterns introduced in Sect. 2.
BPSim 2.0 introduces Scenario entity that shall contain all parameters needed
to run a simulation experiment [11]. A scenario could be used to capture input
parameter specification, simulation results and historical data from past business
processes execution [11].
The ElementParameter is the concrete class definition of all parameters [11].
It indicates the reference to BPMN elements retrieved from the business pro-
cess model interface and identifies different elements using a unique identifier.
ElementParameter is expanded with several parameters [11] such as TimeParam-
eter, ControlParameter that defines control flow of BPMN element, CostParam-
eter, PriorityParameter and ResourceParameter. Expressions could be attached
to ElementParamter to specify added functions such as getResource to select
a collection of available resources, getResourceByRole to select a collection of
available resources based on role and Resource to select an alternative list of
available resources.
ResourceParameter [11], groups all parameters related to resources required
by business process model element, e.g. task. The parameter determines
214 N. Afifi et al.
This subsection introduces how the refined inter- and intra-relationships among
Workflow Resource Patterns, introduced in Sect. 2, are supported by RBPSim,
the new extended BPSim meta-model.
Creation Patterns: Role-based, Capability-based, History-based and Organiza-
tional Distribution are determined by specifying the selectionConstraints and
selectionCriteria in ResourceSelection to eligible resources having specific role,
capabilities, history or organizational unit respectively through the connection
between ResourceSelection and Role, ResourceExecutionHistory and HumanRe-
sources.
Direct Distribution is modeled as Role-based Distribution by making sure
that only one resource is a member of the role within Role. Direct Distribution
Complements Push patterns through setting offeringCriteria to “offer single”
and allocationType to “allocate single” respectively. Deferred Distribution is
determined by setting the allocationTiming to “early Distribution?. Separation
of Duties, Retain Familiar and Case Handling are determined through defining
the required selectionConstraints to the required selection pattern and use the
execution history to select the appropriate resource.
Push Patterns: The complement relationships between Creation and Push
patterns, shown in Table 1, define the offering constraints through specifying
the offeringCriteria in ResourceOffering to offer single or multiple resources and
the offeringTiming to distribution on enablement. ResourceAllocation defines
work items allocation on cyclic basis, randomly, or shortest queue using the
allocationType.
Authorization specifies the resources privileges through resourcePrivileges
that enumerates the required authorization to handle work items execution.
Authorization complements Pull patterns through granting the following
ResourcePrivileges to resource: view offered, reject offered, view allocated, execu-
tion preference and order allocated. Detour patterns complements Authoriza-
tion through identifying delegate, reallocate stateful, reallocate stateless, sus-
pend, resume, skip and redo. Authorization complements Auto-Start Pat-
terns by defining commence created, commence allocated, pilled and chained
execution. Visibility Patterns are determined through visibility to allocated
and unallocated. Multiple Resource Patterns complements Authorization
through simultaneous execution.
Extending BPSim Based on Workflow Resource Patterns 217
3.4 Example
The example explains a simple business process for “Car Maintenance”, see
Fig. 5. When cars arrive, an administration employee receives and records car
information, the selection, offering and allocation of resource is following those
patterns: resources selection requires Role-based Distribution, offering follow Dis-
tribution by Offer-Multiple Resources, distribution timing is based on Distribu-
tion on Enablement and allocation is based on Round Robin. The car is then sent
to the mechanical department where an engineer is selected based on capabili-
ties with experience of 3 years and possession of a certificate following Role-based
Distribution and Capability-based Distribution. Finally, an accountant receives
and records payments for the work done. The accountant pulls work items from
his working queue based on work item priority.
4 Related Work
Several approaches, e.g. [18–20], discuss the need for business process simulation
tools as integral part of business process management.
BPSim was introduced in [11] as standardization effort of process simula-
tion scenario definition. Detailed analysis of BPSim limitations and chances was
discussed in [13]. One of the critical limitations mentioned was BPSim resource
model elaboration for simulation purposes that require extending BPSim meta-
model regarding resources perspective. RAL (Resource Assignment Language)
was introduced in [21], a metamodel that could be used with BPMN 2.0 for bet-
ter resources assignment. Discussions in [13] and in [21] declared that BPSim is
partially supported by current tools and requires additional extensions regarding
the vendors. Both researches indicated that neither BPSim nor BPMN are fully
supporting resources perspective. Work introduced in [5] specifies the resources
perspective in BPMN 2.0 metamodel and presents an extension to support the
resource modeling with respect to Workflow Resource Patterns. The extension
was extended in [22] to provide a conceptual modeling of the BPMN extension
using XML that can be processed by any BPMN tools.
A detailed metamodel for resource allocation constraint was introduced
in [23]. The introduced metamodel could be reflected in BPMN to overcome
drawbacks of supporting resources allocation. Authors in [24] discuss a standard
that might be extended to a process analytics framework. The proposed stan-
dard would limit the challenges when using simulation for business processes.
[3,25] discuss the limitations of business process simulation usage in-reality and
Extending BPSim Based on Workflow Resource Patterns 219
categorize the risks of current simulation settings. These pitfalls were further
extended and refined in [26]. Amongst such risks is the inadequate modeling of
resource behavior and availability. Our work is contributing to the avoidance of
this risk by means of pattern-based enforcement of resource behavior constraints.
Moreover, in [25], the authors use CPN tools by example to address the pitfalls
identified where they only addressed modeling of resource availability without
constraining resource allocation as we discuss in this paper.
A simulation environment based on YAWL workflow tool and CPN tools was
discussed in [2] allowing experiment to start from an intermediate stage rather
than empty execution state. Resources were only examined related to availabil-
ity, utilization and count. A hybrid approach for business process modeling and
simulation with limited resources modeling related to concurrency aspects were
addressed in [27]. The approach examined concurrent execution of tasks through
resources multitasking to avoid blocking resources for long time. Our work is con-
sidered complementary to both approaches. However, the concurrent execution
by human resources is of very limited use especially when it comes to knowledge
intensive tasks.
The work in [28] discusses supporting business process simulation model con-
struction by using event logs. They provide means to extract Resources behavior
from logs. The modeling of resources behaviors was supported by roles, sched-
ules, handling procedures, resources unavailability and grouping resources based
on similar activities handling. Yet, constraints on resource behavior was not
identified by the authors. Thus, our work complements theirs in that regard.
L-Sim, a BPMN-based simulator was presented in [6]. L-Sim supports a
subset of pull and push patterns. OXProS - an Open and Extensible Process
Simulator for BPMN was presented in [1], the tool supports modeling role-
based and chained execution distribution in workflow resources patterns. Other
resource patterns were left for future work. A blueprint architecture was proposed
in [29] for a business process simulation engine that focuses on BPMN models.
Resources modeling was only concerned with resource availability. Authors in [30]
presents an open source and extensible BPMN process simulator that provide
basic resource support provided by BPMN. Resources are defined globally using
a configuration file provided in XML to be involved in multiple processes.
A survey on business process simulation tools was conducted in [4]. Amongst
the evaluation criteria was the ability to model the resources perspective.
Amongst the evaluated tools were Arena2 (a general-purpose simulation tool)
and CPN tools3 . The study results indicated that CPN tools provides better
support for resources representation than Arena. However, both require quite
an effort to model resources behavior in a way that might be inaccessible to
business process experts.
2
https://www.arenasimulation.com/.
3
http://cpntools.org/.
220 N. Afifi et al.
References
1. Garcia-Banuelos, L., Dumas, M.: Towards an open and extensible business process
simulation engine. In: CPN Workshop (2009)
2. Wynn, M.T., Dumas, M., Fidge, C.J., ter Hofstede, A.H.M., van der Aalst, W.M.P.:
Business process simulation for operational decision support. In: ter Hofstede, A.,
Benatallah, B., Paik, H.-Y. (eds.) BPM 2007. LNCS, vol. 4928, pp. 66–77. Springer,
Heidelberg (2008). https://doi.org/10.1007/978-3-540-78238-4 8
3. van der Aalst, W.M.P.: Business process simulation revisited. In: Barjis, J. (ed.)
EOMAS 2010. LNBIP, vol. 63, pp. 1–14. Springer, Heidelberg (2010). https://doi.
org/10.1007/978-3-642-15723-3 1
4. Jansen-Vullers, M., Netjes, M.: Business process simulation-a tool survey. In: Work-
shop and Tutorial on Practical Use of Coloured Petri Nets and the CPN Tools,
vol. 38, pp. 1–20 (2006)
5. Stroppi, L.J.R., Chiotti, O., Villarreal, P.D.: A BPMN 2.0 extension to define the
resource perspective of business process models. In: XIV Congreso Iberoamericano
en Software Engineering (2011)
6. Waller, A., Clark, M., Enstone, L.: L-SIM: Simulating BPMN diagrams with a
purpose built engine. In: Simulation Conference, pp. 591–597. IEEE (2006)
7. Stroppi, L.J.R., Chiotti, O., Villarreal, P.D.: Defining the resource perspective in
the development of processes-aware information systems. Inf. Softw. Technol. 59,
86–108 (2015)
8. Nakatumba, J., Rozinat, A., Russell, N.: Business process simulation: how to get
it right. In: International Handbook on Business Process Management. Springer
(2009)
9. Russell, N., van der Aalst, W.M.P., ter Hofstede, A.H.M., Edmond, D.: Workflow
resource patterns: identification, representation and tool support. In: Pastor, O.,
Falcão e Cunha, J. (eds.) CAiSE 2005. LNCS, vol. 3520, pp. 216–232. Springer,
Heidelberg (2005). https://doi.org/10.1007/11431855 16
Extending BPSim Based on Workflow Resource Patterns 221
10. Freitas, A.P., Pereira, J.L.M.: Process simulation support in BPM tools: The case
of BPMN. In: Proceedings of 2100 Projects Association Joint Conferences (2015)
11. WfMC: BPSim - business process simulation specification (2016)
12. OMG: Business process model and notation (BPMN version 2.0) (2011)
13. Laue, R., Müller, C.: The business process simulation standard (BPSim): Chances
and limits. In: ECMS, pp. 413–418 (2016)
14. Afifi, N., Awad, A., Abdelsalam, H.M.: Rbpsim: A resource-aware extension of
BPSim using workflow resource patterns. In: 10th Zeus Workshop, vol. 2072,
Dresden, Germanny, pp. 31–38. CEUR-WS (2018)
15. Russell, N., Ter Hofstede, A.H., Edmond, D., van der Aalst, W.M.: Workflow
resource patterns. Technical report, BETA Working Paper Series, WP 127, Eind-
hoven University of Technology, Eindhoven (2004)
16. Russell, N., van der Aalst, W.M., ter Hofstede, A.H.M.: Workflow Patterns: The
Definitive Guide. The MIT Press (2016)
17. Mendling, J., Ploesser, K., Strembeck, M.: Specifying separation of duty con-
straints in BPEL4People processes. In: Abramowicz, W., Fensel, D. (eds.) BIS
2008. LNBIP, vol. 7, pp. 273–284. Springer, Heidelberg (2008). https://doi.org/10.
1007/978-3-540-79396-0 24
18. Tumay, K.: Business process simulation. In: Winter Simulation, pp. 93–98. IEEE
Computer Society (1996)
19. Hlupic, V., Robinson, S.: Business process modelling and analysis using discrete-
event simulation. In: Winter Simulation Conference, pp. 1363–1370. IEEE Com-
puter Society Press (1998)
20. Mathew, B., Mansharamani, R.: Simulating business processes-a review of tools
and techniques. Int. J. Model. Optim. 2(4), 417 (2012)
21. Cabanillas, C., Resinas, M., Ruiz-Cortés, A.: RAL: a high-level user-oriented
resource assignment language for business processes. In: Daniel, F., Barkaoui, K.,
Dustdar, S. (eds.) BPM 2011. LNBIP, vol. 99, pp. 50–61. Springer, Heidelberg
(2012). https://doi.org/10.1007/978-3-642-28108-2 5
22. Stroppi, L.J.R., Chiotti, O., Villarreal, P.D.: Extending BPMN 2.0: method and
tool support. In: Dijkman, R., Hofstetter, J., Koehler, J. (eds.) BPMN 2011.
LNBIP, vol. 95, pp. 59–73. Springer, Heidelberg (2011). https://doi.org/10.1007/
978-3-642-25160-3 5
23. Awad, A., Grosskopf, A., Meyer, A., Weske, M.: Enabling Resource Assignment
Constraints in BPMN. Hasso Plattner Institute, Potsdam (2009)
24. Januszczak, J., Hook, G.: Simulation standard for business process management.
In: Proceedings of the 2011 Winter Simulation Conference (WSC), pp. 741–751.
IEEE (2011)
25. van der Aalst, W.M.P., Nakatumba, J., Rozinat, A., Russell, N.: Business process
simulation. In: Brocke, J., Rosemann, M. (eds.) Handbook on Business Process
Management, vol. 1. International Handbooks on Information Systems. Springer,
Heidelberg (2010). https://doi.org/10.1007/978-3-642-00416-2 15
26. van der Aalst, W.M.P.: Business process simulation survival guide. In: vom Brocke,
J., Rosemann, M. (eds.) Handbook on Business Process Management, vol. 1. Inter-
national Handbooks on Information Systems. Springer, Heidelberg (2015). https://
doi.org/10.1007/978-3-642-45100-3 15/
27. Vasilecas, O., Smaižys, A., Rima, A.: Business process modelling and simulation:
hybrid method for concurrency aspect modelling. Baltic J. Mod. Comput. 1(3–4),
228–243 (2013)
28. Martin, N., Depaire, B., Caris, A.: The use of process mining in business process
simulation model construction. Bus. Inf. Syst. Eng. 58(1), 73–87 (2016)
222 N. Afifi et al.
29. Krumnow, S., Weidlich, M., Molle, R.: Architecture blueprint for a business process
simulation engine. In: EMISA, vol. 172, pp. 9–23 (2010)
30. Pufahl, L., Weske, M.: Extensible BPMN process simulator. In: BPM Demo, vol.
1920. CEUR Workshop Proceedings. CEUR-WS.org (2017)
Towards Implementing REST-Enabled
Business Process Choreographies
1 Introduction
In the era of Internet, enterprises can focus on their core capabilities and out-
source supporting activities to other business partners forming so called net-
worked enterprises. This is achieved through fast and reliable inter-organizational
communication. When it comes to the interaction on the Web, one of the
most adopted architectural styles is Representational State Transfer (REST) [1].
On the modeling side, business process choreographies model the inter-
organizational processes that are performed by business participants to reach
a common goal. In particular, choreography diagrams are introduced as part of
BPMN 2.0 [2] to formally specify the ordering of message exchanges between
two or more participants.
However, there is a large conceptual gap between the modeling level and the
concrete RESTful interaction. The gap can be bridged by translating the global
perspective, from which the interactions between participants are modeled, to
the RESTful services that participants need to provide for implementing the
specified interactions. In [3], the authors partially address this gap by annotat-
ing choreography diagrams with REST implementation-specific information in
a semi-automatic fashion. This hybrid model is called RESTful choreography.
In this paper, we further close the gap between RESTful choreography and its
c Springer International Publishing AG, part of Springer Nature 2018
W. Abramowicz and A. Paschke (Eds.): BIS 2018, LNBIP 320, pp. 223–235, 2018.
https://doi.org/10.1007/978-3-319-93931-5_16
224 A. Nikaj et al.
2 Preliminaries
This section introduces choreography diagrams and RESTful choreography along
with a running example used for motivating our approach.
Business process choreography [2] is a specification language for modeling
interactions between two or more participants. It abstracts from the internal logic
of participants’ own process models and provides a global perspective without
a preferential viewpoint. Figure 1 describes the interaction of three participants
involved in a purchase process. The skateboard manufacturer sends an order to
the ball bearings supplier, which the supplier can confirm or reject. In case of
the former, the supplier sends the ball bearings to the skateboard manufacturer
as requested in the order. Once the ball bearings are delivered the skateboard
manufacturer determines the percentage of defective units. If the percentage is
lower than 5% the manufacturer initiates a payment request to the payment
organization (the third participant). Otherwise, the skateboard manufacturer
sends back the ball bearings leading to an unsuccessful purchase.
The main composing unit of a choreography diagram is the choreography
task (the rounded rectangles in Fig. 1), which represent either sending a single
message or a send message followed by a reply message. In a single choreog-
raphy task, the participant who sends the message is called the initiator and
the message is referred to as initiating message, while the recipient may send
back an optional response message. Similarly to activities in business processes,
the choreography tasks are ordered via sequence flows. To express parallel or
exclusive branches a subset of business process gateways is used. In this paper
we consider only three type of gateways: parallel, exclusive and event based
gateways.
1
https://www.airbnb.com/.
2
http://www.easychair.org/.
Towards Implementing REST-Enabled Business Process Choreographies 225
REST [1] is an architectural style for web services. Services provide their
resources to clients via unique identifiers (URI) allowing only a specified set of
operations (REST verbs) with a stateless nature. Key feature of RESTful interac-
tions is the usage of hypermedia as the engine of application state (HATEOAS).
Links provided in the response guide the clients through subsequent interactions.
These architectural constraints contribute to a simplified system architecture,
better performance and improved evolvability of services without affecting the
interaction.
In order to bridge the conceptual gap between process choreographies and
their implementation as RESTful interactions, [4] introduces RESTful chore-
ography diagrams. They are a light-weight extension of BPMN choreography
diagrams with REST annotations (see Fig. 1). Choreography tasks are enhanced
to REST tasks, in which the initiating message is annotated with a REST request
method (e.g. POST /order ) and the return message is annotated with response
status codes (e.g. 201 Created, Location: /order/id ). Hyperlinks can be added in
the request and response to guide the participants, who can follow them accord-
ing to the behavior modeled in the RESTful choreography. Moreover, in [3] it is
shown that the generation of REST annotations can be automatized to a good
extend.
Regarding the second requirement, the control flow and the message
exchanges in a RESTful choreography are modeled from a global perspective.
We provide a mapping, detailed in Sect. 3.2, of the choreography control flow
to business process control flow where message exchanges are implemented by
process orchestration activities.
Figure 2 provides an overview of our approach for implementing the RESTful
choreography via ChoreoGuide. Our approach takes as input a RESTful chore-
ography and its resource model. The target model is a BPMN [2] process model
with embedded REST information that can be deployed to a business process
engine capable of providing a RESTful API.
purpose of specifying the main REST resources, their state and their relation to
each other and to choreography participants as well. It is highly influenced by the
domain area and specified by the involved participants alongside the RESTful
choreography diagram at design time.
Figure 3 illustrates the static view of the choreography resource model for
our running example modeled in Eclipse Modeling Framework3 . Every Resource-
Model, independently of the use case, is always composed of at least two Partic-
ipants and at least one RestResource. Depending on the use case, participants
and REST resources are added as classes that inherit the abstract classes Par-
ticipant and RestResource respectively. The attributes and the associations are
filled depending on the use case domain. For example, Order has a creationDate
and BallBearing as an item.
certain conditions of the resource which are aggregated from the concrete value
assignments of the resource attributes. For example, resource order is in the
state placed only if the quantity of the ball bearings is greater than 0. We use
OCL [5] to formally express such conditions that must hold during the entire
execution of the choreography. Below is the respective OCL expression for the
aforementioned condition.
context o r d e r
inv nonEmptyOrder : s e l f . s t a t e = O r d e r S t a t e : : PLACED
implies s e l f . item . q u a n t i t y > 0
The OCL expression refers to the UML class diagram. In this expression we
look at the class Order. Then we identify an invariant (nonEmptyOrder ) that
must hold during the instantiation of the order objects. The invariant in this
case is an implication where the premise is the resource in state PLACED and
the consequence is an expression that must hold for that state. In fact, this is
the pattern we propose to check for any resource object the validity of its state:
context <<r e s o u r c e >>
inv <<r e s o u r c e s t a t e >>
implies <<c o n d i t i o n >>
In order to use the OCL expression, each RESTResource child class must
have an attribute state of the type Enumeration that contains all the possible
states of that particular resource, e.g., Payment can be in state initiated, con-
firmed, or failed. The condition of resource state can be arbitrary complex and
long (to the extend allowed by OCL). OCL provides the possibility to navigate
the class diagrams and express complex relations between resources. Using a
standard like OCL allows the developers to choose the desired tools that sup-
port evaluating the expressions. Another key benefit is that OCL is free from
side effects. This means that checking whether an OCL expression holds does
not affect the running program. When it comes to choreographies, this is very
useful because ChoreoGuide should only reply whether the state change request
is valid or not without interfering with the internal logic of the participants’
process.
business process constructs (see Fig. 4): a receive task; a service task; an exclusive
gateway which branches the control flow in two parts; a “fail” message interme-
diate throwing event; a “successful” message intermediate throwing event; and
a send task that forwards the initiator’s request. Specifically, as illustrated in
Fig. 4, the receiving task is accessed from the participants by the REST request
specified in the REST task, e.g., POST /order/ HTTP/1.1. In case of a POST
or PUT a resource or a representation of it is delivered as a message payload to
the server. With POST, a new resource is created in an initial state and PUT
is used to update a resource with a new state. After the resource is received
from the server, an automatic check is performed in the subsequent step by a
service task. This task automatically checks whether or not the state-change
request is valid by evaluating the respective OCL expression in the resource
model. In case the OCL expression is evaluated to false, the server replies to
the client (the participant who sent the request) with a HTTP/1.1 423 Pre-
condition Required. Otherwise, the server replies with a HTTP/1.1 201 Created
(for POST) or HTTP/1.1 200 OK (for PUT). RESTful choreography diagrams
model only the successful interaction between the client and the server. Hence,
the response of a valid request used in the process orchestration is derived from
the choreography. Finally, the link for getting the new resource state is forwarded
to the recipient.
Fig. 4. Business process construct for POST and PUT (in brackets) REST tasks
230 A. Nikaj et al.
GET is used to read the state of the resource and DELETE to remove the
resource from the location specified in the URI. In case of GET task, the state
of the resource does not change and, therefore, there are no conditions to be
checked. The corresponding process orchestration (see Fig. 5) has three consec-
utive nodes linked by sequence flows: a receiving task for the request, a message
intermediate throwing event for replying to the initiator by sending the resource
(or a partial representation of it) and a send message task for notifying the
recipient that the resource has been read.
When a DELETE request is sent, there are no conditions to be checked
(similarly to GET task) in terms of attribute values. Hence, the correspond-
ing process orchestration is the same as in the GET case (see Fig. 5), but a
“resource deleted” message notification is sent back in lieu of a reply containing
the resource. It is worth mentioning that the delete request is not forwarded
to the participant prior to the server reply. This is because this delete request
is part of the intended behavior specified by the RESTful choreography. If any
participant would arbitrary request to delete a resource, ChoreoGuide will not
accept the request unless it is part of the designed behavior.
Fig. 5. Business process construct for GET and DELETE (in brackets) REST tasks
followed by the construct for exclusive gateway
4 Implementation Architecture
This section presents the architecture of the proposed approach and discusses
how the generated ChoreoGuide orchestration process is deployed to an existing
process engine to validate the feasability of the approach. Figure 6 shows the
main components of the architecture, (1) the parser responsible for parsing the
RESTful choreography and resource model, (2) the generator that derives the
ChoreoGuide orchestration process and deploys it, and (3) the process engine
232 A. Nikaj et al.
The Chimera engine requires process model to define the data classes it
operates on by providing a data model. For ChoreoGuide we use the sub-classes
of Resource from the resource model of the choreography to be deployed. The
choreography resource model is used to define data objects and their attributes.
OCL conditions are checked in the process engine. The OLC conditions refer to
attributes of resources defined in the resource model. This corresponds to the
data objects managed by the ChoreoGuide instance. Therefore, the conditions
are used as annotations on the sequence flows following the exclusive gateway.
When they evaluate to true the choreography continues, otherwise a response
with HTTP status code 428 is send to the requester.
5 Related Work
There is a plethora of related work covering the execution of the choreography’s
control flow. Zaha et al. are among the first authors to put forward the issue
of local enforceability in [8] in the context of web services standards like SOAP
and WSDL [9]. They propose an approach for analyzing if the relation between
service interactions described from a global perspective could be enforced locally
by generating BPEL [10] templates.
In [11], Barros et al. focus on the correlation between the exchanging mes-
sages. The correlation between messages is an important factor when it comes
to the implementation of choreographies, and thus, to their enforceability. They
introduce a framework for classifying correlation scenarios in the context of SOA
and argue that BPEL does not support certain patterns. Decker et al. [12] extend
BPEL web service composition standard [10] to bring process orchestration closer
to choreographies. BPEL4Chor implements process choreographies by compos-
ing existing BPEL service orchestrations. This bottom-up approach is based on
web services standards like SOAP and WSDL.
In these related work the local enforceability is checked from the control
flow perspective. The authors assume that the data used in the choreography is
234 A. Nikaj et al.
6 Conclusion
In this paper we introduce a novel approach towards implementing business
process choreographies. Existing work, focus on deriving public processes for
each participant and enforce their execution. We propose a central RESTful
service that takes the participants “by the hand” and guides them through every
step along the choreography. The service employs the REST architectural style
to take advantage of the REST constraints.
One key benefit of our approach is that the participants are not required to
run a process engine or a complex system to interact with each other. This is par-
ticularly important for the interaction with human clients that can participate
in the choreography via a simple web browser. Our approach fully employs the
principle of HATEOAS, i.e. hyperlinks are indeed the engine of the choreography
state.
Another important benefit is that ChoreoGuide not only enforces the con-
trol flow but validates the correctness of the resource state changes using the
choreography resource model. In addition, having such a model helps the partic-
ipant to agree on a common data structure and avoid problems coming from the
misinterpretation of the inter-organization data. Future work will focus on the
automatic deployment of ChoreoGuide into a specific business process engine.
Towards Implementing REST-Enabled Business Process Choreographies 235
References
1. Fielding, R.T.: Architectural Styles and the Design of Network-based Software
Architectures. PhD thesis, AAI9980887 (2000)
2. OMG: Business Process Model and Notation (BPMN), Version 2.0, January 2011.
http://www.omg.org/spec/BPMN/2.0/
3. Nikaj, A., Weske, M., Mendling, J.: Semi-automatic derivation of restful choreogra-
phies from business process choreographies. Softw. Syst. Model., January 2018
4. Nikaj, A., Mandal, S., Pautasso, C., Weske, M.: From choreography diagrams
to RESTful interactions. In: Norta, A., Gaaloul, W., Gangadharan, G.R., Dam,
H.K. (eds.) ICSOC 2015. LNCS, vol. 9586, pp. 3–14. Springer, Heidelberg (2016).
https://doi.org/10.1007/978-3-662-50539-7 1
5. OMG: Object Constraint Language (OCL), Version 2.0, May 2006. http://www.
omg.org/spec/OCL/2.0/
6. Nikaj, A., Batoulis, K., Weske, M.: REST-enabled decision making in business
process choreographies. In: Sheng, Q.Z., Stroulia, E., Tata, S., Bhiri, S. (eds.)
ICSOC 2016. LNCS, vol. 9936, pp. 547–554. Springer, Cham (2016). https://doi.
org/10.1007/978-3-319-46295-0 34
7. Pautasso, C., Wilde, E.: Push-enabling RESTful business processes. In: Kappel,
G., Maamar, Z., Motahari-Nezhad, H.R. (eds.) ICSOC 2011. LNCS, vol. 7084, pp.
32–46. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-25535-9 3
8. Zaha, J.M., Dumas, M., Hofstede, A.T., Barros, A., Decker, G.: Service interaction
modeling: bridging global and local views. In: 10th IEEE International Enterprise
Distributed Object Computing Conference, EDOC 2006, pp. 45–55. IEEE (2006)
9. Alonso, G., Casati, F., Kuno, H., Machiraju, V.: Web Services. Springer, Heidelberg
(2004). https://doi.org/10.1007/978-3-662-10876-5
10. Jordan, D., Evdemon, J., Alves, A., Arkin, A., Askary, S., Barreto, C., Bloch, B.,
Curbera, F., Ford, M., Goland, Y., et al.: Web services business process execution
language version 2.0. OASIS standard 11, 10 (2007)
11. Barros, A., Decker, G., Dumas, M., Weber, F.: Correlation patterns in service-
oriented architectures. In: Dwyer, M.B., Lopes, A. (eds.) FASE 2007. LNCS, vol.
4422, pp. 245–259. Springer, Heidelberg (2007). https://doi.org/10.1007/978-3-
540-71289-3 20
12. Decker, G., Kopp, O., Leymann, F., Weske, M.: Bpel4chor: Extending BPEL for
modeling choreographies. In: IEEE International Conference on Web Services,
ICWS 2007, pp. 296–303. IEEE (2007)
13. Pautasso, C.: BPMN for REST. In: Dijkman, R., Hofstetter, J., Koehler, J. (eds.)
BPMN 2011. LNBIP, vol. 95, pp. 74–87. Springer, Heidelberg (2011). https://doi.
org/10.1007/978-3-642-25160-3 6
14. Weber, I., Xu, X., Riveret, R., Governatori, G., Ponomarev, A., Mendling, J.:
Untrusted business process monitoring and execution using blockchain. In: La
Rosa, M., Loos, P., Pastor, O. (eds.) BPM 2016. LNCS, vol. 9850, pp. 329–347.
Springer, Cham (2016). https://doi.org/10.1007/978-3-319-45348-4 19
15. Rimba, P., Tran, A.B., Weber, I., Staples, M., Ponomarev, A., Xu, X.: Comparing
blockchain and cloud services for business process execution. In: 2017 IEEE Inter-
national Conference on Software Architecture, ICSA 2017, Gothenburg, Sweden,
3–7 April 2017, pp. 257–260 (2017)
Disambiguation of DMN Decision Tables
1 Introduction
In the Decision Model and Notation (DMN) specification, decision tables are a
standardized way of expressing decision logic within a decision model [8]. There-
fore, the majority of decision models consist of tables and much research has been
conducted based on this representation, such as analysis and simplification [5,6],
mining [3], and verification of their integration with process models [1,2].
In principal, decision tables provide a clear view of how decisions are taken
because of its tabular representation of rules that map inputs to outputs. In
the simplest case, for a given input, at most one rule matches and its output
is returned. However, the DMN standard allows to design more complex tables
where for a given input more than one rule matches. In these cases, there are
standardized possibilities of resolving such conflicts. For example, one could just
return the outputs of all matching rules. The problem with such tables is that
their input-output behavior is not clearly visible anymore. This is because each
input could match any number of rules, whose outputs are then aggregated in
some way.
In this paper, we describe an algorithm that transforms any type of DMN
decision table that uses S-FEEL syntax into a behaviorally equivalent table
with exclusive rules, i.e., a table that has the same input-output behavior but
for which a given input matches at most one rule. Such tables are unambiguous
because they have a clear input-output behavior and are therefore more com-
prehensible and also better suited for analysis tasks such as the ones described
Fig. 1. A DMN decision table with overlapping rules and a rule order hit policy
The table has five rules, each of which matches for certain combinations of
input values and relates them to an output value. For example, rule 1 matches
for the input (15, 10) and relates it to the output A. However, the same input
is also matched by rule 2, which relates it to a different value, namely B. This
means that rules 1 and 2 are overlapping, leading to ambiguities regarding the
input values for which they both match: Which output value should be chosen
for these input values?
Such ambiguities are resolved by the hit policy of the decision table, repre-
sented by a single letter in its upper left corner. For example, the letter R in
the upper left corner in Fig. 1 means that this table has a rule order hit policy.
Under this policy, ambiguities due to multiple matching rules are resolved by
just putting the outputs of all the matching rules in a list sorted by the order
238 K. Batoulis and M. Weske
of those rules in the table. Therefore, the table would actually relate the input
(15, 10) to the output [A, B].
This leads to the problem of determining the input-output behavior of a DMN
decision table. More precisely, what are the possible output values of the table,
and which input values generate which of these output values? For example, the
possible output values for the table in Fig. 1 are the following:
A, B, C, D, E, [A, B], [C, D], [C, E], [D, E], [C, D, E],
which is a far from obvious by just looking at the table. Similarly, it is not
obvious which of these output values is generated from given input values. For
instance, given the input (65, 90), one can see that it matches rule 3, but since
this is a table with overlapping rules, one cannot be sure that this is the only
matching rule. One has to keep scanning the rules until all matches are found—
an error-prone process. In fact, also rule 5 matches for the input and the output
is [C, E].
To summarize, DMN decision tables with overlapping rules are unintuitive
and hard to understand. Moreover, because of their obscure input-output behav-
ior they are unsuited for analysis tasks such as checking their correct integration
with business process models [1,2,7]. Hence, in this paper, we propose an algo-
rithm that transforms any type of DMN decision table into one that has only
exclusive rules—a process we call disambiguation. Since the rules are exclu-
sive, the possible output values of the table are simply given by the outputs of
the individual rules—no complicated combination of multiple matching rules is
required—leading to a clear input-output behavior of that table.
• rule order : all matching rules’ outputs returned in a list ordered by the
appearance of the rules in the table.
• output order : all matching rules’ outputs returned in a list ordered by the
priority of the rules.
• collect: the outputs of all matching rules are collected and then aggregated
in a predefined way, for example by summing up the output values or taking
the maximum.
Our approach supports all of these hit policies, and translates tables with a
non-unique policy to an equivalent table with a unique policy.
The algorithm for disambiguating decision tables presented in this paper builds
on an analysis approach for finding overlapping and missing rules in a DMN
decision table described in [5,6]. This approach is based on a geometric inter-
pretation of DMN tables. Under this interpretation, the rules of a table are
represented as hyperrectangles with n dimensions, where n corresponds to the
number of inputs of the table. For example, the geometric interpretation of the
table in Fig. 1 is given in Fig. 2 as a set of 2-dimensional hyperrectangles (i.e.,
rectangles).
Each rectangle corresponds to a rule of the table and is labeled with a corre-
sponding identifier. For instance, the rectangle r2 represents the second rule of
the table. This rule states that the Income must be between 10 and 60 and the
Assets between 10 and 25, which is exactly the area covered by r2.
The problem of finding overlapping rules is now formulated as that of finding
intersecting rectangles. Similarly, finding missing rules is equivalent to finding all
the “white” areas that are not covered by any of the rectangles. These problems
can be solved by the sweep line approach [4], which analyzes one dimension
after another by sweeping a line through it. For example, Fig. 2 shows a line
being swept through the Income dimension, from left to right. The problem of
finding overlapping rules, for instance, is then solved as follows: Sweep a line
through the first dimension, collecting as many rules as possible that overlap in
this dimension. Then, check which of the collected rules also overlap in the other
dimensions, by sweeping a line through those dimensions.
3 Table Disambiguation
This section is dedicated to our algorithm for disambiguating any DMN decision
table. For any given input to a DMN decision table with a set of rules R, any
subset of R can match, if the table is not a unique-hit table. Given the geometric
interpretation of the decision table in the running example, it is quite easy to
see which rule(s) match(es) for which input.
For example, for all inputs where Income ∈ [10, 30] and Assets ∈ [10, 25]
rules 1 and 2 match. For Income ∈ (30, 60] and Assets ∈ [10, 25], however, only
rule 2 matches. Of course, there can also be inputs for which no rules match.
These are all the white “gaps” between the rectangles in Fig. 2. This means that
the table is incomplete.
Based on the line sweeping technique, we developed an algorithm that will—
for any DMN table—find out for which inputs which rules match. In case the
table is not a unique-hit table, it can afterwards be translated to one that is
unique. This means that the rules are then exclusive, leading to a more under-
standable, disambiguated decision table.
Algorithm 1 shows the pseudocode of our algorithm. The initial call to the
algorithm will look like this: f indM atchingRules(R, [ ], 0, N, [ ]), so that R is the
set of rules of the table, matchingIntervals is an initially empty list, i is an index
initially set to 0, N is the number of inputs of the table, and matchingRules is
an initially empty list.
Then, the algorithm starts sweeping through the first dimension. This
requires sorting the endpoints of the intervals covered by each rule in this dimen-
sion in ascending order (line 2). Then the endpoints are iterated over (lines 4–12).
Whenever the current endpoint is a lower one (line 8), the rule belonging to this
endpoint is considered active, and is therefore added to the list of activeRules
(line 9). Contrarily, if the current endpoint is an upper one (line 10), the rule
belonging to this endpoint is considered dead, and is therefore deleted from the
list of activeRules (line 11).
For example, if i = 0, then sortedEndpoints = [r1(0), r2(10), r3(20), r1(30),
r5(40), r2(60), r4(80), r3(95), r4(∞), r5(∞)]. Thus, in the first iteration r1
Disambiguation of DMN Decision Tables 241
Algorithm 1. findMatchingRules
Data: rules, matchingIntervals, i, N, matchingRules
1 if i < N then
2 sortedEndpoints = rules.getSortedRulesEndpoints(i)
3 activeRules = [ ]
4 foreach endpoint ∈ sortedEndpoints do
5 if !activeRules.isEmpty() then
6 matchingIntervals[i] = [lastEndpoint, endpoint]
7 findMatchingRules(activeRules, matchingIntervals, i + 1, N ,
matchingRules)
8 if endpoint.isLowerBound() then
9 activeRules.add(endpoint.getRule())
10 else
11 activeRules.delete(endpoint.getRule())
12 lastEndpoint = endpoint
13 else
14 if matchingRules.canBeMergedWith({activeRules, matchingIntervals}) then
15 matchingRules.mergeWith({activeRules, matchingIntervals})
16 else
17 matchingRules.add({activeRules, matchingIntervals})
1
The interval does not include the upper endpoint because for the value 10 also r2
would be active.
242 K. Batoulis and M. Weske
Fig. 3. Line being swept through the second dimension in the interval [0, 10)
and since matchingRules is so far empty, it is added to the list (line 17):
matchingRules = [{[r1], [[0, 10), [0, 30]]}].
The algorithm then returns from the recursive call and deletes r1 from
activeRules because the current endpoint is r1(30), such that it is now empty
and all endpoints of the second dimension have been processed. Therefore,
Fig. 4. Line being swept through the second dimension in the interval [10, 20)
Disambiguation of DMN Decision Tables 243
the algorithm returns to the first dimension, where the current endpoint is
r2(10) (cf. Fig. 2). This means that r2(10) is added to activeRules, so that
in the next loop iteration the line in Fig. 2 is swept until it hits the end-
point r3(20). Hence, a recursive call is made where activeRules = [r1, r2] and
matchingIntervals = [[10, 20)]. This means that the line is swept across the
second dimension in that interval (cf. Fig. 4), so that the algorithm will detect
that
• r1 matches for the intervals [[10, 20), [0, 10)],
• both r1 and r2 match for the intervals [[10, 20), [10, 25]], and
• r1 also matches for the intervals [[10, 20), (25, 30]].
Back in the first dimension, the current endpoint is r3(20), which is added
to activeRules, and the line is swept until it hits the closing endpoint r1(30).
Therefore, the line is swept across the second dimension in the interval [20, 30].
The current state of all matchingRules would then be visualized as in Fig. 5a.
In this figure the rectangles are labeled with rules that match for this rectan-
gle. For example, we found out that rules 1 and 2 both match for the rectangle
[[10, 20), [10, 25]], such that it is labeled r1, r2.
Based on this representation, it becomes obvious that there is room for
improvement, because there are rectangles that are actually very similar, in
the sense that they only differ in one dimension, but in this dimension they are
adjacent. For instance, the two rectangles that are labeled with r1, r2 are the
same with respect to the Assets dimension, i.e., they both cover the interval
244 K. Batoulis and M. Weske
[10, 25]. In the Income dimension they differ, but they are adjacent, because one
covers the interval [10, 20) and the other one covers the interval [20, 30]. Thus,
it seems that there is the possibility of merging pairs of rectangles into a single
rectangle.
In general, two hyperrectangles with n dimensions can be merged if they
overlap in n − 1 dimensions and are adjacent in the remaining dimension, and if
the corresponding rules have the same output. This observation has already been
described in [6]. However, [6] only deals with unique-hit tables. This means that
a hyperrectangle only represents one rule, such that the corresponding output is
unambiguous. In our case, however, more than one rule can belong to the same
rectangle, so that the output of that rectangle depends on the hit policy of the
underlying decision table.
For example, since the table in Fig. 1 has a rule order hit policy, the rectangles
in Fig. 5a that are labeled r1, r2 will produce the output [A, B], so that they can
be merged with each other, but not with any of the remaining rectangles. The
result of all possible merges given a rule order hit policy is illustrated in Fig. 5b.
Now assume that the table has a first hit policy instead. In this case, all of the
rectangles in the lower left corner in Fig. 5a will produce the same output (A).
Therefore, the result after merging would be as shown in Fig. 5c.
Regarding our algorithm this means that whenever a new rectangle is found,
it first checks if it can be merged with other rectangles found so far (line 14). This
includes determining the output that is produced by the newly found rectangle
based on the hit policy of the table. If merges are possible, the newly found
and the existing rectangles are merged and then the combined result is put
into the list of matchingRules instead of its parts (line 17). Therefore, after
the line has been swept across the second dimension in the interval [20, 30], the
list of matchingRules will actually look as visualized in Fig. 5b, because three
rectangle pairs can be merged.
Eventually, the algorithm will have analyzed all sortedEndpoints, and the
final list of matchingRules will have 20 elements in total, the last one being
{[r4, r5], [(95, inf ), [85, inf )]}. This list is visualized in Fig. 6. Given the visual-
ization of this list, it becomes clear that we can now construct a new table that
exactly contains the rules represented as rectangles in the figure. This table is
shown in Fig. 7. It has 15 exclusive rules and is therefore a unique hit table as
indicated by the letter U in the upper left corner. It has the same input-output
behavior as the table in Fig. 1. However, due to the fact that all of its rules
are exclusive it is more easily comprehensible and the set of possible outputs of
the table is immediately visible. Furthermore, this representation of the table is
more suitable for certain analysis tasks as described in Sect. 2.
4 Evaluation
way, we generated tables that have a random number of overlapping rules and
can therefore be disambiguated by our algorithm.
We applied the algorithm to all the tables in all ten sets and then averaged
the execution times over the ten sets. The results are shown in Fig. 8 as a three-
dimensional surface plot. The first observation is that the execution times depend
more heavily on the number of rows than the number of columns. This is because
the more rules there are the more intervals need to be checked for matching
rules. However, the maximum average execution time was actually reached for
the 50 × 6-tables with 151 s; and also other tables with similar numbers of rules
and columns lead to comparably high execution times.
The reason for that is the high number of overlapping rules of those tables.
More specifically, tables with a high number of rules but a low number of columns
have much more rules that overlap across all dimensions. This leads to a high
number of activeRules in Algorithm 1 that need to be compared to all the other
rules for potential merges (line 14). The average total number of overlaps of the
rules of the tables in our data set is shown in Fig. 9. Here, the maximum number
of overlaps is 1138 for the 50 × 6-tables, which coincides with the maximum in
Fig. 8.
Note that the number of overlaps for the tables with a high number of
columns is nearly always zero. This is because the higher the number of columns,
the harder it is for multiple rules to overlap in all of those dimensions. This
shows that our disambiguation approach is especially suited for tables with a
high number of rules and a lower number of columns.
Fig. 8. Execution times in seconds for multi-hit tables with up to 50 rows and 30
columns
Disambiguation of DMN Decision Tables 247
5 Related Work
Transforming decision tables to equivalent ones that are disambiguated, is highly
related to the topic of normalization of decision tables, which has been brought
up in [9] and re-used in [10]. Normalization of decision tables is based on nor-
malization rules in relational databased design and also aims at increasing the
understandability of the tables. To this end, [9,10] describe three normal forms
of decision tables.
The first form is mainly concerned with the layout of the table and the
structure of its logical expressions. For example, rules of a decision should be
represented as rows in the table (DMN also allows a rules-as-columns layout).
Moreover, the individual input conditions must always be connected conjunc-
tively, never disjunctively. Second normal form requires that, unless an input
is explicitly marked as irrelevant for a particular rule, all conditions of a rule
must be relevant for the output of that rule. For example, in rule 4 in the table
in Fig. 1, the Assets input is irrelevant and therefore marked with “-”. Finally,
third normal form demands that the input of the tables should be independent
from each other. Thus, it should not be possible to infer the value of an input
condition from the values of the other input conditions.
Conforming to these normal forms increases the readability and understand-
ability of the table. However, they assume that the table already consists of exclu-
sive rules, i.e., conforms to the unique-hit policy. Hence, our approach starts one
step earlier by first disambiguating tables with overlapping rules into tables with
exclusive rules. Subsequently, the normalization rules of [9,10] can be applied
to further improve the tables.
As already described in Sect. 2 the algorithm for table disambiguation is
based on algorithms for detecting overlapping and missing rules as well as sim-
plifying tables by merging similar rules [5,6]. The table simplification algorithm
248 K. Batoulis and M. Weske
6 Conclusion
References
1. Batoulis, K., Haarmann, S., Weske, M.: Various notions of soundness for decision-
aware business processes. In: Mayr, H.C., Guizzardi, G., Ma, H., Pastor, O. (eds.)
ER 2017. LNCS, vol. 10650, pp. 403–418. Springer, Cham (2017). https://doi.org/
10.1007/978-3-319-69904-2 31
2. Batoulis, K., Weske, M.: Soundness of decision-aware business processes. In: Car-
mona, J., Engels, G., Kumar, A. (eds.) BPM 2017. LNBIP, vol. 297, pp. 106–124.
Springer, Cham (2017). https://doi.org/10.1007/978-3-319-65015-9 7
3. Bazhenova, E., Haarmann, S., Ihde, S., Solti, A., Weske, M.: Discovery of fuzzy
DMN decision models from event logs. In: Dubois, E., Pohl, K. (eds.) CAiSE 2017.
LNCS, vol. 10253, pp. 629–647. Springer, Cham (2017). https://doi.org/10.1007/
978-3-319-59536-8 39
4. Bentley, J.L., Ottmann, T.A.: Algorithms for reporting and counting geometric
intersections. IEEE Trans. Comput. 28(9), 643–647 (1979)
5. Calvanese, D., Dumas, M., Laurson, Ü., Maggi, F.M., Montali, M., Teinemaa,
I.: Semantics and analysis of DMN decision tables. In: La Rosa, M., Loos, P.,
Pastor, O. (eds.) BPM 2016. LNCS, vol. 9850, pp. 217–233. Springer, Cham (2016).
https://doi.org/10.1007/978-3-319-45348-4 13
6. Calvanese, D., Dumas, M., Laurson, Ü., Maggi, F.M., Montali, M., Teinemaa, I.:
Semantics, analysis and simplification of dmn decision tables. Information Systems
(2018)
7. Hasić, F., Smedt, J.D., Vanthienen, J.: Augmenting processes with decision intel-
ligence: principles for integrated modelling. Decis. Support Syst. 107, 1–12 (2017)
Disambiguation of DMN Decision Tables 249
1 Introduction
At the 25th International Purchasing and Supply Education and Research Association
Conference 2016 in Dortmund, the research question “How can blockchain technology
improve process flows and transparency between buyers and suppliers?” was explored
in a research framework [1]. One study found that the current use of modern IT tools,
especially in purchasing and related supply chain management, is not very advanced [2].
Obviously, companies are not well prepared for the requirements of networking through
Industry 4.0 [2]. The purpose of this article is to examine how the use of blockchain
technology can affect the existing processes and systems of purchasing raw materials.
We focus on the procurement of raw materials, since we want to incorporate the require‐
ments of the networking raw material process machines in Industry 4.0. In our concept,
we also focus on Enterprise Resource Planning systems (ERPs) and a holistic view of
the procurement process. Thus our research guiding question is:
How can business processes in purchasing be designed more efficiently and effectively using
blockchain technology?
The article is structured as follows. After the introduction, Sect. 2 explains the theoretical
background of blockchain technology as well as purchasing in the supply chain concept.
In addition, the methodological approach of the article as well as related work is
described. We then use interviews in Sect. 3 to analyze the current issues and challenges,
the perceived benefits, and barriers associated with the blockchain technology. To do
2 Theoretical Background
The blockchain technology was already described in a white paper by Satoshi Nakamoto
in 2008 and has been in use since 2009 with the crypto-currency Bitcoin [3]. A block‐
chain is a stringing (concatenation) of data, which are combined into individual blocks
and stored on all users’ computers. This succession of data into blocks results in a
sequence that reflects the course of transactions like a chain. All the data blocks are
protected against subsequent changes by means of cryptographic methods, so that over
time a gapless chain of linked data blocks is created. The inclusion of a new record in
the blockchain requires the passage of a so-called consensus mechanism that runs across
the network of all participants and is used to reach an agreement among all participants
in the blockchain network about the correct state of data on the blockchain. This ensures
that the data is the same on all nodes in the network [4]. The best-known consensus
mechanism is proof-of-work, in which the computer has to execute a complicated math‐
ematical algorithm with great effort. Only after successful execution a new data block
can be generated in the blockchain, which must be checked by the other computers in
the peer-to-peer network before inclusion in the blockchain [5]. In addition to the data,
each block contains a timestamp as well as the hash value of the previous block. The
blocks are protected against subsequent changes by means of cryptographic methods,
so that a continuous chain of linked data blocks is formed over time [5].
To answer our research question, we first conduct a qualitative content analysis of inter‐
views. With qualitative content analysis it is possible to classify words into content
categories [8]. Based on the classification of current problems and challenges as well as
advantages and obstacles, we have developed a case study. This case study addresses
the realities of a business and attempts to solve previously identified problems and chal‐
lenges using blockchain technology. The evaluation is then carried out by interviewing
experts through a standardized questionnaire.
3 Analysis of Interviews
The class process contains both the external view with the cooperation with business
partners in a logistics chain as well as the internal view for the integration of the primary
and secondary activities in a company. A key issue of the current processes was
mentioned by the interviewees in terms of paper-based operations (e.g. shipping docu‐
ments, customs papers, export declarations, warehouses). Due to the high relevance of
the class process both in terms of problems/challenges and the perceived benefits of the
blockchain technology for the processes as well as the class transparency we subdivided
them further into subclasses. According to Becker and Kahn [15], a process is “the
content-related, temporal and logical sequence of activities that are necessary for
processing a business-relevant object.” The subclasses for the class process and for the
class transparency are based on the empirical conceptual approach Nickerson et al. [14]
have developed. In the class process, the subclass time plays an important role according
to the evaluation of the interviews. The perceived benefits of using blockchain tech‐
nology focus on timely processing, exemplary comments are “… to have access to the
right information at the right time” [16]. Furthermore, an improvement in the quality of
the processes is expected, such as e.g. “…better tracking of orders, reducing errors and
better fraud detection” [17]. The avoidance of fraud by transaction processing in real
time and the immutability of the data on the blockchain is another supposed advantage
as well as the potential of automated process processing by smart contracts on the
blockchain. Finally, the interviewees emphasized the importance of security in the
processes by “… logistics industry wants to see improved connectivity, efficiency and
security thanks to blockchain” [18]. The class transparency contains an external and an
internal view. The external view of transparency refers to the exchange of information
with business partners within a logistical process in a logistical chain [19]. The internal
Using Blockchain Technology for Business Processes in Purchasing − Concept 257
view in this context means both an insight into the own order status of a logistical process
as well as an insight into the entire logistical chain with the possibility of backward and
forward traceability of an order. Blockchain offers interviewees the advantage of being
able to provide proof of possession as well as proof of transport [20]. Closely associated
with this is the benefit of status tracking of the status of the flow of goods, highlighted
by e.g. “Logistics service providers, for example, can document all incidents along the
supply chain completely, unchangeable and visible to everyone” [21]. At the same time,
documentation of a history of the logistical process, highlighted e.g. “With the distrib‐
uted database, network participants can directly engage in transactions and see the
history of all transactions” [22]. Parnell [32] emphasizes the importance of real-time
processing with “real-time sharing of information about process improvements and
maintenance”. The immutability of the data in a blockchain is emphasized by “The
information captured in each transaction is agreed upon by all members of the business
network; once there is a consensus, it becomes a permanent record that cannot be
changed” [32]. In the class obstacles the interviewees see a significant aspect in the
difficulty of adapting a blockchain-based application. Exemplary comments on this are
“Customer engagement is about creating information systems that are truly accepted.
This is less a matter of technology and more a matter of having an approach that inspires
all parties involved, not just customers to work in incremental steps with tangible results
towards a solution.” [16]. Closely related to the difficulty of adaptation is the lack of
confidence that exemplifies “… and adopting a new mindset around a decentralized
network with no central control” [32].
4 Case Study
To answer the research question, we conduct a case study that, according to Ridder [24],
has the advantage of a more detailed description and more detailed analysis. On this
basis, the questions about the “how” and “why” can be answered more easily [24]. Our
case study is suitable in our research subject because a current phenomenon (blockchain
in purchasing) in a real and practical context (the company Schmitz Cargobull AG) is
examined [25]. According to Brüsemeister [26], a case-by-case study is also useful if it
provides access to a hitherto little-explored social area, which points to the use of block‐
chain technology in the integration of “machine-to-ERP-to-blockchain-to-ERP-to-
machine” applications.
The company in our case study is Europe’s leading manufacturer of semi-trailers
and trailers for temperature-controlled freight, general cargo and bulk goods with an
annual production of around 58,000 vehicles and around 5,700 employees [27]. Within
the value chain of Schmitz Cargobull AG, the procurement of raw materials, primary
products, consumables and tools occupies a significant position. A logistics manager
explained in one of our interviews that one goal of supply chain management is the
optimization of transparency about capacities in the production and procurement
network [23]. Since 2002, the entire order processing has been carried out via the AXIT
logistics platform AX4. Through the connection of the own SAP system as well as all
suppliers the data exchange in a procurement process can be automated [28]. In the
258 S. Tönnissen and F. Teuteberg
notification of the manufacturing process to the customer via a status message. The
fourth step prepares the delivery to the customer and writes the data of the goods delivery
to the blockchain, which are taken over by the customer directly into the ERPs and form
the basis for the posting of the physical goods receipt. This goods receipt at the customer
finds in the fifth step and end with the corresponding status message to the supplier
(Fig. 1).
Step 1: (1) Demand request, (2) Purchase requisition, (3) Data transfer ERP, (4)
Demand request, (5) Activity indicator. Step 2: (6) offer, (7) data submission offers
to blockchain (8) takeover offer, (9) offer review, (10) determine favorable offer,
(11) order on the blockchain, (12) smart contract checks order, (13) order confirma‐
tion, (14) smart contract checks order confirmation, (15) take over status in ERP.
Step 3: (16) scheduling, production order, (17) appointment on blockchain, (18)
smart contract recognizes appointment, (19) take over appointment, (20) finish, (21)
write status in blockchain, (22) smart contract recognizes status and takes over in
ERP, (23) takeover status. Step 4: (24) Determine delivery data, (25) Write delivery
data in blockchain, (26) smart contract recognizes delivery data and takes over in
ERP, (27) Transfer of goods receipt data, (28) Physical goods receipt. Step 5: (29)
Posting Goods Receipt, (30) blockchain Indicator, (31) smart contract recognizes
status Goods Receipt, (32) Transfer to ERP, and Basis for Invoicing.
The blockchain will continue to be used as middleware for exchanging status-related
information between the customer’s ERPs and suppliers. Thus, the transparency in the
process of procurement of raw materials for both the buyer and the supplier is signifi‐
cantly increased. The status-relevant information is written to the blockchain in real time
and is available in a timely manner (Tables 2 and 3).
260 S. Tönnissen and F. Teuteberg
70% of Industry 4.0 strategies are aimed at increasing productivity and efficiency [2].
To improve efficiency in our concept, we set up the subsequent smart contracts on the
blockchain in the processes.
The result of our case study shows a process that is characterized by the consideration
of existing ERPs both at the manufacturer and the supplier. The blockchain is switched
as a middleware between these two ERPs and thus provides the participants with a
consistent and transparent database with the guarantee, the immutability and traceability
of the data of the logistical processes in purchasing. With the blockchain as a common
database, media breaks will be avoided in the future. The connection of the existing
Using Blockchain Technology for Business Processes in Purchasing − Concept 261
ERPs is done by blockchain gateways, which are set up on the part of the operators of
the ERPs. Thus, the sovereignty over the data import as well as data export remains in
the hands of the operator. With this conception, the adaptation of the blockchain into
the operational processes could succeed and the hurdle of the initially low confidence
could be overcome. The connection of further suppliers could succeed on this basis
without major obstacles, and thus lay the basis for further acceptance of the blockchain.
Our case study shows significant improvements in the process, since the number of
interactions by a person could be significantly reduced from the needs assessment
through a networked machine at the manufacturer to the receipt of the required raw
materials. Thus, both the time required for the processing of the processes can be reduced
and the process reliability and quality can be increased. The use of smart contracts
contributes significantly to the automation of process steps and thus ensures greater
security in process processing. All companies involved in the process receive transpar‐
ency about the overall process via the status reports in real-time on the blockchain, and
can purposefully control their material supply on the basis of this.
6 Expert Interviews
For the evaluation of our concept, we consulted experts from the operational practice
and asked to fill out a standardized questionnaire (questionnaire and profile of the
participants see also https://tinyurl.com/ybbuw7fd). The experts participated in a work‐
shop on blockchain technology on November 11th 2017. We invited 11 participants for
the survey. The standardized questionnaire was made available to the participants on
November 12th 2017. The respondents gave the following answers to the questions. For
the question “How realistic do you think the process model with ERP and blockchain
is?”, 60% of respondents think it is a realistic use case. 30% of respondents are rather
critical or uncertain about the assessment. The realistic estimate previously given with
60% can be found in comments such as “Technical feasibility I consider realistic” or
“This process happens in all companies in the world every day”. In addition, there are
critical voices such as “However, there will be distrust in the technique that probably
outweighs” or “The system also depends on the number of companies that work with it.
I cannot imagine a widespread distribution in the next 10 years”. The question “What
is your confidence in the blockchain technology?” was answered by 50% for rating 3.
In our odd rating scale, rating 3 represents a neutral middle category. The high number
of responses could therefore suggest an uncertain judgment of the participants, since the
middle scale point can also be used as a flight category [33]. The comments of the
participants confirm the suspicion of the uncertain judgment. Exemplary comments are
“The confidence is neither particularly high nor particularly deep, because I could not
gather any experience with it” or “Due to my low level of knowledge about this tech‐
nology, I find it difficult to make a statement about a manipulation of the individual
steps”. Answering the question “Do you think that blockchain technology has reached
a necessary maturity level for a” real “mission?” led to a balanced result as 50% of
respondents answered yes and 50% answered no. The proponents commented on their
decisions, among others with “I think the technology can already be used in a limited
262 S. Tönnissen and F. Teuteberg
field of application for special applications” or “Today almost all production facilities
are equipped with the appropriate interfaces to introduce these systems nationwide”.
The interlocutors justify their opinion with “I think this step to the introduction is
currently too big. It would have first created general trust in the technology” and
“because I cannot assess how far the degree of maturity has progressed”. In our stand‐
ardized questionnaire, we asked the participants for an assessment as to whether an
effective purchase agreement pursuant to contract law had been concluded. The result
was 80%, no. The reasons for this assessment are exemplary “The contract law is in my
opinion not for machines or IT systems”, “Because the behavior is too digitalized” or
“offer and acceptance cannot be decided by a digital form”. The participants, who claim
that an effective sales contract was made with 20%, justified this with “if on eBay the
automatic bidder system was activated by the buyer, then also a sales contract came
about” or “one could assume that a purchase contract has been concluded by offer and
acceptance”.
We can summarize that the processes of purchasing using blockchain technology have
a potential that can solve today’s problems and challenges. The blockchain technology
has the potential to meet the expectations of the interviews with regard to improving
processes and increasing transparency. By using the smart contracts, numerous process
steps could be automated and efficiency gains achieved through real-time processing of
blockchain. By integrating networked machines, the integration of Industry 4.0 concepts
succeeds. The validation of our concept by a survey of experts from different companies
has shown that the predicted use case with the integration of a blockchain in the existing
ERPs of customers and suppliers is realistic. However, interviewees also found barriers
to successful adaptation of such a solution. It is unclear how adaptation could take place
via a global logistics chain. The trust required between the business partners was inter‐
preted by the interviewees as a clear barrier to integration.
The case study is unrepresentative due to the company’s choice, but demonstrates a
strong ability to collaborate with business partners based on a high prevalence of ERPs
in manufacturing companies. In addition, existing IT concepts for connecting ERPs
across company boundaries could also be able to solve the challenges of purchasing.
Blockchain technology competes with electronic data interchange systems such as EDI,
EDIFACT, AXIT, etc. This entry could help to discuss whether existing IT systems for
enterprise networking or business-to-business data exchange are inferior to blockchain
technology. Here further investigations regarding a cost, benefit and risk assessment are
necessary.
The dissemination of networking of machines with ERPs and subsequently with
business partners is facilitated by standardizing both the information technologies
required for this purpose, e.g. gateways for the connection to the blockchain as well as
data structures and the document structures relevant for a global supply chain process
favors. These requirements for global standardization cannot be met to this day [11].
This makes it difficult to adapt a corresponding blockchain-based application to existing
Using Blockchain Technology for Business Processes in Purchasing − Concept 263
References
1. Foerstl, K., Schleper, M.C., Henke, M.: Purchasing and supply management: from efficiency
to effectiveness in an integrated supply chain. J. Purchasing Supply Manag. 23(2017), 223–
228 (2017)
2. The purchasing and logistics working group of the Schmalenbach-Gesellschaft für
Betriebswirtschaft e.V. (2017): Digitalisierung und Vernetzung in Einkauf und Supply Chain
Management. In: Krause, S., Pellens, B. (Hrsg.): Betriebswirtschaftliche Implikationen der
digitalen Transformation. ZfbF-Sonderheft 72/17
3. Yli-Huumo, J., Ko, D., Choi, S., Park, S., Smolander, K.: Where is current research on
blockchain technology? - A systematic review. PLoS ONE 11(10), e0163477 (2016)
4. Swan, M.: Blockchain, Blueprint for a New Economy. O’Reilly, Farnham (2015)
5. Holotiuk, F., Pisani, F.; Moormann, J.: The impact of blockchain technology on business
models in the payments industry. In: Leimeister, J.M., Brenner, W. (Hrsg.): Proceedings der
13. Internationalen Tagung Wirtschaftsinformatik (WI 2017), St. Gallen (2017)
6. Bungard, P.: CSR und Geschäftsmodelle, Auf dem Weg zum zeitgemäßen Wirtschaften.
Springer, Heidelberg (2018). https://doi.org/10.1007/978-3-662-52882-2
7. Porter, M.: Competitive Advantage: Creating and Sustaining Superior Performance. The Free
Press, New York (1985)
8. Elo, S., Kyngäs, H.: The qualitative content analysis process. J. Adv. Nurs. 62(1), 107–115
(2007)
9. Kshetri, N.: Blockchain’s roles in meeting key supply chain management objectives. Int. J.
Inf. Manag. 39(2018), 80–89 (2018)
10. Biswas, K., Muthukkumarasamy,V., Tan, W.L.: Blockchain based wine supply chain
traceability system. In: Future Technologies Conference (FTC) (2017)
11. Korpela, K., Hallikas, J., Dahlberg, T.: Digital supply chain transformation toward blockchain
integration. In: Proceedings of the 50th HICSS (2017)
12. Hackius, N., Petersen, M.: Blockchain in logistics and supply chain: trick or treat? In:
Proceedings of the Hamburg International Conference of Logistics (HICL) (2017)
13. Google Inc.: How search works. http://www.google.com/intl/ALL/search/howsearchworks/.
Accessed 02 Feb 2018
14. Nickerson, R.C., Varshney, U., Muntermann, J.: A method for taxonomy development and
its application in information systems. Eur. J. Inf. Syst. 22(3), 336–359 (2013)
15. Becker, J., Kahn, D.: Der Prozess im Fokus. In: Becker, J., Kugeler, M., Rosemann, M. (Hrsg.)
(2012): Prozessmanagement. Ein Leitfaden zur prozessorientierten Organisationsgestaltung,
7, Auflage. SpringerGabler, Heidelberg (2012). 10.1007/978-3-642-33844-1_1
16. logifint (2017): Blockchain & logistics: an interview with Louis De Bruin. http://
logisticsandfintech.com/blockchain-logistics-interview-louis-de-bruin. Accessed 30 Jan
2018
264 S. Tönnissen and F. Teuteberg
17. Burns, J.: Dnata and its partners test the use of blockchain technology.
www.aircargoweek.com/dnata-partners-test-blockchain-technology/. Accessed 31 Jan 2018
18. Henderson, J.: Blockchain technology “set to revolutionise” logistics industry (2017).
www.supplychaindigital.com/technology/blockchain-technology-set-revolutionise-logistics-
industry. Accessed 31 Jan 2018
19. Lamming, R.C., Caldwell, N.D., Harrison, D.A., Phillips, W.: Transparency in supply
relationships: concept and practice. J. Supply Chain Manag. 37(4), 4–10 (2001)
20. Nördinger, S.: Darum sollten sich Industrieunternehmen mit Blockchain beschäftigen (2017).
https://www.produktion.de/specials/revolution-blockchain/darum-sollten-sich-
industrieunternehmen-mit-blockchain-beschaeftigen-264.html. Accessed 30 Jan 2018
21. Gläser, T.: Logbuch ohne zentrale Autorität (2017). https://www.it-zoom.de/dv-dialog/e/
logbuch-ohne-zentrale-autoritaet-18512/. Accessed 30 Jan 2018
22. Poll, D.: Logistik: Vorteile durch Blockchain in der Supply Chain (2017). https://
www.produktion.de/technik/logistik/logistik-vorteile-durch-blockchain-in-der-supply-
chain-115.html. Accessed 02 Feb 2018
23. Schonefeld, P.: Planung im Kreislauf (2013). https://beschaffung-aktuell.industrie.de/supply-
chain-management/planung-im-kreislauf/. Accessed 02 Feb 2018
24. Ridder, H.-G.: The theory contribution of case study research designs. Bus. Res. 10(2), 281–
305 (2017)
25. Yin, R.: Case Study Research: Design and Methods. Sage, Thousand Oaks (2002)
26. Brüsemeister, T.: Qualitative Forschung. Ein Überblick, Wiesbaden (2008)
27. Schmitz Cargobull AG: Geschäftsjahr 2016/2017. 10.000 Fahrzeuge mehr produziert als
geplant. Marktführer Schmitz Cargobull steigert Umsatz auf mehr als zwei Milliarden Euro
(2017). https://www.cargobull.com/de/Detail_news-523_213_383.html. Accessed 02 Feb
2018
28. AXIT: Schmitz Cargobull Success Story (2016). https://www.axit.de/images/successstories
PDFs/AX4_Success_Schmitz-Cargobull_d.pdf. Accessed 02 Feb 2018
29. Nördinger, S.: Darum passen Blockchain und Industrie 4.0 zusammen (2017). https://
www.produktion.de/specials/revolution-blockchain/darum-passen-blockchain-und-
industrie-4-0-zusammen-278.html. Accessed 30 Jan 2018
30. Raskin, M.: The law and the legality of smart contracts. Geo. L. Tech. Rev. 305 (2017) https://
papers.ssrn.com/sol3/papers.cfm?abstract_id=2959166. Accessed 30 Aug 2017
31. Scholz-Reiter, B., Jakobza, J.: Supply Chain Management – Überblick und Konzeption. HMD
207/1999 (1999)
32. Parnell, G.: How blockchain can strengthen supply chain links (2017). https://theloadstar.
co.uk/guest-blog-blockchain-can-strengthen-supply-chain-links. Accessed 30 Jan 2018
33. Duller, C.: Einführung in die Statistik mit Excel und SPSS. Ein anwendungsorientiertes Lehr-
und Arbeitsbuch. 3. Auflage. Springer Gabler (2013). https://doi.org/10.1007/
978-3-642-37859-1
Developing a Multiple-Objective Demand
Response Algorithm for the Residential
Context
Abstract. Energy grids are facing various challenges, such as new appliances
and volatile generation. As grid reliability and cost benefits are endangered,
managing appliances becomes increasingly important. Demand Response
(DR) is one possibility to contribute to this task by shifting and managing
electrical loads. DR can address multiple objectives. However, current research
lacks of algorithms addressing these objectives sufficiently. Thus, we aim to
develop a DR algorithm that considers multiple DR objectives. For evaluation,
we implemented the algorithm and formulated demonstration cases for a sim-
ulation. The evaluated algorithm contributes for example to users and energy
providers by realizing various benefits.
1 Introduction
With this algorithm several stakeholders can benefit such as users as they can reach
several goals (e.g., cost minimization and welfare maximization), energy providers as
they can give more incentives to users and achieve a more efficient generation or
predictability, energy grids can get more stable (e.g., ensuring reliability and market
performance) and flexible on contingencies, appliances (e.g., EVs) and infrastructure
(e.g., storages and PVs) will be embedded in a more fertile way.
We already did first steps in designing such an algorithm in a former study [13].
However, further research is needed, to implement and evaluate it. Therefore, our paper
is structured as follows: After describing the optimization problem (Sect. 2) the
research process, adopted from the “algorithm engineering” (AE) approach, is
described (Sect. 3). Afterwards, the algorithm of our former study is described [13]
(Sect. 4). According to the AE approach, we implement the algorithm (Sect. 5), which
means the instantiation of individual parameters within the algorithm. Afterwards, we
derive an experimental setting and simulated four test cases with input data from the
DR field (Sect. 6). We finish with a conclusion, discuss limitations and give an outlook
for further research (Sect. 7).
solution could be found. Addressed goals are heterogeneous, such as cost reduction
(e.g., [18–20]) or peak load reduction (e.g., [21–23]). Some algorithms try to achieve
two goals simultaneously (e.g., reducing costs while respecting users’ comfort, e.g.,
[24–27]). These goals are often realized based on an externally given cost function of
an energy supplier and the comfort level, that is considered as an additional condition
to the objective function. [28] consider multiple goals, however, these are not suitable
for our research goal, as multiple and difficult to solve optimization functions are used.
Algorithms in particular try to combine goals, for example, to minimize costs and
maximize the comfort level of users (e.g., [12, 25, 27]). Batchu and Pindoriya [29]
identified four indicators, addressing major DR objectives: lower energy consumption,
peak load reduction, load profile flattening and cost reductions. The optimization
problem is constructed with multiple objective functions, optimizing towards load
profile flattening, cost reduction, comfort maximization and peak load reduction.
Multiple objective functions mostly return in a complex problem and solving is diffi-
cult. The Berkeley Lab ([28]) addressed multiple goals, however, the resulting opti-
mization problem will get very complex and challenging to solve. They therefore
propose a Greedy algorithm to find a solution.
gi ðxÞ ¼ 0 i ¼ 1; . . .; m ð2Þ
hj ðxÞ\0 j ¼ 1; . . .; n ð3Þ
one T (7) through shifting the whole Tl by an appropriate constant mn;a , i.e., hk ¼
k þ mn;a with 0 mn;a x dn;a .
Furthermore, let ch;min
n;a be the min and ch;max
n;a be the max borders for a load xhn;a with
h 2 T; a 2 An so we can specify, in which borders the intensity of load a can be shifted,
n;a xn;a cn;a
i.e., ch;min h h;max
(8). We note that the given load profiles have to satisfy the
inequality cn;a ln;a ch;max
h;min hk
n;a for all k 2 Tl to get a feasible solution.
Let an;a be the starting and bn;a be the ending time slot for an appliance a. with the
restricted time interval T ½an;a ; bn;a (9). Note that the interval length must be at least the
length of the load profile dn;a . to get a feasible solution, i.e., bn;a an;a dn;a .
In order to turn on and off the constraint i for each appliance a individually (e.g., an
EV has other constraints than a washing machine) let cia 2 f0; 1g be a binary variable
that shows if a constraint is turned on ðcia ¼ 1Þ or not ðcia ¼ 0Þ for the appliance a.
The objective function describes the total cost (4), while the cost in a time slot h is a
function depending on h and the total load xh i.e., ch ¼ cðh; xh Þ xh .
Xx Xx
minxhn;a c ¼ h¼0
ch ¼ h¼0
c h; xh xh ð4Þ
Xx
h¼0
xhn;a ¼ ln;a 8 n 2 N; a 2 An ð5Þ
Xb
xhn;a ln;a c3a ¼ 0 8 n 2 N; a 2 An ð8Þ
n;a
h¼an;a
8 h 2 T; n 2 N; a 2 An ;
xhn;a 0; cia 2 f0; 1g ð9Þ
i ¼ 1; 2; 3
3 Research Methodology
DSR should focus the real world problem, respectively the application area (e.g., [36]),
we assume the AE approach as more suitable (cf., [37]).
The AE approach consists of five different (main)steps (see Fig. 1): modelling the
application area as basis for the design phase, followed by the analysis of the designed
algorithm, implementation of the algorithm and finally an experiment. All of these steps
have a strong relationship to the application, receiving application depending infor-
mation (e.g., input data or the mathematical model) or contributing to the application
(e.g., delivering a library to use or giving a performance guarantee).
The optimization model in Sect. 2.1 only aims at cost reduction. Hence, we need to
identify suitable DR objectives to design an optimization model. [29] identified several
indicators that need to be integrated in the optimization model, as these indicators
match with most of the DR objectives. These additional objectives need to be included
in the optimization function (e.g., done by [29] in a generalized form for a single
home). The authors use multiple optimization functions, which results in an even more
complex problem and the algorithm gets inefficient, for example, regarding the cal-
culation time [32]. We therefore need less objective functions.
One approach within the stated model could be ‘penalty costs’. The basic idea is
that all objectives influence a cost factor, which then, based on a kWh price, gives the
costs in a certain timestamp for placing the load. The objective function is formulated
270 D. Behrens et al.
P
in the following way, over the optimization horizon: minxhn;a c ¼ xh¼0 cðhÞ xh , where
h is the vector of additionally needed information to calculate the penalty costs [13].
First of all, the timeslot (time-of-use) is needed to consider time dependent costs
(e.g., ToUP). In order to achieve a flattening of the load profile, the fluctuation needs to
be minimized. Therefore, we implement rising penalty costs for deviations from the
arithmetic mean of the last X timestamps. This means, if the placement of a certain load
in timeslot Y rises or lowers the consumption in that timeslot more than a certain
percentage compared to the arithmetic mean of the last X timeslots, the cost factor is
raised and penalty costs are added. The same procedure is chosen to integrate a more
effective peak load reduction. If the placement in timeslot Y will create a new peak, the
cost factor should be raised by a certain percentage.
At this point, two upcoming questions must be answered: how can the additional
information be gathered (see following section), and how do the penalty costs look like.
The penalty costs need certain parameters, which need to be instantiated (filled with
concrete values). For example, how much are the penalty costs increasing when cre-
ating a new peak? However, because the selection of the concrete values does not
influence the performance of the algorithm according its runtime, we do this in the
following analyzation phase. The objective function now looks the following:
Xx Xx
minxhn;a c ¼ h¼0
ch ¼ h¼0
c ð hÞ xh ð10Þ
Both the function cðhÞ and the vector h still need to be defined. h needs to involve
information about the timestamp h, the load in this timestamp xh , the arithmetic mean
of the last X timestamps MEANðh XÞ and the actual peak load PEAK. Moreover,
parameters for (a) the cost function (a ToUP with two different intervals is chosen (low
price (LP) and high price (HP))) with additional information, such as costs in different
intervals and interval lengths, (b) increasing cost rate q for r% of deviation from the
mean, and (c) increasing cost rate s for a new peak are needed.
cðhÞ ¼ c h; xh ; MEANðh 8Þ; PEAK
c ¼ LP : 0 h a; b h x
¼ c1 !
c ¼ HP : a\h\b
8
>
> c ¼ c q : 0 r \1 r
>
< c ¼ c q : 1 r xh MEAN ðh X Þ
\2 r ð11Þ
þ c2 !
>
> c ¼ c q : 2 r x h \3 r
>
:
... ...
c ¼ c s : x ¼ PEAK
h
þ c3 !
c ¼ c : else
We now need to investigate, if the algorithm terminates and how the runtime is.
Both depend on the chosen solution strategy. However, if a suitable heuristic (e.g.,
Greedy heuristic) is chosen, we can guarantee termination. Calculating the runtime, n
Developing a Multiple-Objective Demand Response Algorithm 271
5 Implementation
6 Experiment
6.1 Input Data
Conducting the experiment according the stated AE approach needs suitable input data
from the application area. Recorded data from naturalistic living units (analyzed for
example in [41, 42]) or artificial data, which has a sufficient quality ([43, 44]) seems
both possible. In the naturalistic data case, we must add additional information, for
example, the usage time of appliances. For generating artificial data, the LoadPro-
fileGenerator (LPG) [45] can be used that has predefined load profiles and uses a
behavior model for each user in a living unit to simulate the data.
As artificial data has a good quality compared to recorded data [43], we use the
LPG. We choose a household with a family (two children, man at work) and a washing
machine, a dryer and a dishwasher as deferrable appliances. The household has a ToUP
with 21 cent per kWh during off-peak hours (0–5 and 22–24 h) and 39 cents in peak
hours (5 to 22 h). As instantiations for our algorithm, we decided to choose a couple of
variations, which can be seen in the Table 2. To enable a good comparison of our
algorithm, we implemented further algorithms, aiming at different objectives: the first
only aims at cost minimization that is achieved by the ToUP (here called ToU). The
second just aims to realize a load profile flattening. It uses a notional convex cost
function (as used for example in [46], here called ECF).
References
1. Lawrence, T.M., Watson, R.T., Boudreau, M.-C., Mohammadpour, J.: Data flow require-
ments for integrating smart buildings and a smart grid through model predictive control.
Procedia Eng. 180, 1402–1412 (2017)
2. Seidel, S., Recker, J., vom Brocke, J.: Sensemaking and sustainable practicing: functional
affordances of information systems in green transformations. Manag. Inf. Syst. Q. 37, 1275–
1299 (2013)
3. Siano, P.: Demand response and smart grids—a survey. Renew. Sustain. Energy Rev. 30,
461–478 (2014)
4. Hu, Q., Li, F.: Hardware design of smart home energy management system with dynamic
price response. IEEE Trans. Smart Grid. 4, 1878–1887 (2013)
5. Jovanovic, R., Bousselham, A., Bayram, I.S.: Residential demand response scheduling with
consideration of consumer preferences. Appl. Sci. 6, (2016)
6. Koolen, D., Sadat-Razavi, N., Ketter, W.: Machine learning for identifying demand patterns
of home energy management systems with dynamic electricity pricing. Appl. Sci. 7, 1160
(2017)
7. Kosek, A.M., Costanzo, G.T., Bindner, H.W., Gehrke, O.: An overview of demand side
management control schemes for buildings in smart grids. In: 2013 IEEE International
Conference on Smart Energy Grid Engineering. SEGE (2013)
8. Merkert, L., Harjunkoski, I., Isaksson, A., Säynevirta, S., Saarela, A., Sand, G.: Scheduling
and energy – industrial challenges and opportunities. Comput. Chem. Eng. 72, 183–198
(2015)
9. Steen, D., Le, T., Bertling, L.: Price-based demand-side management for reducing peak
demand in electrical distribution systems – with examples from gothenburg. Chalmers
Publication Library (CPL) (2012)
10. Hillemacher, L.: Lastmanagement mittels dynamischer Strompreissignale bei Haushaltskun-
den. (2014)
11. Vidal, A.R.S., Jacobs, L.A.A., Batista, L.S.: An evolutionary approach for the demand side
management optimization in smart grid. In: 2014 IEEE Symposium on Computational
Intelligence Applications in Smart Grid (CIASG), pp. 1–7 (2014)
12. Salinas, S., Li, M., Li, P.: Multi-objective optimal energy consumption scheduling in smart
grids. IEEE Trans. Smart Grid. 4, 341–348 (2013)
13. Behrens, D., Schoormann, T., Knackstedt, R.: Developing an algorithm to consider multiple
demand response objectives. Eng. Technol. Appl. Sci. Res. (2018)
14. Al-Sumaiti, A.S., Ahmed, M.H., Salama, M.M.A.: Smart home activities: a literature review.
Electr. Power Compon. Syst. 42, 294–305 (2014)
15. Balijepalli, V.S.K.M., Pradhan, V., Khaparde, S.A., Shereef, R.M.: Review of demand
response under smart grid paradigm. In: ISGT2011-India, pp. 236–243 (2011)
16. Gerwig, C., Behrens, D., Lessing, H., Knackstedt, R.: Demand side management in
residential contexts - a literature review. In: Lecture Notes in Informatics, pp. 93–107 (2015)
17. Ketter, W., Collins, J., Reddy, P.P., Flath, C.M.: The Power Trading Agent Competition.
Social Science Research Network, Rochester (2011)
18. Ali, S.Q., Maqbool, S.D., Ahamed, T.P.I., Malik, N.H.: Pursuit algorithm for optimized load
scheduling. In: 2012 IEEE International Power Engineering and Optimization Conference
Melaka, Malaysia, pp. 193–198 (2012)
19. Keerthisinghe, C., Verbič, G., Chapman, A.C.: Evaluation of a multi-stage stochastic
optimisation framework for energy management of residential PV-storage systems. In: 2014
Australasian Universities Power Engineering Conference (AUPEC), pp. 1–6 (2014)
276 D. Behrens et al.
20. Zhao, W., Cooper, P., Perez, P., Ding, L.: Cost-driven residential energy management for
adaption of smart grid and local power generation (2014)
21. Huang, Y., Mao, S., Nelms, R.M.: Smooth electric power scheduling in power distribution
networks. In: 2012 IEEE Globecom Workshops, pp. 1469–1473 (2012)
22. McNamara, P., McLoone, S.: Hierarchical demand response for peak minimization using
dantzig–wolfe decomposition. IEEE Trans. Smart Grid. 6, 2807–2815 (2015)
23. Verschae, R., Kawashima, H., Kato, T., Matsuyama, T.: A distributed coordination
framework for on-line scheduling and power demand balancing of households communities.
In: 2014 European Control Conference (ECC), pp. 1655–1662 (2014)
24. De Angelis, F., Boaro, M., Fuselli, D., Squartini, S., Piazza, F., Wei, Q., Wang, D.: Optimal
task and energy scheduling in dynamic residential scenarios. In: Wang, J., Yen, G.G.,
Polycarpou, M.M. (eds.) ISNN 2012. LNCS, vol. 7367, pp. 650–658. Springer, Heidelberg
(2012). https://doi.org/10.1007/978-3-642-31346-2_73
25. Bassamzadeh, N., Ghanem, R., Kazemitabar, S.J.: Robust scheduling of smart appliances
with uncertain electricity prices in a heterogeneous population. Entergy Build. 84 (2014)
26. Kim, S.-J., Giannakis, G.B.: Efficient and scalable demand response for the smart power
grid. In: 2011 4th IEEE International Workshop on Computational Advances in Multi-
Sensor Adaptive Processing (CAMSAP), pp. 109–112 (2011)
27. Song, L., Xiao, Y., van der Schaar, M.: Non-stationary demand side management method for
smart grids. In: 2014 IEEE International Conference on Acoustics, Speech and Signal
Processing (ICASSP), pp. 7759–7763 (2014)
28. Berkeley Lab: Distributed Energy Resources Customer Adoption Model (DER-CAM) |
Building Microgrid. https://building-microgrid.lbl.gov/projects/der-cam
29. Batchu, R., Pindoriya, Naran M.: Residential demand response algorithms: state-of-the-art,
key issues and challenges. In: Pillai, P., Hu, Y.F., Otung, I., Giambene, G. (eds.) WiSATS
2015. LNICST, vol. 154, pp. 18–32. Springer, Cham (2015). https://doi.org/10.1007/978-3-
319-25479-1_2
30. Cortés-Arcos, T., Bernal-Agustín, J.L., Dufo-López, R., Lujano-Rojas, J.M., Contreras, J.:
Multi-objective demand response to real-time prices (RTP) using a task scheduling
methodology. Energy 138, 19–31 (2017)
31. Behrens, D., Schoormann, T., Knackstedt, R.: Towards a taxonomy of constraints in
demand-side-management-methods for a residential context. In: Abramowicz, W. (ed.) BIS
2017. LNBIP, vol. 288, pp. 283–295. Springer, Cham (2017). https://doi.org/10.1007/978-3-
319-59336-4_20
32. Behrens, D., Ruether, C., Schoormann, T., Ambrosi, K., Knackstedt, R.: Effects of
constraints in resdential demand-side-management algorithms - a simulation-based study. In:
International Conference on Operation Research. Springer, Berlin (2017)
33. Hevner, A.R., March, S.T., Park, J., Ram, S.: Design science in information systems
research. Manag. Inf. Syst. Q. 28, 75–105 (2004)
34. Sanders, P.: Algorithm engineering – an attempt at a definition. In: Albers, S., Alt, H., Näher,
S. (eds.) Efficient Algorithms. LNCS, vol. 5760, pp. 321–340. Springer, Heidelberg (2009).
https://doi.org/10.1007/978-3-642-03456-5_22
35. Ketter, W., Peters, M., Collins, J., Gupta, A.: competitive benchmarking: an is research
approach to address wicked problems with big data and analytics. Manag. Inf. Syst. Q. 40,
1057–1080 (2016)
36. Füller, K., Ramanath, R., Böhm, M., Krcmar, H.: Decision support for the selection of
appropriate customer integration methods. Wirtsch. Proc. 2015 (2015)
37. Nieße, A., Tröschel, M., Sonnenschein, M.: Designing dependable and sustainable smart
grids – how to apply algorithm engineering to distributed control in power systems. Environ.
Model Softw. 56, 37–51 (2014)
Developing a Multiple-Objective Demand Response Algorithm 277
38. Mohsenian-Rad, A.-H., Wong, V.W., Jatskevich, J., Schober, R., Leon-Garcia, A.:
Autonomous demand-side management based on game-theoretic energy consumption
scheduling for the future smart grid. IEEE Trans. Smart Grid. 1, 320–331 (2010)
39. Nelder, J.A., Mead, R.: A simplex method for function minimization. Comput. J. 7, 308–313
(1965)
40. Land, A.H., Doig, A.G.: An automatic method of solving discrete programming problems.
Econometrica 28, 497–520 (1960)
41. Monacchi, A., Egarter, D., Elmenreich, W., D’Alessandro, S., Tonello, A.M.: GREEND: an
energy consumption dataset of households in Italy and Austria. ArXiv14053100 (2014)
42. Behrens, D., Schoormann, T., Knackstedt, R.: Datensets für Demand-Side-Management –
Literatur-Review-Basierte Analyse und Forschungsagenda. In: Mayr, H.C., Pinzger, M.
(eds.) Lecture Notes in Informatics (2016)
43. Cao, H.Â., Beckel, C., Staake, T.: Are domestic load profiles stable over time? An attempt to
identify target households for demand side management campaigns. In: IECON 2013 - 39th
Annual Conference of the IEEE Industrial Electronics Society, pp. 4733–4738 (2013)
44. Hoogsteen, G., Molderink, A., Hurink, J.L., Smit, G.J.M.: Generation of flexible domestic
load profiles to evaluate demand side management approaches. In: 2016 IEEE International
Energy Conference (ENERGYCON), pp. 1–6 (2016)
45. Noah Pflugradt: LoadProfileGenerator. http://www.loadprofilegenerator.de/
46. Behrens, D., Gerwig, C., Knackstedt, R., Lessing, H.: Selbstregulierende Verbraucher im
Smart Grid: Design einere Infrastruktur mit Hilfe eines Multi-Agenten-Systems. In:
Proceedings of the Multikonferenz Wirtschaftsinformatik 2014 (2014)
47. Miao, H., Huang, X., Chen, G.: A genetic evolutionary task scheduling method for energy
efficiency in smart homes. Int. Rev. Electr. Eng. 7, 5897–5904 (2012)
48. Soliman, H.M., Leon-Garcia, A.: Game-theoretic demand-side management with storage
devices for the future smart grid. IEEE Trans. Smart Grid. 5, 1475–1485 (2014)
49. Alam, M.R., St-Hilaire, M., Kunz, T.: Cost optimization via rescheduling in smart grids—a
linear programming approach. In: 2013 IEEE International Conference on Smart Energy
Grid Engineering (SEGE), pp. 1–6 (2013)
50. Maqbool, S.D., Ahamed, T.P.I., Ali, S.Q., Pazheri, F.R., Malik, N.H.: Comparison of pursuit
and e-greedy algorithm for load scheduling under real time pricing. In: 2012 IEEE
International Conference on Power and Energy (PECon), pp. 515–519 (2012)
51. Bergstra, J.S., Bardenet, R., Bengio, Y., Kégl, B.: Algorithms for hyper-parameter
optimization. In: Advances in Neural Information Processing Systems (2011)
Toward Resilient Mobile Integration
Processes
Abstract. The widespread use of smart mobile devices and the digital
transformation foster context-aware applications (short: apps) that bridge
the gap between the socio-technical and business worlds in our everyday
lives. While the number of proprietary, mobile app solutions is increasing, a
mobile integration system architecture, e. g., connecting mobile and cloud
apps, and studies of the trade-off between resource-limits and resilient ser-
vice qualities on mobile computing platforms are missing.
In this work we define the core components of a resilient, resource-aware,
mobile enterprise application integration (EAI) system, which requires
extensions beyond the current non-mobile architectures. We evaluate the
proposed architecture and especially its runtime resource and monitoring
characteristics in form of a prototypical mobile realization.
1 Introduction
Current smart mobile devices are in the center of our everyday lives by provid-
ing a huge set of applications (called apps). In a process-like manner, these apps
connect the user’s context (e. g., location, car engine status) with cloud services
(e. g., ERP predictive maintenance), which are called context-aware applications
(e. g., [20,22]). To build more advanced, context-aware mobile applications, there
has been some recent work, e. g., on mobile process [20] and service engines [19]
for data collection scenarios. However, the current mobile apps have two short-
comings: (1) the apps are isolated and communicate via proprietary RPC imple-
mentations, leading to n-square connectivity and data format problems [11] that
are solved by enterprise application integration (EAI) [11], and (2) the apps
might neglect the trade-off between communication qualities (e. g., low latency,
no data loss) and resource-awareness or stability (e. g., CPU, battery consump-
tion) as also identified as challenge in [1]. Figure 1(a) shows the predominant
integration situation in current mobile app solutions (cf. (1)). Most of the apps
are directly connected to multi-cloud apps or services [24]. The redundant com-
munication layers of these apps (cf. marked as com.) are proprietary, and thus
c Springer International Publishing AG, part of Springer Nature 2018
W. Abramowicz and A. Paschke (Eds.): BIS 2018, LNBIP 320, pp. 278–291, 2018.
https://doi.org/10.1007/978-3-319-93931-5_20
Toward Resilient Mobile Integration Processes 279
have varying integration qualities (e. g., security) and increase the overall mem-
ory footprint on the device (e. g., multiple RPC implementations). The inter-app
communication via IPC on the same device is rarely used and the underlying
mobile OS concepts differ largely by their vendors. The mobile platforms offer
services (e. g., storage, network) and context information (e. g., contacts, GPS).
As a solution, we argue that a central integration system on the mobile devices,
shown in Fig. 1(b), solves the problems from (1), reduces the communication logic
within the applications and (2) allows for a central regulation of the resource
consumption for all communication tasks. In contrast to existing service (e. g.,
[20]) and automation solutions (e. g., IFTTT1 ) EAI systems allows for complex
routing and transforming patterns (e. g., [16,17]), whose compositions we call
integration scenarios.
Fig. 1. Current vs. desired EAI solution and monitor on mobile computing systems
For example, Fig. 1(c) denotes a vehicle logistics integration scenario, which
requires central integration logic as part of a cloud solution and a local, context-
aware scenario part on the device. Essentially, the local part collects data from
the environment (e. g., vehicle speed, and defect codes) in a high frequency, and
sends aggregates to the cloud solution with a configurable timing (e. g., every
30 s). The cloud solution adds master data to the incoming messages, determines
the receiving analytical applications and transforms the content accordingly. Fur-
thermore, monitoring data from the EAI system (e. g., number of processed mes-
sages or threads) and the mobile platform (e. g., resource capacities and usage)
are transferred and displayed in specific integration monitors according to [16]
(e. g., message, resource monitors). However, the current EAI architectures are
not designed for mobile deployments and miss resource-aware features, leading
to the research questions, addressed in this work:
Q1 Which device data are relevant for a stable integration solution and what
are suitable reactions on exceptional situations (e. g., offline, low battery)?
Q2 Which mechanism allows for building a context-aware integration system
with a resource-aware monitoring for multiple devices and applications?
1
IF This Than That, visit. 02/2018: https://ifttt.com/.
280 D. Ritter and M. Holzleitner
To answer these questions, we determine and classify mobile device data and
countermeasures (cf. Q1) tailored to mobile devices in Sect. 2. From the resulting
impact analysis, we derive and define an adaptable mechanism in Sect. 3 focusing
on resilient monitoring (cf. Q2), however, applicable to the runtime system.
The term resilient was also used in [1] in the context of component failures,
while we extend it for mobile devices as a characteristic to withstand several
exceptional context or resource-events (cf. Q1), to continue to operate normally
as long as possible. Then we describe the required extensions to the existing EAI
architectures in Sect. 4 (cf. Q3) and evaluate our approach in Sect. 5 (cf. Q4).
We discuss related work in Sect. 6 and conclude in Sect. 7.
In this section, we classify the common device data of current smart mobile
devices, denoting resource or context information and potential issues, and set
them into context to the countermeasures treating these issues in form of excep-
tion and integration adapter patterns from existing integration pattern cata-
logs [16,17].
We differentiate between the device data (e. g., from device specifications) and
the data of EAI systems (e. g., #msgs/sec) [11]. Due to our focus on integra-
tion system monitoring, we assume that the mobile monitor gathers configurable
metrics of the local integration system for local (i. e., on device) or remote visu-
alization and alerting (e. g., cloud monitor [16]). Hence, we concentrate on the
context and resource data, which have impact on the stability of the system (cf.
Q1) and assume that all app metrics can be subsumed by resource data (e. g.,
increasing #msgs/sec → CPU usage or temperature). We distinguish between
the capacity and the actual usage. Similarly, the EAI system data could be
classified.
Context Data and Device Resources. Although our approach is not limited
to the network (cf. Fig. 1(b)), it is the only context data that we discuss in this
work. The network on a mobile device requires wireless network access either in
form of bluetooth, a WIFI module or mobile data via a GSM module. This also
determines the three states that a device can be in: {offline, online, mobile data}
for the metric that we call mobile data state as metric (M0). Note that there is
other data like SSID (WIFI provider) or RSSI (for signal strength), which we do
not consider in this work, however, could be incorporated, if required.
Toward Resilient Mobile Integration Processes 281
2.2 Countermeasures
and thus, reduces the CPU and battery usage (cf. M1,2,5,6) as well as reduces
the data volume transferred per time unit (cf. M0). Since throttling reduces the
messages by keeping them in main memory and emits only a certain ratio per
time, the memory load might increase (cf. M3). This can be mitigated by using
the message sampler pattern [17] by removing messages according to a pattern
(CM-5: “sample messages”). This violates many service qualities (e. g., data
loss), and thus the sampler should be used in critical or urgent situations only.
A special case of the sampler in the context of “store to disk” (CM-1) is a
parameterized sampling on existing (already stored data), which we call data
store retention (CM-6: “Store Retention”), denoting another new pattern.
In contrast to the sampler, the retention deletes messages on disk to insert new
messages according to a policy (e. g., sample outdated messages on disk).
Escalation: Alert and Stop. In more severe situations, it might be necessary
to either stop a part of the system (e. g., monitoring, integration scenarios) like
the stop local pattern [17] (CM-7: “stop processing”) or even to stop the
complete app [17] (stop all; CM-8: “stop app”). Since this leads to planned
downtimes of parts or the complete system, an alert should precede the shut
down to allow for a user intervention. The alert has no direct impact on the
system, however, could be used to treat intervention and recoverable situations
brought to the user in local or cloud monitors (CM-9: “alert user”).
CM Interdependencies. Although most of the countermeasures do not have
side-effects, there are some that have to be discussed. The “stop app” (CM-8)
impacts all other actions by stopping them together with the app. This does not
further interfere with the other countermeasures after restart, since the situation
will be newly evaluated and for example, the stored messages will be redelivered
(CM-3). Furthermore, sample (CM-5) and redeliver from disk executed at the
same time will only deliver parts of the stored data, which is acceptable for
discrete resource and context data, variable with respect to its resolution (e. g.,
CPU, GPS coordinates), as common for context-aware apps.
In this section, we more formally define the behavior of our resilient EAI system
monitor based on the countermeasures for resources and context from Sect. 2.
This denotes the conceptual basis for a resource-aware, Event-Condition-
Action [5] approach that executes countermeasures according to the state of
the system.
The allowed states of a resilient mobile system can be derived from the context
situations found on mobile devices and the respective countermeasures (cf. both
from Sect. 2) and represented as deterministic finite automata (DFA). Thereby
we differentiate between a resource and a network automaton. The resource state
Toward Resilient Mobile Integration Processes 283
transitions are shown in Fig. 2(a), which denotes a DFA with the countermea-
sures (cf. CM-x) as states and the resource metric transitions as events (e. g.,
CPU temperature). When the app starts, the system is initially in normal pro-
cessing, which changes as soon as memory, CPU or battery warning thresholds
are crossed. Thereby, battery and CPU warnings include usage and temperature
events. Whenever a new state is reached through an event transition, the coun-
termeasure is executed by the system. Since this influences the overall system,
the normal processing state can be reached from any countermeasure, as soon as
all metrics transition into a normal state. The app can be shut down from any
state and by external events. Common countermeasures to keep the system in
allowed states are CM-4: “throttle processing”, CM-5: “sample messages” and
CM-7: “stop processing”. When service qualities are about to degraded by the
system (e. g., sample), CM-9: “alert user” is triggered to allow for interventions.
Similarly, the network states are represented in Fig. 2(b). Notably, mostly CM-
1–3 (“store to disk”, “stop storing to disk”, “redeliver from disk, clean disk”)
are required. Instead of sampling, the data store retention (CM-6: “Store Reten-
tion”) is applied to free space for more recent messages.
(a) Basic event calculus predicates from [9] (b) Reactive reasoning, similar to [13]
R1: Device online to offline R3: Battery, CPU temp. → CPU warning
initiallyP (anet ).initiallyP (adisk f ree ). initiallyN (abat ).initiallyN (acpu ).
holds at(norm, t) ⇒ intiates(e( , no net()), store, t)∧ holds at(norm, t) ∧ ¬holds at(abat , t) ⇒
term.(e( , no net()), norm, t) ∧ term.(e( , no net()), anet , t). init.(e( , batwarn ()), abat , t).
holds at(store, t) ⇒ term.(e( , msgrcv ()), adisk f ree , t). holds at(norm, t) ∧ ¬holds at(acpu , t) ⇒
init.(e( , cpuwarn ()), acpu , t).
R2: Device offline to online
holds at(store, t) ∧ holds at(adisk f ree , t) ⇒ holds at(abat , t) ∧ ¬clipped(t − 3, abat , t) ⇒
init.(e( , net()), norm, t) ∧ term.(e( , net()), store, t). term.(e( , cpuwarn ()), norm, t) ∧
holds at(store, t) ∧ ¬holds at(adisk f ree , t) ⇒ init.(e( , cpuwarn ()), throttle, t).
holds at(acpu , t) ∧ ¬clipped(t − 3, acpu , t) ⇒
init.(e( , net()), retry, t) ∧ term.(e( , net()), store, t).
term.(e( , cpuwarn ()), norm, t) ∧
holds at(retry, t) ⇒ init.(e( , msgssent ()), adisk f ree , t).
init.(e( , cpuwarn ()), throttle, t).
holds at(retry, t) ∧ holds at(adisk f ree , t) ⇒
init.(e( , net()), norm, t) ∧ term.(e( , net()), retry, t).
Example. The above table briefly shows excerpts of the overall theories for
three cases that we require in the evaluation Sect. 5: rules R1–R3. The system is
in a normal state at startup time t1 , if holds at(norm, t1 ). Most notably, a single
event, e. g., happens(e( , cpu warn()), t2 ) ∈ T, does not change the system’s
state. In contrast to R1 and R2, the system in R3 moves into another state, only
if this fluent holds for the last three time units t − 3 with ¬clipped(t − 3, acpu , t),
and thus prevents from frequent state changes.
We define an implicitly included app stopped fluent, which marks the termi-
nation of the app and all its integration processes when holds at(app stopped,
t), since a trace can be stopped at any time (not shown).
Toward Resilient Mobile Integration Processes 285
For the execution of the countermeasures, we reason about the defined EC the-
ories and fire, if the system is in an actionable state (e. g., throttle, retry).
Adapted to the resource-constraint devices, the reasoning should require mini-
mal resources. While the common abductive reasoning (i. e., tries to generate
a trace) is not applicable to our case, the deductive or backward reasoning
(i. e., continuously evaluates conjunctions of holds at predicates) is computa-
tionally expensive [21], and thus problematic on mobile devices. Hence we apply
a reactive reasoning from [4], where the fluents’ validity intervals are revised or
extended, when new events occur. This requires a cached EC (CEC) [4], which
caches the maximal validity interval (MVI), in which a fluent holds uninterrupt-
edly. Upon a new event occurrence CEC uses the cached MVIs to compute the
new result. We adopt an approach similar to CEC shown in Fig. 3(b), with a
given initial state (initially), a mobile event trace stream (happens) and app
event effects (initiates or terminates). The reasoner evaluates the MVIs of the
evolving fluents (i. e., holds at) and allows for the execution of countermeasures
on an event stream with a time-based window of five minutes.
three new, vital components: context listener (collecting events: happens), con-
text interpreter (evaluates conditions: holds at), and a countermeasure (execu-
tor). While the listener—bound to the OS services—receives a stream of resource
and context events, it converts them to EC theories according to Sect. 3, which
are fed into the interpreter, evaluating the series of events and reacting by a list
of countermeasures that are applied by the executor. Thereby the monitoring
sub-system is represented by integration processes in a separate integration pro-
cess instance, making the resilience characteristics applicable for non-monitoring
integration scenarios.
Realization. With more than two billion apps and
a market share of approximately 90%2 , the predomi-
nant mobile OS is Android with Java as its major app
programming language. Consequently, we ported the
open source integration system Apache Camel version
2.17.0 [7] to Android as app and added our mon-
itor . Since the used EC approach relies on Horn
clause fragments of first-order logic (with negation), we
extended the system by our 150 kB low footprint dat-
alog system [18], to address the context-aware exten-
sions and the reasoning parts from Sect. 3. All exper-
iments are conducted on a Samsung Galaxy S7 Edge
Fig. 5. Camel app. with 2.3 Ghz QuadCore and 1.6 Ghz QuadCore ARM
Cortex-A53, 4 GB main memory, 32 GB disk and 16
GB microSD, and Android Nougat OS. While the end-to-end communication
tests via WIFI use Apache JMeter, the monitoring cases are covered by a based
cloud monitor3 . Figure 5 shows integration scenarios in form of Camel routes
defined on the mobile device, where each route can be configured locally and
local services and apps (e. g., OBD-II reader) can be used as adapters. While
the system is running, all monitoring and resilience features can be switched-on
and off by the user.
5 Evaluation
In this section, we evaluate our approach realized as mobile app (cf. Sect. 4)
according to the following hypotheses, mirroring research question Q4:
H1 The monitor approach does not have a significant impact on the message pro-
cessing (e. g., → no increase of processing latency or memory consumption),
H2 The service qualities are preserved (e. g., → avoid data loss in external mon-
itor, up to a certain point: cf. trade-off resources vs. qualities).
system and its execution with monitoring and resilience. Only when mon-
itoring is switched on, we assume that the mobile event trace is collected for
visualization according to Fig. 3(b), while the countermeasures are determined
and executed only if the resilience is enabled. Figure 6(a) denotes the results of
the end-to-end latency of messages of different sizes sent from Apache JMeter to
a mobile Apache Camel route over WIFI until the response is received. Although
the WIFI was exclusively used during the benchmark without any other networks
in reach, the mean of the five runs slightly varies for the measurement in the 16kB
message size case. For the other message sizes, the results show no significant
deterioration of the latency with monitoring or resilience compared to the nor-
mal processing (i. e., no monitoring and resilience) with a 0.95 confidence interval
(cf. error bars). Similarly, there is no significant impact on the memory
consumption with increasing messages sizes between the normal process-
ing and our extensions as shown in Fig. 6(b). We indicate this through
the secondary y-axis, which shows the consumption difference diff =
max(memnormal , memmonitor , memresilience )-memnormal , with a relatively sta-
ble diff even with a growing amount of data during message processing. The
memory consumption of the monitoring and resilience remain stable, even for
an increasing number of senders (not shown).
Battery. Furthermore, the battery and CPU temperatures are considered criti-
cal service qualities, for which we show the effect of an increasing message load in
the integration system that let the CPU and Battery temperature rise together
with an increased CPU usage as shown in Fig. 7(b). When the countermeasure
(e. g., throttling) is executed according to rule R-3, then the effect is mitigated
without data loss (cf. from 2:35am UTC). Since the temperature reduces below
the threshold, the state switches from throttling to normal processing.
Discussion. Briefly, we experimentally showed that the normal EAI processing
is invariant to the new EAI system extensions proposed in this work (cf. H1).
Furthermore, we illustrated two cases, in which service qualities were preserved:
no data loss and battery temperature. However, with respect to the trade-off:
with longer offline executions (cf. Fig. 7(a)), leading to an increased disk usage,
data loss cannot be prevented beyond the disk limit. Since the monitoring and
resilience sub-systems are constructed by using Apache Camel routes, and thus
integration processes, the results of this work mainly discussed for the moni-
toring case, are directly applicable to the actual EAI runtimes (e. g., #threads,
#msgs/sec).
6 Related Work
Although there is a growing market for mobile automation software (e. g., IFTTT
or Integromat Mobile), we are not aware of any work on resource-aware, resilient
EAI systems on mobile devices. This observation is supported by recent work on
mobile process [20], service engines [19] for data collection scenarios and surveys
in the field [1], especially mentioning the lack of a resource-aware monitoring and
energy efficiency of current solutions. For a more systematic analysis, we con-
ducted a literature review following [8] across disciplines with the combinations
of the keywords: android, application, context-aware, device, mobile, monitor-
ing, resources, resource-aware, rule-based, quality of service (from [15]). The
horizontal search resulted in 63 articles out of which we selected nine, few from
the years between 2004–2005, but mostly 2012–2016, indicating recent interest
in this topic.
Service Engines. According to the analysis, most of the related work is from
the SOA domain, designing mobile service engines comparable to [19]. The clos-
est known related work from [3] defines a dual system resource approach (i. e.,
Toward Resilient Mobile Integration Processes 289
remote and local service). As the work from [25], the approach is limited to
network-awareness (i. e., minimize communication) and as countermeasure they
shut down the network (cf. M0, Sect. 2). The early work from [6] describes
another mobile SOA monitoring approach of an embedded system in a car. How-
ever, none of them describe a mobile EAI system and they do not monitor the
execution on the mobile system, and thus do not describe a resilient approach.
They were not evaluated from a functional runtime (e. g., latency) and resilience
perspective.
Miscellaneous. The work from [2,23], mainly focus on mobile network obser-
vation and can react to issues with the network (cf. M0, Sect. 2). [12] defines a
central monitoring of a distributed system, by generating the network of moni-
toring trees that optimizes multiple monitoring tasks and balances the resource
consumption at different nodes. Although the focus is different (e. g., no resource-
restrictions) the work can be seen complementary to ours. Similar to our app-
roach, [27] introduces a monitoring system that is evaluated on an android device.
However, the monitoring focuses exclusively on the user’s health in form of a
fitness app. The work by [26] uses a rule-based approach to monitor a pro-
cess system on android similar to [20]. Again, both approaches do not specify
resource-constraints or countermeasures. [14] defines a rule-based mobile phone
system, however, for context-aware, mobile applications.
7 Discussion
Toward a resilient mobile application integration and monitoring, we proposed
solutions along the research questions Q1–4, by (a) an analysis of data, impact
factors and countermeasures, (b) a formal mobile monitoring foundation based
on event calculus that allows for the necessary reasoning over the changing sys-
tem state for a resilient behavior of the monitoring as well as the runtime, (c) an
architecture extension for a resource-aware EAI system, and (d) evaluated the
approach with a mobile app prototype. Although this work mainly defines the
contributions (a)–(c) around a resilient monitoring use-case, the result is applica-
ble to integration runtime processing as well. Since the integration scenarios can
become complex, future work should address on-device modeling. Furthermore,
the presented approach requires the definition of thresholds, whose definition for
all metrics could be difficult and might require adaptations along changing sys-
tem characteristics. This opens another research direction toward parametrized
countermeasures with constraints for automatic system adaptation and opti-
mization, as already indicated by CM-6.
References
1. Aceto, G., Botta, A., De Donato, W., Pescapè, A.: Cloud monitoring: a survey.
Comput. Netw. 57(9), 2093–2115 (2013)
2. Arora, H., Greenwald, L.: Toward the use of local monitoring and network-wide
correction to achieve QoS guarantees in mobile ad hoc networks. In: IEEE SECON,
pp. 128–138 (2004)
3. Chen, S., et al.: Context-aware resource management middleware for service ori-
ented applications based on cross-layer monitoring. In: IEEE SMC (2015). Page
submitted
4. Chittaro, L., Montanari, A.: Efficient temporal reasoning in the cached event cal-
culus. Comput. Intell. 12(3), 359–382 (1996)
5. Dittrich, K.R., et al.: The active database management system manifesto: a rule-
base of ADBMS features. In: Sellis, T. (ed.) RIDS, pp. 1–17 (1995)
6. Gehlen, G., Mavromatis, G.: A rule based data monitoring middleware for mobile
applications. In: IEEE VTC, vol. 5, pp. 2910–2914 (2005)
7. Ibsen, C., Anstey, J.: Camel in Action. Manning Publications Co., Greenwich
(2010)
8. Kitchenham, B.: Procedures for Performing Systematic Reviews, vol. 33 (2004)
9. Kowalski, R., Sergot, M.: A logic-based calculus of events. In: Schmidt, J.W.,
Thanos, C. (eds.) Foundations of Knowledge Base Management. Topics in Infor-
mation Systems, pp. 23–55. Springer, Heidelberg (1989). https://doi.org/10.1007/
978-3-642-83397-7 2
10. Ligus, S.: Effective Monitoring and Alerting. O’Reilly Media Inc. (2012)
11. Linthicum, D.S.: Enterprise Application Integration. Addison-Wesley, Boston
(2000)
12. Meng, S., Kashyap, S.R., Venkatramani, C., Liu, L.: Resource-aware application
state monitoring. IEEE TPDS 23(12), 2315–2329 (2012)
13. Montali, M., Maggi, F.M., Chesani, F., Mello, P., van der Aalst, W.M.: Monitoring
business constraints with the event calculus. ACM TIST 5(1), 17 (2013)
14. Nalepa, G.J., Bobek, S.: Rule-based solution for context-aware reasoning on mobile
devices. Comput. Sci. Inf. Syst. 11(1), 171–193 (2014)
15. Ritter, D., Holzleitner, M.: Integration adapter modeling. In: CAiSE, pp. 468–482
(2015)
16. Ritter, D., May, N., Rinderle-Ma, S.: Patterns for emerging application integration
scenarios: a survey. Inf. Syst. 67, 36–57 (2017)
17. Ritter, D., Sosulski, J.: Exception handling in message-based integration systems
and modeling using BPMN. IJCIS 25(02), 1650004 (2016)
18. Ritter, D., Westmann, T.: Business network reconstruction using datalog. In:
Barceló, P., Pichler, R. (eds.) Datalog 2.0 2012. LNCS, vol. 7494, pp. 148–152.
Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-32925-8 15
19. Schobel, J., Pryss, R., Wipp, W., Schickler, M., Reichert, M.: A mobile service
engine enabling complex data collection applications. In: Sheng, Q.Z., Stroulia, E.,
Tata, S., Bhiri, S. (eds.) ICSOC 2016. LNCS, vol. 9936, pp. 626–633. Springer,
Cham (2016). https://doi.org/10.1007/978-3-319-46295-0 42
20. Schobel, J., Pryss, R., Schickler, M., Reichert, M.: A lightweight process engine for
enabling advanced mobile applications. In: Debruyne, C., et al.: (eds.) On the Move
to Meaningful Internet Systems: OTM 2016 Conferences. OTM 2016. LNCS, vol.
10033, pp. 552–569. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-
48472-3 33
Toward Resilient Mobile Integration Processes 291
21. Shanahan, M.: The event calculus explained. In: Wooldridge, M.J., Veloso, M.
(eds.) Artificial Intelligence Today. LNCS (LNAI), vol. 1600, pp. 409–430. Springer,
Heidelberg (1999). https://doi.org/10.1007/3-540-48317-9 17
22. Smanchat, S., Ling, S., Indrawan, M.: A survey on context-aware workflow adap-
tations. In: MoMM, pp. 414–417. ACM (2008)
23. Tiemeni, G.L.N.: Performance estimation of wireless networks using traffic gener-
ation and monitoring on a mobile device. Ph.D. thesis (2015)
24. Vukolić, M.: The Byzantine empire in the intercloud. ACM SIGACT News 41(3),
105–111 (2010)
25. Wagh, K.S., Thool, R.C.: Monitoring web resources and improve QoS of mobile
host using fuzzy logic. 4, 1779–1784 (2014)
26. Wipp, W.: Workflows on Android: A Framework Supporting Business Process Exe-
cution and Rule-Based Analysis. Ph.D. thesis, Ulm University (2016)
27. Zhang, W., Wang, X.: A lightweight user state monitoring system on android
smartphones. In: Toumani, F., et al. (eds.) ICSOC 2014. LNCS, vol. 8954, pp.
259–269. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-22885-3 23
Social Media and Web-Based Business
Information Systems
Tight and Loose Coupling in Evolving Platform
Ecosystems: The Cases of Airbnb and Uber
Andreas Hein ✉
( )
, Markus Böhm , and Helmut Krcmar
Technical University of Munich, Munich, Germany
{andreas.hein,markus.boehm,krmcar}@in.tum.de
Abstract. The emergence of digital platforms changes the way how companies
interact with their ecosystem. Successful platforms like Apple’s App Store utilize
an ecosystem of third-party developers to drive innovation. Those platforms are
expanding the sphere of influence beyond internal resources and capabilities by
taking advantage of a scalable ecosystem of external developers. However, until
now it is unclear on how those companies establish a platform ecosystem. This
article draws on two case studies in the form of ridesharing and accommodation
platforms to illustrate how they transitioned through four evolutionary phases
with the help of tight and loose coupling partnerships.
1 Introduction
The success of digital platforms is unheard [1]. Examples range from application plat‐
forms like Apple’s App or Google’s Play Store to the currently emerging IoT platform
market. In either case, the surrounding ecosystem of partners, developers, and users
strongly determines the success of the platform. In contrast to traditional businesses,
where the company produces products or services, a platform orchestrates the interaction
in a two-sided market of suppliers (e.g., App developers) and users (e.g., App users) [2–
4]. The ecosystem fosters cross-side network effects between both sides, where each
side profits from additional exchange [2, 3]. By capturing a margin of the created value,
the platform has a self-interest in growing this ecosystem. Current practices in starting
an ecosystem that incorporates tight coupling of selected parties through individual
partnership to a scalable utilization of boundary resources like Application Program‐
ming Interfaces (APIs) and Software Development Kits (SDKs) resulting in a loosely
coupled partnerships [5, 6]. Based on those practices, ecosystems go through several
phases starting with the birth, expansion, leadership, and self-renewal [7]. Those rapidly
emerging digital platform ecosystems pose challenges to traditional companies. They
face the question of how to start an ecosystem effectively to remain competitive.
This article sheds light on this question by elaborating on the ecosystem evolution
of the recent phenomenon of transactional service platforms like Uber and Airbnb. Both
provide a servitization in the form of mobility or accommodation services based on
software artifacts. Uber, for example, faced little competition and managed to grow an
2 Theoretical Background
This article draws on the theoretical groundwork of Moore [7] on business ecosystems
and the theory of Weick [8] on tightly and loosely coupled systems. We combine both
lenses to elaborate on how platform ecosystems evolved.
The theory of business ecosystem is shaped by influences from the field of Anthropology
and co-evolution [9], as well as from Biology and natural ecosystems [10]. Co-evolution
emphasizes that interdependent species or actors evolve in nearly endless reciprocal
cycles. The evolution of one species, thus, influences the natural selection of another
species in the same ecosystem and vice versa. On the other hand, natural ecosystems
respond with sensitivity to environmental change [9, 10]. Moore combines both aspects
to describe how businesses evolve under ever-changing environmental influences in an
ecosystem of partners, suppliers, and customers. The theory proposes four evolutionary
stages [7]:
Birth: The first evolutionary phase describes risk-taking entrepreneurs that focus on
customer needs and how to satisfy those needs. An example is the rise of the personal
computer in the 1970s, which started in 1975 by the sole effort of hobbyists in hard-to-
use products. Only in the late 1970s, companies like Apple utilized an ecosystem of
business partners in the field of independent software developers, training institutes or
computer stores. In this ecosystem, Apple tightly controlled the computer design and
the operating software, while encouraging and co-evolving with other partners to
contribute complementary software or hardware [7].
The first phase harbors the cooperative challenge of collaborating with the demand
and supply side to create new value propositions that are the basis for future innovations.
At the same time, the young ecosystem needs to protect its value proposition from
imitation by tightly integrating the parties involved in the ecosystem [7].
Expansion: In the second phase, business ecosystems expand to new territories. This
phase is characterized by direct battles for market share between competing or over‐
lapping ecosystems. Companies that aim to expand their ecosystem need to have a
Tight and Loose Coupling in Evolving Platform Ecosystems 297
scalable business concept in place. An example for capturing territory can be found in
the personal computer ecosystem when IBM entered and replaced Apple’ dominant
position. IBM opened its architecture to hardware supplier fostering additional comple‐
mentary products. Also, they licensed MS-DOS and ensured the compatibility and port‐
ability of popular applications like Wordstar [7].
The cooperative challenge is to establish a scalable ecosystem of supply and demand
to achieve maximum market coverage. From a competitive perspective, the company
needs to establish a de-facto industry standard by dominating key market segments [7].
Leadership: The ecosystem leadership phase expresses stable structures and a high
concentration of bargaining power by a dominating company. The leader creates prod‐
ucts or services that are of critical value for the whole ecosystem and protects this posi‐
tion through patents or constant innovation. The central ecological position of the
ecosystem leader is often instantiated by clearly defined interfaces to partners reflecting
the industry standard. Though those interfaces, the ecosystem leaders encourage partners
in the ecosystem to take over activities and to accelerate the growth of the whole. In the
case of the personal computer ecosystem, missing leadership and a too open architecture
opened a window of opportunity for competitors like Lotus, Intel, and Microsoft. Micro‐
soft and Intel proofed to exercise control over critical components in the form of the
operating systems and chips needed for the personal computer [7].
The cooperative challenges imply to provide a compelling vision for the future that
encourages the ecosystem to work together to strengthen the central position of the
leader. On the other hand, the competitive challenge is to maintain central bargaining
power in the ecosystem [7].
Tight and loose coupling describes the degree of dependency between actors within and
between organizations or systems. Orton and Weick [6] differentiate between tight,
loose, and decoupled systems:
298 A. Hein et al.
Tight Coupling: In a tightly coupled system, elements are strongly dependent on each
other, and they do not act independently. Researchers speak of responsiveness without
distinctiveness [6, 8]. Typical characteristics of tightly coupled systems are an increased
understanding of each other’s needs, a close relationship, a low degree of information
asymmetry, and the ability to tailor products or services to strategic needs [8, 11]. A
practical example for tight coupling is a strategic partnership like seen in the IT
outsourcing, where the partnership is determined by precise rules like Service-Level-
Agreements and a high degree of mutual dependency [12]. Loose Coupling: The
concept of loose coupling refers to independent elements that are distinct or separate
from one another, yet responsive. Scientists call this order distinctive responsiveness.
The advantage of a system is that elements maintain flexibility to react to change from
outside but adhere to standards in the system. Thus, the system maintains stability as
change from external sources cause no ripple effects. As a result, the system is simul‐
taneously closed and open, flexible and stable, as well as rational and spontaneous [6].
Decoupling: In opposition to neither distinctive nor responsive elements, which is per
definition no system, decoupled elements are distinct to each other [6].
3 Research Approach
The research design follows a multiple case study with subject to accommodation and
ridesharing platforms. The method is particularly suitable as it captures and describes
the complexity of real events [13]. The cases show boundaries, features, and limitations,
by putting the business ecosystems [7] and tight/loose coupling [6, 8] theories into a
specific context including the respective environment and firms [14]. Benbasat, Gold‐
stein and Mead [15] provide an orientation on whether the analyzed topic is a phenom‐
enon and the usage of a case study is appropriate. First, the context is crucial to observe
the phenomenon of emerging platform ecosystems opposed to an isolated view on plat‐
forms. Secondly, the tremendous success of platform businesses like Airbnb and Uber
shows the significance and actuality of the research topic and link to the contemporary
event of emerging platform ecosystems [13]. Additionally, neither control nor manip‐
ulation of the subject or event took place. This was guaranteed as the case study describes
the phenomenon in the view of a neutral observer [16]. After conducting the two case
studies, we draw cross-case conclusions on how a platform ecosystem evolves by
showing similarities and differences between the cases.
The data for the two cases are based on archival data and empirical studies, ranging
from the emergence of the respective platform ecosystem to the present. Archival data
includes financial data, independent reports on milestones, partnerships, key events (like
introducing an API), and the general development of the platform and its ecosystem.
We used additional empirical studies to validate financial figures, events, as well as
strategic decisions. Further, we adhered to the principle of data triangulation to make
sure that each fact or decision is cross-validated [17]. Another mechanism to take care
of business model specifics was including the main competitor as a benchmark in the
ecosystem. Overall, we collected 24 archival data sources and 12 empirical studies on
Tight and Loose Coupling in Evolving Platform Ecosystems 299
the case of accommodation platform ecosystems, as well as 28 archival data sources and
16 empirical studies on the ridesharing platform ecosystem case (see Table 1).
4 Results
The results show that both platform ecosystems used tight coupling to evolve from birth
to the expansion phase. In a second stage, both adhered to a more open approach through
loose coupling resulting in platform leadership and self-renewal.
Birth: Founded in 2009, Uber triggered the trend of on-demand ridesharing companies.
Instead of competing directly with the taxi ecosystem but rather with the concept of
owning a car, Uber managed to collect $ 1.75 million during their first two rounds of
seed investment in 2009 and 2010. The company started in the San Francisco Bay Area
and provided “high-end” sedans with reputable drivers that can be easily booked via
iPhone taking advantage of the GPS sensors. Uber did not want to compete for the lowest
price but charged 1.5x of the regular taxi rate offering a comfortable and convenient way
of moving from A to B. Another distinguishing feature is the use of state-of-the-art
technology incorporating the iPhone sensors to provide an easy to use experience. Uber
managed to establish a position in the mobility ecosystem by matching the demand of
convenient mobility services with the supply of professional chauffeurs.
Most people tend to think that Airbnb had always been the number one platform
when it comes to short-term lodging. The truth is that Airbnb was surrounded by other
vacation rental companies like HomeAway or Flip-key founded in 2005 and 2007 from
the very beginning. In 2008, the company called AirBed & Breakfast entered a highly
competitive ecosystem offering short-term living quarters in the center of San Francisco.
During the initial phase, Airbnb managed to keep up with its competitors by providing
additional value-adding services resulting from tight-coupled partnerships. One of them
was proposing free and professional photography from the housings on the platform to
reduce, on the one hand, the effort for the host to join the platform and to increase, on
the other hand, the reliability and trustworthiness for the guest. At the end of 2011,
Airbnb established its position in the ecosystem and raised over $ 1 billion in funding,
which was nearly half the amount of HomeAway with 2.2 billion dollars.
300 A. Hein et al.
Expansion: After introducing the iOS and Android Apps in 2010, Uber launched a
national expansion starting in the San Francisco Bay Area, expanding to New York City
in 2011. Until the year 2012, Uber collected over $ 1.9 billion in funding and could
develop their business without any direct competition in the ecosystem. In the same year,
Lyft emerged as the first direct competitor. Lyft dates back to Zimride, which was a
long-distance ridesharing company. Lyft has the same focus as its competitor with on-
demand ridesharing in a peer-to-peer manner. Ending in the year 2012, Uber used the
three years without competition to establish its position with a valuation of $ 1.9 billion
compared to $ 275 million of Lyft. Based on this leading edge, Uber established various
tightly coupled partnerships to secure and expand on its current position in the
ecosystem. One exemplary partnership that strengthened the supply side (drivers) was
providing special interest rates on car loans for GM and Toyota vehicles [18]. Newly
bought cars that engage in the Uber platform can be seen as assets to lower the interest
rates for potential drivers ensuring scalability on the supply side. Other partnerships
involve PayPal and American Express to offer additional payment services reducing the
entrance barriers for passengers or the corporation with Concur for business travelers
or flight procedures within United Airlines to expand the market reach [19]. With the
help of those tightly managed partnerships, Uber could extend its position in the
ecosystem and establish itself as a de-facto standard until 2014 with a valuation of $ 41
billion compared to 2.5 $ billions of Lyft.
Airbnb needed to be more careful as they acted in a highly competitive ecosystem.
One example is that Airbnb offered an active API endpoint including an affiliation
program for developers [20]. However, new competition in the form of Wimdu and
9Flats emerged. Airbnb could not establish enough authority and consequently did not
officially introduce the programming interface. Besides closing the API end-point,
Airbnb also launched a new flagging feature to defend its content by reporting ques‐
tionable behavior. The response is somewhat understandable, as both new competitors
were known to copy established business models including their content [21]. Airbnb
focused on additional tight coupling partnerships to grow the ecosystem and to progress
further. Examples are the partnerships with Lloyds of London to offer insurances for
hosts included in every platform booking or the affiliation with American Express intro‐
ducing new payment functionalities [22]. All those partnerships aim to improve the
relationship between the host and the guest, in turn, affecting the reputation of Airbnb.
At the end of the year 2015, Airbnb harvested the fruits of their expansion with $ 25.5
billion followed by HomeAway with an evaluation of $ 3.9 billion. Besides, also the
number of listings (Airbnb: from 50,000 listings in 2011 to 550,000 in 2014) and active
drivers (Uber: from near zero in 2012 to 160,000 in 2014) rose sharply [23, 24].
Leadership and Self-Renewal: Starting the next phase, Uber established enough
authority in the ecosystem to respond to the demand of third-party developers by opening
up the platform through Application Programming Interfaces (APIs) and Software
Development Kits (SDKs). Starting in 2014, third-party developers could integrate core
mobility services provided by Uber, like ordering a ride with the help of predefined code
snippets. Further, Uber opened the micro-service architecture of the platform to help
developers using driver, car, and ride data to develop new applications, and to enhance
Tight and Loose Coupling in Evolving Platform Ecosystems 301
the driving experience during the trip [25]. One example of tweaking the open mobility
services is the use of calling an Uber via an SMS over the phone. Those apps emerge
out of the loosely coupled relationship between developers in the ecosystem and are
especially useful in areas without mobile internet. Uber is not only using the ecosystem
to expand the market reach but also takes advantage of the innovation capabilities of
external developers. They control the ecosystem through the strategical use of compo‐
nents in the form of APIs like the Trip Experience API or the UberRUSH API strength‐
ening their core service. One example is the integration of contextual information during
the ride that Uber provides to third-party developers. They use the information to create
value-adding applications that benefit both Uber and the passenger. Consequently, Lyft
followed this strategy and introduced an open API in 2016. As of 2017, Uber has estab‐
lished a vibrant ecosystem of developers providing five different APIs, 22 SDKs in
several programming languages and three software libraries [26]. They transitioned from
a tight coupling approach to target special needs like supporting potential drivers owning
a car to a more loosely coupled approach to utilize a whole ecosystem of developers to
foster constant innovation.
In the accommodation industry, it was after 2015 and onward, as Airbnb’s competitor
HomeAway first opened their platform through active API endpoints including docu‐
mentation, guides, and an affiliate program. The market need for the content data of
those accommodation platforms was omnipresent. In this period, several companies
tried to scrape the data from the Airbnb platform, like AirDNA analyzing the content
of Airbnb and selling it for investment purposes or several attempts to reverse engineer
the not yet publicly open Airbnb API endpoints. All those efforts point toward the direc‐
tion that Airbnb controlled the core components of the platform ecosystem. Finally, the
combination of all those market signals and the economic advance and authority
regarding market valuation to the leading competitors influenced Airbnb’s decision
towards a more open and loosely coupled approach. Arrived at the year 2017, Airbnb
emerged as a winner of the shark tank of short-term lodging platforms and finally taking
control over their ecosystem by introducing an official API program including docu‐
mentation and partnership programs.
5 Discussion
The two cases show that platform ecosystems follow the evolutionary pattern of business
ecosystems. In addition, the theoretical lens of tight and loose coupling helps to explain
how platforms exercise scalability through network effects to drive innovation. The
results indicate that the platforms use tight coupling partnerships in the birth and expan‐
sion phase to become ecosystem leaders, followed by a transition towards openness and
loosely coupled partnerships to drive external innovation (see Fig. 1). Instead of domi‐
nating each step in the value chain as, for example, Original Equipment Manufacturers
do in the automotive industry; platform ecosystems are in their mature states more
loosely coupled and open, which fosters open innovation effects.
302 A. Hein et al.
Fig. 1. The use of tight and loose coupling during the evolution of platform ecosystems
During the birth phase, both platforms exercised a high degree of control over their
ecosystem. Airbnb needed to deal with competition and utilized tight coupling partner‐
ships to increase the scope of the platform, thus, maximizing the value for customers.
An example is the collaboration with professional photographers to increase the trust
between guests and hosts. Uber had no direct competition in the first place but managed
to differentiate its value proposition through a more convenient and reliably way of
inner-city traveling. The phase of expansion dealt with an emerging competition and
encouraged securing and expanding mechanisms. The appearance of additional compe‐
tition forced Airbnb to retain control over their content by closing all open API
endpoints. Also, the platform established tight coupling partnerships like insuring apart‐
ments with companies to further increase the trust between hosts and guests. On the
other hand, Uber used tight coupling partnerships on the supply side by collaborating
with GM to provide special interest rates for potential drivers out-competing other plat‐
forms. Both, GM and Uber developed mutual trust, where Uber convinced GM that
drivers could use the car as a profit generation resource, while GM convinced Uber to
provide safe and reliable vehicles. Another example is the cooperation with Concur to
expand to business travelers. The left part of Fig. 1 illustrates how both platforms took
advantage of tight coupling partnerships to indirectly influence the value proposition
between the demand and supply side. Airbnb offered a professional photo shooting for
free, creating value for the host through the polished appearance of the room, as well as
value for the guest as all pictures made by Airbnb were verified reducing the risk of
being tricked. The same is true for the special car financing loans for Uber drivers.
Passengers enjoy brand-new cars and drivers have the opportunity to earn an inde‐
pendent living. The platforms indirectly enhanced cross-side network effects between
the supply and demand-side through strategic or tightly coupled partnerships. The aim
of those partnerships is either to strengthen the value proposition between the two sides
as seen in the photograph or car financing example or to expand the ecosystem as with
Concur. Tight coupling partnerships lead to a high degree of control over the strategic
Tight and Loose Coupling in Evolving Platform Ecosystems 303
intention and output of the collaboration. In return, both parties need to allocate resources
to align on a mutual strategy and outcomes.
According to Uber, the primary intention behind opening the platform was based on the
three pillars of utility, revenue, and distribution [27]. Uber sees utility as a continuous
improvement process of making the platform profit from new ideas within the
ecosystem. They utilize a compelling vision for the future of mobility-as-a-service going
hand in hand with Moore’s description of ecosystem leadership. The second pillar aims
to incentivize the development of new applications in the form of affiliate programs that
encourage ecosystem partners to contribute toward core components. Lastly, the distri‐
bution pillar provides value during additional distribution (e.g., improving the experi‐
ence during the ride). Uber provides stability by leading the co-innovation through
targeted APIs like the Trip Experience API enabling developers to embed complemen‐
tary content, games or other functionality that enhance the experience during trips. As
Airbnb just opened their platform, it is at a too early stage to predict what complementary
services emerge around the ecosystem. However, the early approaches of setting up an
affiliate program point in a somewhat similar direction. As a prerequisite for opening
the platform, both platforms established control and authority over their ecosystem to
304 A. Hein et al.
deal with competition. The examples of Uber and Airbnb show the utilization of tight
coupling partnerships to build the core ecosystem. Both platforms used value-adding
partnerships to enhance the scope and range of the platforms core services, as well as
securing mechanisms to stay on the leading edge regarding competition [5]. After
expanding to a dominant position, both platforms opened their ecosystem around crucial
platform services that evolved toward a market standard. They integrated external
developers in loosely coupled partnerships through boundary resources like APIs and
SDKs. Other platforms from industries like Social Networks and App Stores follow a
similar approach. From the majority of loosely coupled developers’ point of view, it
makes sense to invest time and resources in an already established market standard,
instead of multi-homing between several competing platforms with an increased effort.
A technical commonality is that both companies transitioned from a monolithic to a
modular micro-service infrastructure that can be adapted to key products or services.
With this approach, the platform makes sure that newly created products extend the main
product or service. Uber, for example, opened the ecosystem for third party development
to add additional value to passengers during the ride fostering complementary innova‐
tion. To achieve this goal, both platforms use scalable interfaces between the modular
micro-service infrastructure and the ecosystem. SDKs, libraries, and documentation
help to simplify the development process by covering the most used programming
languages to include pre-defined functions and snippets helping developers to get
started. By transitioning from a tightly coupled ecosystem to a more arms-length
approach, the platform governance needs to be adapted as well as shown in the example
of the incentive programs. Other aspects of the ecosystem range from coping with the
on- and off-boarding of developers in a scalable way to the automatic input and output
control of applications within the ecosystems [28, 29].
This research provides insights into evolutionary phases and coupling mechanisms in
the context of platform ecosystems. We show that platforms follow the evolutionary
phases as proposed by Moore [7]. However, there are differences when introducing the
mechanisms of tight and loose coupling. The two cases indicate that platforms use tight
coupling partnerships in the beginning to enhance the core service or value proposition
strengthening cross-side network effects between the supply and demand-side. Thus,
tight coupling is one promising mechanism to answer the open question on how plat‐
forms strategize to foster cross-side network effects [30]. Another theoretical contribu‐
tion points towards the increasing role of value co-creation in a platform ecosystem [31].
We highlight that platforms first prepare the ecosystem and then enable third-party
developers to actively innovate around core services. For practitioners, we show that
timing is crucial when it comes to the question when to open a platform. Airbnb learned
that authority and control over their ecosystem were important and that the platform’s
content can be scraped by competitors if not protected properly. Thus, it was important
to first strengthen core services through strategically selected tight coupling partnerships
and then opening up to utilize the innovation capability of a loosely coupled ecosystem.
Tight and Loose Coupling in Evolving Platform Ecosystems 305
In this sense, a fruitful area for future research could be the identification of value co-
creation practices with emphasize on the tangible character of tight coupling partner‐
ships and loosely-coupled or intangible ecosystem relationships. Also, new technology
like blockchain might influence the future development of platform ecosystems. The
research also faces limitations. The data for the two case studies are mainly based on
archival data and empirical research taking the role of the neutral observer. We thus
limit the explanatory power of this contribution on answering the motives or exact prac‐
tices of both platforms when using tight and loose coupling mechanisms. One way to
mitigate this problem is conducting in-depth case studies to elaborate on the why and
how of tight and loose coupling practices. In total, this research helps to explain how
platforms accelerate network effects in their ecosystem evolution. They first establish
ecosystem leadership with the help of strategical or tight coupling partnerships that
strengthen the value proposition of their core service. Then, they unleash the network
externalities by opening the platform to the ecosystem and enabling third parties in
loosely coupled partnerships through clearly defined interfaces to innovate around those
core services.
Acknowledgements. This work is part of the TUM Living Lab Connected Mobility (TUM
LLCM) project and has been funded by the Bavarian Ministry of Economic Affairs and Media,
Energy and Technology (StMWi) through the Center Digitisation. Bavaria (ZD.B), an initiative
of the Bavarian State Government.
References
1. Evans, P.C., Gawer, A.: The rise of the platform enterprise: a global survey (2016)
2. Eisenmann, T., Parker, G., Alstyne, V.M.W.: Strategies for two-sided markets. Harvard Bus.
Rev. 84, 92–101 (2006)
3. Rochet, J.C., Tirole, J.: Platform competition in two-sided markets. J. Eur. Econ. Assoc. 1,
990–1029 (2003)
4. Hein, A., Böhm, M., Krcmar, H.: Platform Configurations within Information Systems
Research: A Literature Review on the Example of IoT Platforms, pp. 465–476.
Multikonferenz Wirtschaftsinformatik, Lüneburg, Germany (2018)
5. Ghazawneh, A., Henfridsson, O.: Balancing platform control and external contribution in
third-party development: the boundary resources model. Inf. Syst. J. 23, 173–192 (2013)
6. Orton, J.D., Weick, K.E.: Loosely coupled systems: a reconceptualization. Acad. Manag.
Rev. 15, 203–223 (1990)
7. Moore, J.F.: The Death of Competition: Leadership and Strategy in the Age of Business
Ecosystems. Harper Business, New York (1997)
8. Weick, K.E.: Educational organizations as loosely coupled systems. Adm. Sci. Q. 21, 1–19
(1976)
9. Bateson, G.: Mind and Nature: A Necessary Unity. Dutton, New York (1979)
10. Gould, S.J.: The Structure of Evolutionary Theory. Harvard University Press (2002)
11. Danneels, E.: Tight–loose coupling with customers: the enactment of customer orientation.
Strateg. Manag. J. 24, 559–576 (2003)
12. Steensma, H.K., Corley, K.G.: On the performance of technology-sourcing partnerships: the
interaction between partner interdependence and technology attributes. Acad. Manag. J. 43,
1045–1067 (2000)
306 A. Hein et al.
13. Yin, R.K.: Case Study Research: Design and Methods. Sage Publications, Thousand Oaks
(2014)
14. Merriam, S.B.: Case Study Research in Education: A Qualitative Approach. Jossey-Bass, San
Francisco (1988)
15. Benbasat, I., Goldstein, D.K., Mead, M.: The case research strategy in studies of information
systems. MIS Q. 11, 369–386 (1987)
16. Siggelkow, N.: Persuasion with case studies. Acad. Manag. J. 50, 20–24 (2007)
17. Klein, H.K., Myers, M.D.: A set of principles for conducting and evaluating interpretive field
studies in information systems. MIS Q. 23, 67–93 (1999)
18. Tam, D.: Uber teams up with GM, Toyota to get more cars on the road, C-net. https://
www.cnet.com/news/uber-teams-up-with-gm-toyota-to-get-more-cars-on-the-road/. Accessed
01 Nov 2017
19. Hempel, J.: Uber and the future of American Express, Fortune. http://fortune.com/
2014/06/09/uber-and-the-future-of-american-express/. Accessed 1 Nov 2017
20. DuVander, A.: AirBNB API Will Pay Developers with Affiliate Program,
ProgrammableWeb. https://www.programmableweb.com/news/airbnb-api-will-pay-
developers-affiliate-program/2011/10/13. Accessed 01 Feb 2018
21. Tsotsis, A.: Airbnb Freaks Out over Samwer Clones, TechCrunch. https://techcrunch.com/
2011/06/09/airbnb/. Accessed 1 Nov 2017
22. Empson, R.: Airbnb Makes Good, Will Now Cover Up To $1M in Property Damages,
TechCrunch. https://techcrunch.com/2012/05/22/airbnb-1m-guarantee/. Accessed 01 Nov
2017
23. Hall, J.V., Krueger, A.B.: An analysis of the labor market for Uber’s driver-partners in the
United States. ILR Review 0019793917717222 (2015)
24. Ferenstein, G.: Uber and Airbnb’s incredible growth in 4 charts, VentureBeat. https://
venturebeat.com/2014/06/19/uber-and-airbnbs-incredible-growth-in-4-charts/. Accessed 01
Feb 2018
25. Haddad, E.: Service-Oriented Architecture: Scaling the Uber Engineering Codebase as We
Grow, Uber Engineering. https://eng.uber.com/soa/. Accessed 01 Nov 2017
26. Programmable Web: Uber API, Programmable Web. https://www.programmableweb.
com/api/uber. Accessed 1 Nov 2017
27. Wintrob, G.: Talking to Uber about the three pillars of its API platform, TechCrunch. https://
techcrunch.com/2016/06/16/talking-to-uberhe-three-pillars-of-ubers-api-platform/. Accessed 01
Feb 2018
28. Hein, A., Schreieck, M., Wiesche, M., Krcmar, H.: Multiple-case analysis on governance
mechanisms of multi-sided platforms. In: Multikonferenz Wirtschaftsinformatik, Ilmenau,
Germany, pp. 1613–1624 (2016)
29. Schreieck, M., Hein, A., Wiesche, M., Krcmar, H.: The challenge of governing digital
platform ecosystems. In: Linnhoff-Popien, C., Schneider, R., Zaddach, M. (eds.) Digital
Marketplaces Unleashed, pp. 527–538. Springer, Heidelberg (2018). https://doi.org/
10.1007/978-3-662-49275-8_47
30. McIntyre, D.P., Srinivasan, A.: Networks, platforms, and strategy: emerging views and next
steps. Strateg. Manag. J. 38, 141–160 (2017)
31. Lusch, R.F., Nambisan, S.: Service innovation: a service-dominant logic perspective. MIS
Q. 39, 155–175 (2015)
On Feeding Business Systems with Linked
Resources from the Web of Data
Abstract. Business systems that are fed with data from the Web of
Data require transparent interoperability. The Linked Data principles
establish that different resources that represent the same real-world enti-
ties must be linked for such purpose. Link rules are paramount to trans-
parent interoperability since they produce the links between resources.
State-of-the-art link rules are learnt by genetic programming and build
on comparing the values of the attributes of the resources. Unfortunately,
this approach falls short in cases in which resources have similar values
for their attributes, but represent different real-world entities. In this
paper, we present a proposal that leverages a genetic programming that
learns link rules and an ad-hoc filtering technique that boosts them to
decide whether the links that they produce must be selected or not.
Our analysis of the literature reveals that our approach is novel and our
experimental analysis confirms that it helps improve the F1 score by
increasing precision without a significant penalty on recall.
1 Introduction
The feasibility of many emerging business relies on the availability and inter-
operability of suitable on-line datasets [1]. The Web of Data provides business
with islands of data (availability) that can be transparently used as required by
the business models (interoperability). The Web of Data is ruled by the Linked-
Data principles that support the idea that resources within different datasets
that represent the same real-world entities must be linked so as to facilitate data
interoperability [4]. Link rules are intended to help link resources automatically
however they may be linked by different kind of relations, e.g., spatial or tem-
poral coverage, we focus only on owl:sameAs that relates resources representing
the same real-world entity.
2 Related Work
The earliest techniques to learn link rules were devised in the field of traditional
databases, namely: de-duplication [7,17], collective matching [2,3,6,14,21], and
entity matching [15]. They set a foundation for the researchers who addressed
On Feeding Business Systems with Linked Resources from the Web of Data 309
the problem in the context of the Web of Data, where data models are much
richer and looser than in traditional databases.
Some of the proposals that are specifically-tailored to web data work on a
single dataset [10,16], which hinders their general applicability; there are a few
that attempt to find links between different datasets [5,8,9,13], but they do not
take the neighbours of the resources being linked into account, only the values of
the attributes; that is, they cannot make resources with similar values for their
attributes apart in cases in which they represent different real-world entities. An
additional problem is that they all assume that data are modelled by means of
OWL ontologies. Unfortunately, many common datasets in the Web of Data do
not rely on OWL ontologies, but on simple RDF vocabularies that consists of
classes and properties whose relationships and constraints are not made explicit.
The previous problems motivated several authors to work on techniques that
are specifically tailored to work with RDF datasets. Most such proposals rely on
genetic programming [11,12,19,20] in which chromosomes encode the link rules as
trees, which facilitates performing cross-overs and mutations. They differ regard-
ing the expressivity of the language used to encode the link rules and the heuris-
tics used to implement the selection, replacement, cross-over, and mutation oper-
ators, as well as the performance measure on which the fitness function relies.
Isele and Bizer [11,12] contributed with a supervised proposal called Genlink. It is
available with the Silk framework [24], which is gaining impetus thanks to many
real-world projects [23]. It uses a tournament selection operator, a generational
replacement operator, custom cross-over and mutation operators, and its fitness
function relies on the Matthews correlation coefficient. It can use a variety of cus-
tom string transformation functions and the Levenshtein, Jaccard, Numeric, Geo-
graphic, and Date string similarity measures. An interesting feature is that the
size of the link rules must not be pre-established at design time, but it is dynami-
cally adjusted during the learning process. Ngomo and Lyko [19] contributed with
a supervised proposal called Eagle, which is available with the LIMES frame-
work [18]. It uses a tournament selection operator, a μ + β replacement opera-
tor, tree cross-over and mutation operators, and its fitness function relies on the
F1 score. It does not use transformation functions, but the Levenshtein, Jaccard,
Cosine, Q-Grams, Overlap, and Trigrams string similarity functions. The maxi-
mum size of the link rules must be pre-established at design time. Nikolov et al. [20]
contributed with an unsupervised proposal. It uses a roulette-wheel selection oper-
ator, an elitist replacement operator, a tree cross-over operator, a custom muta-
tion operator, and a pseudo F1 fitness function. Transformations are not taken into
account, but the library of similarity functions includes Jaro, Levenshtein, and I-
Sub. The maximum size of the link rules rules is also set at design time. There is a
diverging proposal by Soru and Ngomo [22]. It supports the idea of using common
machine-learning techniques on a training set that consists in a vectorisation of
the Cartesian product of the resources in terms of the similarity of their attributes.
Transformation functions cannot be used and the string similarity functions
available are Q-Grams, Cosine, and Levenshtein. Whether the size of the rules
must be pre-set or not depends on the underlying machine learning technique.
310 A. Cimmino and R. Corchuelo
Unfortunately, none of the proposals for RDF datasets take the neighbourhoods
into account (Fig. 1).
Clearly, the state of the art does not account for a proposal to link resources in
RDF datasets that takes their neighbours into account. Ours is tailored to work
with such datasets and it is novel in that it is not intended to generate link rules,
but leverages the rules that are learnt with other proposals and boosts them in
order to analyse the neighbours, which our experimental analysis confirms that
has a positive impact on precision without degrading recall.
3 Our Proposal
Our proposal consists in three components, namely: the first one learns link
rules, the second one filters out the links that they produce, and the third relies
on two voting strategies to select most reliable filtered links.
The link rule learner is based on Genlink [12], which is a state-of-the-art
genetic programming that has proven to be able to learn good link rules for
many common datasets. Genlink is feed with a set of examples representing
On Feeding Business Systems with Linked Resources from the Web of Data 311
resources that should be linked, then it learns a set of link rules from them. As
a result Genlink returns a set of link rules (all of them aim at linking the same
kind of resources). A quality score ranks the output rules, usually the best is
selected and the rest discarded. Doing this task several times for different kind
of resources a set of link rules is generated, one of these rules is selected to be
improved using our approach and the rest are the supporting rules.
The filter is an ad-hoc component that works as follows: it takes a link rule
and executes it to produce a set of candidate links; then, it analyses the neigh-
bours of the resources involved in each candidate link by boosting the remaining
rules; links in which the corresponding neighbours are similar enough are pre-
served as true positive links while the others are discarded as false positive links.
The selector is an ad-hoc component that works as follows: it takes the filtered
links and the supporting rules used to filter them; then, it performs a voting
strategy regarding how many rules filtered the same link and another voting
strategy regarding how many links were filtered by each rule; finally, a subset of
the filtered links is selected and preserved as the rest are discarded.
Below, we present the details of the filter and selector, plus an ancillary
method that helps measure how similar the neighbours of two resources are.
Example 1. Listing 1 presents two sample datasets that are based on the DBLP
and the NSF datasets. The resources are depicted in greyed boxes whose shapes
encode their classes (i.e., the value of property rdf :type), the properties are rep-
resented as labelled arrows, and the literals are encoded as strings. The genetic
component learns the following link rules in this scenario, which we represent
using a Prolog-like notation for the sake of readability:
r1 : link (A, R) if rdf :type(A) = dblp:Author , rdf :type(R) = nsf :Researcher ,
NA = dblp:name(A), NR = nsf :name(R),
levenstein(lfname(NA ), lfname(NR )) > 0.80.
where levenstein and jaccard denote the well-known string similarity functions
(normalised to interval [0.00, 1.00]), lfname is a function that normalises people’s
names as “last name, first name”, and lowercase is a function that changes a
string into lowercase.
Intuitively, link rule r1 is applied to a resource A of type dblp:Author and a
resource R of type nsf :Researcher ; it computes the normalised Levenshtein simi-
larity between the normalised names of the author and the researcher; if it is greater
than 0.80, then the corresponding resources are linked. Link rule r2 should now be
easy to interpret: it is applied to a resource A of type dblp:Article and a resource
P of type nsf :Paper and links them if the normalised Jaccard similarity amongst
the lowercase version of the title of article A and the title of paper P is greater
than 0.65.
312 A. Cimmino and R. Corchuelo
It is not difficult to realise that link rule r1 links resources dblp:weiwang and
nsf :weiwang1 or dblp:binliu and nsf :binwliu, which are true positive links, but
also dblp:weiwang and nsf :weiwang2 , which is a false positive link. In cases like
this, the only way to make a difference between such resources is to analyse their
neighbours, be them direct (e.g., dblp:weiwang and dblp:article2 ) or transitive
(e.g., nsf :weiwang1 and nsf :paper2 ).
Listing 1 presents the method to filter links. It works on a link rule r , a set
of supporting link rules S , a source dataset D1 , a companion dataset D2 , and
a threshold θ that we explain later. It returns K , which is the subset of links
produced by base link rule r that seem to be true positive links.
The method first initialises K to an empty set, stores the source and the
target classes of the base link rule in sets C1 and C2 , respectively, and the links
that result from applying it to the source and the companion datasets in set L1 .
The main loop then iterates over the set of supporting link rules using vari-
able r . In each iteration, it first computes the sets of source and target classes
involved in link rule r , which are stored in variables C1 and C2 , respectively;
next, it finds the set of paths P1 that connect the source classes in C1 with the
source classes in C1 in dataset D1 ; similarly, it finds the set of paths P2 that
connect the target classes in C2 with the target classes in C2 in dataset D2 . By
On Feeding Business Systems with Linked Resources from the Web of Data 313
path between two sets of classes, we mean a sequence of object properties that
connect resources with the first set of classes to resources with the second set of
classes, irrespective of their direction. Simply put: the idea is to find the way to
connect the resources linked by the base link rule with the resources linked by
the supporting link rule, which is done by the intermediate and the inner loops.
The intermediate loop iterates over the set of pairs of paths (p1 , p2 ) from the
Cartesian product of P1 and P2 . If there is at least a pair of such paths, it then
means that the resources involved in the links returned by base link rule r might
have some neighbours that might be linked by supporting link rule r .
The inner loop iterates over the collection of links (a, b) in set L1 . It first
finds the set of resources A that are reachable from resource a using path p1 in
source dataset D1 and the set of resources B that are reachable from resource b
using path p2 in the companion dataset D2 . Next, the method applies supporting
link rule r to the source and the companion dataset and intersects the resulting
links with A×B so as to keep resources that are not reachable from a or b apart;
the result is stored in set E . It then computes the similarity of sets A and B ;
intuitively, the higher the similarity, the more likely that resources a and b refer
to the same real-world entity. If the similarity is equal or greater than threshold
θ, then link (a, b) is added to set K ; otherwise, it is filtered out. When the main
loop finishes, set K contains the collection of links that involve neighbours that
are similar enough according to the supporting rules.
We do not provide any additional details regarding the algorithms to find
paths or resources since they can be implemented using Dijkstra’s algorithm to
find the shortest paths in a graph. Computing the similarity coefficient is a bit
more involved, so we devote a subsection to this ancillary method below.
Example 2. In our running example, link rule r1 is the base link rule, i.e.,
we are interested in linking authors and researchers, and we use link rule
r2 as the support link rule, i.e., we take their articles and papers into
account. Their source classes are C1 = {dblp:Author } and C1 = {dblp:
Article}, respectively, and their target classes are C2 = {nsf :Researcher }
and C2 = {nsf :Paper }, respectively. Link rule r1 returns the following
links: L1 = {(dblp:weiwang, nsf :weiwang1 ), (dblp:weiwang, nsf :weiwang2 ), (dblp:
binliu, nsf :binwliu)}; note that the first and the third links are true positive
links, but the second one is a false positive link. Link rule r2 returns the fol-
lowing links: L2 = {(dblp:article1 , nsf :paper3 ), (dblp:article2 , nsf :paper2 ), (dblp:
article4 , nsf :paper2 ), (dblp:article5 , nsf :paper5 )}, which are true positive links.
The sets of paths between the source and target classes of r1 and r2 are
P1 = {dblp:writtenBy} and P2 = {nsf :leads, nsf :supports}. Now, the links
in L1 are scanned and the resources that can be reached from the resources
involved in each link using the previous paths are fetched.
Link l1 = (dblp:weiwang, nsf :weiwang1 ) is analysed first. The method finds
A = {dblp:article1 , dblp:article2 , dblp:article3 , dblp:article4 } by following resource
dblp:weiwang through path dblp:writtenBy; similarly, it finds B = {nsf :
paper1 , nsf :paper2 , nsf :paper3 } by following resource nsf :weiwang1 through path
nsf :leads, nsf :supports. Now supporting link rule r2 is applied and the results
314 A. Cimmino and R. Corchuelo
are intersected with A×B so as to keep links that are related to l1 only; the result
is E = {(dblp:article1 , nsf :paper3 ), (dblp:article2 , nsf :paper2 ), (dblp:article4 , nsf :
paper2 )}. Then, the similarity of A and B in the context of E is computed, which
returns 0.67; intuitively, there are chances that l1 is a true positive link.
Link l2 = (dblp:weiwang, nsf :weiwang2 ) is analysed next. The method finds
A = {dblp:article1 , dblp:article2 , dblp:article3 , dblp:article4 } by following resource
dblp:weiwang through path dblp:writtenBy; next, it finds B = {nsf :paper4 }
by following resource nsf :weiwang2 through path nsf :leads, nsf :supports. Now
supporting link rule r2 is applied and the result is intersected with A × B , which
results in E = ∅. In such a case the similarity is zero, which intuitively indicates
that it is very likely that l2 is a false positive link.
Link l3 = (dblp:binliu, nsf :binwliu) is analysed next. The method finds A =
{dblp:article5 } by following resource dblp:binliu through path dblp:writtenBy;
next, it finds B = {nsf :paper5 } by following resource nsf :binwliu through path
nsf :leads, nsf :supports. Now supporting link rule r2 is applied and the result is
intersected with A × B , which results in E = {(dblp:article5 , nsf :paper5 )}. The
similarity is now 1.00, i.e., it is very likely that link l3 is a true positive link.
Assuming that θ is set to, e.g., 0.50, the filterLinks method would return
K = {(dblp:weiwang, nsf :weiwang1 ), (dblp:binliu, nsf :binwliu)}. Note that the
previous value of θ is intended for illustration purposes only because the running
example must necessarily have very little data.
Listing 2 shows our method to compute similarities. Its input consists of sets A
and B , which are two sets of resources, and E , which is a set of links between
them. It returns the Szymkiewicz-Simpson overlapping coefficient, namely:
|A ∩ B |
overlap(A, B ) =
min{|A|, |B |}
elements are subsets of X that are equal according to E , and intersect, which
given two reduced sets of resources X and Y and a set of links E returns the
intersection of X and Y according to E . Their definitions are as follows:
reduce(X , E ) = {W | W ∝ W ⊆ X ∧ W × W ⊆ E )}
intersect(X , Y , E ) = {W | W ∝ W ⊆ X ∧ ∃W : W ⊆ Y ∧ W × W ∈ E }
where X ∝ φ denotes the maximal set X that fulfils predicate φ, that is:
X ∝ φ ⇔ φ(X ) ∧ ( ∃X : X ⊆ X ∧ φ(X ))
The method to compute similarities then works as follows: it first reduces the
input sets of resources A and B according to the set of links E ; it then computes
the intersection of both reduced sets; finally, it computes the similarity using
Szymkiewicz-Simpson’s formula on the reduced sets.
Example 3. Analysing link l1 = (dblp:weiwang, nsf :weiwang1 ) results in sets A =
{dblp:article1 , dblp:article2 , dblp:article3 , dblp:article4 }, B = {nsf :paper1 , nsf :
paper2 , nsf :paper3 }, and E = {(dblp:article1 , nsf :paper3 ), (dblp:article2 , nsf :
paper2 ), (dblp:article4 , nsf :paper2 )}. If E is interpreted as an equality relation
by computing its reflexive, symmetric, transitive closure, then it is not difficult
to realise that dblp:article2 and dblp:article4 can be considered equal, because
dblp:article2 is equal to nsf :paper2 and nsf :paper2 is equal to dblp:article4 . Thus,
set A is reduced to A = {{dblp:article1 }, {dblp:article2 , dblp:article4 }, {dblp:
article3 }} and set B is reduced to B = {{nsf :paper1 }, {nsf :paper2 }, {nsf :
paper3 }}. As a conclusion, |A ∩ B | = |{{dblp:article1 , nsf :paper3 }, {dblp:
article2 , dblp:article4 , nsf :paper2 }}| = 2, |A | = 3, and |B | = 3; so the simi-
larity is 0.67.
When link l2 = (dblp:weiwang, nsf :weiwang2 ) is analysed, A = {dblp:
article1 , dblp:article2 , dblp:article3 , dblp:article4 }, B = {nsf :paper4 }, and E = ∅.
Since the equality relation E is then empty, the similarity is zero because the
intersection between the reductions of sets A and B is empty.
In the case of link l3 = (dblp:binliu, nsf :binwliu), A = {dblp:article5 }, B =
{nsf :paper5 }, and E = {(dblp:article5 , nsf :paper5 )}. As a conclusion, |A ∩ B | =
|{{dblp:article5 , nsf :paper5 }}| = 1, |A | = 1, and |B | = 1, where A and B
denote, respectively, the reductions of sets A and B ; so the similarity is 1.00.
which given a link (a, b) returns the set of link rules that have selected it. The
previous functions are formally defined as follows:
The method to select links first projects the set of correspondences K onto
the set of links M and the set of supporting link rules S . It then computes g as
a percentage, according to μ, of the maximum number of link rules that have
selected each candidate link; it also computes d as a percentage, according to
ρ, of the maximum number of links that a support link rule has selected as
candidates. Next, it computes the set of top links U as the set of links in M that
have been selected by at least g link rules; similarly, it computes the set of top
link rules V as the set of link rules in S that have selected at least d links. The
resulting set of links L is computed as the subset of links in M that are either
top links or have been selected by top link rules.
4 Experimental Analysis
In this section, we first describe our experimental environment and then comment
on our results.
Computing facility: We run our experiments on a virtual computer that was
equipped with four Intel Xeon E5-2690 cores at 2.60 GHz and 4 GiB of RAM.
The operating system was CentOS Linux 7.3.
On Feeding Business Systems with Linked Resources from the Web of Data 317
Prototype:1 We implemented our proposal with Java 1.8 and the following com-
ponents: the Genlink implementation from the Silk Framework 2.6.0 to generate
link rules, Jena TDB 3.2.0 to work with RDF data, ARQ 3.2.0 to work with
SPARQL queries, and Simmetrics 1.6.2, SecondString 2013-05-02, and JavaS-
tringSimilarity 1.0.1 to compute string similarities.
Evaluation datasets:2 We used the following datasets: DBLP, NSF, BBC,
DBpedia, IMDb, RAE, Newcastle, and Rest. We set up the following scenarios:
(1) DBLP–NSF, which focuses on the top 100 DBLP authors and 130 princi-
pal NSF researchers with the same names; (2) DBLP–DBLP, which focuses on
the 9 076 DBLP authors with the same names who are known to be different
people; (3) BBC–DBpedia, which focuses on 691 BBC movies and 445 DBpedia
films that have similar titles; (4) DBpedia–IMDb, which focuses on 96 DBpedia
movies and 101 IMDb films that have similar titles; (5) RAE–Newcastle, which
focuses on 108 RAE publications and 98 Newcastle papers that are similar; and
(6) Rest–Rest, which focuses on 113 and 752 restaurants published by OAEI. The
scenarios DBLP-DBLP, Rest-Rest and RAE-Newcastle rely on datasets from the
literature which have not being curated nor modified, the gold standard of the
former two is released with the data and the rest have being extracted from
LinkLion3 . BBC-DBpedia, DBLP-NSF and DBPEDIA-IMDb uses a subset of
the original datasets due to their size, but again data has not being modified
and gold standard have being generated by hand thanks to the size reduction.
Baseline: Our baseline was the Genlink implementation from the Silk Frame-
work 2.6.0, which is a state-of-the-art genetic programming to learn link rules.
Measures: On the one hand we explored a large portion of the parameter space
to establish optimal values for θ, μ, ρ in each scenario. On the other hand we
measured the number of links returned by each proposal (Links), precision (P ),
recall (R), and the F1 score (F1 ) using 2-fold cross validation. We also computed
the normalised differences in precision (ΔP ), recall (ΔR), and F1 score (ΔF1 ),
which measure the ratio from the difference found between the baseline and our
proposal and the maximum possible difference for each performance measure.
Results: The results are presented in Listing 2. We analyse them regarding
precision, recall, and the F1 score below.
The results regarding precision clearly show that our technique improves
the precision of the rules learnt by GenLink in every scenario. In average, the
difference in precision is 77%. The worst improvement is 57% in the DBpedia–
IMDb scenario since these datasets are clearly unbalanced: movies in DBpedia
have related actors, writers, or directors that IMDb may not contain at all; this
obviously makes it impossible for our proposal to find enough context to make a
decision. The best improvement is 100% in the DBLP–DBLP scenario since there
1
The prototype is available at https://github.com/AndreaCimminoArriaga/
TeidePlus.
2
The datasets are available at https://goo.gl/Pu76SU.
3
http://www.linklion.org/.
318 A. Cimmino and R. Corchuelo
are 9 069 authors with very similar names, which makes it almost impossible for
GenLink to generate rules with good precision building solely on the attributes of
the resources. The p-value computed by Iman-Davenport’s test is 0.00; since it is
clearly smaller than the standard confidence level, we can interpret it as a strong
indication that there is enough evidence in our experimental data to confirm the
hypothesis that our proposal works better than the baseline regarding precision.
The normalised difference of recall ΔR shows that our proposal generally
retains the recall of the link rules learnt by GenLink, except in the DBpedia–
IMDb and the Rae–Newcastle scenarios. The problem with the previous scenarios
was that there are many incomplete resources, that is, many resources without
neighbours. For instance, there are 43 papers in the Newcastle dataset that
are not related to any authors. The incompleteness of data has also a negative
impact on the recall of the base link rules. In our prototype, we are planning on
implementing a simple check to identify incomplete resources so that the links
in which they are involved are not discarded as false positives, but identified as
cases on which our proposal cannot make a sensible decision. Iman-Davenport’s
test returns 0.25 as the corresponding p-value; since it is larger than the standard
confidence level, it may be interpreted as a strong indication that the differences
in recall found in our experiments are not statistically significant. In other words,
the cases in which data are that incomplete do not seem to be common-enough
for them to have an overall impact on our proposal.
We also studied ΔF1 , which denotes the normalised difference in F1 score.
Note that it is 58% in average and that the corresponding Iman-davenport’s
p-value is 0.00, which can be interpreted as a strong indication that the differ-
ence is significant from a statistical point of view. Overall, this result confirms
that our proposal helps improve precision without degrading recall and that the
improvement in precision is enough for the F1 score to improve significantly.
5 Conclusions
References
1. Alili, H., Belhajjame, K., Grigori, D., Drira, R., Ghezala, H.H.B.: On enriching
user-centered data integration schemas in service lakes. In: Abramowicz, W. (ed.)
BIS 2017. LNBIP, vol. 288, pp. 3–15. Springer, Cham (2017). https://doi.org/10.
1007/978-3-319-59336-4 1
2. Ananthakrishna, R., Chaudhuri, S., Ganti, V.: Eliminating fuzzy duplicates in data
warehouses. In: VLDB, pp. 586–597 (2002)
3. Bhattacharya, I., Getoor, L.: Collective entity resolution in relational data. TKDD
1(1), 1–36 (2007)
4. Bizer, C., Heath, T., Berners-Lee, T.: Linked data: principles and state of the art.
In: WWW (Invited talks) (2008). https://www.w3.org/2008/Talks/WWW2008-
W3CTrack-LOD.pdf
5. Cruz, I.F., Antonelli, F.P., Stroe, C.: AgreementMaker: efficient matching for large
real-world schemas and ontologies. PVLDB 2(2), 1586–1589 (2009)
6. Dong, X., Halevy, A.Y., Madhavan, J.: Reference reconciliation in complex infor-
mation spaces. In: SIGMOD, pp. 85–96 (2005)
7. Hernández, M.A., Stolfo, S.J.: The merge/purge problem for large databases. In:
SIGMOD Conference, pp. 127–138 (1995)
8. Holub, M., Proksa, O., Bieliková, M.: Detecting identical entities in the semantic
web data. In: Italiano, G.F., Margaria-Steffen, T., Pokorný, J., Quisquater, J.-J.,
Wattenhofer, R. (eds.) SOFSEM 2015. LNCS, vol. 8939, pp. 519–530. Springer,
Heidelberg (2015). https://doi.org/10.1007/978-3-662-46078-8 43
9. Hu, W., Qu, Y.: Falcon-AO: a practical ontology matching system. J. Web Sem.
6(3), 237–239 (2008)
10. Huber, J., Sztyler, T., Nößner, J., Meilicke, C.: CODI: combinatorial optimization
for data integration. In: OM, pp. 134–141 (2011)
11. Isele, R., Bizer, C.: Learning expressive linkage rules using genetic programming.
PVLDB 5(11), 1638–1649 (2012)
12. Isele, R., Bizer, C.: Active learning of expressive linkage rules using genetic pro-
gramming. J. Web Sem. 23, 2–15 (2013)
13. Jiménez-Ruiz, E., Cuenca Grau, B.: LogMap: logic-based and scalable ontology
matching. In: Aroyo, L., et al. (eds.) ISWC 2011. LNCS, vol. 7031, pp. 273–288.
Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-25073-6 18
14. Kalashnikov, D.V., Mehrotra, S., Chen, Z.: Exploiting relationships for domain-
independent data cleaning. In: SDM, pp. 262–273 (2005)
15. Köpcke, H., Rahm, E.: Frameworks for entity matching: a comparison. Data Knowl.
Eng. 69(2), 197–210 (2010)
16. Lacoste-Julien, S., Palla, K., Davies, A., Kasneci, G., Graepel, T., Ghahramani,
Z.: SIGMa: Simple greedy matching for aligning large knowledge bases. In: KDD,
pp. 572–580 (2013)
17. Monge, A.E., Elkan, C.: The field matching problem: algorithms and applications.
In: KDD, pp. 267–270 (1996)
18. Ngomo, A.C.N., Auer, S.: LIMES: a time-efficient approach for large-scale link
discovery on the Web of data. In: IJCAI, pp. 2312–2317 (2011)
19. Ngomo, A.-C.N., Lyko, K.: EAGLE: efficient active learning of link specifications
using genetic programming. In: Simperl, E., Cimiano, P., Polleres, A., Corcho, O.,
Presutti, V. (eds.) ESWC 2012. LNCS, vol. 7295, pp. 149–163. Springer, Heidelberg
(2012). https://doi.org/10.1007/978-3-642-30284-8 17
320 A. Cimmino and R. Corchuelo
20. Nikolov, A., d’Aquin, M., Motta, E.: Unsupervised learning of link discovery con-
figuration. In: Simperl, E., Cimiano, P., Polleres, A., Corcho, O., Presutti, V. (eds.)
ESWC 2012. LNCS, vol. 7295, pp. 119–133. Springer, Heidelberg (2012). https://
doi.org/10.1007/978-3-642-30284-8 15
21. Rastogi, V., Dalvi, N.N., Garofalakis, M.N.: Large-scale collective entity matching.
PVLDB 4(4), 208–218 (2011)
22. Soru, T., Ngomo, A.C.N.: A comparison of supervised learning classifiers for link
discovery. In: SEMANTICS, pp. 41–44 (2014)
23. Szekely, P., et al.: Building and using a knowledge graph to combat human traf-
ficking. In: Arenas, M., et al. (eds.) ISWC 2015. LNCS, vol. 9367, pp. 205–221.
Springer, Cham (2015). https://doi.org/10.1007/978-3-319-25010-6 12
24. Volz, J., Bizer, C., Gaedke, M., Kobilarov, G.: Silk: a link discovery framework for
the web of data. In: LDOW (2009)
Increasing the Explanatory Power of Investor
Sentiment Analysis for Commodities
in Online Media
1 Introduction
In particular, prior research (e.g., [2, 3, 6]) often uses generic or stock-specific diction‐
aries (e.g., Loughran and McDonald [9]) without term weighting. This practice is in
stark contrast to the important role of term weighting in automatic text retrieval [10].
Thus, explanatory power might be impaired, which then negatively affects the interpre‐
tation of findings reported in empirical studies. We address this gap in the literature by
building on machine learning-based sentiment classification research [11–14].
The objective of this research is to evaluate classifier designs that complement plain
dictionary-based methods used in prior research with machine learning (ML). We report
on a comprehensive study using nine different classifiers and novel domain-specific
corpora. The main finding is that such designs achieve up to 25% higher explanatory
power regarding returns. Statistical tests show that the dictionary of Loughran and
McDonald [15], used in prior research (e.g., [2]), is detrimental to high explanatory
power for commodity returns. Our findings suggest that investor sentiments of other
dictionaries (i.e., Henry [16]) should be used in combination with machine learning-
based term weighting when studying the relationship between investor sentiments and
market variables. To enable such research in the first place, we have created a novel
corpus of labelled online media articles for training machine learning classifiers. Further,
we provide datasets of online media articles on commodities; these datasets can be used
for future empirical studies and are freely available – unlike the TRN. In contrast to the
TRN, our dataset also includes blog articles, which are more subjective than news arti‐
cles, and thus carry more sentiment. Furthermore, we study not only individual commod‐
ities but also the asset class of commodities.
Our paper proceeds as follows. We first discuss the theoretical background to our
research. Then, we describe our methodology for the design of classifiers for investor
sentiment in online media texts. We report on our evaluation of classifier designs.
Finally, we discuss the implications and limitations of our study before concluding the
paper.
2 Background
This section describes designs for classifiers of investor sentiments in online media texts
using dictionary-based approaches that prevail in the extant IS literature and more
advanced machine learning approaches, which allow for term weighting.
First, we designed dictionary-based investor sentiment classifiers: We used diction‐
aries of Henry [16], and Loughran and McDonald [9], separately and in combination of
both. The dictionary by Loughran and McDonald [9] was developed in relation to
companies’ 10-K reports. The dictionary is publicly available (http://www3.nd.edu/
~mcdonald/Word_Lists.html) and contains 2,329 negative words and 354 positive
words. The dictionary by Henry [16] was developed in the context of earnings press
releases and contains 85 negative words and 104 positive words. To combine both
dictionaries, we use the union of both sets, which results in a set of 400 positive and
2,366 negative words. According to Feuerriegel and Neumann [2], the dictionary of
Henry [16] is better suited for sentiment identification and studying relationships to
returns than the dictionary of Loughran and McDonald [9].
Second, we designed dictionary-based investor sentiment classifiers with prepro‐
cessing: Extending the above approach for dictionary-based investor sentiment classifica‐
tion, we replicated the approach of Feuerriegel and Neumann [2] as closely as possible: (1)
Stop words were removed using a list of 571 words from Lewis et al. [25], cited in Feuer‐
riegel and Neumann ([2], p. 6). (2) Negations were taken into account in the sense of
inverting the polarity of positive or negative words from the dictionary used. Following
Feuerriegel and Neumann ([2], p. 6) and Dadvar et al. [26], cited in Feuerriegel and
Neumann ([2], p. 6), inversions of sentiment polarity words were conducted within a
sentence in the scope of three words after the word “no” and for all words in a sentence after
an occurrence of one of the following words or phrases: “rather”, “hardly”, “couldn’t”,
“could not”, “wasn’t”, “was not”, “didn’t”, “did not”, “wouldn’t”, “would not”, “shouldn’t”,
“should not”, “weren’t”, “were not”, “don’t”, “do not”, “doesn’t”, “does not”, “haven’t”,
“have not”, “won’t”, “will not”, “hadn’t”, “had not”, and “never”. (3) The Porter stemmer
was used to stem each word in all dictionaries, word lists, and texts from online media used
in this research. (4) We did not use synonym merging as proposed by Feuerriegel and
Neumann ([2], p. 6) because the list of synonyms merged is not disclosed in their paper.
Third, our machine learning-based investor sentiment classifier design follows
standard approaches and configurations for textual investor sentiment classification [13,
27]. We used the whole online media texts, represented by unigrams, for classifier
training on the document level. Higher level n-grams such as bigrams were not consid‐
ered because the purpose of this classifier design is to evaluate machine learning-based
term weighting of sentiment dictionaries containing only single words. Thus, we restrict
the feature set to unigrams that are present in a dictionary of sentiment-laden words with
respect to the classes positive and negative. The dictionaries used are the ones also used
in the plain dictionary-based approaches above to allow for comparison. Using the
dictionaries for feature selection seems reasonable because the complete feature set of
unigrams is most likely not relevant for investor sentiment classification and explaining
relationships to commodity returns. This assumption was corroborated by a pre-experi‐
ment in our evaluation setup that did not improve explanatory power of investor
Increasing the Explanatory Power of Investor Sentiment Analysis 325
4 Evaluation
This section reports on evaluating the explanatory power of classifier designs in terms
of correlation of classified investor sentiments regarding commodity returns. Figure 1
depicts the full process as well as the corpora and datasets used: (1) Texts are collected
from feeds and classified by an asset class classifier to filter the ones related to commod‐
ities, (2) a commodity classifier categorizes the remaining texts into specific commod‐
ities, (3) investor sentiment classifiers are trained on a novel investor sentiment corpus
for categorizing the texts into positive or negative classes, (4) the classifiers are applied
to the texts in the dataset of commodity texts and an aggregated daily score is calculated,
and (5) the correlation between investor sentiment and commodity returns is determined
as our main evaluation metric.
Fig. 1. The full process of training, applying, and evaluating investor sentiment classifiers.
We created a novel investor sentiment corpus for training and evaluating investor senti‐
ment ML classifiers. The corpus consists of texts from online media about commodities
326 A. Klein et al.
that were manually categorized on the document level into a positive or negative investor
sentiment class. To identify suitable texts, online articles related to finance and
economics were manually retrieved using Google’s search engine with a custom search
filter such that the searched websites were the same as the websites used for the evalu‐
ation of the explanatory power of investor sentiments regarding commodity returns. For
representing different market phases properly in our corpus, articles were sampled over
several years. Commodity texts discussing multiple commodities (e.g., gold, crude oil)
each with different investor sentiment classes were not considered for the corpus. The
reason is that a uniform document level sentiment could not be assigned for the training
of the document level classifiers. In our experience, such texts occur seldom. Thus, the
classifiers should work reasonably well on real world datasets. Altogether, the investor
sentiment corpus consists of 309 positive and 212 negative online media texts. Consid‐
ering the laborious task of annotation, the corpus can be regarded as rather large.
and news articles. It must be noted, however, that the posting activity of the news feeds
was in general higher than that of the blogs. The list of feeds was created by reviewing
the web sites used in the EU-project FIRST [32], as well as by reviewing rankings of
the top financial and economic news web sites and weblogs. Reviewed lists of websites
include “50 Blogs Every Serious Trader Should Read” (http://traderhq.com/best-
trading-blogs/) and “Top 100 Futures Trading Blogs” (http://commodityhq.com/
investor-resources/top-100-futures-trading-blogs/).
As a first step to obtain daily investor sentiments, all commodity texts were classified
by the investor sentiment classifiers on the document level. We aggregated the investor
sentiment classifications to a daily score: ln((1 + p)/(1 + n)) [33], where p is the number
of positive text documents and n is the number of negative text documents referring to
a particular commodity on a particular day (i.e., from market close to market close). On
Mondays and after a bank holiday, the score incorporates also texts of the preceding
weekend or bank holiday.
For relating investor sentiments to returns, commodity log returns, which are based
on the daily closing prices of exchange-traded funds (ETFs), and a buy-and-hold strategy
were used. The tickers of the ETFs used in this work are (1) DBC for the commodities
asset class, (2) GLD for gold as an asset, and (3) USO for crude oil.
N = 284 for the commodities asset class, gold, and crude oil.
Table 1 presents our results. The correlation coefficients of all classifiers are positive
and statistically significantly different from zero at the 1% level for the commodities
asset class, gold, and crude oil, except for (1) all dictionary-based classifiers that use
only the Loughran and McDonald dictionary [9] and (2) the machine learning-based
classifiers that use only the Loughran and McDonald dictionary [9] for dictionary-based
feature selection.
Table 1 provides the following findings: For individual commodities, a combination
of the dictionaries of Henry (H) [16] and Loughran and McDonald (LMcD) [9] with
machine learning-based term weighting (ML-Dict.) observed the highest correlation
with returns. For the commodities asset class, machine learning in combination with the
dictionary of Henry [16] for feature selection observed the highest correlation. Thus, for
high explanatory power of the classified investor sentiments for returns, the dictionaries
should be complemented with machine learning-based term weighting.
Consistent with using ML for weighting the terms of dictionaries, the dictionary of
LMcD used on its own for dictionary-based classification observed the lowest correla‐
tion with returns among all dictionaries. This finding is confirmed by a one-tailed Steig‐
er’s Z-test [34]. That is, for the commodities asset class, the correlation with returns of
investor sentiment classifiers using ML with the dictionary of H or a combination of H
and LMcD is statistically significantly higher than using ML with the dictionary of
328 A. Klein et al.
Table 1. Accuracy of different classifiers of investor sentiment and correlation between classified
investor sentiments and commodity returns
Classifier of Accuracy and Correlation coefficients
investor micro-avg. F1 Commodities Gold Crude oil
sentiment (%)
Dict. (LMcD) 55.1 0.075 0.139* 0.127*
Dict. P. (LMcD) 52.4 0.118* 0.221*** 0.112
Dict. (H) 72.9 0.177** 0.222*** 0.200***
Dict. P. (H) 71.8 0.168** 0.170** 0.187**
Dict. (H, LMcD) 69.5 0.180** 0.233*** 0.194***
Dict. P. (H, 66.6 0.173** 0.209*** 0.170**
LMcD)
ML- Dict. 70.8 0.111 0.135* 0.158**
(LMcD)
ML- Dict. (H) 74.6 0.226*** 0.207*** 0.205***
ML- Dict. (H, 77.7 0.205*** 0.266*** 0.229***
LMcD)
Notes: Accuracy equals micro-averaged F1-measure (i.e., a weighted combination of both, recall and precision), thus
providing an overall measure of classification performance. Correlations are Spearman’s rank correlations between classified
investor sentiments and daily commodity returns. Correlations are statistically significant at the *** = 0.1%, ** = 1%, * = 5%
level. The length of the time series is N = 284. Abbreviations are: ML-Dict = machine learning with dictionary-based feature
selection; Dict. = dictionary-based with dictionaries of H = Henry [16] and LMcD = Loughran and McDonald [9]; Dict. P.
= dictionary-based with preprocessing [2].
LMcD alone (p = 0.024 and p = 0.036 respectively). Further, the correlation with returns
is also statistically significantly higher than using a plain dictionary-based approach
using LMcD on its own without preprocessing (p = 0.017 and p = 0.011 respectively)
or with preprocessing (p = 0.063 and p = 0.042 respectively). The statistical evidence
for individual commodities points into the same direction but is a bit weaker. Thus, the
dictionary of LMcD should not be used stand-alone for high explanatory power
regarding returns of commodities.
Another insight of the analysis in Table 1 is that adding pre-processing to the
dictionary-based classifiers in most cases did not increase the correlation with returns.
Statistical evidence, again using the one-tailed Steiger’s Z-test, supports this finding for
the dictionary of H and also LMcD regarding the correlation to gold returns (p = 0.039
and p = 0.035).
Consistent with above findings, the ML-based sentiment classifier observing the
highest explanatory power for gold and crude oil returns also has the highest overall
classification performance (in terms of accuracy and F1 measure) regarding the human-
made gold standard corpus. By using ML-based classifiers, the accuracy can be
improved considerably compared to plain dictionary-based approaches used in previous
research. Furthermore, we did not observe that preprocessing methods [2] help to
improve classifiers’ accuracy.
Increasing the Explanatory Power of Investor Sentiment Analysis 329
5 Discussion
5.1 Contributions
Our study has several implications for future research. First, online media texts should
be investigated further in terms of explanatory or predictive effects on market variables.
Second, complementing dictionaries with machine-learning based term weighting
should be investigated further. To this respect, the dictionary of Henry [16] and a
combination of the dictionaries of Henry [16] and Loughran and McDonald [9] seems
to be a good starting point. However, creating dictionaries more fitted to the vocabulary
of commodity-related texts seems worthwhile because the dictionaries of Henry [16]
and Loughran and McDonald [9] relate to earnings press releases and company reports,
respectively. For creating a commodity-specific dictionary, corpus-based methods have
already been proposed [35] and could be applied. Furthermore, pre-processing of texts
like stemming, accounting for negations, and removing stop words does not systemat‐
ically improve the explanatory power of investor sentiment classifiers regarding returns.
To fuel research, we provide our datasets of commodity texts under https://wi2.uni-
hohenheim.de/analytics. Thus, fellow researchers can also investigate machine learning-
based methods and apply them in empirical studies.
330 A. Klein et al.
5.4 Limitations
The limitations of our research are noted as follows. First, the time span covered by our
dataset is limited to about 13 months. A larger dataset would allow for a more extensive
evaluation and increase the external validity of our findings. However, this is a common
limitation of related studies that use their own dataset instead of a commercial one [14,
36]. Second, the size of the corpus of human annotated texts could be larger to improve
the validity of the trained investor sentiment classifiers. But again, the size of related
corpora is similar (e.g., [37]) because human annotation is a costly and time-consuming
task. Third, the dictionaries used in this research are not specific to commodities.
Publicly available sentiment dictionaries refer to stocks or are not even specific to finance
(e.g., the General Inquirer’s Harvard IV dictionary). We plan to overcome this limitation
with our own dictionary in a more extensive study. Fourth, we used SVM for machine
learning based on previous findings that show SVM to be the best performing algorithm
for sentiment text classification [13, 29]. Nevertheless, other classifiers, e.g., ensemble
or deep learning classifiers, could be applied. Finally, our research does not measure
actual utility of investor sentiment for selecting and timing commodities in an investor’s
portfolio. A portfolio simulation will be required in future work to address this issue.
6 Conclusion
Prior IS and finance research has studied investor sentiment from news using classifiers
of commercial vendors or dictionary-based classifiers, which do not weigh the impor‐
tance of terms regarding the problem of explaining returns. Nevertheless, weighing
terms or other features is standard practice in machine learning-based classification (e.g.,
classification of spam). Thus, the explanatory power of classifiers that do not weigh
terms might be impaired, directly affecting the results of empirical studies. To address
this gap in the literature, this research evaluates nine classifier designs for better under‐
standing the impact of classifier design on the explanatory power regarding returns. We
find that complementing prevailing dictionary-based classifiers with machine learning
for automatically weighing terms increases the explanatory power for the asset class of
commodities by a 25% stronger correlation with returns and by a 14% stronger
Increasing the Explanatory Power of Investor Sentiment Analysis 331
correlation for individual commodities (of gold and crude oil). This result should help
to improve investors’ decision making and also help other IS researchers in studying
relationships between investor sentiments and market variables. Furthermore, we
corroborate prior findings relating to a commercial news-only dataset to hold also for
more opinionated online media texts in the sense that a statistically significantly positive
relationship exists between investor sentiment from online media texts and commodity
returns. Thus, freely available online media texts seem to be worthwhile to be studied
by investors. For fuelling further IS research in the field, we provide datasets of online
media texts.
References
16. Henry, E.: Are investors influenced by how earnings press releases are written? J. Bus.
Commun. 45(4), 363–407 (2008)
17. Brown, G.W., Cliff, M.T.: Investor sentiment and asset valuation. J. Bus. 78(2), 405–440
(2005)
18. Baker, M., Wurgler, J.: Investor sentiment and the cross-section of stock returns. J. Finan.
61(4), 1645–1680 (2006)
19. Pan, J., Poteshman, A.M.: The information in option volume for future stock prices. Rev.
Finan. Stud. 19(3), 871–908 (2006)
20. Joachims, T.: Text categorization with support vector machines: learning with many relevant
features. In: Nédellec, C., Rouveirol, C. (eds.) ECML 1998. LNCS, vol. 1398, pp. 137–142.
Springer, Heidelberg (1998). https://doi.org/10.1007/BFb0026683
21. Klein, A., Altuntas, O., Riekert, M., Dinev, V.: Combined approach for extracting object-
specific investor sentiment from weblogs. In: Proceedings of the 11th International
Conference on Wirtschaftsinformatik (WI 2013), Leipzig, Germany, pp. 691–705 (2013)
22. Shleifer, A., Summers, L.H.: The noise trader approach to finance. J. Econ. Perspect. 4(2),
19–33 (1990)
23. Shleifer, A., Vishny, R.W.: The limits of arbitrage. J. Finan. 52(1), 35–55 (1997)
24. Barberis, N., Thaler, R.: A survey of behavioral finance. In: Handbook of the Economics of
Finance, vol. 1, pp. 1053–1128 (2003)
25. Lewis, D.D., Rose, T.G., Li, F., Yang, Y., Rose, T.G., Li, F.: RCV1: a new benchmark
collection for text categorization research. J. Mach. Learn. Res. 5, 361–397 (2004)
26. Dadvar, M., Hauff, C., de Jong, F.: Scope of negation detection in sentiment analysis. In:
Proceedings of the Dutch-Belgian Information Retrieval Workshop, pp. 16–20 (2011)
27. Riekert, M., Leukel, J., Klein, A.: Online media sentiment: understanding machine learning-
based classifiers. In: Proceedings of the 24th European Conference on Information Systems
(ECIS 2016), Istanbul, Turkey (2016)
28. Fan, R., Chang, K., Hsieh, C.: LIBLINEAR: a library for large linear classification. J. Mach.
Learn. Res. 9, 1871–1874 (2008)
29. Tang, H., Tan, S., Cheng, X.: A survey on sentiment detection of reviews. Expert Syst. Appl.
36(7), 10760–10773 (2009)
30. Hastie, T., Tibshirani, R., Friedman, J.: The Elements of Statistical Learning: Data Mining,
Inference, and Prediction, Second Edition. Springer Series in Statistics, 2nd edn. Springer,
Heidelberg (2009). https://doi.org/10.1007/978-0-387-84858-7
31. Leopold, E., Kindermann, J.: Text categorization with support vector machines. How to
represent texts in input space? Mach. Learn. 46(1–3), 423–444 (2002)
32. Lombardi, P., Aprile, G., Gsell, M., Winter, A., Reinhardt, M., Queck, S.: D1.1 definition of
market surveillance, risk management and retail brokerage usecases (2011)
33. Antweiler, W., Frank, M.Z.: Is all that talk just noise? The information content of Internet
stock message boards. J. Finan. 59(3), 1259–1294 (2004)
34. Steiger, J.H.: Tests for comparing elements of a correlation matrix. Psychol. Bull. 87(2), 245–
251 (1980)
35. Pröllochs, N., Feuerriegel, S., Neumann, D.: Generating domain-specific dictionaries using
Bayesian learning. In: Proceedings of the 23rd European Conference on Information Systems
(ECIS 2015), pp. 1–14 (2015)
36. Bollen, J., Mao, H., Zeng, X.: Twitter mood predicts the stock market. J. Comput. Sci. 2(1),
1–8 (2011)
37. Das, S.R., Chen, M.Y.: Yahoo! For Amazon: sentiment extraction from small talk on the
Web. Manage. Sci. 53(9), 1375–1388 (2007)
Comparative Analysis of the Informativeness
and Encyclopedic Style of the Popular
Web Information Sources
1 Introduction
Very often the decision making depends on the information that is found in various
Internet sources. Enterprises increasingly use various external big data sources in order
to extract and integrate information into their own business systems [1]. Meanwhile, the
Internet is flooded with meaningless blogs, computer-generated spam, and texts that
convey no useful information for business purposes. Firms, organizations, and indi‐
vidual users can publish texts that have different purposes. The quality of information
about the same subject in these texts can greatly depend on different aspects. For business
purposes, however, organizations and individual users need a condensed presentation
of material that identifies the subject accurately, completely and authentically. At the
same time, the subject matter should be displayed in a clear and understandable manner.
In other words, the decision making should prefer texts of an encyclopedic style,
which is directly related to the informativeness concept i.e. the amount of useful
2 Related Work
Nowadays, the problem of determining informativeness of a text has become one of the
most important tasks of NLP. In recent years, many articles devoted to the solution of
the problem have appeared.
Usually, informativeness of a text is considered at three levels of the linguistic
system: (1) at the sentence level, (2) at the level of the discourse and (3) at the level of
the entire text. In traditional linguistics, the definition of the sentence informativeness
is based on the division of an utterance into two parts - the topic (or theme) and the
comment (or rheme, rema). [6]. At the discourse level, the traditional approach involves
the anchoring of information units or events to descriptives and interpretives within a
narrative frame [7].
1
http://u.cs.biu.ac.il/~koppel/BlogCorpus.htm.
Comparative Analysis of the Informativeness and Encyclopedic Style 335
3 Methodology
We argue that the encyclopedic style of an article can be represented explicitly by the
various linguistic means and semantic characteristics in the text. We have defined four
336 N. Khairova et al.
possible semantic types of facts in a sentence, which, in our opinion, might help to
determine the encyclopedicness of the text. Figure 1 shows the structural scheme for
distinguishing four types of facts from simple English sentences.
Fig. 1. Structural scheme for distinguishing four types of facts from simple English sentences:
subj-fact, obj-fact, subj-obj fact and complex fact.
We called the first type of facts as subj-fact. We defined this type of the facts in an
English sentence as the smallest grammatical clause that includes a verb and a noun. In
this case, the verb represents Predicate of an action and the noun represents Subject2 of
an action. According to our model of fact extraction from English texts [3], the semantic
relations that denote the Subject of the fact can be determined as the following logical-
linguistic equation:
γ1 (z, y, x, m, p, f, n) = yout
( ((f
can
∨ fmay ∨) fmust ∨ fshould ∨ fcould ∨ fneed ∨ fmight ∨
f would
∨ f )(n ∨ n ) p ∨ p ∨ pIII xf mout ∨ (xl (mis ∨ mare ∨ mhavb mhasb ∨
out not out I ed
(1)
m hadb
∨ mwas ∨ mwere ∨ mbe ∨ mout )zby ),
where the subject variable zprep defines the syntactic feature of the preposition after the
verb in English phrases; the subject variable y (yap ∨ yaps ∨ yout = 1) defines whether
there is an apostrophe in the end of the word; the subject variable x defines the position
of the noun with respect to the verb; the subject variable m defines whether there is a
form of the verb “to be” in the phrase and the subject variable p defines the basic forms
of the verb in English.
Additionally, in this study, we have appended two subject variables f and n to account
for modality and negation. The subject variable f defines the possible forms of modal
verbs:
2
We use ‘Predicate’, ‘Subject’ and ‘Object’ with the first upper-case letters to emphasize the
semantic meaning of words in a sentence.
Comparative Analysis of the Informativeness and Encyclopedic Style 337
Using the subject variable n we can take into account the negation in a sentence:
nnot ∨ nout = 1
Definition 1. The subj-fact in an English sentence is the smallest grammatical clause
that includes a verb and a noun (or personal pronoun) that represents the Subject of the
fact according to the Eq. (1).
The Object of a verb is the second most important argument of a verb after the subject.
We can define grammatical and syntactic characteristics of the Object in English text
by the following logical-linguistic equation:
γ2 (z, y, x, m, p, f, n) = yout (nnot ∨ nout )(fcan ∨ fmay ∨ fmust ∨ fshould ∨ fcould ∨ fneed ∨
( ) ( )
fmight ∨ fwould ∨ fout )(zout x1 mout pI ∨ ped ∨ pIII ∨ xf zout ∨ zby (mis ∨ mare ∨ (2)
m havb
∨m hasb
∨m hadb
∨m
was
∨m
were
∨ m ∨ m )(p ∨ p ),
be out ed III
Definition 2. The obj-fact in an English sentence is the smallest grammatical clause that
includes a verb and a noun (or personal pronoun) representing the Object of the fact
according to the conjunction of grammatical features in Eq. (2).
The third group of facts includes main clauses, in which one noun has to play the
semantic role of the Subject and the other has to be the Object of the fact.
Definition 3. The subj-obj fact in an English sentence is the main clause that includes
a verb and two nouns (or personal pronouns), one of them represents the Subject of the
fact accordance with the Eq. (1) and the other represents the Object of the fact accordance
with the Eq. (2).
We also defined the complex fact in texts as a grammatical simple English sentence
that includes a verb and a few nouns (or personal pronouns). In that case, the verb also
represents Predicate, but among nouns, one of these has to play the semantic role of the
Subject, other has to be the Object and the others are the attributes of the action.
Definition 4. The complex fact in an English sentence is the simple sentence that
includes a verb and a few nouns (or personal pronouns), one of them represents the
Subject, another represents the Object of the fact in accordance with the Eqs. (1) and (2)
respectively and the others represent attributes of the action.
These can be the attributes of time, location, mode of action, affiliation with the
Subject or the Object, etc. According to our previous articles, the attributes of an action
in English simple sentence can be represented by nouns that were defined by the logical-
linguistic equations [12].
Additionally, we distinguish a few semantic kinds of nouns that can be extracted by
labels of POS-tagging. Moreover, additionally, we distinguish a few semantic types of
nouns that can be extracted by labels of POS-tagging. These are NNP* (plural and single
proper noun), CD (numeral), DT (determiner, which marked such words as “all”, “any”,
“each”, “many” etc.), FW (foreign word), LS (list item marker), MD (modal auxiliary).
The approach is based on our hypothesis that occurrence of the proper names, numerals,
determiner words, foreign words, items markers, modal auxiliary words in a text can
influence its encyclopedicness. For instance, the occurrence of proper nouns, numerals,
338 N. Khairova et al.
foreign words and items markers in a sentence can explicitly represent that a statement
is formulated more precisely and concisely. On the contrary, we can guess that occur‐
rence of modal auxiliary words in a sentence makes the statement vaguer and more
implicit.
Fig. 2. Graphical representation of Universal Dependencies for the sentence “The Marines
reported that ten Marines and 139 insurgents died in the offensive”. Source: DependenSee
For our analysis we used 7 out of 40 grammatical relations between words in English
sentences, which UD v1 contains4. In order to pick the subj-fact out, we distinguish three
types of dependencies: nsubj, nsubjpass and csubj. Nsubj label denotes the syntactic
subject dependence on a root verb of a sentence, csubj label denotes the clausal syntactic
subject of a clause, and nsubjpass label denotes the syntactic subject of a passive clause.
Additionally, in order to pick the obj_fact out, we distinguish three types of
3
http://chaoticity.com/dependensee-a-dependency-parse-visualisation-tool/.
4
http://universaldependencies.org/en/dep/.
Comparative Analysis of the Informativeness and Encyclopedic Style 339
dependencies: obj, iobj, dobj and ccomp. Obj denotes the entity acted upon or which
undergoes a change of state or motion. The labels iobj, dobj and ccomp are used for
more specific notation of action object dependencies on the verb.
Considering that the root verb is the structural center of a sentence in Dependency
grammar, we distinguish additional types of facts that we can extract from a text. The
root grammatical relation points to the root of the sentence [15]. These are such fact
types that are formed from the root verb (root_fact, subj_fact_root, obj_fact_root,
subj_obj_fact_root, complex_fact_root) and the other ones, in which the action Predi‐
cate is not a root verb (obj_fact_notroot, subj_obj_fact_notroot, complex_fact_notroot).
For completeness of the study, we attribute sentences with copular verbs to a special
type of facts, which we called copular_fact. We should do this for the following reason.
Despite the fact that, as is widely known, the copular verb is not an action verb, such a
verb can often be used as an existential verb, meaning “to exist”.
Our dataset comprises four corpora, two of which include articles from English Wiki‐
pedia. We consider texts from Wikipedia for our experiments for a few reasons. First,
we assume that since Wikipedia is the biggest public universal encyclopedia, conse‐
quently Wikipedia’s articles must be well-written and must follow the encyclopedic
style guidelines. Furthermore, Wikipedia articles can be divided into a different quality
classes [16–18], hence the best Wikipedia’s articles have a greater degree of encyclo‐
pedicness than most other texts do. These hypotheses allow us to use the dataset of
Wikipedia articles in order to evaluate the impact our proposed linguistic features on
the encyclopedic style of texts.
The first Wikipedia corpus, which we called “Wikipedia_6C”, comprises 3000
randomly selected articles from the 6 quality classes of English Wikipedia (from the
highest): Featured articles (FA), Good articles (GA), B-class, C-class, Start, Stub. We
exclude A-class articles since this quality grade is usually used in conjunction with other
ones (more often with FA and GA) as it was done in the previous studies [17, 18]. The
second Wikipedia corpus, which is called “WikipediaFA”, comprises 3000 only the best
Wikipedia articles that randomly are selected from the best quality class - FA.
In order to process plain texts of described above corpora, we use Wikipedia database
dump from January 2018 and special framework WikiExtractor,5 which extracts and
cleans text from a Wikipedia database dumps.
In addition, in order to compare the encyclopedic style of texts from Wikipedia and
texts from other information sources, we have produced two further corpora. The first
one is created on the basis of The Blog Authorship Corpus [5]. The corpus collected
posts of 19,320 bloggers gathered from blogger.com one day in August 2004. The blog‐
gers’ age is from 13 to 47 years6. For our purposes, we extract all texts of 3000 randomly
selected bloggers (authors) from two age groups: “20s” bloggers (ages 23–27) and “30s”
bloggers (ages 33–47). Each age group in our dataset has the same number of bloggers
5
https://github.com/attardi/wikiextractor.
6
Groups description available on the page http://u.cs.biu.ac.il/~koppel/BlogCorpus.htm.
340 N. Khairova et al.
(1500 each). Since bloggers have a different number of posts of various size, in our
corpus we consider all texts of one blogger as a separate item. Hence we got in total
3000 items for our “Blogs” corpus.
The second supplementary corpus, which is called “News”, is created on the basis
of articles and news from popular mass media portals, such as “The New York
Times”7 and “The Guardian”8. For a more comprehensive analysis, we extracted an
equal number of news from different topics. For our experiment, we selected 10 topics
for each news source. So, we extracted 150 recent news from each topic of each news
source (“The New York Times” and “The Guardian”) in January 2018. In total, we have
got 3000 items for our News corpus.
Thus, we have had four corpora with the same number of items for our experiments.
Table 1 shows the distributions of the analyzed texts according to the categories. By
categories, we mean topics of newspaper posts in the “News” corpus, age groups of
bloggers in the “Blogs” corpus and the special manual mark of assessment quality of
Wikipedia articles in “Wikipedia_6C” corpus and “Wikipedia_FA” corpus.
Definition 5. The frequency of each feature in a corpus is defined as the number of the
feature occurrence in the corpus divided by the number of words in –the corpus multi‐
plied by million.
In order to assess the impact of the various types of facts in a sentence and some
types of words in a text on the degree of encyclopedic text, we focus on two experiments.
Both of them classify texts from Blogs, News and Wikipedia. The difference lies in the
selected Wikipedia corpus. In the first experiment, we used texts from Wikipedia_6C
7
https://www.nytimes.com/.
8
https://www.theguardian.com.
Comparative Analysis of the Informativeness and Encyclopedic Style 341
corpus, which includes Wikipedia articles of different quality. We called this experiment
as BNW6 model. In the other experiment we use texts from ‘Wikipedia_FA’ corpus,
which only consists of the best Wikipedia articles. We called second experiment as
BNWF model.
The used Random Forests classifier of the Data Mining Software Weka 39 allows
determining the probability that an article belongs to one of the three corpora. Table 2
shows detailed accuracy by two models respectively.
Additionally, the used Data Mining Software Weka 3 allows constructing a confu‐
sion matrix that permits to visualize the performance of the model. Such matrices for
both classification models are shown in Table 3. Each row in the matrix allows repre‐
senting the number of instances in a predicted class, while each column represents the
instances in an actual class. It makes possible to present which classes were predicted
correctly by our model.
BNW6 BNWF
Blogs News Wikip Blogs News Wikipedia_
edia FA
2691 274 34 Blogs 2683 283 33 Blogs
199 2575 227 News 206 2655 140 News
10 276 2714 Wikiped 24 183 2793 Wikipedi
ia_6C a_FA
Obviously, that the best Wikipedia articles must be well-written and consequently
must follow the encyclopedic style guidelines. This is confirmed by higher coefficients
of recall and precision of BNWF classification model than BNW6 one.
The Random Forest classifier can show the importance of features in the models. It
provides two straightforward methods for feature selection: mean decrease impurity and
mean decrease accuracy. Table 4 shows the most significant features, which are based
on average impurity decrease and the number of nodes using that feature.
9
https://www.cs.waikato.ac.nz/ml/weka/index.html.
342 N. Khairova et al.
Table 4. The most significant features of our models based on average impurity decrease (AID)
and the number of nodes using the features (NNF)
Feature BNW6 BNWF
AID NNF AID NNF
root_fact 0.53 7526 0.52 6354
subj_fact_root 0.47 7614 0.48 6287
subj_fact_notroot 0.45 6537 0.45 5786
obj_fact_notroot 0.42 5678 0.41 4772
obj_fact_root 0.4 5155 0.4 5354
subj_obj_fact_root 0.38 5270 0.39 4479
complex_fact_root 0.39 3994 0.38 3882
complex_fact_notroot 0.35 5541 0.35 4413
copular_fact 0.34 3924 0.33 3254
CD 0.33 4262 0.32 3668
DT 0.32 4646 0.31 4061
FW 0.29 2083 0.31 2087
MD 0.31 4388 0.3 3702
NNP* 0.29 4893 0.29 4112
LS 0.22 775 0.23 664
References
1. Cai, L., Zhu, Y.: The challenges of data quality and data quality assessment in the big data
era. Data Sci. J. 14, 1–10 (2015)
2. Béjoint, H.: Modern Lexicography: An Introduction, pp. 30–31. Oxford University Press
(2000)
3. Khairova, N., Petrasova, S., Gautam, A.: The logical-linguistic model of fact extraction from
English texts. In: International Conference on Information and Software Technologies,
Communications in Computer and Information Science, CCIS 2016, pp. 625–635 (2016)
4. Nivre, J., et al.: Universal dependencies v1: a multilingual treebank collection In: Proceedings
of the Tenth International Conference on Language Resources and Evaluation (LREC 2016),
Paris, France, May. European Language Resources Association (ELRA) (2016)
5. Schler, J., Koppel, M., Argamon, S,. Pennebaker, J.: Effects of age and gender on blogging.
In: Proceedings of 2006 AAAI Spring Symposium on Computational Approaches for
Analyzing Weblogs, pp. 191–197 (2006)
6. Leafgren, J.: Degrees of explicitness: information structure and the packaging of Bulgarian
subjects and objects. John Benjamins, Amsterdam & Philadelphia (2002)
7. Berman, R.A., Ravid, D.: Analyzing narrative informativeness in speech and writing. In:
Tyler, A., Kim, Y., Takada, M. (eds.) Language in the Context of Use: Cognitive Approaches
to Language and Language Learning. Cognitive Linguistics Research Series. pp. 79–101.
Mouton de Gruyter, The Hague (2008)
8. Rennie, J.D.M., Jaakkola, T.: Using term informativeness for named entity detection. In:
Proceedings of SIGIR 2005, pp. 353–360 (2005)
9. Kireyev, K.: Semantic-based estimation of term informativeness. In: Human Language
Technologies: The 2009 Annual Conference of the North American Chapter of the ACL, pp.
530–538 (2009)
10. Wu, Z., Giles, L.C.: Measuring term informativeness in context. In: Proceedings of NAACL
2013, Atlanta, Georgia, pp. 259–269 (2013)
11. Shams, R.: Identification of informativeness in text using natural language stylometry.
Electronic Thesis and Dissertation Repository, 2365 (2014)
12. Huang, A.H., Zang, A.Y., Zheng, R.: Evidence on the information content of text in analyst
reports. Acc. Rev. 89(6), 2151–2180 (2014)
13. Sokolova, M., Lapalme, G.: How much do we say? Using informativeness of negotiation text
records for early prediction of negotiation outcomes. Group Decis. Negot. 21(3), 363–379
(2012)
344 N. Khairova et al.
14. Lex, E., Voelske, M., Errecalde, M., Ferretti, E., Cagnina, L., Horn, C., Granitzer, M.:
Measuring the quality of web content using factual information. In: Proceedings of the 2nd
joint WICOW/AIRWeb Workshop on Web Quality, pp. 7–10. ACM (2012)
15. De Marneffe, M.C., Manning, C.D.: Stanford typed dependencies manual, pp. 338–345.
Technical report. Stanford University (2008)
16. Lewoniewski, W.: Enrichment of information in multilingual wikipedia based on quality
analysis. In: Abramowicz, W. (ed.) BIS 2017. LNBIP, vol. 303, pp. 216–227. Springer, Cham
(2017). https://doi.org/10.1007/978-3-319-69023-0_19
17. Węcel, K., Lewoniewski, W.: Modelling the quality of attributes in wikipedia infoboxes. In:
Abramowicz, W. (ed.) BIS 2015. LNBIP, vol. 228, pp. 308–320. Springer, Cham (2015).
https://doi.org/10.1007/978-3-319-26762-3_27
18. Lewoniewski, W., Węcel, K., Abramowicz, W.: Quality and importance of wikipedia articles
in different languages. In: Dregvaite, G., Damasevicius, R. (eds.) ICIST 2016. CCIS, vol.
639, pp. 613–624. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46254-7_50
19. McEnery, T., Hardie, A.: Corpus Linguistics: Method, Theory and Practice, pp. 48–52.
Cambridge University Press, Cambridge (2012)
Applications, Evaluations, and
Experiences
Satellite Imagery Analysis
for Operational Damage Assessment
in Emergency Situations
Abstract. When major disaster occurs the questions are raised how to
estimate the damage in time to support the decision making process and
relief efforts by local authorities or humanitarian teams. In this paper we
consider the use of Machine Learning and Computer Vision on remote
sensing imagery to improve time efficiency of assessment of damaged
buildings in disaster affected area. We propose a general workflow that
can be useful in various disaster management applications, and demon-
strate the use of the proposed workflow for the assessment of the damage
caused by the wildfires in California in 2017.
1 Introduction
In emergency situations like devastating wildfires, floods, earthquakes or
tsunami, decision makers need to get information about possible damages in
residential area and infrastructure most rapidly after the event. One of the most
valuable sources to get such information from, is the Earth Observation systems,
which include satellite and aerial remote sensing, since it can be captured shortly
after the disaster without all the risks related to the ground observations. The
combination of this information with statistical and ground observation data
contributes to even a better valuation of physical and human losses caused by
disaster [2].
There are several international programs that are arranged to support the
information exchange during and after disasters such as UNOSAT (UNITAR
Operational Satellite Applications Program) [4], Space disaster charter [1],
Humanitarian Openstreetmap Team (HOT) [3] or Tomnod [5] which is the Dig-
italglobe satellite company crowdsourcing platform. All these are useful initia-
tives providing tools and activation workflows for emergency mapping done by
specialists or by volunteers in a collaborative way [8,9].
c Springer International Publishing AG, part of Springer Nature 2018
W. Abramowicz and A. Paschke (Eds.): BIS 2018, LNBIP 320, pp. 347–358, 2018.
https://doi.org/10.1007/978-3-319-93931-5_25
348 G. Novikov et al.
Fig. 1. Time distribution of the building features created by OSM users for Kathmandu
region (source - osm-analytics.org)
Following the news of this incident we found several related media publica-
tions that provide assessments of damages. We assume that the work was doing
manually on satellite images - comparing to the date of incident (Dec. 4) it’s
taken about six days to prepare maps for the article (Dec. 11 LA Times article
update) [15]. That most probably is caused by the amount of work needed to
find appropriate data sources and make a damage map.
Satellite Imagery Analysis in Emergency Situations 349
1. Determine the case of interest. The deep learning methods work significantly
better when the objects of interest are similar to each other, so the case should
be narrow, for example “burned houses” or “flooded roads”.
2. Create a training dataset. The deep learning methods need a training data
so they could learn the objects of interest and their properties. The training
dataset consist of the real data (in our case, two aerospace images, one taken
before the catastrophic event, and the other - after the event) and the labels
that annotate and outline every damaged structure of the type.
350 G. Novikov et al.
3. Train and validate a deep learning method using the dataset. The method (or
model) extracts the information from the training dataset. Its ability to detect
the damaged objects of interest is validated using a subset of the training
data. This pre-trained model will be used in every case of the forthcoming
emergency of the given type.
4. Obtain information of a new emergency case. This is where our method starts
working. The data should be of the same or similar type (spatial resolution,
target objects, color depth) as that used for training, this is a critical require-
ment for the model to work properly.
5. Fine-tune the model for the new case. Despite the similarity of the data, the
model may be unable to process them correctly due to small differences, for
example different sunlight. The fine-tuning can be done using automatically
annotated data from the new case, or using the manual markup for a small
subset of the data.
6. Run the automatic image processing. Now that the model is trained for the
case, we can make an automated processing of the rest of the data and have
them ready for the decision making.
Using this approach, we need to spend some time for creation of the reference
training dataset, but normally it should be made before the emergency event.
Then, after the event, the amount of work needed is much less that allows us to
propose a fast working and thus efficient solution.
4 The Experiment
To validate and demonstrate the workflow, we have chosen the case of wildfires
in two areas of Ventura and Santa Rosa counties, California, USA, where many
houses were burned to the ground in 2017. The advantage of this choice is justi-
fied by the availability of hi-resolution data provided by Digitalglobe within their
Open Data Program [13]. In the following section we will follow our workflow on
the case, and describe both general principles of the deep learning application
to the imagery and our steps in the particular case.
The training area is chosen in the Ventura, Santa Barbara, California, that was
severely affected by the Thomas Fire in the December, 2017 (see Fig. 3).
The preferable source for high-resolution data is the Digitalglobe Open Data
Program. This program is one of the few sources of free high-resolution data,
and the data is distributed early after the event [13]. However, in the case of
South California the existing Openstreetmap (OSM) mapping which was used
as the initial input for the markup is based on Google Maps/Bing Maps imagery
that is better orthorectified, so that the image coordinates differ, as it can be
seen in the Fig. 4. This makes existing map not as good source of the ground
truth.
Fig. 2. Satellite images of buildings before and after the fire event
Fig. 3. Training area in Ventura and the resulting markup (Openstreetmap, Stamen
design; Digitalglobe. Bounding box coordinates: -119.2278 34.2836 -119.1916 34.3065)
352 G. Novikov et al.
Due to these reasons, we had to use the Google Maps imagery, that is similar
to the Digitalglobe open data in terms of image characteristics (similar spatial
resolution, also 3-band 8-bit per band images), but both pre-event and post-event
images are available, and the better orthorectification leads to good alignment
with the OSM.
The crowdsourced mapping data from OSM were not full and did not contain
the information about burned buildings, so it was necessary to improve the OSM
markup before using it as the ground truth for training the deep convolutional
network. We facilitated the manual work by using of OSM as the source of initial
vector features, selecting all the ones tagged as “building”. All the extracted fea-
tures than were checked through the visual inspection and annotated with the
appropriate additional tag “disaster” = “damaged area” if the one was destroyed
by the fire. To complete training dataset we used cartographic tools as Open-
streetmap ID which is open source editor for online mapping for OSM [6]. The
final markup contains 760 not damaged buildings and 320 ruined buildings (see
Fig. 3) was exported in GeoJSON format using OSM API and additionally pro-
cessed using our Python stack tool to convert and rasterize vector data into
1-band pixel masks.
4.3 Train and Validate a Deep Learning Method Using the Dataset
Validation on the part of the Ventura dataset that was not used for training
gave appropriate results, see Fig. 5. Pixel-wise F 1-score for the class of burned
buildings is 0.859 and for the class of unburned buildings is 0.818.
Fig. 5. Results of the change detection on the validation subset of data in Ventura.
Left: image taken before fires, center: image taken after fire, right: segmentation results
black - non-damaged, gray - damaged buildings
Overall training of the model has taken less than an hour on a Tesla P100
GPU using Keras with Tensorflow backend [16,18]. The trained model is now
ready for the new cases of the massive fire.
We consider the fire in Santa Rosa, California (Tubbs Fire, October 2017) as the
“new case” of the same type of events (see Fig. 6). The Open Data program has
images both before and after the fire event, so we can use them for the test.
Fig. 6. A map of the new zone of the Tubbs fire in Santa Rosa, California
354 G. Novikov et al.
As the data in this case have similar characteristics, we tried the image
segmentation with the model as is, without any changes. The result is unsatis-
factory, however it does have some correlation with the real data. This can be
caused by differences in season, solar angle, image preprocessing difference, or by
some difference in she structure of the residential areas themselves. For example,
buildings in Santa Rosa are closer to each other. The example of the results see
in Fig. 7.
Fig. 7. Results of the change detection on the test subset of data in Santa Rosa with-
out fine-tuning. Left: image taken before fires, center: image taken after fire, right:
segmentation results black - non-damaged, gray - damaged buildings
Fig. 8. Small dataset for fine-tuning of the net to the new data. White - unburned
buildings, black contours - burned buildings
Satellite Imagery Analysis in Emergency Situations 355
time and effort. The preparation of the dataset took about an hour of manual
markup by one person.
Switching from one part of dataset to another, the results of the model were
greatly deteriorated, but the dense marking of just less than 300 houses on new
images allowed to improve the quality on the whole new data and reach almost
the same result for 10 min of additional training.
Fig. 9. An example of the test area image before and after the fire
After the fine-tuning, the change detection method can give very good results
on image segmentation, and even give a good chance to distinguish between
separate houses that is very important in the task of damage assessment when
it is necessary to estimate the accurate number of damaged structures.
Note that the segmentation approach is more robust than the detection one
because it allows to estimate the area of the target changes, that can be necessary
in other emergency cases like floods, blockages, fire damage to crops or forests etc.
5 Time Efficiency
The manual markup of our Ventura training area (Fig. 3) should take about
1.5–2 days by a qualified specialist, assuming that mapping of buildings features
356 G. Novikov et al.
Fig. 10. Results of the image segmentation before and after fine-tuning
takes averagely 30 s per feature. But more realistic is the time evaluation of HOT
mapping as it represents the real rate of community. Besides, the HOT tools are
built on the top of OSM Data Base and are planned not for mapping of the
state of the objects like “burnt buildings” but to improve the basic maps when
the cartographic data is missed and needed by humanitarian and emergency
organizations.
The full workflow for the new area, where we had to make the only a small train-
ing subset took about 3 h including model training and automatic processing.
That gives us less time needed for information retrieval for the emergency
management.
6 Further Research
At the current stage we have developed a workflow and a method of the damaged
areas segmentation. In the further research we will continue the development of
the segmentation method to increase its accuracy and robustness to the data
characteristics changes.
The method can be also extended to the problem of instance segmentation,
that is distinguishing between separate objects, counting the objects, and con-
verting them to the map that can be used online and in GIS applications.
We will apply the approach to the Open Data in the case of new events of
this domain, the other types of disasters such as floods and tsunami, and will
extend the training dataset to extrapolate this approach to the other cases and
territories.
7 Conclusion
We have formulated the problem based on the research of the tools and frame-
works for disaster mapping. Based on the problem, we proposed a workflow
involving deep learning and use of open data for the emergency mapping.
Satellite Imagery Analysis in Emergency Situations 357
We have created the training and test datasets for California fires, which
means the raster mask of the vector features of damaged and non damaged
buildings in the area and the appropriate pre- and post-event imagery to develop
a change detection method and validate the approach.
We developed a method of change detection, based on convolutional neural
networks, that is able to make semantic segmentation of the area subjected to
massive fires, mapping burned buildings. The method can be easily extended to
the new areas and data sources with a little training for the data peculiarities
(fine tuning). We are going to elaborate on this approach by increasing its com-
putational capabilities with large scale sparse convolutional neural networks [24],
using a new loss function, specially tailored for change detection in sequences
of events [25,26], imposing a confidence measure on top of the change detector
based on non-parametric conformal approach [27,28].
The workflow turned up to give substantial profit in terms of time needed
for emergency mapping and in the future we will extend it to the other cases.
References
1. About the Disasters Charter. International Disasters Charter. 2000–2018. https://
disasterscharter.org/web/guest/about-the-charter
2. Eguchi, R.T., et al.: The January 12, 2010 Haiti earthquake: a comprehensive
damage assessment using very high resolution areal imagery. In: 8th International
Workshop on Remote Sensing for Disaster Management. Tokyo Institute of Tech-
nology, Japan (2010)
3. About – Humanitarian OpenStreetMap Team. https://www.hotosm.org/about
4. Maps and Data – UNITAR. http://www.unitar.org/unosat/maps
5. Tomnod (2018). https://www.tomnod.com/
6. OpenStreetMap/iD: The easy-to-use OpenStreetMap editor in JavaScript (2018).
https://github.com/openstreetmap/iD
7. Meier, P.: Human computation for disaster response. In: Michelucci, P. (ed.) Hand-
book of Human Computation, pp. 95–104. Springer, New York (2013). https://doi.
org/10.1007/978-1-4614-8806-4 11
8. Lang, S., et al.: Earth Observation for Humanitarian Assistance GI Forum 2017,
vol. 1, pp. 157–165 (2017)
9. Dittus M., Quattrone G., Capra L.: Mass participation during emergency response:
event-centric crowdsourcing in humanitarian mapping. In: CSCW, 2017, pp. 1290–
1303 (2017)
10. Voigt, S., Kemper, T., Riedlinger, T., Kiefl, R., Scholte, K., Mehl, H.: Satellite
image analysis for disaster and crisis-management support. IEEE Trans. Geosci.
Remote Sens. 45(6), 1520–1528 (2007)
11. Joyce, K.E., Belliss, S.E., et al.: A review of the status of satellite remote sens-
ing and image processing techniques for mapping natural hazards and disasters.
Progress Phys. Geogr. 33(2), 183–207 (2009)
12. Barnes, C.F., Fritz, H., Yoo, J.: Hurricane disaster assessments with image-driven
data mining in high-resolution satellite imagery. IEEE Trans. Geosci. Remote Sens.
45(6), 1631–1640 (2007)
358 G. Novikov et al.
The preventive CBM strategy enables planning and execution of maintenance actions
based on the actual maintenance demand of machines. It is based on the principle of
continuously monitoring the machine condition using sensors. Two major tasks are
performed by CBM, namely the identification of fault types (fault diagnostics) and the
estimation of the remaining useful life (prognostics).
The selection of the most suitable ML techniques for specific fault diagnostics tasks
is often elusive and based on gut feeling. The performance of each technique is highly
dependent on the specific problem, in particular considering data characteristics as well
as the external inputs. Individual ML techniques, however, exhibit several general char‐
acteristics, which make them appropriate to a greater or lesser extent for a specific
problem. By assessing and weighting these specific characteristics for a certain problem,
the selection of a suitable algorithm can be supported. This paper aims to provide a
qualitative assessment of ML techniques for fault diagnostics, which facilitates the
selection of a suitable technique for a specific problem at hand. Based on a structural
literature research, ML techniques are identified which have been applied to fault
2 State-of-the-Art
In order to present a short overview of the covered concepts first CBM and ML are
introduced. The last section shows how ML has been applied to CBM in the past.
Evolving from simpler maintenance techniques like corrective and preventive mainte‐
nance, CBM has been the focus of research for a larger period of time [1]. It was intro‐
duced to maximize the uptime of machines by avoiding unplanned breakdowns in
combination with lowering maintenance costs. CBM tries to detect, predict and respond
to machinery failures before they occur, by assessing the machinery’s condition, based
on operational data [2]. CBM is initiated by the fact that most faults are preceded by
certain symptoms, indicating a degradation process. However, CBM analysis has to cope
with several difficulties. These difficulties include dealing with a large number of
continuously observed process variables (sometimes even more than 1500 process vari‐
ables). In addition, difficulties are imposed by insufficient, unreliable or incomplete data
introduced by sensor malfunctions or incomplete labeling, which is sometimes
performed by hand [3]. Therefore, systems need to be able to deal with missing values.
To come to maintenance decisions, based on the degradation process, CBM consists
of two major elements, diagnostics and prognostics. The diagnostics process is used to
detect and diagnose faults [4], while prognostics predicts faults and deterioration of
components [5]. In the following, the paper will focus on diagnostics.
Diagnostics. Diagnostics, whose final process goal is accurate fault detection and diag‐
nosis (FDD), can be viewed as the process of obtaining the relationship between cause
and effect of a fault [6]. Hence, it includes the tasks: fault detection, isolation and
identification. Fault detection continuously analyzes whether and when a fault occurred
[7]. The faulty component is then identified during fault isolation [8]. According to
Schwabacher and Goebel [9], fault identification “requires determining the specific fault
mode, rather than just reporting which sensor has an unusual value” (p. 107). The fault
mode is an estimate of nature and extent of a fault [8]. Determining fault modes is
important in order to take failure-tolerant actions and thereby suitably restore machines’
normal condition [7]. Ultimately, diagnostics tries to solve classification tasks, because
Qualitative Assessment of Machine Learning Techniques 361
it attempts to separate data into normal and faulty states, including the type and location
of faults [10].
During machine operation, several processes are running on a machine. Condition
monitoring (CM) monitors outputs from sensors created during the machine’s processes
out of known inputs [11]. Methods for fault diagnostics can be grouped into three
classes [7].
The first class is called model-based, because its methods use a mathematical model
of the physical system. Methods in this class obtain their models by using physical
principles or system identification techniques [7]. Using these methods, only small
amounts of online data - recorded during CM - are needed to be compared to the expected
output generated by the model, using the same inputs as the machine’s process [11]. The
second class, signal-based methods, analyzes continuous signals, measured from sensors
by looking for anomalies without using an input-output model [7]. These methods can
be applied to either online or historic data [11].
Diagnostics has to be performed using learn-by-example methods, in cases where
the analyzed system is too complex to allow model creation or is not displaying anoma‐
lies within one continuous measured signal. These methods are part of the third class of
knowledge-based methods. Recent knowledge-based diagnostics techniques analyze
whether consistency exists between current process data and implicit knowledge
obtained from huge amounts of multisource historical process data [11]. Due to recent
research advances in this field, the focus of this paper is set on this area of methods.
Supervised Learning. Supervised learning requires labeled input data [11]. This para‐
digm tries to find a generalized description that maps all possible data points to the
appropriate label. To achieve this, the technique tries to find patterns and rules that
describe correlations between certain characteristics of the data and faults [11].
As mentioned earlier, diagnostics tasks are classification problems. Classification
decides based on previously learned examples to which of N given classes a data point
belongs. In discrete classification problems, each data point belongs to exactly one
class [13].
Supervised ML can be split into two groups: classification and regression ([13, 16,
17]). As opposed to classification, regression aims at predicting one variable from known
values of at least one other variable using a mathematical function [18]. Since in FDD
the task is to decide if a component is in healthy or faulty state a decision boundary
separating the states has to be identified. To generate boundaries from historic data, ML
classification techniques can be used [13]. Regression, on the other hand, can be used
for prognostics, however is not suited for diagnostics [19].
Neural Network (DNN) developed in the last decade with increasing application in the
last years can be applied for both supervised and unsupervised learning.
most important pros and cons of the presented ML techniques are discussed based on
Table 1.
Table 1. Scoring of the eight ML techniques based on the presented criteria: ‘--’ represents the
worst (0), ‘o’ the neutral (3) and ‘++’ the best (5) rating; accuracy is weighted with a factor of 5,
while all other categories are weighted with a factor of 1.
Even though SVMs are in general slow during training, there are good classifiers,
achieving over all high accuracy. Similar to ANNs, SVMs have a very poor Explanation
Facility, since its choices of features are sometimes hard to trace [21]. In general, only
366 T. Habrich et al.
RFs and DNNs achieve better accuracy. SVMs, RFs and DNNs have the ability to deal
with overfitting, while they also have acceptable Robustness. In general, RFs achieve
the highest overall rating, because they lack major drawbacks. DNNs consist of several
parameters that need to be set and optimized iteratively leading to a long learning
process. For this reason, DNN performs poor in the Parameter Handling category.
Conclusively, a technique superior in every aspect was not identified. In order to select
appropriate techniques for a given data set, it is necessary to identify the properties of
the data, such as the amount of noise and the type of attributes to select the best technique.
4 Validation
In order to demonstrate the benefits of the previously established rating of the eight ML
techniques, appropriate techniques are selected for an application case and later applied
to real world data. To be able to make an appropriate selection of techniques it is neces‐
sary to identify the problem characteristics of the application case with regard to the
above-identified criteria. In this paper, this process is demonstrated based on an appli‐
cation case in the in-vitro diagnostics (IVD) industry.
The questionnaire indicated that the accuracy of a classifier is important for the
considered company (III), for this reason SVM, RF and DNN are selected. A second
reason for the selection of these three techniques is that they are able to achieve above
average general rating. The transparency of classification (IV) is not a very important
factor in the context of CBM in IVD as long as the ML classifier provides acceptable
performance. Therefore, techniques like ANN or SVM can be applied to the problem,
which would have been impossible if high transparency was required. The two
remaining methods from supervised learning are K-NN and DT. In this case study, DT
is additionally selected to compare its performance to the combined classifier RF.
Evaluation of Techniques. To evaluate the techniques ANN, SVM, RF, DT and DNN,
a test setup was created in close cooperation with STRATEC. Data was acquired from
a reliability run of ten peristaltic pumps. Potential signals to monitor the condition of
the pump were evaluated. The most promising approach was monitoring of motor
currents. Based on the motor currents, the failures of tubing inside the peristaltic pump
should be analyzed and predicted. Therefore, the motor signals were preprocessed using
feature extraction. Extracted features, which were used for model development, include
the range of the motor current over time, as well as the standard deviation and kurtosis.
Additionally the skewness of the data as well as a crest factor were determined and added
to the data set. Subsequently, in this binary classification problem, class 0 represents the
‘no fault’ class and class 1 represents the ‘faulty’ class. Since 10 pumps were tested
under the same conditions, 9 out of 10 data sets were combined to form the training set
(approx. 13000 rows). Pump 10 was excluded from the training set and was therefore
used to test the classifier (approx. 1500 rows).
After data preparation, the selected ML techniques were applied on the training data
using processes inside RapidMiner. The models where then tested on the test data to
acquire the classification results. RapidMiner implemented four out of five techniques
while the ‘Deep Learning’ Module within RapidMiner was designated to act as two layer
DNN using stochastic gradient descent as optimizer.
Most parameters were kept with their default value. However, some selected param‐
eters were altered. The LibSVM module with a svm type of nu-SVC and a sigmoid
kernel was used as SVM. An ANN was configured using one hidden layer size with a
size of 5, 500 training cycles and a learning rate of 0.35. The RF parameters were set to
gain_ratio as the split criterion, a number of 21 trees and a maximum depth of 20. A
decision tree was built, using information_gain as the split criterion und a pruning
confidence of 0.1. Deep Learning was set to non-reproducible, using rectifier as the
activation function and a number of 10 training epochs.
In order to assess the performance of the ML techniques, the performance measure‐
ments accuracy and recall are considered. While accuracy reflects the total number of
misclassifications (independent of the class), the weighted recall and precision penalize
false positive alarms by weighting false positives with a factor of 2. This is imposed by
requirement (II) obtained earlier.
The results presented in Fig. 1 meet the expectations, given the assessment of the
eight techniques in Sect. 3. The RF achieves outstanding performance producing only
one misclassified example. The ANN performs very well in this problem. This is because
368 T. Habrich et al.
an ANN can yield very good results independent of the problem. Its scoring presented
in Sect. 3 was mostly lowered by its low robustness, the danger of overfitting and its
low transparency. None of these drawbacks is of importance for the problem at hand.
As presented in Fig. 1 DT is inferior to all other techniques, although this was to be
expected. In general, it can be noted that all techniques have been optimized to achieve
a very high recall of class 0 (no fault). Thereby, fewer data points are classified as
‘faulty’, even though some of them are actually ‘faulty’. The ‘Deep Learning’ module
yielded mixed results during application, whereas an acceptable solution could be
obtained after several iterations. However, it performs worse than the RF, SVM and
ANN in this problem.
Subsequently, the best techniques for application in IVD are selected. Since all
methods except the DT have a very high recall in class 0 (no fault), the four methods of
SVM, ANN, RF and DNN are applicable to CBM in IVD. In order to combine the
advantages of each technique, an ensemble of these classifiers is proposed in this context.
5 Conclusion
In this paper, eight well-known ML techniques have been assessed according to different
criteria obtained from literature. To display the results, they have been combined in
Table 1, which allows for easy reuse of the ratings for different tasks. Table 1 extends
previous work done by Kotsiantis [20] by investigating further sources (in combination
with an update of relevant aspects of the already assessed techniques) as well as adding
additional requirements relevant in the context of fault diagnostics. Furthermore, four
additional techniques have been included in the assessment. The eight techniques have
been identified performing a literature research beforehand and adjusted to fit the specific
requirements for fault diagnostics. These techniques include both supervised techniques
(ANN, SVM, DT, RF, K-NN, K-Means) as well as unsupervised techniques (SOM,
DNN). It was shown that, with regard to the identified criteria, Random Forests exhibit
the best overall performance. However, Neural Networks, Support Vector Machines,
Qualitative Assessment of Machine Learning Techniques 369
Self-Organizing Maps and Deep Neural Network score well. Depending on the specific
application, different criteria could be weighted differently leading to other results. The
application of the developed rating is demonstrated using a case study from the In-Vitro
Diagnostics industry. During the case study, requirements for CBM from IVD have been
obtained, using a self-completion questionnaire. These requirements were used to select
five supervised techniques. The techniques are applied to a real-world data set from IVD.
Their performance was presented and compared based on quantitative measures.
Our approach can assist other researchers at selecting the best-suited ML tech‐
nique(s) in other application areas of CBM. This is beneficial, even though in this paper
the ratings were limited to only eight ML techniques, while there are plenty of techniques
available. Generating additional ratings for those techniques could be part of future
research. Furthermore, there might be more criteria available and usable to distinguish
ML techniques. In summary, while our rating is very useful in the context of CBM, there
might be other fitting criteria for other ML application areas.
References
1. Mahantesh, N., Aditya, P., Kumar, U.: Integrated machine health monitoring: a knowledge
based approach. Int. J. Syst. Assur. Eng. Manage. 5(3), 371–382 (2014)
2. Shin, J.H., Jun, H.B.: On condition based maintenance policy. J. Comput. Des. Eng. 2(2),
119–127 (2015)
3. Venkatasubramanian, V., Rengaswamy, R., Yin, K., Kavuri, S.N.: A review of process fault
detection and diagnosis part I: quantitative model-based methods. Comput. Chem. Eng.
27(3), 293–311 (2003)
4. Jardine, A.K.S., Lin, D., Banjevic, D.: A review on machinery diagnostics and prognostics
implementing condition-based maintenance. Mech. Syst. Signal Process. 20(7), 1483–1510
(2006)
5. Peng, Y., Dong, M., Zuo, M.: Current status of machine prognostics in condition-based
maintenance: a review. Int. J. Adv. Manuf. Technol. 50(1–4), 297–313 (2010)
6. Lee, J., Wu, F., Zhao, W., Ghaffari, M., Liao, L., Siegel, D.: Prognostics and health
management design for rotary machinery systems – reviews, methodology and applications.
Mech. Syst. Signal Process. 42(1–2), 314–334 (2014)
7. Gao, Z., Cecati, C., Ding, S.: A survey of fault diagnosis and fault-tolerant techniques-part
I: fault diagnosis with model-based and signal-based approaches. IEEE Trans. Industr.
Electron. 62(6), 3757–3767 (2015)
8. Vachtsevanos, G., Lewis, F., Roemer, M., Hess, A., Wu, B.: Intelligent Fault Diagnosis and
Prognosis for Engineering Systems. Wiley, Hoboken (2006)
9. Schwabacher, M., Goebel, K.: A survey of artificial intelligence for prognostics. The
intelligence report. In: Association for the Advancement of Artificial Intelligence AAAI Fall
Symposium 2007, pp. 107–114. AAAI Press, Arlington (2007)
10. Venkatasubramanian, V.: Prognostic and diagnostic monitoring of complex systems for
product lifecycle management: challenges and opportunities. Comput. Chem. Eng. 29(6),
1253–1263 (2005)
11. Dai, X., Gao, Z.: From model, signal to knowledge: a data-driven perspective of fault
detection and diagnosis. IEEE Trans. Industr. Inf. 9(4), 2226–2238 (2013)
12. Wu, F., Wang, T., Lee, J.: An online adaptive condition based maintenance method for
mechanical systems. Mech. Syst. Signal Process. 24(8), 2985–2995 (2010)
370 T. Habrich et al.
13. Marsland, S.: Machine Learning: An Algorithmic Perspective, 2nd edn. Chapman and Hall/
CRC Press, Boca Raton (2014)
14. ISO/IEC 2382.: Information Technology – Vocabulary. International Organization for
Standardization, Geneva (2015)
15. Zhao, X., Li, M., Xu, J., Song, G.: Multi-class semi-supervised learning in machine condition
monitoring. In: 2009 International Conference on Information Engineering and Computer
Science, pp. 1–4. IEEE Press, Wuhan (2009)
16. Bishop, C.M.: Pattern Recognition and Machine Learning. Springer, New York (2006)
17. Chapelle, O., Scholkopf, B., Zien, A.: Introduction to semi-supervised learning. In: Chapelle,
O., Scholkopf, B., Zien, A. (eds.) Semi-Supervised Learning, pp. 1–12. The MIT Press,
Cambridge (2006)
18. Ahmad, R., Kamaruddin, S.: A review of condition-based maintenance decision-making. Eur.
J. Industr. Eng. 6(5), 519–541 (2012)
19. Susto, G.A., Schirru, A., Pampuri, S., McLoone, S., Beghi, A.: Machine learning for
predictive maintenance: a multiple classifier approach. IEEE Trans. Industr. Inf. 11(3), 812–
820 (2015)
20. Kotsiantis, S.B.: Supervised machine learning: a review of classification techniques.
Informatica 31(3), 249–268 (2007)
21. Chandola, V., Banerjee, V., Kumar, V.: Anomaly detection: a survey. ACM Comput. Surv.
41(3), 1–58 (2009)
22. Vom Brocke, J., Simons, A., Niehaves, B., Riemer, K., Plattfaut, R., Cleven, A.:
Reconstructing the giant: on the importance of rigour in documenting the literature search
process. In: ECIS, vol. 9, pp. 2206–2217 (2009)
23. Widmer, G., Kubat, M.: Learning in the presence of concept drift and hidden contexts. Mach.
Learn. 23(1), 69–101 (1996)
24. Arel, I., Rose, D.C., Karnowski, T.P.: Deep machine learning – a new frontier in artificial
intelligence research. IEEE Comput. Intell. Mag. 5(4), 13–18 (2010)
A Comparative Evaluation of Log-Based
Process Performance Analysis Techniques
Abstract. Process mining has gained traction over the past decade and
an impressive body of research has resulted in the introduction of a vari-
ety of process mining approaches measuring process performance. Hav-
ing this set of techniques available, organizations might find it difficult to
identify which approach is best suited considering context, performance
indicator, and data availability. In light of this challenge, this paper aims
at introducing a framework for categorizing and selecting performance
analysis approaches based on existing research. We start from a system-
atic literature review for identifying the existing works discussing how to
measure process performance based on information retrieved from event
logs. Then, the proposed framework is built starting from the information
retrieved from these studies taking into consideration different aspects
of performance analysis.
1 Introduction
Businesses are at a turning point where they have to incorporate digitalization
or fade away. Digital technologies continue to set their transformative marks
on virtually all industry domains and have allowed the expansion of businesses
to markets previously inaccessible. The forces of innovation and creativity have
enabled young businesses to challenge incumbents in practically every sector.
However, one thing has not changed. Businesses will always seek to improve
their processes because “every good process eventually becomes a bad process”
[16]. This is even more relevant in a fast-changing digital era.
The first step to process improvement is to understand where processes can
be improved. In the past, given the lack of data availability and high cost of
data processing, performance analysis methods identified improvement opportu-
nities based on manual analysis, and at times combined with random sampling
(e.g., six sigma [28]). Relying on such manually driven methods, process analysts
assessed the performance of processes so to find opportunities for improvement.
Today, much of the data is captured digitally and, over the past decade, analysis
of large sets of data has improved remarkably. No longer are businesses restricted
c Springer International Publishing AG, part of Springer Nature 2018
W. Abramowicz and A. Paschke (Eds.): BIS 2018, LNBIP 320, pp. 371–383, 2018.
https://doi.org/10.1007/978-3-319-93931-5_27
372 F. Milani and F. M. Maggi
to select the most prioritized processes, limit the scope, or confine the selection
of data due to limitations of time-consuming analysis or tools. In addition, the
accessibility to open source tools has never been easier, in particular for data
driven analysis of business processes. Therefore, in a digital era, businesses can-
not hope to survive with manually driven methods. The process analysis must
also be digitally transformed by tapping into data-driven analysis methods.
One group of techniques for data driven performance analysis uses event
logs of processes to assess performance. Indeed, nowadays, business processes
are often supported by IT systems that log their execution. For instance, an
order-to-invoice process might include activities such as register, validate,
approve, fill order and send invoice. Each order has a unique id and every
activity is recorded in the event log with information about the time when it was
executed (timestamp) and other additional data such as the resource (person or
device) that executed the activity. As such, the process is inherently captured in
the log. With process mining techniques [2], the performance of such processes
can be assessed and analyzed in great detail based on event logs.
The body of research and tools within the field of process mining has grown
significantly during this decade. However, the availability of tools and approaches
developed for specific aspects of process performance does not make it easier for
businesses to employ them. In fact, it poses a challenge. There is no way for
businesses to easily get an overview of what performance indicators can be mea-
sured, what input data is required for such analysis, or what industry specific
implementations are available. In light of this context, we propose a framework
for the selection of log-based performance analysis techniques. We do so by con-
ducting a systematic literature review to identify the body of existing work. We
analyze the results and focus on identifying existing process performance indi-
cators, required input data, and approaches available. Based on the results, we
build a framework for the selection of suitable performance analysis approaches.
The structure of this paper is as follows. Section 2 summarizes the research
protocol for the systematic survey. In Sect. 3, the research questions are discussed
and the framework is presented in Sect. 4. Finally, Sect. 5 concludes the paper.
– RQ2: What input data is required for measuring process performance? For per-
formance analysis, it is important that the “right” set of data is captured. To
this end, it is important do understand what kind of data is required as input
for process performance analysis techniques.
– RQ3: What are the main approaches/algorithms and tools available for analy-
sis of process performance? The final research question aims at capturing the
various methods that can be applied for performance analysis.
To find relevant studies, we sought studies within the domain of “process min-
ing”. However, as process mining covers many aspects of business process analy-
sis, such as process discovery [1], we included “performance” to focus the search.
The boolean search string (“process mining” AND “performance”) was used. The
search was applied to Web of Science (WoS) and Scopus databases. These elec-
tronic databases were selected as they are the primary databases for computer sci-
ence research. The search was conducted in January 2018 and resulted in a total
of 330 studies from Scopus and 194 from WoS. After having removed duplicates,
349 studies remained. The first filtering was aimed at removing studies that were
clearly out of scope (based on title), shorter than 6 pages, not accessible, or not
written in English. The abstract and introduction of the remaining 101 studies
were examined. Peer reviewed studies introducing or extending an existing app-
roach for performance analysis, or directly dealing with performance of a process
were included. After this filtering, the list of studies was reduced to 48. These stud-
ies were examined in full following the inclusion criteria of being within the field of
log-based performance analysis (IC1), proposing a new approach (IC2) or extend-
ing an existing approach (IC3) for measuring process performance, and applying
the presented method(s) to an industrial case study (IC4).
The final list of studies, following the above criteria, constitutes of 32
studies. For each study in the final list, standard metadata was extracted. In order
to address the first research question, data about performance category, metric,
and unit of measurement was extracted. For the second research question, informa-
tion about what input data is required to do the analysis was extracted. Finally, for
the third research question, information about tools used and underlying method
was extracted.
3 Results
In this section, we examine the final list of studies from the perspective of the
defined research questions. The first being the different aspects that can be
assessed, followed by the data input required and available tools and methods.
Time. The time aspect can be divided into four categories. These are process,
fragment, activity and waiting duration (see Fig. 1).
Process Duration. Process duration is the time distance between the start event
of a process and the end event. Several techniques measure process duration. For
instance, [34] examines the process duration of a peer-review process to identify
bottlenecks. Similarly, Engel et al. [11] use electronic exchange messages to analyze
the duration of inter-organizational processes. Different aspects of process dura-
tion can also be analyzed. For instance, [23] examines the influence of contextual
factors (such as weekday or season) on process duration. Ballambettu et al. [5]
propose a method for identifying key differences of process variants that could
affect process duration. Suriadi et al. [39] look at the processes of the emergency
departments at four different hospitals. They compare these processes and their
process duration to identify differences. Piessens et al., [32] recognize that some
event logs contain advanced constructs such as cancelations, multiple concurrent
instances, or advanced synchronization constructs. They use these constructs to
gain accurate assessment of process duration.
activity. However, a process have variants where activity durations vary across dif-
ferent variants. To address variability in activity duration based on variants, in [5],
the authors propose a method that allows for identifying key differences of activity
duration across process variants.
Waiting Duration. Waiting time in processes is one of the main wastes [10] and,
as such, it has to be reduced for process improvement. As it is one of the main
approaches to improve processes, waiting time analysis is the focus of many tech-
niques. Jaisook and Premchaiswadi [19] investigate hospital logs to examine the
average duration a patient spends waiting in a private hospital. The authors do so
by using the built-in functionality of Disco [14]. Similarly, Perimal-Lewis et al. [31]
rely on Disco to examine the processes of an emergency department. In so doing,
they apply process mining to identify deviating activities in regards to waiting
duration. The results highlight bottlenecks in the performance of the processes.
Park et al. [29] propose a framework for analyzing block manufacturing processes
by assessing the total waiting time. In a similar manner, in [34], the authors exam-
ine the total waiting time of a peer-review process. In [5,6], the authors present a
framework for analyzing similar processes across several installations. They pro-
pose a method to analyze a collection of logs from different performance perspec-
tives, one of which is waiting duration. An aspect of waiting duration is delay anal-
ysis. Delays refer to cases where the completion time is later than the planned com-
pletion time. Senderovich et al. [38] analyze a process log from the perspective of
operational deviations resulting in tardiness (delays) from a process duration per-
spective. Park et al. [30] analyze delays in a make-to-order manufacturing firm.
They define two delay indicators, activity and processing delay and found that
some delays can be explained by seasonality.
of “work queues” from event logs at the level of an entire process or of individ-
ual activities. In [38], the authors target the analysis of resource-driven processes
based on event logs. In particular, they focus on processes for which there exists
a predefined assignment of activity instances to resources that execute activities.
The goal is to decrease the tardiness and lower the flow time.
More advanced performance analysis techniques have been recently presented
in [26,40]. In [26], the authors present a technique to understand how bottle-
necks form and dissolve over time via the notion of Staged Process Flow. In
[40], Suriadi et al. present a framework based on the concept of event interval.
The framework allows for a systematic approach to sophisticated performance-
related analysis (e.g., resource productivity trends, recurring working patterns
of resources, waiting time distributions over time, and resource performance
comparison), even with information-poor event logs.
Other studies overlay the performance measures on top of a process model by
replaying the log on the process model [3,32] and calculating aggregate perfor-
mance measures for each element in the process model during the replay. Tech-
niques for enhancing the quality of performance analysis based on log replay
have been proposed [9]. A related technique supported by contemporary perfor-
mance analysis tools is log animation. Log animation displays in a movie-like
fashion how cases circulate through the process model over time [8,9,13].
The analyzed studies mainly use ProM (15 studies) and Disco (6 studies). In
the remaining studies, the authors developed their own applications.
4 Framework
In this section, we synthesize the above results in a framework aimed at assist-
ing businesses to find the most suitable approach for performance analysis. Busi-
nesses, often not acquainted to the academic domain within this field, might find
it challenging to navigate through the studies. As such, our framework might help
in identifying the first steps.
The framework considers three types of techniques. Most studies aim at
descriptive performance analysis of a single log. Concerning this type of tech-
niques, we consider two aspects. The first aspect is derived from the first research
question about performance perspectives. As such, the performance perspectives
are time, resources, and quality. The second aspect refers to the data available
in the input logs. Depending on what data is available, different types of the
performance can be analyzed. Note that the log must include at least case id,
activity, and timestamp (minimum required data). Some studies compare logs
of similar processes from several sites or use logs pertaining to collaborative pro-
cesses. Such approaches are more complex but might be highly relevant for some
businesses. Finally, we noted case studies contextualizing performance analysis
to a certain domain. As such case studies are also valuable for businesses, we
include them in the framework. When combining all these techniques, we gain a
framework as shown in Table 2.
A business seeking to conduct data-driven performance analysis, should first
select the type of technique. Descriptive analysis will show the current state and
Table 2. Framework
highlight cases and/or areas in the process where there are opportunities for
improvement. For descriptive analysis, the minimum requirement is an input log
capturing mandatory data (case id, activity, and timestamp). With this data, it
is possible to perform process and fragment duration analysis. If the log contains
timestamps for start and end of activities, it will be possible to conduct activity
and waiting duration analysis. For delay analysis, it might be required to have
scheduling data. For human resource performance analysis, the log must contain
data about who performed which activity. However, resource does not need only
to be human. For non-human resource analysis, the log must clearly show how
much of the materials was used for each case or activity. For quality analysis,
the log must also contain, for each case, data about if the case had a desired or
undesired outcome. This might be in case of a defect or complaint made.
Our review reveals approaches used for complex performance analysis. Com-
plex analysis covers comparative performance analysis between several installa-
tions, such as treatment processes at different hospitals or ERP systems installed
at several client organizations. These approaches not only analyze the perfor-
mance of each variant, but also compare them so to identify reasons for one
being more efficient than the other. Another type of complex analysis is the one
related to the performance of collaborative processes that can be inter- or intra-
organizational. In addition, there are techniques to extract process characteris-
tics from event logs to the aim of discriminating between positive and negative
performance, techniques for the analysis of the evolution of process performance
over time, and for sophisticated performance-related analysis with information-
poor event logs.
Finally, the framework contains case studies from different industry domains.
Most methods have validated their results on real-life industry logs and in so
doing, also gained some insight that is specific for that industry. For instance,
financial, healthcare, and manufacturing processes have been used to validate
results. This will be valuable to businesses operating within the same industry
or that have similar processes as those used for testing the results in the analyzed
studies.
5 Conclusion
Business process performance analysis has been conducted for many decades to
identify opportunities for process enhancement. In this light, performance anal-
ysis based on process mining techniques offer great value for businesses and our
systematic literature review identifies tools for them to use. However, it might
be difficult for businesses to navigate within this field. Therefore, we propose a
framework to aid them in finding suitable methods. The framework considers
the complexity of the analysis, performance perspectives, required input data,
and tool availability. In particular, we show that performance is analyzed from
time (process, fragment, activity, and waiting duration), resources, and quality
aspects. Although process flexibility is also a performance indicator [10], cur-
rently there are no approaches for its analysis. Therefore, an important avenue
A Comparative Evaluation of Log-Based Process 381
for future work in the process performance analysis field is the development of
techniques for analyzing this performance perspective.
References
1. van der Aalst, W.M.P.: Process discovery: an introduction. In: Process Mining, pp.
125–156. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-19345-
35
2. van der Aalst, W.M.P.: Process Mining - Data Science in Action, 2nd edn. Springer
(2016)
3. van der Aalst, W.M.P., Adriansyah, A., van Dongen, B.F.: Replaying history on
process models for conformance checking and performance analysis. Wiley Inter-
disc. Rew Data Min. Knowl. Discov. 2(2), 182–192 (2012)
4. Arpasat, P., Porouhan, P., Premchaiswadi, W.: Improvement of call center cus-
tomer service in a Thai bank using Disco fuzzy mining algorithm. In: ICT and
Knowledge Engineering, pp. 90–96 (2015)
5. Ballambettu, N.P., Suresh, M.A., Bose, R.P.J.C.: Analyzing process variants to
understand differences in key performance indices. In: Dubois, E., Pohl, K. (eds.)
CAiSE 2017. LNCS, vol. 10253, pp. 298–313. Springer, Cham (2017). https://doi.
org/10.1007/978-3-319-59536-8 19
6. Jagadeesh Chandra Bose, R.P., Gupta, A., Chander, D., Ramanath, A., Dasgupta,
K.: Opportunities for process improvement: a cross-clientele analysis of event data
using process mining. In: Barros, A., Grigori, D., Narendra, N.C., Dam, H.K. (eds.)
ICSOC 2015. LNCS, vol. 9435, pp. 444–460. Springer, Heidelberg (2015). https://
doi.org/10.1007/978-3-662-48616-0 31
7. Cho, M., Song, M., Yoo, S.: A systematic methodology for outpatient process
analysis based on process mining. In: Ouyang, C., Jung, J.-Y. (eds.) AP-BPM 2014.
LNBIP, vol. 181, pp. 31–42. Springer, Cham (2014). https://doi.org/10.1007/978-
3-319-08222-6 3
8. Conforti, R., Dumas, M., La Rosa, M., Maaradji, A., Nguyen, H., Ostovar, A.,
Raboczi, S.: Analysis of business process variants in Apromore. In: BPM Demos,
pp. 16–20 (2015)
9. van Dongen, B.F., Adriansyah, A.: Process mining: fuzzy clustering and perfor-
mance visualization. In: Rinderle-Ma, S., Sadiq, S., Leymann, F. (eds.) BPM 2009.
LNBIP, vol. 43, pp. 158–169. Springer, Heidelberg (2010). https://doi.org/10.1007/
978-3-642-12186-9 15
10. Dumas, M., La Rosa, M., Mendling, J., Reijers, H.A.: Fundamentals of Business
Process Management, 2nd edn. Springer (2018)
11. Engel, R., Krathu, W., Zapletal, M., Pichler, C., Bose, R.J.C., van der Aalst,
W.M.P., Werthner, H., Huemer, C.: Analyzing inter-organizational business pro-
cesses. Inf. Syst. e-Bus. Manag. 14(3), 577–612 (2016)
12. Ganesha, K., Supriya, K.V., Soundarya, M.: Analyzing the waiting time of patients
in hospital by applying heuristics process miner. In: ICICCT, pp. 500–505 (2017)
382 F. Milani and F. M. Maggi
13. Günther, C.W., van der Aalst, W.M.P.: Fuzzy mining – adaptive process simplifi-
cation based on multi-perspective metrics. In: Alonso, G., Dadam, P., Rosemann,
M. (eds.) BPM 2007. LNCS, vol. 4714, pp. 328–343. Springer, Heidelberg (2007).
https://doi.org/10.1007/978-3-540-75183-0 24
14. Günther, C.W., Rozinat, A.: Disco: discover your processes. In: BPM Demos, pp.
40–44 (2012)
15. Hachicha, M., Fahad, M., Moalla, N., Ouzrout, Y.: Performance assessment archi-
tecture for collaborative business processes in BPM-SOA-based environment. Data
Knowl. Eng. 105, 73–89 (2016)
16. Hammer, M.: What is business process management? In: Brocke, J., Rosemann,
M. (eds.) Handbook on Business Process Management 1. International Handbooks
on Information Systems, pp. 3–16. Springer, Heidelberg (2010). https://doi.org/
10.1007/978-3-642-00416-2 1
17. Hompes, B.F.A., Buijs, J.C.A.M., van der Aalst, W.M.P.: A generic framework
for context-aware process performance analysis. In: Debruyne, C., et al. (eds.) On
the Move to Meaningful Internet Systems: OTM 2016 Conferences. OTM 2016.
LNCS, vol. 10033, pp. 300–317. Springer, Cham (2016). https://doi.org/10.1007/
978-3-319-48472-3 17
18. Huang, Z., Lu, X., Duan, H.: Resource behavior measure and application in busi-
ness process management. Expert Syst. Appl. 39(7), 6458–6468 (2012)
19. Jaisook, P., Premchaiswadi, W.: Time performance analysis of medical treatment
processes by using Disco. In: ICT and Knowledge Engineering, pp. 110–115 (2015)
20. Kitchenham, B.: Procedures for performing systematic reviews. Keele University,
Keele, UK, 33, pp. 1–26 (2004)
21. de Leoni, M., van der Aalst, W.M.P., Dees, M.: A general framework for correlating
business process characteristics. In: Sadiq, S., Soffer, P., Völzer, H. (eds.) BPM
2014. LNCS, vol. 8659, pp. 250–266. Springer, Cham (2014). https://doi.org/10.
1007/978-3-319-10172-9 16
22. de Leoni, M., van der Aalst, W.M.P., Dees, M.: A general process mining framework
for correlating, predicting and clustering dynamic behavior based on event logs.
Inf. Syst. 56, 235–257 (2016)
23. Leyer, M.: Towards A context-aware analysis of business process performance. In:
PACIS 2011: Quality Research in Pacific Asia, p. 108 (2011)
24. Mans, R.S., Schonenberg, M., Song, M., van der Aalst, W.M.P., Bakker, P.J.:
Application of process mining in healthcare-a case study in a Dutch hospital. In:
BIOSTEC, pp. 425–438 (2008)
25. Nakatumba, J., van der Aalst, W.M.P.: Analyzing resource behavior using process
mining. In: Rinderle-Ma, S., Sadiq, S., Leymann, F. (eds.) BPM 2009. LNBIP,
vol. 43, pp. 69–80. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-
642-12186-9 8
26. Nguyen, H., Dumas, M., ter Hofstede, A.H.M., La Rosa, M., Maggi, F.M.: Business
process performance mining with staged process flows. In: Nurcan, S., Soffer, P.,
Bajec, M., Eder, J. (eds.) CAiSE 2016. LNCS, vol. 9694, pp. 167–185. Springer,
Cham (2016). https://doi.org/10.1007/978-3-319-39696-5 11
27. Nogayama, T., Takahashi, H.: Estimation of average latent waiting and service
times of activities from event logs. In: Motahari-Nezhad, H.R., Recker, J., Wei-
dlich, M. (eds.) BPM 2015. LNCS, vol. 9253, pp. 172–179. Springer, Cham (2015).
https://doi.org/10.1007/978-3-319-23063-4 11
28. Pande, P.S., Neuman, R.P., Cavanagh, R.R.: The Six Sigma Way. McGraw-Hill
(2000)
A Comparative Evaluation of Log-Based Process 383
29. Park, J., Lee, D., Zhu, J.: An integrated approach for ship block manufacturing
process performance evaluation: case from a Korean shipbuilding company. Int. J.
Prod. Econ. 156, 214–222 (2014)
30. Park, M., Song, M., Baek, T.H., Son, S.Y., Ha, S.J., Cho, S.W.: Workload and
delay analysis in manufacturing process using process mining. In: Bae, J., Suriadi,
S., Wen, L. (eds.) AP-BPM 2015. LNBIP, vol. 219, pp. 138–151. Springer, Cham
(2015). https://doi.org/10.1007/978-3-319-19509-4 11
31. Perimal-Lewis, L., Teubner, D., Hakendorf, P., Horwood, C.: Application of process
mining to assess the data quality of routinely collected time-based performance
data sourced from electronic health records by validating process conformance.
Health Inform. J. 22(4), 1017–1029 (2016)
32. Piessens, D., Wynn, M.T., Adams, M.J., van Dongen, B.F.: Performance analysis
of business process models with advanced constructs. In: Australasian Conference
on Information Systems (2010)
33. Pika, A., Wynn, M.T., Fidge, C.J., ter Hofstede, A.H.M., Leyer, M., van der
Aalst, W.M.P.: An extensible framework for analysing resource behaviour using
event logs. In: Jarke, M., et al. (eds.) CAiSE 2014. LNCS, vol. 8484, pp. 564–579.
Springer, Cham (2014). https://doi.org/10.1007/978-3-319-07881-6 38
34. Premchaiswadi, W., Porouhan, P.: Process modeling and bottleneck mining in
online peer-review systems. SpringerPlus 4(1), 441 (2015)
35. Reijers, H.A., Song, M., Jeong, B.: On the performance of workflow processes with
distributed actors: does place matter? In: Alonso, G., Dadam, P., Rosemann, M.
(eds.) BPM 2007. LNCS, vol. 4714, pp. 32–47. Springer, Heidelberg (2007). https://
doi.org/10.1007/978-3-540-75183-0 3
36. Senderovich, A., Weidlich, M., Gal, A., Mandelbaum, A.: Queue mining – predict-
ing delays in service processes. In: Jarke, M., et al. (eds.) CAiSE 2014. LNCS,
vol. 8484, pp. 42–57. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-
07881-6 4
37. Senderovich, A., Weidlich, M., Gal, A., Mandelbaum, A.: Queue mining for delay
prediction in multi-class service processes. Inf. Syst. 53, 278–295 (2015)
38. Senderovich, A., Weidlich, M., Yedidsion, L., Gal, A., Mandelbaum, A., Kadish, S.,
Bunnell, C.A.: Conformance checking and performance improvement in scheduled
processes: a queueing-network perspective. Inf. Sys. 62, 185–206 (2016)
39. Suriadi, S., Mans, R.S., Wynn, M.T., Partington, A., Karnon, J.: Measuring patient
flow variations: a cross-organisational process mining approach. In: Ouyang, C.,
Jung, J.-Y. (eds.) AP-BPM 2014. LNBIP, vol. 181, pp. 43–58. Springer, Cham
(2014). https://doi.org/10.1007/978-3-319-08222-6 4
40. Suriadi, S., Ouyang, C., van der Aalst, W.M.P., ter Hofstede, A.H.M.: Event inter-
val analysis: why do processes take time? Dec. Supp. Syst. 79, 77–98 (2015)
41. Wang, Y., Caron, F., Vanthienen, J., Huang, L., Guo, Y.: Acquiring logistics pro-
cess intelligence: methodology and an application for a chinese bulk port. Expert
Syst. Appl. 41(1), 195–209 (2014)
42. Wombacher, A., Iacob, M.: Start time and duration distribution estimation in
semi-structured processes. In: SAC, pp. 1403–1409 (2013)
43. Wongvigran, S., Premchaiswadi, W.: Analysis of call-center operational data using
role hierarchy miner. In: ICT and Knowledge Engineering, pp. 142–146 (2015)
44. Yampaka, T., Chongstitvatana, P.: An application of process mining for queueing
system in health service. In: JCSSE, pp. 1–6 (2016)
Blockchain for Business Applications:
A Systematic Literature Review
1
International Hellenic University, Thessaloniki, Greece
{i.konstantinidis,g.siaminos,c.timplalexis,
p.zervas,v.peristeras}@ihu.edu.gr
2
Fraunhofer Institute for Applied Information Technology FIT,
53754 Sankt Augustin, Germany
[email protected]
1 Introduction
when 51% of the users approve the provided solution. Then, the block of the transaction
is added to the blockchain. The blockchain is a list of blocks that includes every single
transaction that has ever been made [4, 5]. The blocks are visible to all users, but they
cannot be edited.
Blockchain is becoming increasingly popular and use cases are showing up at a large
variety of industries. The European Commission presented a report in April 2016 [6],
supporting that blockchain technology has the potential to radically overhaul existing
business models. At the same report, it was estimated that smart contracts, based on
blockchain technology, could reduce infrastructure costs of banks from 13.8 to 18.4
billion euros annually by 2022. On October 2017, Bloomberg published an article which
stated that Goldman Sachs and Google are among the most active blockchain investors
[7]. Moreover, 10 of the largest U.S.A. banks have invested $267 million in six block‐
chain companies and one consortium.
In this paper, we conduct a systematic literature review aiming to explore the busi‐
ness areas that blockchain technology is applied. We discuss existing or future use cases
found in the literature and we analyze the impact that blockchain could have on multiple
industries. We also take into consideration possible concerns that may arise from the
expansion of blockchain applications to various sectors.
The remainder of the paper is structured as follows. In Sect. 2, we describe the
research method, we formulate our research questions and we analyze the procedure
that led us to our final set of primary studies. In Sect. 3, we report the results of the
literature review and discussion follows at Sect. 4. Finally, in Sect. 5, we provide our
concluding remarks.
2 Research Method
The goal of our research is to point out industries in which blockchain technology use
cases are met. Almost ten years after it was first introduced, blockchain has expanded
its use at a large variety of services beyond cryptocurrencies. This relevant experience
gives us the chance to discuss issues emerged over the last years. For this, we formulated
research questions, (RQ1) What are the business sectors in which blockchain applica‐
tions are being used or developed? (RQ2) What are the obstacles and challenges of
blockchain technology? The first question aims at discovering in the literature, appli‐
cations of blockchain technology that are currently used as well as research results for
other potential uses. The second question tries to showcase issues and challenges related
to the expansion of blockchain application areas.
In this paper, a systematic literature review approach was followed according to the
guidelines proposed by Kitchenham [8]. In order to cover a large spectrum of relevant
publications, we decided to search the following widely recognized and extensively used
electronic libraries: ACM Digital Library, IEEE Xplore Digital Library, Science Direct,
386 I. Konstantinidis et al.
and Springer Link. The keyword strings that were used are: “BLOCKCHAIN AND
APPLICATIONS”, “BLOCKCHAIN AND USE CASES”. At the next step, we decided
in which fields of the papers we would apply the search terms. In order to get a reasonable
number of results, we searched for the keyword strings in the paper title, abstract, and
keywords. We confined our search to publications written in the English language and
selected as a content type only journals and conference papers, rejecting book chapters
or webpages. No restrictions regarding the paper release date were used. The procedure
for selecting the literature was conducted in November 2017, so it contains papers that
were published or indexed up to that date. The final number of papers that we gathered
after removing duplicates is 385. Then, we manually excluded papers that their title
seemed irrelevant to our research, reducing the papers to 125. The above procedure was
repeated by scanning the papers’ abstracts, reaching 70 papers. We read the whole text
of these papers, culminating in 44 of them in order to extract information and answer
our research questions (Fig. 1).
3 Results
In this section, we present the results of our literature review. The results are organized
by research question and each attribute of the concept matrix is examined separately
(Table 1).
Table 1. Concept matrix RQ1: business sectors in which blockchain applications are being used
or developed and number of papers referring to them
Cryptocurrencies E-government Healthcare Supply chain Energy Banking
Total # 8 14 10 7 8 5
Papers
Blockchain for Business Applications 387
identity management system built on top of the Bitcoin and Ethereum blockchain
respectively. The authors of [29, 33] propose innovative blockchain platforms in order
to overcome the issue of tax fraud by increasing transparency. In particular, a new
blockchain protocol, Pajakoin, is created as a simple, transparent and secure Value-
Added Tax system [29], while in [33] a potential blockchain database is introduced,
towards managing dividend flows, aiming to diminish as much as possible the double
spending problem in the public taxation sector. The digitalization of the core govern‐
mental activities is likely to happen through using the blockchain platforms. A use case
where academic certificates are stored in a secure way is analyzed in [26], an approach
that might be very useful in the public sector. In [27], a system where private data can
be shared with many organizations by the order of the user is proposed. The users of
this system have full control of their data and new information is automatically updated
in every organization that has access to those data. In [28], the authors suggest a novel
distributed online lottery protocol that applies techniques developed for voting appli‐
cations for the purpose of reducing security risks while avoiding the trusted third party.
Finally, an application of blockchain is mentioned in [43] where the technology is used
to track politicians’ activities and serve as a transparency tool to citizens’ hands.
with 100 residential households with artificial agents, implemented on a private block‐
chain, while in [3] it is highlighted that blockchain technology at local power grids allows
the distribution, metering and billing of the electricity to be administered by the
community itself without a reliant third-party intervention. In [11, 23] blockchain-based,
intelligent, trusted measurement and monitoring of energy-related assets in a Smart Grid
or a microgrid is suggested.
Other Business Sectors. Apart from these results, blockchain can play a major role in
other business sectors that are not so extensively analyzed in our literature. It can be a
solution for a pay-as-you-go car insurance application. More specifically, all the data
streamed from the vehicle monitoring engine are stored in the blockchain database,
which guarantees that the data are tamper-proof and traceable providing a quicker and
better customer experience and less operating cost as well as avoiding frauds [44].
Moreover, with the exploitation of smart contracts, blockchain can be used in farming
insurance by gathering weather data, where farmers need insurance protection against
390 I. Konstantinidis et al.
the consequences of bad weather. Smart contract could also be used, in combination
with smart sensors, for home insurance with automatic reimbursements for damages
[43]. Blockchain technology has also potentials for construction management. It can
provide a reliable infrastructure for building information management, legal arguments
and secure storage of sensor data during all life-cycle stages without using a centralized
building information model avoiding the need of the trusted third party [45]. There are
a lot of expectations that blockchain technology will be incredible disruptive to auto‐
motive industry as well. The future connected vehicles that will be part of IoT will need
a comprehensive security architecture to protect the transferred data. In [46] the authors
introduce us to an automotive security platform utilizing blockchain to tackle the impli‐
cated security and privacy challenges of future connected vehicles. Education will also
be potentially benefited by the blockchain implementation. Academic credentials must
be universally recognized and verifiable. Blockchain solutions in education could
streamline verification procedures – thus reducing fraudulent claims of un-earned
educational credits. The University of Nicosia [26] where the academic certificates of
the students were stored in blockchain, is the first example moving towards this direction.
Table 2. Concept matrix for RQ2: obstacles and challenges of blockchain technology
Papers Privacy Security Latency Computational cost
Total 6 6 3 4
Security. Even though blockchain is used in many sectors, there are still some security
concerns that need to be addressed. In [2], the “Transaction Malleability” attack is
described. This type of attack occurs when someone changes the unique transaction ID
before the transaction is confirmed. As a result, the transaction is modified and crypto‐
currencies are being transferred to the attacker’s account. The authors of [3] support that
there are challenges at the blockchain’s individual nodes, whose keys may be stolen,
leading to malicious transactions. In [10], it becomes clear that the most crucial issue
proof of work mechanism faces is, when miners try to control more than 50% of the
network’s computing power in order to prevent transactions from gaining confirmations,
which is known as the 51% attack. It is summarized that further research needs to be
done towards this direction, in order to find solutions that will increase blockchain’s
confidence. A solution for the 51% attack is suggested at [29], where a VAT system is
developed on a centralized blockchain. In [16] the authors express their concern about
the fact that all nodes participating in a blockchain are connected to a P2P network but
generally security leaks have been reported about P2P networks. In [14] a product
ownership management system has been developed on the Ethereum blockchain but
since Ethereum is still under development its security is not fully verified.
Blockchain for Business Applications 391
Privacy. Privacy is a main issue that is still under research. In [13], it is mentioned that
even though blockchain technology can provide transparency in the clinical trial and
precision medicine, this could lead to privacy concerns. The anonymity of the blockchain
users’ identity cannot be fully accomplished using cryptographic keys. It is also stated
that in the traditional blockchain about 60% of the users’ identities had been compro‐
mised via big data analytics of data sets across the Internet. Furthermore, in [16] the
authors mention that financial systems, such as the banking systems, must provide high
privacy in contrast to the current blockchain technology, which has a low privacy level.
In [19] the authors claim that the obstacles of blockchain’s deployment in the healthcare
sector are also psychological, as there are data sharing concerns among medical organ‐
izations. Moreover, in some applications, privacy issues could lead to trust problems.
In [31] it is claimed that in a smart grid infrastructure (energy sector) privacy is a facet
of trust in the sharing economy. In addition, Bitcoin users can send digital coins to a
specific address that belongs to themselves. In [49] it is mentioned that these addresses
of the same person can be linked.
Latency. One of the biggest limitation all blockchain consensus protocols have in
common, is that every fully participating node in the network must process every
transaction. Decentralization is a core and innovative characteristic of blockchain
technology but, unfortunately, drives to latency challenges. In [26], it is referred that
Bitcoin blockchain theoretical transaction power is seven transactions per second. As
a result, in a financial system where speed and executions in high rates are obliga‐
tory, the P2P network that blockchain provides is clearly far from being applicable
[16]. The e-commerce sector is also beyond the usage of blockchain technology. The
time needed for a transaction to be verified in bitcoin blockchain, which is the domi‐
nant and “traditional” blockchain protocol, is almost ten minutes something opposite
to the real-time transactions that the retail businesses require. However, many plat‐
forms based on alternative blockchain protocols are developed to overcome this
latency obstacle and speed up the confirmation process [50].
4 Discussion
After thoroughly gathering information from our literature, in this last section we present
blockchain’s hype, applications grouped by use cases, and concerns, as topics that caught
our attention during conducting our review and made an impression upon us.
31
Data Storage. Blockchain technology and its special features could bring big advan‐
tages to data storage systems, as it will provide cheaper, faster, more secure and decen‐
tralized storage than the existing cloud storage platforms. Decentralized storage works
by distributing the data across a network of nodes, in a similar way to the distributed
ledger technology characteristic of blockchain. Blockchain applications that already
encompass decentralized storage are Storj, Maidsafe and IPFS [43]. The use of block‐
chain as a database applies to different economic sectors. E-government, healthcare,
banking and supply chain sectors try to implement various blockchain projects to opti‐
mize their operational procedures. MedRec, ARIA [3], University of Nicosia certificate
storage, Bitnation, E-resident and Everledger [26] are only some of those projects that
their usage could bring enormous profits. On the contrary, having a great potential is not
the same as having great success. There are still some insurmountable obstacles that
prevent the adoption of the technology with its current structure. New protocols and
platforms are created every single day to correct the previous ones. There is no doubt
that blockchain in the near future will make interaction between people and organiza‐
tions faster and cheaper.
and doubts, even to CEO’s, let alone to everyday consumers. It takes time for a new
technology, especially to something as foundational as blockchain, to incorporate itself
into the fabric of modern society. In fact, “the more likely blockchain is to disrupt the
global financial system, the less likely is to succeed” [48].
Another challenge that has to be tackled about blockchain is establishing standard‐
ization and regulatory framework. In April 2016, European Union in one of its reports
points out that “The future of blockchain requires the development of a common
language with specific rules for interaction, which will be achieved through standardi‐
zation processes.” [8]. In September 2016 ISO accepted Australia’s proposal to manage
the Secretariat of ISO/TC 307 for new international standards on blockchain and in
March 2017 the Roadmap Report was released [52, 53]. Currently, 29 participating
member-states and 13 observing member-states are developing 4 standards that are at
proposal or preparatory stage.
For the time being, the existing legal framework applies to activities related with the
blockchain technology. The regulators are monitoring blockchain-based activities,
acquiring knowledge in order to make the law keep up with the evolution of the tech‐
nology. Regulators are gradually starting to understand the blockchain use cases, while
at the same time the innovators are trying to find the regulatory principles that apply to
their activity. Current blockchain applications are considered rather immature to guide
the legislation towards a specific direction. The development of the existing applications
and the creation of new ones are expected to showcase the legal gaps that need to be
regulated. In the meantime, existing legal principles can sufficiently face blockchain-
related criminal activity whereas experience gained over time will provide guidance and
dictate the need for new regulations that cover all the legal scenarios that may arise by
the use of blockchain technology.
5 Conclusion
Blockchain technology is challenging the status quo in several areas in a radical way
providing a decentralized database of any transaction involving transfer of value i.e.
(money, goods, property, assets or even votes). This generic nature is what makes
blockchain technology attractive to many business areas today. However, we conclude
our review highlighting the risks, effects as well as the unintended consequences of
blockchain technology on established markets. The disruption of blockchain technology
in business sectors is increasing in pace. Therefore, we believe that further critical
research is needed to exploit its capabilities and understand the limitations when applied
in a large scale.
396 I. Konstantinidis et al.
References
17. Munsing, E., Mather, J., Moura, S.: Blockchains for decentralized optimization of energy
resources in microgrid networks. In: 2017 IEEE Conference on Control Technology and
Applications (CCTA) (2017). https://doi.org/10.1109/CCTA.2017.8062773
18. Mettler, M.: Blockchain technology in healthcare: the revolution starts here. In: 2016 IEEE
18th International Conference on e-Health Networking, Applications and Services (2016)
19. Kshetri, N.: Blockchain’s roles in strengthening cybersecurity and protecting privacy.
Telecommun. Policy 41, 1027–1038 (2017). https://doi.org/10.1016/j.telpol.2017.09.003
20. Sikorski, J., Haughton, J., Kraft, M.: Blockchain technology in the chemical industry:
machine-to-machine electricity market. Appl. Energy 195, 234–246 (2017)
21. Castellanos, J., Coll-Mayor, D., Notholt, J.: Cryptocurrency as guarantees of origin:
simulating a green certificate market with the Ethereum Blockchain. In: 2017 IEEE
International Conference on Smart Energy Grid Engineering (SEGE) (2017)
22. Aitzhan, N.Z., Svetinovic, D.: Security and privacy in decentralized energy trading through
multi-signatures, blockchain and anonymous messaging streams. IEEE Trans. Dependable
Secure Comput. 1 (2016). https://doi.org/10.1109/TDSC.2016.2616861
23. Imbault, F., Swiatek, M., de Beaufort, R., Plana, R.: The green blockchain: managing
decentralized energy production and consumption. In: 2017 IEEE International Conference
on Environment and Electrical Engineering and 2017 IEEE Industrial and Commercial Power
Systems Europe (EEEIC/I&CPS Europe) (2017)
24. Hou, H.: The application of blockchain technology in e-government in China. In: 2017 26th
International Conference on Computer Communication and Networks (ICCCN) (2017)
25. Dubovitskaya, A., Xu, Z., Ryu, S., Schumacher, M., Wang, F.: How blockchain could
empower ehealth: an application for radiation oncology. In: Begoli, E., Wang, F., Luo, G.
(eds.) DMAH 2017. LNCS, vol. 10494, pp. 3–6. Springer, Cham (2017). https://doi.org/
10.1007/978-3-319-67186-4_1
26. Ølnes, S.: Beyond bitcoin enabling smart government using blockchain technology. In:
Scholl, H.J., Glassey, O., Janssen, M., Klievink, B., Lindgren, I., Parycek, P., Tambouris, E.,
Wimmer, Maria A., Janowski, T., Sá Soares, D. (eds.) EGOVIS 2016. LNCS, vol. 9820, pp.
253–264. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-44421-5_20
27. Alboaie, S., Cosovan, D.: Private data system enabling self-sovereign storage managed by
executable choreographies. In: Chen, L.Y., Reiser, H.P. (eds.) DAIS 2017. LNCS, vol. 10320,
pp. 83–98. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-59665-5_6
28. Grumbach, S., Riemann, R.: Distributed random process for a large-scale peer-to-peer lottery.
In: Chen, L.Y., Reiser, H.P. (eds.) DAIS 2017. LNCS, vol. 10320, pp. 34–48. Springer, Cham
(2017). https://doi.org/10.1007/978-3-319-59665-5_3
29. Wijaya, D.A., Liu, J.K., Suwarsono, D.A., Zhang, P.: A new blockchain-based value-added
tax system. In: Okamoto, T., Yu, Y., Au, M.H., Li, Y. (eds.) ProvSec 2017. LNCS, vol. 10592,
pp. 471–486. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-68637-0_28
30. Zhao, Z., Chan, T.-H.H.: How to vote privately using bitcoin. In: Qing, S., Okamoto, E., Kim,
K., Liu, D. (eds.) ICICS 2015. LNCS, vol. 9543, pp. 82–96. Springer, Cham (2016). https://
doi.org/10.1007/978-3-319-29814-6_8
31. Sun, J., Yan, J., Zhang, K.: Blockchain-based sharing services: what blockchain technology
can contribute to smart cities. Finan. Innov. 2 (2016)
32. Polyzos, G., Fotiou, N.: Blockchain-assisted information distribution for the Internet of
Things. In: 2017 IEEE International Conference on Information Reuse and Integration (IRI)
33. Hyvärinen, H., Risius, M., Friis, G.: A blockchain-based approach towards overcoming
financial fraud in public sector services. Bus. Inform. Syst. Eng. 59, 441–456 (2017). https://
doi.org/10.1007/s12599-017-0502-4
398 I. Konstantinidis et al.
34. Zhao, H., Zhang, Y., Peng, Y., Xu, R.: Lightweight backup and efficient recovery scheme for
health blockchain keys. In: 2017 IEEE 13th International Symposium on Autonomous
Decentralized System (ISADS) (2017). https://doi.org/10.1109/ISADS.2017.22
35. Guo, Y., Liang, C.: Blockchain application and outlook in the banking industry. Finan. Innov.
2 (2016). https://doi.org/10.1186/s40854-016-0034-9
36. Zhao, J., Fan, S., Yan, J.: Overview of business innovations and research opportunities in
blockchain and introduction to the special issue. Finan. Innov. 2 (2016)
37. Mengelkamp, E., Notheisen, B., Beer, C., Dauer, D., Weinhardt, C.: A blockchain-based
smart grid: towards sustainable local energy markets. Comput. Sci. Res. Dev. 33, 207–214
(2017). https://doi.org/10.1007/s00450-017-0360-9
38. Augot, D., Chabanne, H., Chenevier, T., George, W., Lambert, L.: A user-centric system for
verified identities on the bitcoin blockchain. In: Garcia-Alfaro, J., Navarro-Arribas, G.,
Hartenstein, H., Herrera-Joancomartí, J. (eds.) ESORICS/DPM/CBT -2017. LNCS, vol.
10436, pp. 390–407. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-67816-0_22
39. Liu, P.T.S.: Medical record system using blockchain, big data and tokenization. In: Lam, K.-
Y., Chi, C.-H., Qing, S. (eds.) ICICS 2016. LNCS, vol. 9977, pp. 254–261. Springer, Cham
(2016). https://doi.org/10.1007/978-3-319-50011-9_20
40. Azouvi, S., Al-Bassam, M., Meiklejohn, S.: Who Am I? secure identity registration on
distributed ledgers. In: Garcia-Alfaro, J., Navarro-Arribas, G., Hartenstein, H., Herrera-
Joancomartí, J. (eds.) ESORICS/DPM/CBT -2017. LNCS, vol. 10436, pp. 373–389. Springer,
Cham (2017). https://doi.org/10.1007/978-3-319-67816-0_21
41. Bashir, M., Strickland, B., Bohr, J.: What motivates people to use bitcoin? In: Spiro, E., Ahn,
Y.-Y. (eds.) SocInfo 2016. LNCS, vol. 10047, pp. 347–367. Springer, Cham (2016). https://
doi.org/10.1007/978-3-319-47874-6_25
42. Vandervort, D., Gaucas, D., Jacques, R.S.: Issues in designing a bitcoin-like community
currency. In: Brenner, M., Christin, N., Johnson, B., Rohloff, K. (eds.) FC 2015. LNCS, vol.
8976, pp. 78–91. Springer, Heidelberg (2015). https://doi.org/10.1007/978-3-662-48051-9_6
43. Lamberti, F., Gatteschi, V., Demartini, C., Pranteda, C., Santamaria, V.: Blockchain or not
blockchain, that is the question of the insurance and other sectors. IT Professional 1 (2017)
44. Vo, H., Mehedy, L., Mohania, M., Abebe, E.: Blockchain-based data management and
analytics for micro-insurance applications. In: Proceedings of the 2017 ACM on Conference
on Information and Knowledge Management - CIKM 2017 (2017)
45. Turk, Ž., Klinc, R.: Potentials of blockchain technology for construction management.
Procedia Eng. 196, 638–645 (2017). https://doi.org/10.1016/j.proeng.2017.08.052
46. Steger, M., Dorri, A., Kanhere, S., Römer, K., Jurdak, R., Karner, M.: Secure Wireless
Automotive Software Updates Using Blockchains: a proof of concept. Adv. Microsyst.
Autom. Appl. 2017, 137–149 (2017)
47. Rimba, P., Tran, A., Weber, I., Staples, M., Ponomarev, A., Xu, X.: Comparing blockchain
and cloud services for business process execution. In: 2017 IEEE International Conference
on Software Architecture (ICSA) (2017). https://doi.org/10.1109/ICSA.2017.44
48. Forbes article. https://www.forbes.com/sites/quora/2017/09/21/whats-holding-blockchain-
back-from-large-scale-adoption/#267d559d2309
49. Conoscenti, M., Vetrò, A., De Martin, J.C.: Blockchain for the Internet of Things: a systematic
literature review, pp. 1–6 (2016)
50. Xu, Y., Li, Q., Min, X., Cui, L., Xiao, Z., Kong, L.: E-commerce blockchain consensus
mechanism for supporting high-throughput and real-time transaction. In: Wang, S., Zhou, A.
(eds.) CollaborateCom 2016. LNICST, vol. 201, pp. 490–496. Springer, Cham (2017). https://
doi.org/10.1007/978-3-319-59288-6_46
Blockchain for Business Applications 399
51. Delmolino, K., Arnett, M., Kosba, A., Miller, A., Shi, E.: Step by step towards creating a safe
smart contract: lessons and insights from a cryptocurrency lab. In: Clark, J., Meiklejohn, S.,
Ryan, P.Y.A., Wallach, D., Brenner, M., Rohloff, K. (eds.) FC 2016. LNCS, vol. 9604, pp.
79–94. Springer, Heidelberg (2016). https://doi.org/10.1007/978-3-662-53357-4_6
52. ISO. https://www.iso.org/committee/6266604.html
53. Roadmap for blockchain standards – Report. http://www.standards.org.au/OurOrganisation/
News/Documents/Roadmap_for_Blockchain_Standards_report.pdf
ICT-Based Support for the Collaboration
of Formal and Informal Caregivers –
A User-Centered Design Study
1
Technology and Inclusion Lab, Furtwangen University, Furtwangen, Germany
{madeleine.renyi,christophe.kunze}@hs-furtwangen.de
2
School of Business Administration and Economics,
Department of Accounting and Information Systems,
Osnabrück University, Osnabrück, Germany
[email protected]
Abstract. Given the demographic change and the resulting need for compre‐
hensive care strategies, collaborative care plays an important role in ensuring care
for everyone in need in the future. Despite the fact that collaboration software
tools can significantly relieve caregivers by improving various work processes,
their use in informal care networks has not yet become common practice. For this
purpose, a mixed method user-centered design study including literature review,
market research, system analysis, case studies as well as workshops and inter‐
views were conducted. On the basis of this, the needs and requirements for mobile
collaboration support were identified. The design of the requested application
must strike the balance between the request for simple functionality and the
diversity of information needed in order to offer an added value for all participants.
1 Introduction
The e-health business sector is steadily driving innovation in the healthcare market. A
special field of application of e-health is the support of long-term care. In principle, long-
term care is divided into four sectors [1]: (1) family and informal care sector, (2) state
or public sector, (3) voluntary and non-governmental-organization (NGO) sector and
(4) care market or private sector, whereby the respective shares depend on country,
family situation and health status of the person in need. As elderly care basically repre‐
sents an individual mixture of different care services, in the remainder of this contribu‐
tion, we refer to elderly care as care-mix. Bäuerle and Scherzer [2] define care-mix as
the composition of various support services and daily activities at an individual level,
whereby professional and voluntary services as well as help from relatives are combined
and networked.
There are several reasons illustrating that care-mix arrangements are important for
ensuring nursing care, for instance, the increasing burden on caring relatives
(compatibility of work and long-term care) and the expected increasing lack of skilled
employees in professional care [3]. In addition to nursing activities, care-mix also
includes services such as driving, meals on wheels or neighborhood assistance. In this
context, approaches like organized neighborhoods and caring communities are
becoming more and more the center of attention.
The future role of technical support systems in care-mix settings has not been inves‐
tigated in depth so far. Görres et al. [3] argue that the potential of computer supported
organized work is currently far from being fully exploited. The goal of the technology
deployment must be to coordinate and control the individual offers in order to optimize
care processes. As pointed out by Pinelle and Gutwin [4], present care-mix configura‐
tions are usually weakly structured and loosely coupled work processes in which the
involved employees have only limited access to information from other organizations.
Furthermore, the current low level of networking between the various actors impedes
the introduction of cross-organizational processes and information exchange systems.
The difficulty of reaching a “critical mass” of users, which is necessary for a meaningful
system use, as well as unequally distributed added values are described as typical
barriers [4].
The underlying research questions (RQ) of this contribution therefore are:
RQ1: What are the user expectations and requirements towards a support system for
collaboration in care-mix settings?
RQ2: What are the specific components of a mobile system supporting collaboration
in care-mix settings?
regions Kirchheim unter Teck and Wernau. In Fischerbach and Wernau there are well
organized and highly frequented neighborhood associations. Although in Kirchheim
unter Teck there are no such structures, the local district Rauner participates in a funded
project for the construction of district networking structures. In total 12 interviews and
10 workshops were conducted in these regions.
Interviews: The results of the literature analysis were used to develop guidelines that
enable a standardized interview conduction, which again ensures verifiable and usable
interview results. The interviews were conducted with two caring relatives, two civic
volunteers, one professional nurse, two coordinators of organized neighborhood assis‐
tance, three coordinators of care services, one coordinator of a welfare association and
one district manager. The interviews lasted between 30 and 70 min. All interviews were
recorded, transcribed and analyzed using the computer software MAXQDA (http://
www.maxqda.de/). The data-analyzation method is based on the concept of Mayring [7].
3 Related Work
The capabilities of supporting home-care settings through digital and connecting techno-
logies are discussed in most developed countries. The findings from literature relate to
ICT-Based Support for the Collaboration of Formal and Informal Caregivers 403
the fields documentation, (activity and social) awareness and collaboration, whereas the
presented systems include a wide range of recent technologies like mobile, ubiquitous
and ambient computing. Fitzpatrick and Ellingsen [10] provide insights of 25 years
(1987–2012) of computer supported collaborative work research in healthcare.
Design: Due to the fact that health care systems vary, home care is managed differently
from country to country. Therefore, research findings of one country may not hold true
for another. And yet, general design guidelines for informal care collaboration may be
useful (cf. [13, 14]).
1
Installation 10-50, last accessed: 13.12.2017, https://play.google.com/store/apps/details?
id=com.begsolutions.simplicare.full.
404 M. Renyi et al.
with family caregivers and care recipients and thus enhance transparency. Examples for
this business model are Mavencare (https://mavencare.com/connected-home-care) and
honor (https://www.joinhonor.com/apps).
Visionaries like Buurtzorg (Netherlands, http://www.buurtzorgnederland.com/) see
nursing and caring from a community perspective. This new organizational model
connects autonomous, self-guided teams with the neighborhood via cross-generational
platforms in order to ensure needs-oriented care structures. This concept has revolu‐
tionized the care market in the Netherlands and has already inspired other countries (e.g.
Belvita, Switzerland, https://belvita.ch/belvita-idee/).
Mixed financing: An example for this model is the Jointly-App (https://join‐
tlyapp.com/). Payment is made per care circle for which all associated accounts are free.
The fee for one circle ($ 2.99) can be individually borne by care providers, companies
(compatibility of work and family), health insurances or informal caregivers. The present
download statistics indicate that there is a certain interest in this business model.2
Beyond previous findings in the literature, this article aims to contribute to the
understanding of collaboration needs in care-mix settings in Germany. According to the
authors, it is indispensable to follow a holistic, mixed-method approach in order to
involve all participants in outpatient care networks. Further, the design of a sustainable,
functioning business model imperatively requires the involvement of the public and
private sector. Thereby, a special focus lies on the location- and device-independent
mobile support of the target group.
4 Results
The findings gained from our interviews and workshops are in line with those from
literature (c.f. related work section).
Technology and Usability: The technical affinity among the participants varies
considerably, which is dependent on age and (professional) experience of the respective
user. By and large, the affinity for technology within the target group is rather low and
ranges from daily (work and private life) to no use at all. And even though all respondents
stated to possess a smartphone, they mainly use it to make phone calls. In fact, phone
calls and e-mails constitute the preferred communication channels for professional care‐
givers. Semi-professionals and informal caregivers favor instant messaging services for
fast communication.
The preliminary assumption that younger and technologically more experienced
actors would be especially open and interested in the use of such platforms could not be
confirmed. Furthermore, the case studies showed that particularly less technically expe‐
rienced caregivers still have difficulties in handling mobile communication systems
while technically versed caregivers almost intuitively operate standard tools to improve
2
Installation 1.000-5.000, last accessed: 13.12.2017, https://play.google.com/store/apps/details?
id=org.carersuk.jointlyapp.
ICT-Based Support for the Collaboration of Formal and Informal Caregivers 405
communication and cooperation. The generally very appealing design as well as the
quite high usability of common tools set the standard for the care-specific software
solutions. If the specific tools do not meet the expectations, it must be reckoned that
these will be rejected by technically affine users. From the second case study it emerged
that the users perceived the different message types (tasks, appointments, text messages)
as rather confusing.
Responsibilities: When family members live in the immediate vicinity, they usually
take over the coordination of the individual care case as caregiving relative. Otherwise,
the role of the coordinator is taken by an official supervisor. The primary responsible
person in a care-mix setting has to keep the overview, is the contact person in case of
questions or problems, manages the care-mix and leads the network of actors.
According to the participants, networking the different actors would enhance the care
process. Usually, the respective actors know nothing about each other or who is respon‐
sible for what. Here, a consistent networking would ameliorate the coordination of
appointments, foster social interaction and strengthen the capacity for emergency
prevention and management.
“Tacit” Case Knowledge: Care-mix settings with professional nursing care usually
benefit from a patient record with a report section in the flat of the care recipient. This
406 M. Renyi et al.
Pin Board for Informal Communication: The pin board is meant to replace the
previous communication via paper notes. In this way information “not worth a call” can
be conveyed in the respective care-mix networks. For a fast overview of all entries in
all networks a separate collected view is needed. Via one click a private chat with the
author of the entry can be opened. Moreover, a comment and a like function encourage
social interaction. Tasks from the pin board can be marked as “done”. Appointments
can be created from pin board entries.
Information Page for Case Knowledge: To gain a basic knowledge of the care recip‐
ient, an information page is provided that contains e.g. the housing situation or prefer‐
ences of the care recipient. This page does not replace care documentation of a care
service provider, and only contains non-sensitive personal data to enhance the collabo‐
ration in the care-mix team.
Calendar for the Coordination of Appointments: The network specific calendar can
be used to organize appointments and tasks concerning the care of the central network
person, which prevents conflicting schedules. By means of an export function, appoint‐
ments can be transferred to the personal calendars.
All respondents explicitly stress that the use of smartphones would be desirable.
Notwithstanding the fact that most of the respondents use android-devices, the aspired
collaboration tool must equally be suitable for all involved actors. Hence, the application
is to be programmed device-independent (cf. client tier in Fig. 1). In this way, no one is
obliged to acquire new and expensive devices, which is what some respondents worry
about and may lead to refusal of the tool.
408 M. Renyi et al.
Another often discussed topic is the integration of sensors, such as life-sign buttons,
medication dispensers or motion detectors in the apartment of the person in need of care,
which provide additional information and convey a comforting feeling for the whole
network and especially the care responsible.
Whereas [11] set to a complete proprietary development, in the design process the
need to support different communication channels and optimally integrate these to one
stream, to reach out to all actors, was identified as a further requirement for the appli‐
cation. Push-notifications generate instant awareness and reduce the impact of media
breaks. Against this backdrop, the solution should be set up as a mixture of proprietary
development (e.g. chat, pin board) and external app launcher (e-mail, phone, internet
browser).
A frequently mentioned aspect is the internet connection. In that regard, rural more
than urban participants complain about poor internet connections. Especially the case
studies showed that a mobile support tool must include offline functionalities. A sepa‐
rated system architecture – containing a middleware tier for transaction services,
mapping and filtering, as well as a data tier – ensures low data traffic and fast loading
times. A synchronization of all essential information in the application tier guarantees
offline services (c.f. Fig. 1).
The mock-up presented in Fig. 2 has already been transformed into a first prototype
and is ready for a first field test. The application tier is implemented using the frontend
frameworks IONIC and Angular. Due to limited manpower, this first prototype does not
communicate with a middleware but directly with the server.
ICT-Based Support for the Collaboration of Formal and Informal Caregivers 409
Fig. 2. Mock-ups of the concept created with IONIC. (left) Start view which presents an overview
of all networks (middle) network specific calendar (right) message overview
Data Security, Roles and Permissions: The concept is designed to enhance collabo‐
ration in care-mix settings. The documentation of health-sensitive data which may stand
in conflict with data security laws is not intended. Nevertheless, it cannot be ruled out
that future users may misuse the tool. So far two roles (network administrators and
members) with different access restrictions are implemented. More complex role and
access models were evaluated as restrictive and confusing.
Design – One for All or All One’s Own? While it is generally agreed among the
stakeholder groups that the collaborative functionalities of a support platform are neces‐
sary, concept workshops showed quite diverse profiles of requirements per group. Other
study activities, however, showed that it is not necessary to have one single design for
all groups neither. According to the results, three designs are appropriate to cover the
needs of all stakeholders. One user interface should be designed for single network users
(less technically-versed persons). A second application should be designed for members
of multiple care-mix networks. (e.g., nurses, family doctors, neighborhood helpers) (c.f.
Fig. 1). The third design should be set up for professional actors like managers of care
services or organized neighborhood help. These stakeholders usually hold special roles
as advisors and contact persons, which requires a different overview of the networks,
persons, capabilities and resources.
Regardless of this, in order to find acceptance, the design must be appealing and
meet the high standards of universal collaboration software.
References
3. Görres, S., Seibert, K., Stiefler, S.: Perspektiven zum pflegerischen Versorgungsmix. Pflege-
report 2016, pp. 3–14. Schattauer GmbH, Stuttgart (2016)
4. Pinelle, D., Gutwin, C.: A groupware design framework for loosely coupled workgroups. In:
Gellersen, H., Schmidt, K., Beaudouin-Lafon, M., Mackay, W. (eds.) ECSCW 2005, pp. 65–
82. Springer, Dordrecht (2005). https://doi.org/10.1007/1-4020-4023-7_4
5. Szabo, E.: Organisationskultur-Forschung im Überblick. In: Organisationskultur und
Ethnographie, pp. 1–46. Deutscher Universitätsverlag, Wiesbaden (1998)
6. Muller, M.J., Kuhn, S.: Participatory design. Commun. ACM 36, 24–28 (1993)
7. Mayring, P.: Qualitative inhaltsanalyse. In: Handbuch Qualitative Forschung in der
Psychologie, pp. 601–613. VS Verlag für Sozialwissenschaften, Wiesbaden (2010)
8. Hutchinson, H., Mackay, W., Westerlund, B., Bederson, B.B., Druin, A., Plaisant, C.,
Beaudouin-Lafon, M., Conversy, S., Evans, H., Hansen, H.: Technology probes: inspiring
design for and with families. In: Proceedings of the SIGCHI Conference on Human Factors
in Computing Systems, pp. 17–24 (2003)
9. Rosson, M.B., Carroll, J.M.: Scenario-based design. In: Sears, A., Jacko, J.A. (eds.) Human-
Computer Interaction, pp. 145–162. CRC Press, Boca Raton (2009)
10. Fitzpatrick, G., Ellingsen, G.: A review of 25 years of CSCW research in healthcare:
contributions, challenges and future agendas. Comput. Support. Coop. Work 22, 609–665
(2012)
11. Bossen, C., Christensen, L.R., Grönvall, E., Vestergaard, L.S.: CareCoor: augmenting the
coordination of cooperative home care work. Int. J. Med. Inform. 82, e189–e199 (2013)
12. Moser, C., Krischkowsky, A., Neureiter, K., Tscheligi, M.: Mediating informal care online:
findings from an extensive requirements analysis. Interact. Des. Archit. J. 24, 33–48 (2015)
13. Abou Amsha, K., Lewkowicz, M.: Shifting patterns in home care work: supporting
collaboration among self-employed care actors. In: De Angeli, A., Bannon, L., Marti, P.,
Bordin, S. (eds.) COOP 2016: Proceedings of the 12th International Conference on the Design
of Cooperative Systems, 23–27 May 2016, Trento, Italy, pp. 139–154. Springer, Cham
(2016). https://doi.org/10.1007/978-3-319-33464-6_9
14. Schorch, M., Wan, L., Randall, D.W., Wulf, V.: Designing for those who are overlooked -
insider perspectives on care practices and cooperative work of elderly informal caregivers.
In: Proceedings of the 19th ACM Conference on Computer-Supported Cooperative Work
and Social Computing, CSCW 2016, pp. 785–797 (2016)
15. Bødker, S., Grönvall, E.: Calendars: time coordination and overview in families and beyond.
In: Bertelsen, O., Ciolfi, L., Grasso, M., Papadopoulos, G. (eds.) ECSCW 2013: Proceedings
of the 13th European Conference on Computer Supported Cooperative Work, 21–25
September 2013, Paphos, Cyprus, pp. 63–81. Springer, London (2013). https://doi.org/
10.1007/978-1-4471-5346-7_4
16. Ganoe, C.H., Somervell, J.P., Neale, D.C., Isenhour, P.L., Carroll, J.M., Rosson, M.B.,
Mccrickard, D.S.: Classroom BRIDGE: using collaborative public and desktop timelines to
support activity awareness. In: Proceedings of the 16th Annual ACM Symposium on User
Interface Software and Technology, pp. 21–30. ACM Press, New York (2003)
17. Bastide, R., Bardy, P., Borrel, B., Boszodi, C., Bouet, M., Gani, K., Gayraud, E., Gourc, D.,
Lamine, E., Manenq, P.H., Schneider, M., Toumani, F.: Plas ‘O’Soins: a software platform
for modeling, planning and monitoring homecare activities. IRBM 35, 82–87 (2014)
Identifying Suitable Representation Techniques
for the Prioritization of Requirements
and Their Interdependencies for Multiple
Software Product Lines
1 Introduction
they are not met, the original benefit of diversifying the software product portfolio has
not been completely realized.
This paper considers the representation techniques in the prioritization of require‐
ments which have interdependencies between SPLs. Focus is given to the interdepen‐
dencies of requirements which are most critical to market-driven software [1] – require‐
ments related to value and cost. It starts with an overview of the current research in the
areas of multiple SPLs and requirements management, including prioritization of
requirements and requirements interdependencies. Following the overview is a
summary of five representation techniques for the prioritization of requirements and
their interdependencies. Finally, the paper concludes with the authors’ assessment of
the five representation techniques based on requirement prioritization, requirement
interdependencies, and the consideration of cost/value interdependencies between SPLs.
2 Overview
This overview delves deeper into the subject areas and foremost research of multiple
SPLs and managing software requirements, including the prioritization of requirements
and assessment of requirement interdependencies.
Much consideration has been given to requirements management for single software
products, but far less consideration has been given to requirements between software
product lines [7] or to particular domains [8].
A software product line (SPL) is a set of software products sharing a set of common
features but containing variation points [9]. One of the advantages to SPLs is the reduced
cost of development and testing, with an increased opportunity to address different
stakeholder groups [7]. An example for a single SPL, given by Rosenmüller et al. [10],
is mail client software which relies on a common mail framework. Variations of the mail
client software, to support different protocols, for example, would all rely on the same
common mail framework [10].
Multiple SPLs, in comparison, are composed of many interconnected subsystem
versions and variants [6]. Multiple SPLs commonly refers to vertically tiered software
stacks, with application SPLs and infrastructure SPLs [5], but can also refer to distributed
SPLs, as in the case of sensor software [10]. An example of vertically tiered multiple
SPLs would be individual application software product lines that all rely on the same
database platform, which is itself a software product line [5, 6] (Fig. 1).
The analysis that follows in this paper focuses on requirements representation techni‐
ques and their application to requirements prioritization and requirements interdependen‐
cies between multiple SPLs, specifically vertically tiered SPLs where the upper tier (appli‐
cation) software requires software functionality from the lower (infrastructure) tiers.
Furthermore, the focus in this paper is on mature products on the market, with
development approach that is incremental and market-driven, as in the industrial survey
conducted by Carlshamre et al. [1]. The focus has been restricted to market-driven
414 S. Lewellen and M. Helfert
Fig. 1. Dependency model between an application SPL and an infrastructure SPL [5]
software, because most requirements interdependencies for this type of software are of
the “cost/value” type [1].
It is not possible to address all functional and non-functional requirements in the next
software release, due to constraints from budget, resources or time [7]. Therefore, a
prioritization of requirements should be applied [8].
which have been identified in the foremost literature, (2) identify the requirements inter‐
dependency types which are most critical to the scenario that one SPL has a requirement
dependency to another SPL.
Tables 1 and 2 provide the most referenced sources for requirement interdependency
categories [1, 13], and are in general alignment with each other on interdependency
types.
These requirement interdependency types provide the basis for a more complex
analysis of how interdependencies, sometimes also called requirements relations, are
considered in representation techniques for multiple SPLs. In Fig. 2, the difference
between a requirement interdependency within one SPL and an interdependency
between multiple SPLs is depicted. While most of the foremost literature and represen‐
tation techniques consider the simple case of a requirement requiring another require‐
ment within the same SPL, the case we want to consider is how the foremost represen‐
tation techniques are suitable to the more complex case that there is a requirement inter‐
dependency between product lines.
416 S. Lewellen and M. Helfert
Fig. 2. Requirement interdependency within one SPL as opposed to between multiple SPLs
Because each SPL is, in its software productization, independent of the other one,
the separate product lines have the potential to have different release timelines, market
segments, share of company operational revenue, etc. [5]. Therefore, we can identify
some existing requirement interdependency types to be more critical for requirements
between SPLs.
We identify the “requires” type to be critical due to the definition of vertically tiered
multiple SPLs. If a requirement between SPLs exist, it will always be a requirement
from the top-most-level SPL of the SPL tier beneath it. Additionally, any requirement
between SPLs most likely also has a “cost/value” interdependency type due to the nature
of SPLs, which are designed to increase the value of software options at reduced overall
development cost [1].
With this research, we aim to answer: which of the foremost techniques for repre‐
senting the prioritization and interdependencies of requirements within one SPL is the
most suitable for representing the more complex requirements interdependencies
between multiple SPLs?
The software requirements catalog (SRC) is a method for collecting and considering
software requirements for reuse instead of considering individual software features or
418 S. Lewellen and M. Helfert
components for reuse. The creation of the SRC includes a classification phase, where
the functionality of the requirement to fulfill the goal – the reason for the existence of
the project – is described. The description includes a prioritization rating which reflects
the suitability of the requirement to the project goal [17–19].
There is a qualitative high-medium-low rating scale for evaluating the priority of the
requirement to fulfill the project goal. However, the definition of a “priority” in the SRC
context is a variation on the definition of priority we have discussed previously. In the
context of SRC, priority is a rating of the requirement in only the parameter of its suit‐
ability to fulfill the overall project purpose, and not of its overall criticality [17].
The requirement constraints and dependencies with other software projects are
defined and refined in order to continuously update the requirements catalog [17]. The
tracking of requirements interdependencies, also called traceability, seems to consider
simple relationships, such as the “requires” structural interdependency between projects
with common requirements. However, there is no mention of more complex require‐
ments interdependencies between projects like “cost/value” interdependencies.
In the following section, we present the criteria we used to assess the requirement
prioritization and interdependency representation techniques suitable for multiple SPLs.
We then compare the overall rating for the representation types discussed in Sect. 3.
Identifying Suitable Representation Techniques 419
Priority Consideration. Each representation type in Sect. 3. was evaluated against the
sophistication of requirement priority consideration on a decimal scale from zero to one
using the criteria in Table 3. A rating of zero corresponds to an absence of consideration.
A rating of 0.5 corresponds to a simple scale priority rating, where requirements are
given a standalone rating. A rating of 1.0 corresponds to a comparative prioritization
where the requirements are compared to one another and then receive a relative priority
rating.
Suitability for Multiple SPLs. Each representation type in Sect. 3 was evaluated
against the suitability of the type for multiple SPLs on a decimal scale from zero to one
using the criteria in Table 5. Specifically, the suitability criteria refer to the interde‐
pendency types “requires” and “cost/value”, which play a critical role in the assessment
of requirements interdependencies between SPLs. A rating of zero corresponds to an
absence of consideration for even the most basic structural interdependency type,
“requires”. A rating of 0.5 corresponds to a simple consideration for either “requires”
or “cost/value” interdependency types. A rating of 0.75 corresponds to either consider‐
ation for both “requires” and “cost/value” interdependency types or an in-depth consid‐
eration of either one. A rating of 1.0 corresponds to a representation type that allows for
420 S. Lewellen and M. Helfert
requirement consideration for requirements from external SPLs, which bring potentially
their own “cost/value” and “requires” requirements.
It was determined based on the given criteria that the SRC method and metamodel
ontology both had a simple prioritization method and a simple interdependency tracking
method. They both also had no mention of the “cost/value” requirements interdepend‐
ency type, which is critical to a thorough representation of requirements between SPLs.
The metamodel ontology representation rates slightly higher on the graph for multiple
SPL suitability because it has greater potential for further extensibility through metadata.
The SRC method limits flexibility to requirements interdependency handling between
SPLs because the requirement scoring is from the perspective of the requirement to
fulfill a specific project, and is therefore better suited to smaller software projects with
fewer diverse stakeholders.
Identifying Suitable Representation Techniques 421
Slightly more suited for requirements between SPLs is the fuzzy graph representa‐
tion, which gives very detailed information about requirements interdependencies and
their uncertainties, but does not take prioritization into consideration at all.
Most suitable to requirements between SPLs are the directed graph (digraph) model
and the cost-value diagram. Both are capable of distilling the required requirements
interdependency types in a compact and straightforward way. Both could potentially be
extended to represent relationships between SPLs. However, the cost-value diagram
adds an additional perspective of potential investment areas (ex. high-value/low-cost
requirements), which could valuable if extended to multiple SPL requirements support.
5 Conclusion
References
1. Carlshamre, P., Sandahl, K., Lindvall, M., Regnell, B., och Dag, J.: An industrial survey of
requirements interdependencies in software product release planning. In: Proceedings of the
Fifth IEEE International Symposium, Requirements Engineering (2001)
2. Mougouei, D., Powers, D.M., Moeini, A.: Dependency-aware software release planning. In:
IEEE/ACM 39th International Conference Software Engineering Companion (ICSE-C), 2017
(2017)
3. Bagnall, A.J., Rayward-Smith, V.J., Whittley, I.M.: The next release problem. Inf. Softw.
Technol. 43(14), 883–890 (2001)
4. Carlshamre, P., Regnell, B.: Requirements lifecycle management and release planning in
market-driven requirements engineering processes. In: 11th International Workshop
Database and Expert Systems Applications, Proceedings (2000)
5. Schirmeier, H., Spinczyk, O.: Challenges in software product line composition. In: 42nd
Hawaii International Conference, System Sciences, HICSS 2009 (2009)
6. Damiani, F., Schaefer, I., Winkelmann, T.: Delta-oriented multi software product lines. In:
Proceedings of the 18th International Software Product Line Conference, vol. 1 (2014)
7. Soomro, S., Hafeez, A., Shaikh, A., Musavi, S.H.A.: Ontology based requirement
interdependency representation and visualization. In: Shaikh, F.K., Chowdhry, B.S.,
Zeadally, S., Hussain, D.M.A., Memon, A.A., Uqaili, M.A. (eds.) IMTIC 2013. CCIS, vol.
414, pp. 259–270. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10987-9_24
8. Berander, P., Andrews, A.: Requirements prioritization. In: Aurum, A., Wohlin, C. (eds.)
Engineering and Managing Software Requirements, pp. 69–91. Springer, Heidelberg (2005).
https://doi.org/10.1007/3-540-28244-0_4
9. Sellier, D., Mannion, M., Mansell, J.X.: Managing requirements inter-dependency for
software product line derivation. Requirements Eng. 13(4), 299–313 (2008)
10. Rosenmüller, M., Siegmund, N., Kästner, C., ur Rahman, S.: Modeling dependent software
product lines. In: Proceedings of the GPCE Workshop on Modularization, Composition and
Generative Techniques for Product Line Engineering (McGPLE) (2008)
11. Achimugu, P., Selamat, A., Ibrahim, R., Mahrin, M.N.: A systematic literature review of
software requirements prioritization. Inf. Softw. Technol. 56(6), 568–585 (2014)
12. Kahn, J., Rehman, I.: Comparison of requirement prioritization technique to find the best
priorization technique. Modern Educ. Comput. Sci. 7(11), 53–59 (2015)
13. Dahlstedt, A., Persson, A.: Requirements Interdependencies: State of the Art and Future
Challenges. In: Aurum, A., Wohlin, C. (eds.) Engineering and Managing Software Requirements,
pp. 95–116. Springer, Heidelberg (2005). https://doi.org/10.1007/3-540-28244-0_5
14. Zhang, H., Li, J., Zhu, L., Jeffery, R., Liu, Y., Wang, Q., Li, M.: Investigating dependencies
in software requirements for change propagation analysis. Inf. Softw. Technol. 56(1), 40–53
(2014)
15. Goknil, A., Kurtev, I., van den Berg, K., Veldhuis, J.: Semantics of trace relations in
requirements models for consistency checking and inferencing. Softw. Syst. Model. 10(1),
31–54 (2011)
16. Saeki, M., Kaiya, H.: On Relationships among models, meta models, and ontologies. In:
Proceedings of the Proceedings of the 6th OOPSLA Workshop on Domain-Specific Modeling
(DSM 2006) (2006)
17. Pacheco, C., Garcia, I., Calvo-Manzano, J., Arcilla, M.: Reusing fuctional software
requirements in small-sized software enterprises: a model oriented to the catalog of
requirements. Requirements Eng. 22(2), 275–287 (2017)
Identifying Suitable Representation Techniques 423
18. Robertson, S., Robertson, J.: Mastering the Requirements Process, 6th edn. Addison-Wesley,
Munich (2013)
19. Wiegers, K.: First things first: prioritizing requirements. Software Dev. 7(9), 48–53 (1999)
20. Mougouei, D., Powers, D.: Modeling and selection of interdependent software requirements
using fuzzy graphs. Int. J. Fuzzy Syst. 19(6), 1812–1828 (2017)
21. Karlsson, J., Olsson, S., Ryan, K.: Improved practical support for large-scale requirements
prioritising. Requirements Eng. 2(1), 51–60 (1997)
22. Valiris, G., Chytas, P., Glykas, M.: Making decisions using the balanced scorecard and the
simple multi-attribute rating technique. Perform. Meas. Metrics 6(3), 159–171 (2005)
Author Index