Copy of Software Engineering AI Vol 17 - Roger Lee
Copy of Software Engineering AI Vol 17 - Roger Lee
Series Editor
Janusz Kacprzyk
Polish Academy of Sciences, Warsaw, Poland
© The Editor(s) (if applicable) and The Author(s), under exclusive license
to Springer Nature Switzerland AG 2024
This work is subject to copyright. All rights are solely and exclusively
licensed by the Publisher, whether the whole or part of the material is
concerned, specifically the rights of translation, reprinting, reuse of
illustrations, recitation, broadcasting, reproduction on microfilms or in any
other physical way, and transmission or information storage and retrieval,
electronic adaptation, computer software, or by similar or dissimilar
methodology now known or hereafter developed.
The publisher, the authors and the editors are safe to assume that the advice
and information in this book are believed to be true and accurate at the date
of publication. Neither the publisher nor the authors or the editors give a
warranty, expressed or implied, with respect to the material contained
herein or for any errors or omissions that may have been made. The
publisher remains neutral with regard to jurisdictional claims in published
maps and institutional affiliations.
This Springer imprint is published by the registered company Springer
Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham,
Switzerland
Foreword
The main purpose of this book is to seek peer-reviewed original research
papers on the foundations and new developments in Networking and
Parallel/Distributed Computing Systems. The focus will also be on
publishing in a timely manner, the results of applying new and emerging
technologies originating from research in Networking and
Parallel/Distributed Computing Systems. The findings of this book can be
applied to a variety of areas, and applications can range across many fields.
The papers in this book were chosen based on review scores submitted
by members of the editorial review board and underwent rigorous rounds of
review.
We would like to thank all contributors including all reviewers, and all
editorial board members of this book for their cooperation in helping to
publish this book.
It is our sincere hope that this book provides stimulation and inspiration,
and that it will be used as a foundation for works to come.
Qiuxiang Yang
Simon Xu
Lizhen Fu
July 2023
Editorial Review Board
Kiumi Akingbehin, University of Michigan, USA
Yasmine Arafa, University of Greenwich, UK
Jongmoon Baik, Korea Advanced Institute of Science and Technology,
South Korea
Ala Barzinji, University of Greenwich, UK
Radhakrishna Bhat, Manipal Institute of Technology, India
Victor Chan, Macao Polytechnic Institute, Macao
Morshed Chowdhury, Deakin University, Australia
Alfredo Cuzzocrea, University of Calabria, Italy
Hongbin Dong, Harbin Engineering University, China
Yucong Duan Hainan, University, China
Zongming Fei, University of Kentucky, USA
Honghao Gao, Shanghai University, China
Cigdem Gencel Ambrosini, Ankara Medipol University, Italy
Gwangyong Gim, Soongsil University, South Korea
Takaaki Goto, Toyo University, Japan
Gongzhu Hu, Central Michigan University, USA
Wen-Chen Hu, University of North Dakota, USA
Naohiro Ishii, Advanced Institute of Industrial Technology, Japan
Motoi Iwashita, Chiba Institute of Technology, Japan
Kazunori Iwata, Aichi University, Japan
Keiichi Kaneko, Tokyo University of Agriculture and Technology, Japan
Jong-Bae Kim, Soongsil University, South Korea
Jongyeop Kim, Georgia Southern University, USA
Hidetsugu Kohzaki, Kyoto University, Japan
Cyril S. Ku, William Paterson University, USA
Joonhee Kwon, Kyonggi University, South Korea
Sungtaek Lee, Yong in University, South Korea
Weimin Li, Shanghai University, China
Jay Ligatti, University of South Florida, USA
Chuan-Ming Liu, National Taipei University of Technology, Taiwan
Man Fung Lo, The University of Hong Kong, Hong Kong
Chaoying Ma, Greenwich University, UK
Prabhat Mahanti, University of New Brunswick, Canada
Tokuro Matsuo, Advanced Institute of Industrial Technology, Japan
Mohamed Arezki Mellal, M’Hamed Bougara University, Algeria
Jose M. Molina, Universidad Carlos III de Madrid, Spain
Kazuya Odagiri Sugiyama, Jogakuen University, Japan
Takanobu Otsuka, Nagoya Institute of Technology, Japan
Anupam Panwar, Apple Inc., USA
Kyungeun Park, Towson University, USA
Chang-Shyh Peng, California Lutheran University, USA
Taoxin Peng, Edinburgh Napier University, UK
Isidoros Perikos, University of Patras, Greece
Laxmisha Rai, Shandong University of Science and Technology, China
Fenghui Ren, University of Wollongong, Australia
Kyung-Hyune Rhee, Pukyong National University, South Korea
Abdel-Badeeh Salem, Ain Shams University, Egypt
Toramatsu Shintani, Nagoya Insutitute of Technology, Japan
Junping Sun, Nova Southeastern University, USA
Haruaki Tamada, Kyoto Sangyo University, Japan
Takao Terano, Tokyo Institute of Technology, Japan
Kar-Ann Toh, Yonsei University, South Korea
Masateru Tsunoda, Kindai University, Japan
Trong Van Hung, Vietnam Korea University of Information and
Communications Tech., Viet Nam
Shang Wenqian, Communication University of China, China
John Z. Zhang, University of Lethbridge, Canada
Rei Zhg, Tongji University, China
Contents
Develop a System to Analyze Logs of a Given System Using Machine
Learning
Md. Tarek Hasan, Farzana Sadia, Mahady Hasan and
M. Rokonuzzaman
Study on Locality, Fairness, and Optimal Resource Allocation in
Cluster Scheduling
Cherindranath Reddy Vanguru
Digital Word-of-Mouth and Purchase Intention.An Empirical Study in
Millennial Female and Consumers
Melissa del Pilar Usurin-Flores, Miguel Humberto Panez-Bendezú and
Jorge Alberto Vargas-Merino
Alignment of Business Process and Information System Models
Through Explicit Traceability
Aljia Bouzidi, Nahla Zaaaboub Haddar and Kais Haddar
An Assessment of Fintech for Open Banking:Data Security and
Privacy Strategies from the Perspective of Fintech Users
Amila Munasinghe, Srimannarayana Grandhi and Tasadduq Imam
Using Key Point Detection to Extract Three Dimensional Phenotypes
of Corn
Yuliang Gao, Zhen Li, Seiichi Serikawa, Bin Li and Lifeng Zhang
Optic Cup Segmentation from Fundus Image Using Swin-Unet
Xiaozhong Xue, Linni Wang, Ayaka Ehiro, Yahui Peng and Weiwei Du
From Above and Beyond:Decoding Urban Aesthetics with the Visual
Pollution Index
Advait Gupta, Manan Padsala, Devesh Jani, Tanmay Bisen,
Aastha Shayla and Susham Biswas
Subcellular Protein Patterns Classification Using Extreme Gradient
Boosting with Deep Transfer Learning as Feature Extractor
Manop Phankokkruad and Sirirat Wacharawichanant
Building a Shapley FinBERTopic System to Interpret Topics and
Articles Affecting Stock Prices
Yoshihiro Nishi and Takahashi Hiroshi
Can a Large Language Model Generate Plausible Business Cases from
Agent-Based Simulation Results?
Takamasa Kikuchi, Yuji Tanaka, Masaaki Kunigami, Hiroshi Takahashi
and Takao Terano
Analyzing the Growth Patterns of GitHub Projects to Construct Best
Practices for Project Managements
Kaide Kaito and Haruaki Tamada
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed
Computing , Studies in Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_1
Farzana Sadia
Email: [email protected]
Mahady Hasan
Email: [email protected]
M. Rokonuzzaman
Email: [email protected]
Abstract
Software error detection is a critical aspect of software development.
However, due to the lack of time, budget, and workforce, testing
applications can be challenging, and in some cases, bug reports may not
make it to the final stage. Additionally, a lack of product domain knowledge
can lead to misinterpretation of calculations, resulting in errors. To address
these challenges, early bug prediction is necessary to develop error-free and
efficient applications. In this study, the author proposed a system that uses
machine learning to analyze system error logs and detect errors in real time.
The proposed system leverages imbalanced data sets from live servers
running applications developed using PHP and Codeigniter. The system
uses classification algorithms to identify errors and suggests steps to
overcome them, thus improving the software’s quality, reliability, and
efficiency. Our approach addresses the challenges associated with large and
complex software where it can be difficult to identify bugs in the early
stages. By analyzing system logs, we demonstrate how machine learning
classification algorithms can be used to detect errors and improve system
performance. Our work contributes to a better understanding of how
machine learning can be used in real-world applications and highlights the
practical benefits of early bug prediction in software development.
1 Introduction
In today’s data-driven world, data plays a significant role in almost every
aspect of our lives. As more and more businesses rely on data analysis to
drive decision-making, the importance of accurate and error-free data has
become increasingly apparent [1]. Large-scale data sets are collected and
evaluated across various industries, and the problem of errors within those
data sets has become more pressing. One common source of data for
software applications is system logs, which often contain valuable
information and error messages. However, many researchers and analysts
overlook the importance of proper data cleaning and preparation before
analysis [1].
This paper focuses on analyzing system logs to identify and address
errors using machine learning techniques. The goal is to provide guidelines
and solutions for optimizing software performance and reliability. By
leveraging machine learning algorithms, this study aims to predict errors
and provide solutions to prevent instability or inconsistency in the software
during the development life cycle.
To achieve this goal, the proposed framework includes a search-based
testing approach using deep neural networks. The framework incorporates
strategies for code embedding, refactoring policy, mutation testing, and
evaluating test cases. The system logs are preprocessed and analyzed using
machine learning algorithms to identify patterns and predict errors. The
resulting guidelines and solutions can be applied to any application logs,
and the proposed data mining export code can be easily adapted for use in
other settings.
Ultimately, the goal of this study is to contribute to the development of
error-free software by providing a set of cleansed data for further
investigation and analysis. By focusing on data quality and leveraging
machine learning techniques, we hope to improve software performance,
reliability, and efficiency.
The problem of the paper is related to the issues faced by companies
due to the delivery of software without proper testing [15] and quality
checking, resulting in increased development costs. Additionally, there are
difficulties in creating unique or unfamiliar business transactions in the
testing environment. Furthermore, even if the software is running well,
errors in logs may arise due to not following proper standards or unknown
process cycles. These errors may not stop the application’s execution, but
the author creates a considerable amount of logs in the production server,
which affects performance and stability. Missing data integrity can also
create errors and faults in transactions. Therefore, the paper aims to deal
with these log issues and provide guidelines for their resolution without
human intervention. The authors propose a unified algorithm for data
cleansing, and the focus is on analyzing the logs of systems running in the
production environment. The authors suggest machine learning algorithms
to detect errors and propose solutions to optimize them to create an error-
free application. The author aims to create a framework for analyzing big
data to improve fault detection and problem identification.
The authors aim to develop a unified algorithm that can resolve data
quality issues in unclean logs without the need for human intervention or
master data [2]. The focus of their study is on analyzing the logs of running
systems in the production environment to provide guidelines and solutions
for optimizing errors and creating error-free applications. They run machine
learning algorithms to identify errors in the logs and suggest solutions to
overcome them. The authors also emphasize the need for a framework for
analyzing big data to improve fault detection and problem identification
during data preprocessing [1].
2 Literature Review
In recent years, the use of machine learning for log analysis has gained
significant attention in the field of software engineering [3]. Researchers
have proposed various techniques and models for log analysis to improve
software reliability, performance, and maintainability.
One approach is to use clustering algorithms to group log messages
based on their similarity, which can help to identify common patterns and
anomalies in the log data [4]. Another approach is to use classification
algorithms to detect and categorize different types of log messages, such as
errors, warnings, and informational messages [5].
Researchers have also explored the use of natural language processing
(NLP) techniques for log analysis, such as topic modeling and sentiment
analysis, to gain insights into the causes of log messages and to identify
potential areas for improvement [6].
In addition to machine learning, researchers have also proposed other
approaches for log analysis, such as rule-based systems and pattern-
matching techniques [3]. Rule-based systems use predefined rules to detect
specific types of log messages, while pattern-matching techniques search
for specific patterns or sequences of log messages that may indicate a
problem.
Despite the various approaches proposed for log analysis, there are still
challenges and limitations in this field. One challenge is the complexity and
variability of log data, which can make it difficult to develop accurate
models and algorithms [4]. Another challenge is the lack of labeled data for
training and testing machine learning models, which can limit their
effectiveness [5].
Despite these challenges, the potential benefits of log analysis using
machine learning are significant, including improved software quality,
reduced downtime, and increased productivity [6].
3 Research Design
The aim of this research is to detect and classify system errors by using a
Random Forest feature selection algorithm, to minimize the error rate and
predict possible error solutions based on current errors.
3.1 Data Collection
The system log files will be collected from a live-running application that is
privately available. The data will be selected from a specific time period to
ensure consistency in the data. The dataset will be cleaned using the random
sampling technique to reduce the size of the dataset and make it more
manageable. Those applications are creating a huge number of log files [2]
where authors have seen many error tags. Since the logs are huge and there
is a lack of a big data handling machine. The author picked up 1 week of
data from 2021, which are about 921MB,
3.5 Evaluation
The performance of the algorithm will be evaluated based on accuracy and
precision. The results will be compared to existing methods for detecting
and classifying system errors. The data was not organized because it was
gathered in its raw form, from sources. It contains null values, incomplete
data, some missing values, and a sampling date format that is inconsistent.
Thus, data analysis had to be performed using any relevant language and
algorithm to check the dirtiness of the data, and then a well-structured
dataset could be generated using data cleaning techniques, algorithms, and
procedures, which could then be used for analysis or visualization [8].
The research will provide insights into the causes of system errors and
offer possible solutions to minimize the error rate. The Random Forest
feature selection algorithm will be useful in identifying the most relevant
features for analysis, and the error classification system will provide a
framework for predicting possible error solutions. The results of this
research will contribute to the field of data mining and error analysis.
The study will only consider a specific time period and may not provide
a comprehensive analysis of system errors. The dataset is based on a live-
running application, and the results may not be generalizable to other
applications. The study may also face computational limitations due to the
large size of the dataset (see Fig. 1).
Fig. 1 Data analysis model
4 Findings
To achieve a clean, error-free data set for analysis in machine learning, pre-
processing raw data is a critical step, as outlined by Haider et al. [11]. In
this study, customized algorithms were employed to clean the data, and
Python or R are recommended for this task due to their built-in libraries for
statistical analysis and interpretation, as noted by Hossen and Sayeed [1].
The data analyzed in this study covers a one-week period, during which a
total error count of 4042852 was observed.
Upon analysis of the cleaned data set, it was found that only warning
and notice-type errors occurred, which did not cause the application to stop
executing any script. These types of errors are non-fatal and do not halt
script execution. The cleaned data set contains various columns related to
the data, such as Title (log type), Type (error type), the affected variable and
line number, error, and filename. Table 1 provides an overview of our
cleaned data.
Table 1 Mined data set
Title Type Variable, line File
ERROR Notice payment, 192 /controller/./billing.php
ERROR Notice courses, 119 /controller/./billing.php
ERROR Notice semester, 120 /controller/./billing.php
ERROR Warning LastName, 63 /controller/./landing.php
Table 2 illustrates the error frequency which has been generated using
Anaconda a big data handling tool. The first column contains information
about the error type, then the second column represents the error file
location and the third column shows how many errors occurred on those
files.
Table 2 Error type and frequency
Table 5 is used to detect system total error and non-error count which
has shown in Fig. 2.
Fig. 2 Error rate
Raw data is often incomplete, inconsistent, and redundant, making it
unsuitable for direct data mining. Therefore, advanced analysis techniques
are required to process the data [11]. In this study, the author decided to use
one day of acquired data and apply machine-learning techniques [1]. The
data had a shape of (475680, 5), but since the CSV file did not provide
numeric data, a method for transforming nominal data into numeric features
was used [13]. To detect whether each row of the data represents an error or
not, the raw data had to be converted into a numerical format [14]. Excel
was used to generate the desired pattern, which can be seen more clearly in
Fig. 3 and Table 6.
Table 6 Error pattern mining
The inclusion of the affected file name and line number in the cleaned
data set mentioned in Table 1 is expected to be beneficial for future
research. With this information, an AI system can locate faulty files and
affected variables mentioned there, and suggest solutions for the identified
issues. In addition, a new approach called Deep Check can be proposed for
testing Deep Neural Networks (DNN) using symbolic execution and
program analysis. Deep Check uses a white-box technique to facilitate
symbolic execution and identify critical elements in the DNN [9].
In order to utilize the data for analysis, it was transformed into a
structured format. To ensure its reliability, the information was verified and
checked for any instances of missing data [1].
Our proposed methodology requires the use of supervised learning, with
logistic regression being the preferred model due to our data’s pattern.
Given the large input data, it is more convenient to create a prediction
model. To ensure the accuracy of the preprocessing, mining patterns, and
analysis, we selected a single day’s data to train and test the logistic
regression model. The system must differentiate between errors and non-
errors in the selected data. We split the data into training and test sets, with
20% used for testing and 80% for training, resulting in a total dataset of
(475680, 2), with (380544, 2) for training and (95136, 2) for testing.
5 Proposed Model
The proposed solution suggested by the authors includes building a data-
cleaning function using machine learning algorithms to predict potential
errors in data. To further improve the application’s quality, the authors
suggest creating loosely coupled modules and classes with global variable
declarations and high cohesion. The actual code-writing process should
involve array or variable declarations, type declarations of variables, isset
checks, empty value checks, and type checks before using any variables.
Finally, a Unit test is recommended to optimize unwanted errors.
To incorporate this proposed solution, developers can follow these
guidelines during the development process of an application or feature to
reduce the chances of errors and bugs. By creating loosely coupled modules
and classes with global variable declarations, developers can make sure that
their code is easy to maintain and update. Moreover, by following the
suggested coding practices, such as array or variable declarations, type
declarations of variables, isset checks, empty value checks, and type checks
before using any variables, developers can ensure that their code is error-
free and more reliable. Finally, by conducting a Unit test, developers can
identify and optimize any unwanted errors, further improving the
application’s overall quality.
Overall, incorporating these guidelines into the development process
can optimize the production cost and time, make the application scalable,
reliable, and faster, and ultimately, ensure an error-free log or output (see
Fig. 4).
Fig. 4 Proposed model
6 Conclusions
The purpose of this study was to create a data cleaning function that can
improve data quality by identifying and predicting errors. The research
question focused on whether machine learning techniques can be used to
improve data quality in software applications. The literature review
highlighted the importance of data cleaning in machine learning and the use
of various techniques such as classification models to identify and correct
errors in software applications.
The study analyzed one week’s worth of data and found a total of
4042852 errors, which were mainly warning and notice-type errors. The
data cleaning function developed by the authors was able to identify and
correct these errors, leading to improved data quality. The authors also
provided guidelines for developing error-free software applications,
including loosely coupled module creation, global variable declaration, and
unit testing.
Future work could focus on expanding the scope of the study to include
data from multiple sources and different time periods. The authors could
also explore the use of other machine learning techniques, such as
clustering, to identify and correct errors in software applications. Also,
more data structuring recommendations will be included. Integrity
constraints (IC), such as Functional Dependencies (FD), can be used in
conjunction with Machine Learning to classify the type of error to be
captured in the event of a data set with an inaccurate value [1]. To fully
appreciate the data’s potential for bringing significant benefits to a variety
of businesses, it is necessary to learn from it [11]. Some “Context of source
code processing” will be considered where a mutation in the context plays
as refactoring source code and 1-Time, K-Time mutation will be played an
important role in the concept [12]. In that study, the author will focus on a
few research questions, and the author will try to solve them one by one by
considering best practices.
In conclusion, this study demonstrated the importance of data cleaning
in machine learning and provided a solution for improving data quality in
software applications. By using machine learning techniques and following
the guidelines provided by the authors, software developers can create
error-free and efficient applications that are scalable, reliable, and faster.
Future research in this area could lead to further improvements in data
quality and software development practices.
References
1. Hossen J, Sayeed S (Sep 2018) Modifying cleaning method in big data analytics process using
random forest classifier. In: 2018 7th (ICCCE). IEEE, pp 208–213
2.
Al-Janabi S, Janicki R (Jul 2016) A density-based data cleaning approach for deduplication with
data consistency and accuracy. In: 2016 SAI computing conference (SAI). IEEE, pp 492–501
3.
Mehra S, Verma R (2019) An approach towards log analysis of large-scale systems. Int J Comput
Appl 181(41):18–22
4.
Li Q, Zhao C, He X (2016) A clustering-based log analysis approach for improving software
reliability. J Syst Softw 118:197–212
5.
Xu X, Li Y, Wang Y, Li X, Li B (2018) Log classification based on multi-view feature learning.
Inf Softw Technol 98:126–139
6.
Zhou M, Zhang Z, Li Y, Zhang H (2019) Log analysis using natural language processing
techniques: a survey. J Syst Softw 151:99–115
7.
Dimov T, Orozova D (Jun 2020) Software for data cleaning and forecasting. In: 2020 21st
international symposium on electrical apparatus & technologies (SIELA). IEEE, pp 1–4
8.
Kumar V, Khosla C (Jan 2018) Data cleaning-a thorough analysis and survey on unstructured
data. In: 2018 8th international conference on cloud computing, data science & engineering
(confluence). IEEE, pp 305–309
9.
Wardat M, Le W, Rajan H (May 2021) DeepLocalize: fault localization for deep neural
networks. In: 2021 IEEE/ACM 43rd international conference on software engineering (ICSE).
IEEE, pp 251–262
10.
Gao K, Khoshgoftaar TM, Napolitano A (Dec 2012) A hybrid approach to coping with high
dimensionality and class imbalance for software defect prediction. In: 2012 11th international
conference on machine learning and applications, vol 2. IEEE, pp 281–288
11.
Haider SN, Zhao Q, Meran BK (Jul 2020) Automated data cleaning for data centers: a case
study. In: 2020 39th Chinese control conference (CCC). IEEE, pp 3227–3232
12.
Pour Maryam V, Zhuo L, Lei M, Hadi H (2021) A search-based testing framework for deep
neural networks of source code embedding. Calgary, Canada. https://doi.org/10.1145/1188913.
1188915
13.
Zdravevski E, Lameski P, Kulakov A, Kalajdziski S (Sep 2015) Transformation of nominal
features into numeric in supervised multi-class problems based on the weight of evidence
parameter. In: 2015 federated conference on computer science and information systems
(FedCSIS). IEEE, pp 169–179
14.
Liu H, Ashwin KTK, Thomas JP, Hou X (2016) Cleaning framework for BigData
15.
Hasan MT, Mahal SN, Bakar NMA, Hasan MM, Islam N, Sadia F, Hasan M (2022) An unified
testing process framework for small and medium size tech enterprise with incorporation of
CMMI-SVC to improve maturity of software testing process. In: ENASE, pp 327–334
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed
Computing , Studies in Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_2
Abstract
In the field of computing, evaluating system performance is critical. The
need for new scheduling paradigms, particularly in the context of large
cluster computing systems (Lu et al. in IEEE Trans Parallel Distrib Syst
34:1145–1158, 2023), is paramount. The authors will examine the
feasibility of multiple techniques that strive to impart a fair distribution of
resources across systems, with the endgame being to minimize latency.
Moreover, the authors will explore the feasibility of devising solutions with
varying degrees of enhanced data locality, which has consistently been the
backbone of efficient scheduling. Another direction of the study will
investigate the advantages and disadvantages of two contrasting
approaches: the first of these will look into the efficacy of erasing specific
jobs and freeing up resources for the next set of tasks, whereas the
alternative will attempt to finish pending jobs before tackling the remaining
assignments. Finally, the authors conclude that they abided by professional
norms in extracting and analyzing the results of their experiments
(Wennmann et al. in Investig Radiol 58(4):253–264, 2023).
Keywords Cluster scheduling – Latency – Distribution function – Locality
– Fair sharing
1 Introduction
Facebook is one of the most prominent data collectors on the internet [3].
Facebook stores its data in a data warehouse that was over 2 PB in size and
increased by 15 TB each day. Due to the company’s nature, many
researchers and developers execute various data processing tasks
simultaneously. In the example of Facebook, these jobs can be short ad-hoc
queries through Hive [4] or large multi-hour machine learning jobs.
Therefore, it is crucial to implement a scheduling system that provides
fair sharing, i.e. that all resources are distributed fairly, thus decreasing
latency. With this goal in mind, M. Zaharia et al. developed the Hadoop
Fair Sched uler (HFS) [5]. The authors also found a way of increasing data
locality which improved the system’s overall throughput.
This analysis will have a detailed look into the experiments that the
authors ran. We will also reproduce one of the experiments.
2.3 Macrobenchmarks
In October 2009, the authors measured the caseload distribution on
Facebook. The authors used this information to sample 100 jobs to simulate
realistic inter-arrival times and input sizes for a multi-user workload on a
regular day. This distribution can be seen in Table 1. Three different
workloads have been created to test different scenarios:
Table 1 How job sizes are distributed along map tasks at Facebook
Fig. 1 The cumulative distribution functions of job running times in different bin ranges indicate
that in the IO-heavy workload, fair sharing enhances performance for smaller jobs but may cause
slower processing of larger jobs. However, the introduction of delay scheduling can further enhance
performance, particularly for medium-sized jobs. From the original paper [6]
Fig. 2 Our results for the experiment shown in Fig. 1. The bin numbering is from the original
experiment’s distribution
Fig. 3 The mean enhancement in speed obtained by implementing delay scheduling rather than the
simplistic fair sharing approach for jobs in each bin of the IO-intensive workload is illustrated, while
the black lines represent the respective standard deviations. From the original paper [6]
As mentioned before a statistical means of this experiment was the
calculation of the standard derivation.
(2) CPU-heavy workload: In order to create a task that is predominantly
CPU-bound, the authors executed each input via a costly user-dependent
function, resulting in a significantly slower job execution. Only a small
fraction, specifically 0.01% of the records were produced as output.
With this experiment the authors want to ensure that even under a CPU-
heavy load, such as Machine Learning or an in-depth analysis of the data
like clustering, delay scheduling manages to outperform the FIFO
scheduler, which was in place at that time.
(3) Mixed workload: The mixed workload experiment aims at
presenting a realistic high workload for the scheduling system. The jobs that
are submitted during this experiment contain both CPU-heavy and IO-
heavy workloads. Furthermore, the job pool also contains a variety of short
and long jobs.
2.4 Microbenchmarks
Microbenchmarks try to stress test Delay scheduling in more specific cases,
where locality is hard to achieve. These experiments try to test the quality
of the introduced scheduling method in a more controlled manner.
(1) Hierarchical scheduling: In their paper, the authors introduced a
hierarchical scheduling policy, which prioritises jobs that need to be run on
production for customers like queries etc. These jobs require a higher
priority than experimental machine learning tasks. This experiment aimed at
evaluating this scheduling policy, the Hadoop Fair Scheduler (HFS). It
attempts to assess the speed at which new tasks receive resources depending
on their level of importance.
(2) Delay scheduling with small jobs: Small jobs can pose a challenge to
scheduling systems due to the high amount of throughput they require. In
this experiment, the authors show how small jobs that are handled by a
system utilising delay scheduling compare to those that do not use delay
scheduling. They created three different filter jobs, one with three, one with
ten and one with 100 map tasks.
This experiment was not run on an Amazon EC2, but rather on a private
cluster. The private cluster also had 100 nodes, but 8 cores and 4 disks per
node. Contrary to the EC2 cluster it had 6 instead of 4 map slots and 4
instead of 2 reduce slots per node.
We are unsure why the authors picked a private cluster to run this
experiment. The private cluster has been defined before this experiment but
it was never mentioned as the utilised resource.
(3) Delay Scheduling with Sticky Slots: Earlier in the paper, the authors
introduced so called Sticky Slots. These Sticky Slots occur when a task gets
repeatedly submitted to the same slot. This happens when a job never leaves
its original slot. Sticky slots can have a negative impact on locality of a job.
In their experiment, Zaharia et al. reproduced this locality problem by
creating a 180-GB dataset, which was spread in 2 GB chunks over all 100
nodes in the EC2 Cluster. Then 5 and 50 jobs were submitted which caused
the sticky slot phenomenon to occur.
3.2 Obstacles
One of the first challenges was to figure out the exact experiments the
original authors ran. They refer to the benchmarks described by Zheng Shao
[7]. However, the purpose of this suite is to compare three different data
processing systems, including Hadoop and Hive. Since Hive is built on top
of Hadoop and many of the authors are from Facebook solving a problem
affecting their company,1 it is safe to assume that they used the Hive
benchmarks for evaluation purposes.
Building these benchmarks as Java programs was also the cause of great
frustration. This was probably due to the sparse documentation and the
12 years that have passed since the publication of the original code.
However, after figuring out the exact library versions to use, environment
variables to set, and the Perl programming language, we have grown to like
the benchmark program.
The aforementioned script saves the statistics of each map task into a
separate output file. These are then converted to JSON documents that
only contain the relevant information and can be subsequently plotted using
the rest of our Python scripts.
Fig. 4 The mean acceleration achieved by using delay scheduling compared to the basic fair sharing
method for jobs in each bin of the IO-intensive workload is depicted, with the black lines indicating
the corresponding standard deviations. From the original paper [6]
Fig. 5 Average speedup of delay scheduling over naïve fair sharing for jobs in each bin in the IO-
heavy workload
Fig. 6 Average speedup of delay scheduling over FIFO scheduling for jobs in each bin in the IO-
heavy workload
However if we take a closer look at Fig. 3, the standard deviations are
suspiciously large. In our evaluation, the outlier deviations are gone which
might be the result of the multiple repetitions we had.
Overall, we gained more from abandoning the FIFO scheduler than
Zaharia et al. We theorise, that this might be due to the smaller sized cluster
where the scheduling opportunities are more rare as compared with their
100-node clusters.
4 Conclusion
Even though the authors did not disclose the number of repetitions of their
experiments, we found that this paper is of very high quality. We also have
to note that the authors provided no information about the time of day the
experiment or any other external factors which could have affected the
experiment, especially, when running on AWS hardware.
Zaharia et al. described the experiments with a high level of detail, and
they also explained their outcomes and consequences formidably well. The
graphs created from the experiments understandably showed the results,
and the authors made good use of all charts and tables pointing out essential
features. We reproduced a macrobenchmark with an IO-heavy workload.
The naive fair scheduler was not significantly different from the fair
scheduler with delay scheduling in their experiment.
In our repeated experiments, we came to almost the same conclusions as
the authors did when conducting the experiments giving us reason to
believe that the results in the paper are reproducible.
Overall, we conclude that the authors did an excellent job presenting
delay scheduling and showing through experimenting with different
realistic and intentionally straining experiments. These evaluations showed
how their introduced scheduling method fares in real-world applications
compared with the system in place at that time. The conducted experiments,
their consequences, and interpretations were explained and presented in an
in-depth but understandable manner.
References
1. Lu R, Zhang W, Wang Y, Li Q, Zhong X, Yang H, Wang D (2023) Auction-based cluster federated
learning in mobile edge computing systems. IEEE Trans Parallel Distrib Syst 34(4):1145–1158
2.
Wennmann M, Bauer F, Klein A, Chmelik J, Grözinger M, Rotkopf LT, Neher P, Gnirs R, Kurz
FT, Nonnenmacher T et al (2023) In vivo repeatability and multiscanner reproducibility of MRI
radiomics features in patients with monoclonal plasma cell disorders: a prospective bi-institutional
study. Investig Radiol 58(4):253–264
3.
Dwyer T (2016) Convergent media and privacy, ser. Palgrave global media policy and business.
Palgrave Macmillan
4.
Thusoo A, Sarma JS, Jain N, Shao Z, Chakka P, Anthony S, Liu H, Wyckoff P, Murthy R (2009)
Hive: a warehousing solution over a map-reduce framework. Proc VLDB Endow 2(2):1626–1629
[Crossref]
5.
Sharma S, Bharti RK (2023) New efficient Hadoop scheduler: generalized particle swarm
optimization and simulated annealing-dominant resource fairness. Concurr Comput Pract Exp
35(4):e7528
[Crossref]
6.
Zaharia M, Borthakur D, Sen Sarma J, Elmeleegy K, Shenker S, Stoica I (2010) Delay
scheduling: a simple technique for achieving locality and fairness in cluster scheduling. In:
Proceedings of the 5th European conference on computer systems, ser. EuroSys’10. Association
for Computing Machinery, New York, NY, USA, pp 265–278. [Online]. https://doi.org/10.1145/
1755913.1755940
7.
Shao Z (2009) Hive performance benchmarks. [Online]. https://issues.apache.org/jira/browse/
HIVE-396
8.
Borthakur D (2007) The Hadoop distributed file system: architecture and design. Hadoop Project
Website 11(2007):21
Footnotes
1 The company responsible for Hive and also utilising it to a great extent.
2 With the caveat, that our cluster setup—similarly to the originals—is also not representative of real
life networking conditions: each node is located in the same rack.
3 We used the Hadoop Fair Scheduler with its thresholds set to zero to implement this.
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing , Studies in
Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_3
Abstract
The eWOM manages to be a source of information for current and potential customers, as it
plays a significant and transcendental role in the purchase intention. The objective of this
study was to identify the impact of digital word-of-mouth on the purchase intention of
Peruvian millennial female consumers. A probabilistic sample of 355 women between 26
and 41 years old was determined, and two valid and reliable questionnaires were applied. It
was concluded that digital word-of-mouth indeed has a significant impact on the purchase
intention of millennial female consumers, which was demonstrated by χ2 tests and ordinal
logistic regression (χ2 = 0.000 < 0.05; 164.019 = 0.000 < 0.05; Nagelkerke R2 = 0.425; Wald
coefficient = 52.203 = 0.000 < 0.05), confirming that the purchase intention is explained by
the actions of digital word-of-mouth. This work highlights the importance of managing the
credibility and value of our electronic media, and within action plans adopting actions to
improve the quality and quantity of comments as antecedents to purchase and repurchase
intention.
Background
Regarding the eWOM variable, Ismagilova et al. [17] and Kasabov [18] state that it is
composed of any manifestation within digital platforms, where potential, current or former
consumers expose positive, neutral, or negative experiences about a product, service, brand,
or company.
Cuervo et al. [19], on the other hand, define eWOM as a form of digital communication
that allows understanding consumer behavior and the main motivations that lead them to
seek information online, which, in turn, affects their purchase decision or use of goods and
services. Meanwhile, Babić et al. [20] indicate that the variable in question is the action in
which consumers provide information about goods, services, or brands to other consumers
through virtual environments, and that this represents one of the most revealing and
explanatory developments in contemporary consumer behavior.
Credibility
Vallejo et al. [21] mention that the credibility of information is formed by the set of beliefs
of the consumer, based on the comments and opinions read in digital media, so its role is
decisive in the degree of influence that recommendations from other consumers can have on
the information receiver.
In addition, Abd-Elaziz et al. [22] explain that it is necessary to differentiate the
credibility of the information included in online recommendations from the credibility of the
online seller, since the former, referring to the credibility of eWOM, is based on information
obtained from customer experiences, with a non-commercial purpose. Likewise, Filieri [23]
argues that the credibility of a review is defined as the trust that readers have in what they
are reading. In this sense, consumers do not take into account opinions that are perceived as
unreliable.
Quality
Erkan and Evans [24] define the quality of eWOM as the persuasive strength of the
arguments expressing the information contained in a message or comment issued on digital
platforms. In this sense, Tsao and Hsieh [25] explain that it should be accurate, objective,
and complete, as well as reliable and useful for consumers.
In this framework, Abrego et al. [26] consider that the quality of electronic word of
mouth (eWOM) is a multidimensional construct composed of 4 factors: relevance, accuracy,
comprehensibility, and timeliness.
Quantity
Regarding the amount of eWOM, Alabdullatif and Akram [15] define it as the number of
comments about a product or service, and the greater the number of comments and/or
reviews, the more useful the information a user obtains to recommend or rate a service
becomes. Similarly, Ismagilova et al. [27] report that the quantity of eWOM messages makes
the information more visible to users interested in it, demonstrating the popularity of the
product or service.
Value
Tata et al. [28] explain that, regarding this dimension, distributors or sellers provide filters to
facilitate the search for reviews, allowing customers to filter between positive, negative, or
recent reviews. Therefore, while a set of positive opinions can improve the purchase
decision, a set of negative opinions may provoke a rejection response or another type of
response depending on the overall sentiment of the set of opinions.
In this sense, Purnawirawan et al. [29] mention that if a set of reviews is predominantly
positive or negative, it has a greater influence on perceived usefulness, and when reviews
refer to products, they have a greater influence on attitudes towards unknown brands.
Purchase Intention
Jang and Shin [30] defines it as the willingness of customers to engage in some type of
online transaction to satisfy their purchasing needs, and both utilitarian and hedonic values
can influence consumer purchase intention. Likewise, Lee et al. [31] and Nuseir [32] assert
that this dimension determines the consumer’s buying willingness and is represented when
the consumer considers various conditions and foundations to ultimately make a purchase.
In addition, Dehghani and Tumer [33] explain that this variable is a possibility that lies in
the hands of customers, since it depends greatly on the value of the product and the
recommendations shared by consumers.
Perceived Usefulness
Zappala [34] states that this dimension refers to the degree to which consumers believe that
using the Internet will improve their performance or productivity, thus enhancing the
outcome of their shopping experience. Similarly, Taherdoost [35] defines it as an individual’s
perception of how the use of new technology will increase or improve their performance,
meaning it refers to consumers’ perceptions of the outcome of their experience.
In addition, Cho and Sagynov [36] mention that there is a similarity between the
concepts of perceived usefulness and relative advantage in technology, as electronic
commerce constitutes an innovation within distribution channels. Therefore, the utility
provided to the consumer will be closely linked to its advantages as a sales system.
Tavera and Londoño [37] argue that in e-commerce there is a certain level of uncertainty in
the purchase process, as consumers do not have control over the actions of the online seller
or information about the intentions of the other party in the transaction, which causes people
to show hesitation towards developing virtual purchasing behaviors. Therefore, trust in the
seller is important for making a purchase.
Subjective Standard
Subjective standards are determined by the pressures perceived by others, how individuals
behave in certain ways, and their motivation to follow the views of these people [38].
Alalwan [39] indicates that subjective norm represents the degree to which a user perceives
the expectations of others regarding a certain behavior.
The hypothesis test was performed using an Ordinal Logistic Regression analysis. In this
sense, it was possible to explain the incidence of the independent variable Digital Word-of-
Mouth (eWOM) on the dependent variable Purchase Intention since the variables are of the
ordinal category. In this measure, the model contrast was performed at a significance level of
5% (α = 0.05) and a confidence level of 95%, deciding based on the statistical significance
value (Sig.) obtained in the test, analyzing the following statistics.
Goodness-of-fit measures: Likelihood ratio Chi-square, to examine whether the predictor
variable included in the model has a significant relationship with the response variable. In
this sense, if the p-value obtained is lower than the significance level (p < 0.05), the
assumption that the variables are significantly associated is accepted. On the other hand, if
the p-value obtained is higher than the significance level (p > 0.05), the assumption of
significant association is rejected.
Pseudo-R2: This indicator explains the level of influence of the dependent variable X on
the independent variable Y, being useful to know the fit of the model to the data. In the
estimation, the coefficients of the Cox and Snell, Nagelkerke, and McFadden tests are
offered, but there is no agreement on which pseudo-R2 statistics is better.
Measures of association and predictive efficiency: Wald test, to estimate the significance
of the probability that the coefficient is different from zero when projecting the dependent
variable in the regression model. In this sense, if the p-value obtained is lower than the
contrast level (p < 0.05), the assumption that the coefficient is equal to zero is rejected,
understanding that the independent variable contributes to the prediction of the dependent
variable. On the contrary, if the p-value obtained in the test is higher than the contrast level
(p > 0.05), the assumption that the coefficient is equal to zero is accepted, understanding that
the independent variable does not contribute to the prediction of the dependent variable.
3 Results
3.1 Hypothesis Testing
The Table 1 shows that the significance is less than 0.05; therefore, Ha is accepted, i.e.,
eWOM does have a significant influence on purchase intention. In this sense, when eWOM
exists, it will generate purchase intention; therefore, purchase intention is explained by the
possible actions or influence of eWOM.
Table 1 Chi-square test of eWOM and purchase intention
a
0 cells (0.0%) have expected a count lower than 5. The minimum expected count is 15.82
In the Table 2, it is possible to observe that the significance level is less than 0.05, so all
the specific hypotheses are accepted. In other words, the credibility, quality, quantity, and
value of the eWOM influence the purchase intention of millennial consumers, i.e., the
purchase intention is explained by the possible actions or influence of the eWOM.
Table 2 Chi-square tests of specific hypotheses (unified)
In Table 5, a Wald coefficient = 52.203 associated with a p-value = 0.000 lower than the
contrast level (p < 0.05) was determined; therefore, H0 is rejected, and Ha is accepted,
estimating at a confidence level of 95% that as digital word-of-mouth levels decrease, the
probability in the levels of purchase intention decreases, concluding that: Digital word-of-
mouth has a significant impact on the purchase intention of millennial female consumers.
Table 5 Measures of association and predictive effectiveness of digital word-of-mouth in the purchase intention of
millennial consumer
In Table 6, Wald Chi-Square coefficients associated with a p-value = 0.000 lower than
the contrast level (p < 0.05) were determined for the levels of quality, quantity, and value of
the word-of-mouth. Therefore, it is concluded that as the levels of these dimensions
associated with digital word-of-mouth decrease, the probability of the levels of purchase
intention decreases. Thus, the quality, quantity, and value of digital word-of-mouth
significantly influence the purchase intention of millennial female consumers.
Table 6 Tests of model effects
References
1. Tien D, Amaya A, Liao Y (2019) Examining the influence of customer-to-customer electronic word-of-mouth on
purchase intention in social networking sites. Asia Pac Manag Rev 24(3):238–249
2.
INCAE. Mujeres Millennial: Profesionales, Trabajadoras, Urbanas
3.
IPSOS (2018) Global Trends 2018
4.
DATUM Internacional (2019) ¿En qué se diferencian los Millennial del Perú?
5.
CPI (2021) CPI Perú: Población 2021
6.
IPSOS (2018) IPSOS Millennials: Más X, Que Z
7.
Carbajal N, Chocaca C (2020) EWOM en redes sociales en relación a la intención de compra en la categoría de jeans
urbano femenino en el nivel socioeconómico A y B de los distritos de la zona 8 de Lima Metropolitana. Universidad
Peruana de Ciencias Aplicadas, Lima
8.
Rossell DL (2020) Influencia del EWOM en la intención de compra en los aplicativos de delivery en Lima
Metropolitana. Universidad Peruana de Ciencias Aplicadas, Lima
9.
Levy S, Gvili Y, Hino H (2021) Engagement of ethnic-minority consumers with electronic word of mouth (EWOM) on
social media: the pivotal role of intercultural factors. J Theor Appl Electron Commer Res 16(7):2608–2632
[Crossref]
10.
Zinko R, Patrick A, Furner C, Gaines S, Kim M, Negri M, Orellana E, Torres S, Villarreal C (2021) Responding to
negative electronic word of mouth to improve purchase intention. J Theor Appl Electron Commer Res 16:1945–1959
[Crossref]
11.
Nawaz Z, Zhao C, Nawaz F, Safeer A, Irshad W (2021) Role of artificial neural networks techniques in development of
market intelligence: a study of sentiment analysis of EWOM of a women’s clothing company. J Theor Appl Electron
Commer Res 16(5):1862–1876
[Crossref]
12.
Le-Hoang P (2020) The relationship between online convenience, online customer satisfaction, buying intention and
electronic word-of-mouth. Indep J Manag Prod 11(7):2943–2966
[Crossref]
13.
Khan K, Ali M (2017) Impact of electronic word of mouth on consumer purchase intention in footwear industry of
Pakistan. Kuwait Chap Arab J Bus Manag Rev 6:52–63
[Crossref]
14.
Fortes N, Santos A (2020) A Influência Do EWOM Na Intenção de Compra Online. Revista Ibérica de Sistemas e
Tecnologias de Informação E34:408–420
15.
Alabdullatif AA, Akram MS (2018) Exploring the impact of electronic word of mouth and property characteristics on
customers’ online booking decision. TEM J 7(2):411–420
16.
Anastasiei B, Dospinescu N (2019) Electronic word-of-mouth for online retailers: predictors of volume and valence.
Sustainability 11(3):1–19
[Crossref]
17.
Ismagilova E, Dwivedi YK, Slade E, Williams MD (2017) Electronic word of mouth (EWOM) in the marketing
context. SpringerBriefs in Business; Springer International Publishing, Cham
[Crossref]
18.
Kasabov E (2016) Unknown, surprising, and economically significant: the realities of electronic word of mouth in
Chinese social networking sites. J Bus Res 69(2):642–652
[Crossref]
19.
Cuervo S, Salcedo N, Gutiérrez K, Joaquín M, Ramírez K, Tumbalobos C (2016) Rendimiento del tráfico web en la
elección de un programa de posgrado. Universidad ESAN, Lima
20.
Babić A, Sotgiu F, Valck K, Bijmolt THA (2015) The effect of electronic word of mouth on sales: a meta-analytic
review of platform, product, and metric factors. J Mark Res 53(3):297–318
[Crossref]
21.
Matute J, Polo Y, Utrillas A (2015) Las características del boca - oído electrónico y su influencia en la intención de
recompra online. Revista Europea de Dirección y Economía de la Empresa 24(1):61–75
[Crossref]
22.
Abd-Elaziz ME, Aziz WM, Khalifa GS, Abdel-Aleem M (2015) Determinants of electronic word of mouth (EWOM)
influence on hotel customers’ purchasing decision. Int J Herit Tour Hosp 9(2/2):194–223
23.
Filieri R (2016) What makes an online consumer review trustworthy? Ann Tour Res 58:46–64
[Crossref]
24.
Erkan I, Evans C (2016) The influence of EWOM in social media on consumers’ purchase intentions: an extended
approach to information adoption. Comput Hum Behav 61:47–55
[Crossref]
25.
Tsao W-C, Hsieh M-T (2015) EWOM persuasiveness: do EWOM platforms and product type matter? Electron
Commer Res 15(4):509–541
[Crossref]
26.
Abrego D, Sánchez Y, Medina JM (2017) Influencia de los Sistemas de Información en los Resultados
Organizacionales. Contaduria y Administración 62:303–320
[Crossref]
27.
Ismagilova E, Slade EL, Rana NP, Dwivedi YK (2020) The effect of electronic word of mouth communications on
intention to buy: a meta-analysis. Inf Syst Front 22:1203–1226
[Crossref]
28.
Tata SV, Prashar S, Gupta S (2020) An examination of the role of review valence and review source in varying
consumption contexts on purchase decision. J Retail Consum Serv 52:101734
[Crossref]
29.
Purnawirawan N, Eisend M, De Pelsmacker P, Dens NA (2015) Meta-analytic investigation of the role of valence in
online reviews. J Interact Mark 31(1):17–27
[Crossref]
30.
Jang S-H, Shin J-I (2016) The influence of contextual offer, utilitarian, and hedonic value on purchase intention in
mobile location-based services. Int J Bus Policy Strat Manag 3(1):7–12
[Crossref]
31.
Lee WI, Cheng SY, Shih YT (2017) Effects among product attributes, involvement, word-of-mouth, and purchase
intention in online shopping. Asia Pac Manag Rev 22(4):223–229
32.
Nuseir MT (2019) The impact of electronic word of mouth (e-WOM) on the online purchase intention of consumers in
the Islamic countries—a case of (UAE). J Islam Mark 10:759–767
[Crossref]
33.
Dehghani M, Tumer M (2015) A research on effectiveness of Facebook advertising on enhancing purchase intention of
consumers. Comput Hum Behav 49:597–600
[Crossref]
34.
Zappala S (2017) Impact of e-commerce on consumers and small firms. In: Zappalà S, Gray C (eds) Routledge
35.
Taherdoost H (2018) A review of technology acceptance and adoption models and theories. Procedia Manuf 22:960–
967
[Crossref]
36.
Cho YC, Sagynov E (2015) Exploring factors that affect usefulness, ease of use, trust, and purchase intention in the
online environment. Int J Manag Inf Syst 19(1):21–36
37.
Tavera JF, Londoño BE (2014) Determining factors in technology acceptance of e-commerce in developing countries
(Fatores Determinantes Da Aceitação Tecnológica Do E-Commerce Em Países Emergentes). Revista Ciencias
Estratégicas 22(31):101–119
38.
Ham M, Jeger M, Ivković AF (2015) The role of subjective norms in forming the intention to purchase green food.
Econ Res-Ekonomska istraživanja 28(1):738–748
[Crossref]
39.
Alalwan AA (2018) Investigating the impact of social media advertising features on customer purchase intention. Int J
Inf Manag 42:65–77
[Crossref]
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed
Computing , Studies in Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_4
Aljia Bouzidi
Email: [email protected]
Abstract
In the software development lifecycle, business process models (BPMs)
turn out to play an ever more pivotal role in the development and continued
management of information systems (ISs). However, BPMs and IS models
(ISMs) are traditionally expressed separately. This separation causes drift
between them, impedes their interoperability, and thus builds up misaligned
models. Traceability in software development proves its ability to link
together related artifacts from different sources within a project (for
examples, business modelling, requirements, design models), improves
project outcomes by assisting designers and other stakeholders with
common tasks such as impact analysis, etc. In this paper, we propose an
improvement and an extension of an existing requirement traceability
method in order to tackle the traceability between design, requirement and
BPMs. In fact, the extension consists in adding the UML class diagram
concepts structured according to the Model View Controller (MVC) design
pattern to be traced with BPMN and UML use cases models in a single
unified model. This method is based on the integration mechanism, acts at
the model and the meta-model levels, and can be used to develop a new IS
and/or to examine the misalignment of the existing ISMs and the BPMs
after BPM/ISM evolution.
1 Introduction
Traceability in software development proves its ability to link together
related artifacts from different sources within a project (for examples,
business modelling, requirements, uses cases, design models), improves
project outcomes by assisting designers and other stakeholders with
common tasks such as impact analysis, etc. Thereby creating an explicit
traceability model that has not a standalone guideline, but it has huge
benefits in terms of consistency, quality, and automation. Although its
creation is not a trivial task, an explicit traceability model remains a
reference for a consistent definition of typed traceability links between
heterogeneous model concepts that help to ensure their alignment and their
coevolution.
In this context, in [1], authors adopt the integration mechanism to
propose an accurate requirement engineering method that acts at the meta-
model and model level, and establishes traceability between BMs and ISMs
to bridge the gap between business and requirement modelling in a
straightforward way. Indeed, authors of [1], firstly define an integrated trace
meta-model that ensures a straightforward integration of the business and
software worlds for representing the BPMN [9] and the UML use case
models [10] in the form of a unified single meta-model. It defines also
traceability links between interrelated concepts to correlate overlapped
concepts as new modelling concepts. Then, they define an integrated model
as an instantiation of the proposed integrated trace meta-model. They draw
it as a new diagram baptized BPSUC (Business Process Supported Use
Cases).
Research conducted in this paper is to enhance and extend the work
presented in [1]. The enhancement consists in adding class diagram
concepts structured according to the MVC pattern. Our intervention
considers both the meta-model and the model levels. Hence, in the
integrated trace meta-model proposed by [1], we add new modelling
concepts to express trace links between the class diagram, use case diagram
and the BPMN concepts. Class diagram concepts that have no
corresponding concepts are also added in the integrated trace meta-model.
Proposed traceability concepts and class diagram concepts are instantiated
in BPSUC diagram. Accordingly, BPSUC is now able to design class
diagram elements and the proposed traceability concepts combined with
their corresponding BPMN and use case diagram artifacts.
The remainder of this paper is structured as follows: Sect. 2 gives an
overview of the method presented in [1]. In Sect. 3, we explain our
proposal. Section 4 is devoted to discussing related works. In Sect. 5, we
show the feasibility of our contributions in practice. Finally, Sect. 6
concludes the current paper and outlooks future works.
3 Our Contribution
The research work conducted in this paper is an improvement and extension
of the method of [1]. The extension consists in improving the integrated
trace meta-model, and the BPSUC diagram to include the artifacts of the
UML class diagram structured according to the MVC design pattern. In this
section, we explain how we add the UML class diagram concepts in the
integrated trace meta-model and the BPSUC diagram, and the rectifications
made on them.
4 Implementation
In order to apply our approach into the practice, we have developed a visual
modelling to support the proposed integrated trace meta-model and the
BPMNTraceISM diagram.
The tool acts as an internal plugin within the Eclipse framework.
Indeed, deploying our modelling tool in the form of a plug-in increases its
reusability and availability in any Eclipse platform without dependency on
the system runtime or the workspace containing the modelling tool. In
addition, our editor is a fully functional graphical tool that allows business
engineers and software designers to work together with a single integrated
graphical user interface (GUI) that incorporates both BPMN, use case and
class diagram elements.
The construction process of this editor begins with using the ECore
meta-modelling language to develop the improved integrated meta-model.
Then, we implement a toolbox to design instances of the integrated trace
meta-model classes.
BPTraceISM environment is composed of four main parts: the project
explorer containing an EMF project that includes BPMNTraceISM
diagrams, the modelling space, the toolbox that contains the graphical
elements of a BPMNTraceISM diagram, and the properties tab to edit the
properties of an element selected in the modelling space part.
Figure 3 outlines a simple example of a BPMNTraceISM diagram
modelled by using the editor. The modelling space contains an OUActor
called supplier associated to an UCsF called manage purchase order. In the
business compartment of the UCsF “manage purchase order”, we have a
user task called Accept purchase order. In the class diagram compartment,
we have four classes linked via undirected associations. Each class has a
name and a stereotype. The boundary class Manage purchase order contains
an operation called “acceptPurhaseOrder()”.
Fig. 3 Example of a BPMNTraceISM diagram designed by using the BPtraceISM editor
5 Related Works
In related works we focus on the approaches that define explicit traceability
models separated from source models.
These works include approaches that propose guidelines for creating
traceability models. For instance, a guideline is proposed in [3] for
establishing traceability between software requirements and UML design
models. This guideline includes two main concepts: (1) meta-model that
represents relationships between requirements and the UML diagrams, and
(2) process that is applied according to specific steps. However, this
guideline focuses on establishing traceability at the met-model level, only.
There are other model-based researches that aim to maintain the
traceability. For example, the research in [5] that proposes a change
propagation-based coevolution of transformations. Its premise is that
knowledge of the meta-model evolution can be propagated by means of
resolutions to drive the transformation coevolution. To deal with particular
cases, the authors introduce composition-based techniques that permit
developers to compose resolutions meeting their needs. Adopting the same
purpose, the authors of [7] refer to machine learning techniques to introduce
an approach baptized TRAIL (TRAceability lInk cLassifier). This study
uses the histories of existing traceability links between pairs of artifacts to
train a machine learning classifier to be able to classify the link between
any new or existing pair of artifacts as related or unrelated. Some other
approaches define traceability models for eliciting requirements of complex
systems. For example, in [2, 6]. Furthermore, the authors of [8] propose a
traceability model that traces the model elements at different levels of the
enterprise architecture. It uses general concepts for representing different
artifacts used to model traceability such as “traceability links”, “Aspect”,
“Element”, “requirement”, etc. Likewise, the authors of [4] base on the
deep learning machine and propose a neural network architecture that uses
word embedding and Recurrent Neural Network (RNN) technique to
generate automatically trace links. The final output of RNN is a vector that
represents the semantic information of the artifact. The tracing network then
compares the semantic vectors of two artifacts and outputs the probability
that they are linked. Yet, managing all meta-models of overall levels in a
single traceability model may yield complex models. There are researches
which are dedicated to specific languages. For example, the approach of
[11] that proposes a meta-model traceability in the form of an extension of
the BPMN meta-model. Then, they define trace links between some
elements. Further, the authors [12] base on Natural Language Processing
techniques to define an enhanced framework of software artifact traceability
management. To illustrate their approach, a tool that supports the
traceability between requirements, UML class diagrams, and corresponding
Java code is implemented. Overall, existing works that defined explicit
traceability model are mostly focused on the meta-model model level, only
and ignore the model level. Moreover, existing explicit traceability models
trace either between UML diagrams at the same/different abstraction or
between business model concepts. However, none of the existing
approaches achieved reliable results when dealing with traceability between
BPMN models, UML use case models and UML class diagram.
6 Conclusion
In this paper, we improved and extend an existing traceability method
between BPMN and UML use case models that establish the traceability
considering the meta-model and the model levels.
The extension consists in enhanced an integrated traceability meta-
model by adding UML class diagram concepts structured according to the
MVC design pattern and trace links between the class diagram concepts and
the existing concepts of the integrated trace meta-model. Traceability links
are created basing on semantic mapping between BPMN use case and class
diagram meta-models.
We have also improved the instantiation of the integrated trace meta-
model by creating new icons for representing UML class diagram elements
combined with UML use case model and BPMN diagram elements in one
unified diagram that we call BPMTraceISM.
In our future works, we look forward to define heuristics on the
compliance of a new diagram to constraints specified dynamically by the
developer. These heuristics should be able to automatically detect the
changes made on the BPMN, UML use case and UML class diagrams, and
to indicate the elements that will be affected by the changes.
References
1. Bouzidi A, Haddar N, Haddar K (2019) Traceability and synchronization between BPMN and
UML use case models. Ingénierie des Systèmes d Inf 24(2):215–228
2.
de Carvalho EA, Gomes JO, Jatobá A, da Silva MF, de Carvalho PVR (2021) Employing
resilience engineering in eliciting software requirements for complex systems: experiments with
the functional resonance analysis method (FRAM). Cogn Technol Work 23:65–83
[Crossref]
3.
Eyl M, Reichmann C, Müller-Glaser K (2017) Traceability in a fine grained software
configuration management system. In: Software quality. Complexity and challenges of software
engineering in emerging technologies: 9th international conference, SWQD 2017. Vienna,
Austria, Proceedings 9. Springer International Publishing, pp 15–29
4.
Guo J, Cheng J, Cleland-Huang J (2017) Semantically enhanced software traceability using deep
learning techniques. In: 2017 IEEE/ACM 39th international conference on software engineering
(ICSE) IEEE, pp 3–14
5.
Khelladi DE, Kretschmer R, Egyed A (2018) Change propagation-based and composition-based
co-evolution of transformations with evolving metamodels. In: Proceedings of the 21th
ACM/IEEE international conference on model driven engineering languages and systems, pp
404–414
6.
Lopez-Arredondo LP, Perez CB, Villavicencio-Navarro J, Mercado KE, Encinas M, Inzunza-
Mejia P (2020) Reengineering of the software development process in a technology services
company. Bus Process Manag J 26(2):655–674
[Crossref]
7.
Mills C, Escobar-Avila J, Haiduc S (2018) Automatic traceability maintenance via machine
learning classification. In: 2018 IEEE international conference on software maintenance and
evolution (ICSME). IEEE, pp 369–380
8.
Moreira JRP, Maciel RSP (2017) Towards a models traceability and synchronization approach of
an enterprise architecture. SEKE, pp 24–29
9.
OMG BPMN specification.: Business Process Model and Notation. http://www.bpmn.org/.
Accessed 31 Feb 2023
10.
OMG UML Specification (2017) Unified modeling language (omg uml), superstructure,
version2. In: Object management group, p 70
11.
Pavalkis S, Nemuraite L, Milevičienė E (2011) Towards traceability metamodel for business
process modeling notation. In: Building the e-world ecosystem: 11th IFIP WG 6.11 conference
on e-business, e-services, and e-society, I3E 2011. Kaunas, Lithuania. Revised Selected Papers
11. Springer, Berlin, pp 177–188
12.
Swathine K, Sumathi N, Nadu T (2017) Study on requirement engineering and traceability
techniques in software artefacts. Int J Innovat Res Comput Commun Eng 5(1)
13.
Bouzidi A, Haddar NZ, Ben-Abdallah M, Haddar K (2020) Toward the alignment and
traceability between business process and software models. In ICEIS 2:701–708
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed
Computing , Studies in Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_5
Srimannarayana Grandhi
Email: [email protected]
Tasadduq Imam
Email: [email protected]
Abstract
The use of Fintech for open banking gradually increased as it allows
consumers to share their data with third party providers and access their
products and services. Due to a lack of trust in data security and privacy,
there is a disinclination of consent to share personal data and a fear of
losing control over the data hoarded during open banking. Despite
significant research, there is still no research from the users’ cognitive and
behavioural perspectives to demonstrate the challenges associated with
Fintech use. This study presents a research model for assessing the position
of Fintech in the open banking context, especially from the customers’
perspectives on data security and privacy issues. The article lays out the
protocols and constructs relevant to the model. Using a quantitative
approach, this study will collect data through online survey questionnaires
and employ the structural equation modelling technique to test the model.
This study is expected to help develop appropriate policies to enhance the
use of Fintech by consumers in the Australian banking sector.
1 Introduction
Fintech created new ecosystems with a massive influx of financial data.
This data is considered a salient asset in the modern finance world.
Possessing quality and volume data gives competitive advantages to
financial organisations [1]. Governance of data quality is a paramount
prerequisite to data management due to the mammoth exchange of
information in financial organisations [2, 3]. Due to its power inherency in
data, the governance of such power is crucial in the finance sector.
Governance of power is a global phenomenon, though managed by
country-specific governments. Governments recognise these information
asymmetries as disadvantages in creating viable business opportunities
from both the growth aspect of the evolving Fintech industry as well as fair
user perspectives [1, 4]. Thus, the emergence of an innovative solution
—‘open banking’ to liberalise the data held by banks [1, 5, 6]. Whilst this
move has been frowned upon by the banks due to its marginalised returns in
exchange for their data, using this drive-through mechanism boosted the
Fintech infants’ penetration ability into the competitive lending industry [7].
Despite many benefits in Fintech open banking, challenges persist in
managing and preserving the core of sound data management where users’
positive adoption perceptions are stimulated whilst data security and
privacy remain unaffected [8] and as a measure Fintech blockchain is used
as a safer mode strategy in assuring the open banking application
programming interface (API) security and privacy-related concerns [9].
However, it’s inevitable to expect completely risk-free when exposing data
for system consumption [10], given they are also human-operated and
technology itself advances to beat its own capabilities [11], exposing to
cyberattacks, malware, undue access, fraudulent representation penetration
attempts to APIs [2, 8, 11–13], apart from the likely presence of Fintech
brokers who trade data to big technology [4].
This study identifies three research gaps. First, users greatly trust
Fintech providers to secure personal and sensitive financial data privately
[14], and providers uphold the trust in safeguarding the sensitive data [15].
Whilst there is ample literature that studies the Fintech adoption criteria
using various models [14, 16–18] still, there is a gap in the customer
expectations of data security and privacy which may seem to hinder the
Fintech open banking adoption in Australia and ways to enhance its
adoption that requires further assessment.
Second, consumers have very little control over the data they leave
behind in the process as the responsibility of security and privacy
arrangements seems to alter their form with a change of hands [8, 16, 19].
Whilst customers’ expectations of privacy protection are that the mandated
physical binds and perceived moral binds of data protection laws are
obeyed by Fintech [1, 3, 6], studies explaining the customers’ rights on such
hoarded data beyond their primary objective of release and regulatory
aspects seem to lack. The duty of care phenomenon often has been
conveniently neglected and unsupported [8], which needs further review.
Third, Australia has been an early adopter of open banking amongst the
very few countries with the government supporting the emerging industry
by introducing regulatory sandbox and legislative platforms, especially
toward provoking the uptake. There is, however, a little research on the
impact of such steps on customer trust and adoption intents.
To address the identified research gaps, this study aims to review open
banking customers’ perspective of data security and privacy, the
interconnectedness of the data security and privacy strategies and their
importance in shaping Fintech use.
Based on previous technology adoption studies, this study presents a
research model to pinpoint the underlying behavioural factors affecting
Fintech adoption and use by the consumers of Australian financial
organisations. Especially, it focuses on how Fintech data security and
privacy elements embedded in open banking operations enhance or hinder
the use of Fintech for open banking and address the following research
questions.
RQ1: What data security and privacy challenges are customers of
Australian financial organisations facing while using Fintech for open
banking?
2 Literature Review
In this section, a condensed summary of the articles reviewed is provided to
highlight the open banking concept, its benefits, and the importance of
enhancing the data security and privacy aspects, government regulatory
governance and customer trust factors in the Australian context to enhance
the use of Fintech by customers for open banking. The technology adoption
model, which can characterise salient features like establishing customer
trust, data security and privacy assurance, social influence, facilitating
conditions and the role of government regulations, is also presented toward
developing the proposed framework and criteria linked to the proposition.
Fig. 1 Proposed framework—(extracted from UTAUT, TTF, S–O–R and trust theory)
H1: Data security and privacy have a positive influence on customers’ trust.
5 Research Design
Further to conceiving the research model, it is worth reflecting on the
research design and approach as relevant to address the research questions
based on the proposed model. In this regard, this study considers an
objectivist epistemological view and is concerned with the cause and
outcomes [40], especially intending to investigate the role of behavioural
factors in enabling trust in Fintech and promoting Fintech adoption and use.
Thus, this study will adopt a quantitative research approach to address this
goal. A questionnaire will also be developed for data collection, to measure
the constructs and the relationships among the constructs presented in the
research model, which combines multiple research models [41, 42]. Using
the online data collection tool, the online survey questionnaire will be
distributed to the customers of Australian financial organizations who use
Fintech for open banking. This will help understand their views on the role
of behavioural factors promoting trust in technologies and their intent to
accept and use Fintech for open banking. Upon collecting the data, it will be
analyzed and validated using the statistical tool - Statistical Package for the
Social Sciences (SPSS). Then, the survey data will be used to test the
proposed research model and its nominated propositions by applying the
Structural Equation Modelling (SEM) technique. The results of the survey
data will be further compared with that in existing findings and
recommendations will be made accordingly.
6 Conclusion
This study has reviewed previous literature relating to Fintech data security
and privacy strategies and how customers’ cognitive and behavioural
adoption challenges concerning the adoption of Fintech for open banking.
The literature revealed the importance of human factors in technology
adoption decisions and identified research gaps. It can be seen from the
literature that there are limited studies on the challenges faced by customers
of Australian financial organizations. Even though Australia has been an
early adopter in the regulatory sandbox in Fintech open banking compared
to other countries like the UK, Europe, and the USA, there seems a
consumer adoption lag in Australian Fintech open banking. A research
model is proposed in this paper to assess the Fintech adoption intention for
open banking and actual adoption criteria by Australian Fintech users by
examining the impact of data security and privacy strategies and
government regulations on customer trust. Further, the model also expects
to predict the impact of social influence and facilitating conditions on
customer trust and thereby the actual adoption decision. This study intends
to use a quantitative method, online survey questionnaires for data
collection. The proposed model’s derived propositions will be tested using
the structural equation modelling technique to establish the impact of each
variable.
This study has some limitations. First, this study focuses on the
Australian context, it collects data from the customers of Australian
financial organizations. The findings may not be relevant to other countries.
Even so, the research is important with a lack of similar research on the
Australian context. Potentially, testing the research finding across contexts
can be follow-up future research. Second, the proposed study adopts a
quantitative method which may result in offering inconclusive evidence.
Adopting a mixed method, like data collection through interviews can
enable the researchers to seek in-depth answers from the participants. This
is another possible future undertaking. However, with a solid theoretical
basis dictating the research model and the use of SEM which may offer
important information on the relationships among the factors presented in
the research model, this research will expectantly highlight consumers’
different cognitive behaviours concerning Fintech in open banking, and
thereby help reshape data security and privacy strategies of the banking
sector. The outcomes may also guide enacting government regulations to
cater to different values on Fintech adoption in open banking in Australia.
References
1. Fracassi C, Magnuson W (2021) Data autonomy. Vanderbilt Law Rev 74(2):327–383
2.
Brown E, Piroska D (2022) Governing Fintech and Fintech as governance: the regulatory
sandbox, riskwashing, and disruptive social classification. New Polit Econ 27(1):19–32. https://
doi.org/10.1080/13563467.2021.1910645
3.
Solove DJ (2006) A taxonomy of privacy. Univ Pa Law Rev 154(3):477–564
[Crossref]
4.
Di Porto F, Ghidini G (2020) “I access your data, you access mine”: requiring data reciprocity in
payment services. IIC Int Rev Intellect Prop Compet Law 51:307–329. https://doi.org/10.1007/
s40319-020-00914-1
5.
Aytaş B, Öztaner SM, Şener E (2021) Open banking: opening up the walled gardens. J Paym
Strat Syst 15(4):419–431
6.
Palmieri A, Nazeraj B (2021) Open banking and competition: an intricate relationship. In: EU
and comparative law issues and challenges series, Osijek, JJ, Strossmayer University of Osijek,
pp 217–237
7.
Barr MS, DeHart A, Kang A (2019) Consumer autonomy and pathways to portability in banking
and financial services. University of Michigan Center on Finance, Law & Policy Working Paper,
University of Michigan Law & Economics Research Paper, no. 19-022. https://doi.org/10.2139/
ssrn.3483757
8.
Borgogno O, Colangelo G (2020) Consumer inertia and competition-sensitive data governance:
the case of open banking. J Eur Consum Mark Law 9(4):143–150. https://doi.org/10.2139/ssrn.
3513514
[Crossref]
9.
Wang H, Ma S, Dai HN, Imran M, Wang T (2020) Blockchain-based data privacy management
with nudge theory in open banking. Futur Gener Comput Syst 110:812–823. https://doi.org/10.
1016/j.future.2019.09.010
[Crossref]
10.
PwC (2020) Financial services technology 2020 and beyond: embracing disruption. 48
11.
Wang JS (2021) Exploring biometric identification in FinTech applications based on the
modified TAM. Financ Innov 7(42). https://doi.org/10.1186/s40854-021-00260-2
12.
Arner D, Buckley R, Charamba K, Sergeev A, Zetzsche D (2022) Governing Fintech 4.0:
bigtech, platform finance, and sustainable development. Fordham J Corp Financ Law 27(1):1–71
13.
Jenga L (2022) Open banking. Oxford University Press, Oxford, United States: Incorporated.
http://ebookcentral.proquest.com/lib/cqu/detail.action?do-cID=7034203. Accessed 7 Aug 2022
14.
Stewart H, Jürjens J (2018) Data security and consumer trust in FinTech innovation in Germany.
Inf Comput Secur 26:109–128. https://doi.org/10.1108/ICS-06-2017-0039
[Crossref]
15.
Barbu CM, Florea DL, Dabija DC, Barbu MC (2021) Customer experience in Fintech. J Theor
Appl Electron Commer Res 16(5):1415–1433. https://doi.org/10.3390/jtaer16050080
16.
Chan R, Troshani I, Rao HS, Hoffmann A (2022) Towards an understanding of consumers’
FinTech adoption: the case of Open Banking. Int J Bank Mark 40:886–917. https://doi.org/10.
1108/IJBM-08-2021-0397
[Crossref]
17.
Li G, Dai JS, Park EM, Park ST (2017) A study on the service and trend of Fintech security
based on text-mining: focused on the data of Korean online news. J Comput Virol Hacking Tech
13(4):249–255. https://doi.org/10.1007/s11416-016-0288-9
[Crossref]
18.
Singh S, Sahni MM, Kovid RK (2020) What drives FinTech adoption? A multi-method
evaluation using an adapted technology acceptance model. Manag Decis 58:1675–1697. https://
doi.org/10.1108/MD-09-2019-1318
[Crossref]
19.
Kassab M, Laplante PA (2022) Open banking: what it is, where it’s at, and where it’s going.
Computer 55:53–63. https://doi.org/10.1109/MC.2021.3108402
[Crossref]
20.
Ernst & Young Global Limited (2019) Global FinTech Adoption Index 2019. Ernst & Young
Global Limited, London, UK. https://assets.ey.com/content/dam/ey-sites/ey-com/en_gl/topics/
banking-and-capital-markets/ey-global-fintechadoption-index.pdf. Accessed 7 Aug 2022
21.
Gomber P, Robert J, Kauffman C, Bruce WW (2018) On the Fintech revolution: Interpreting the
forces of innovation, disruption, and transformation in financial services. J Manag Inf Syst
35(1):220–265. https://doi.org/10.1080/07421222.2018.1440766
22.
Ng AW, Kwok BK (2017) Emergence of Fintech and cybersecurity in a global financial centre. J
Financ Regul Compliance 25:422–434. http://dx.doi.org.ezproxy.cqu.edu.au/10.1108/JFRC-01-
2017-00132
23.
Kaabachi S, Ben MS, O’Leary B (2019) Consumer’s initial trust formation in IOB’s acceptance:
the role of social influence and perceived compatibility. Int J Bank Mark 37:507–530. https://doi.
org/10.1108/IJBM-12-2017-0270
[Crossref]
24.
Li Z, Li W, Wen Q, Chen J, Yin W, Liang K (2019) An efficient blind filter: location privacy
protection and the access control in FinTech. Futur Gener Comput Syst 100:797–810. https://doi.
org/10.1016/j.future.2019.04.026
[Crossref]
25.
Lee JM, Kim HJ (2020) Determinants of adoption and continuance intentions toward internet-
only banks. Int J Bank Mark 38:843–865. https://doi.org/10.1108/IJBM-07-2019-0269
[Crossref]
26.
Rogers EM (1995) Diffusion of innovations (4th ed.) New York: Free Press
27.
Kurniasari F, Tajul Urus S, Utomo P, Abd Hamid N, Jimmy SY, Othman IW (2022) Determinant
factors of adoption of Fintech payment services in Indonesia using the UTAUT approach. Asia-
Pac Manag Account J 17:97–125
28.
Peong KK, Peong KP, Tan K, Tan Y (2021) Behavioural intention of commercial banks’
customers towards financial technology services. J Financ Bank Rev JFBR 5:10–27. https://doi.
org/10.35609/jfbr.2021.5.4(2)
29.
Turner M, Kitchenham B, Brereton P, Charters S, Budgen D (2010) Does the technology
acceptance model predict actual use? A systematic literature review. Inf Softw Technol 52:463–
479. https://doi.org/10.1016/j.infsof.2009.11.005
[Crossref]
30.
Roberts-Lombard M, Petzer DJ (2021) Relationship marketing: an S–O–R perspective
emphasising the importance of trust in retail banking. Int J Bank Mark 39:725–750
[Crossref]
31.
Kelman HC (1958) Compliance, identification, and internalization: three processes of attitude
change. J Conflict Resolut 2(March):51–60
[Crossref]
32.
Venkatesh V, Bala H (2008) Technology acceptance model 3 and a research agenda on
interventions. Decis. Sci. 39(2):273–315. https://doi.org/10.1111/j.1540-5915.2008.00192.x
33.
Venkatesh V, Davis FD (2000) A theoretical extension of the technology acceptance model: four
longitudinal field studies. Manage Sci 46(2):186–204
34.
Venkatesh V (2000) Determinants of perceived ease of use: integrating control, intrinsic
motivation, and emotion into the technology acceptance model. Inf Syst Res 11(4):342–365
35.
Venkatesh V, Morris MG, Davis GB, Davis FD (2003) User acceptance of information
technology: toward a unified view. MIS Quarterly 27:425–478
36.
Zhang Q, Li Y, Wang R, Liu L, Tan Y, Hu J (2021) Data security sharing model based on privacy
protection for blockchain-enabled industrial internet of things. Int J Intell Syst 36:94–111
[Crossref]
37.
Ajzen I (1985) From intentions to actions: A theory of planned behavior. In J. Kuhl & J.
Beckman (Eds.) Action-control. From cognition to behavior Heidelberg: Springer 11–39
38.
Lai PC (2017) The literature review of technology adoption models and theories for the novelty
technology. J Inf Syst Technol Manag 14:21–38. https://doi.org/10.4301/S1807-
17752017000100002
[Crossref]
39.
Kassab M, Laplante P (2022) Trust considerations in open banking. IT Prof 24:70–73. https://
doi.org/10.1109/MITP.2021.3136015
[Crossref]
40.
Yilmaz K (2013) Comparison of quantitative and qualitative research traditions: epistemological,
theoretical, and methodological differences. Eur J Educ 48(2):311–325. https://doi.org/10.1111/
ejed.12014
[Crossref]
41.
O’Leary Z (2021) The essential guide to doing your research project, 4th edn. SAGE
Publications Ltd
42.
Wahyuni D (2012) The research design maze: understanding paradigms, cases, methods and
methodologies. J Appl Manag Account Res 10(1):69–80. https://ssrn.com/abstract=210308
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed
Computing , Studies in Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_6
Zhen Li
Email: [email protected]
Seiichi Serikawa
Email: [email protected]
Bin Li
Email: [email protected]
Lifeng Zhang
Email: [email protected]
Abstract
Corn 3D phenotype extraction faces several problems, including low
precision, excessive manual involvement, long processing times, and the
requirement for complex equipment. To address these issues, we propose a
novel key point detection deep learning model called V7POSE-GSConV,
which operates on RGB-D data. This method is built upon the YOLOv7-
POSE key point detection model, allowing us to directly capture key points
of corn plants using RGB data as input. Leveraging corresponding RGB-D
data, this work can derive the comprehensive structure of corn, extract
parameters such as leaf length, leaf angle, plant height, and ear phenotype.
In our experiments, the YOLOv7-POSE model achieved a recognition
accuracy of 99.45% in training stage. To further optimize the model for
efficiency while maintaining accuracy, we introduced GSConV. The results
demonstrate a 2% reduction in the number of model parameters in
V7POSE-GSConV, with no loss in accuracy.
1 Introduction
Corn is one of the most vital staple foods globally, and its products and
derivatives find application in various fields. Increasing corn yield by
breeding has always been the focus of research. By studying phenotypic
parameter differences among corn plants, this discipline enables targeted
crop improvement, demand-driven breeding, and increased crop yields.
Rapidly acquiring a substantial number of plant parameters is a significant
goal in plant phenomics. In the era of information technology, plant
phenomics plays a pivotal role in breeding [1].
Traditional methods for obtaining phenotypic parameters in corn
breeding heavily rely on manual measurements using rulers and protractors.
These methods suffer from inefficiency, measurement errors, damage. A
more advanced approach involves obtaining high-precision plant point
clouds through 3D reconstruction. However, 3D reconstruction method
comes with disadvantages, including expensive equipment, specific site
requirements, susceptibility to weather conditions, complex procedures.
With the evolution of deep learning, key point detection has emerged as a
prominent task. End-to-end key point detection models can directly use
RGB data as input and output key points of corn plants. Utilizing
corresponding RGB-D data, we can directly output corresponding 3D key
points. These 3D key points facilitate the rapid extraction of relevant plant
phenotypes. Key point detection-based methods have advantages such as
cost-effectiveness, fast extraction speeds, and robustness. Consequently, the
development of corn key point identification technology is crucial for the
swift and convenient extraction of corn plant phenotypes.
In the realm of 2D key point detection, Toshev et al. [2] pioneered the
application of Convolutional Neural Networks (CNN) for human key point
detection and introduced the DeepPose algorithm for deep pose estimation.
Deep Neural Networks [3] are initially employed for rough human key
point detection, using DNN regression to predict key point positions in the
human body. Under consistent image sizes and unchanged computational
parameters, cascaded regressors refine the human body key point with
higher accuracy. DeepPose represents a pivotal shift from traditional to
deep learning methods for key point detection in the human body. Wei et
al. [4] introduced Convolutional Pose Machine (CPM), which derives
image features and spatial context directly from the study data. CPM
initially approximates human body key point locations through multi-stage
estimation and progressively refines these positions at each stage. By
combining multiple convolutional networks, CPM resolves the issue of
decreasing neural network gradients as layer count rises. Newell et al. [5]
introduced a Stacked Hourglass network (SHN), which stacks multiple
hourglass networks in a cascaded manner to leverage multi-scale features in
human body key point detection. Lifshitz et al. [6] introduced Deep
Consensus Voting, which utilizes high-density multi-object voting instead
of sparse key point location sets for detecting human body key points. It not
only offers effective detection but also computes the probability of joint key
points related to the detection result. Zhang et al. [7] proposed a fast pose
estimation model called Fast Pose Distillation, a lightweight human key
point detection model utilizing four hourglass models, suitable for low-
compute environments. This network has effectively reduced the network
complexity associated with human key point detection. Thus far, research in
two-dimensional single-player key point detection based on depth methods
is comprehensive, with a primary work focused on improving detection
accuracy and precision.
In classical networks such as YOLOv5 [8] and YOLOv7 [9], there are
also structures specifically designed for key point detection tasks, they all
get outstanding results.
(2) Top point: This denotes the point of the highest point of the corn stalk.
(3) Leaf connection point: This denotes the point of where a leaf is
attached to the main stem of the corn plant.
(4) Highest Point of a Leaf: This denotes the point of the uppermost point
on a leaf.
(5) Angle point: This denotes the point of the one-quarter of the distance
from the base to the apex of the leaf.
(6) Tip point: This denotes the point of the end of a leaf blade.
(7) Stalk point: This denotes the point of the root of a corn ear connects to
the main stem of the corn.
(2) Leaf Angle: Leaf angle is the degree of inclination between the midrib
of a leaf blade and the main stem of the corn plant. It plays a critical
role in shaping the canopy structure and can have a direct impact on
crop yield, as cited in reference [11]. As shown in Fig. 7, We calculate
it by the Angle between these three points, leaf connection point, angle
point and top point.
(3) Leaf Length: Leaf length refers to the length of the leaf. Since the
blade has curvature in the normal state, it is measured by straightening
the blade in the artificial scene. As shown in Fig. 8, we calculate the
blade length by the euclidean distance between leaf connection point,
angle point, highest Point of a leaf and tip point.
(4) Ear Position: Ear position indicates the specific section or segment of
the corn plant where the ear grows and develops. As shown in Fig. 9,
we calculate it by root point and stalk point.
(2) Head Network: The head network is where key point detection is
outputted. It involves the concatenation of feature maps and the ELAN
module as Fig. 12 to extract different scale feature maps. And then use
the concat and ELAN-H module as Fig. 13 to extract the final feature
map for regress.
Fig. 10 YOLOv7-POSE model
3.2 GS-ConV
GSConV [10] as Fig. 14 is a novel approach employed to enhance the trade-
off between model accuracy and computational efficiency. GSConV
leverages a multi-step process to optimize convolutional operations within a
neural network.
Firstly, GSConV initiates by a standard convolutional layer.
Subsequently, it incorporates DWConv (Depthwise Convolution), a deep
convolution operation designed to efficiently process feature maps.
To maximize the benefits of both convolution types, GSConV
concatenates the output of these two convolutional layers to capture
essential features.
Then, GSConV includes a shuffle operation that strategically rearranges
the channel order of the feature maps. This reshuffling ensures that
corresponding channel numbers from the preceding two convolutional
operations are placed adjacent to each other, thereby enhancing information
flow and facilitating more efficient computations.
3.3 V7POSE-GSConV
To address the requirement for model miniaturization, we substituted the
CBS module with the GSConV module in the re-implemented YOLOv7-
POSE model as Fig. 15. This alteration led to a reduction in the number of
model parameters without compromising accuracy. Through this module
replacement, we successfully decreased the model’s parameter count by
2%.
Detection result
Percision (%) Recall (%) Map (%) Parameters number
YOLOv7-POSE 99.40 99.15 99.56 37.0 M
V7POSE-GSConV 99.45 99.10 99.56 36.2 M
Phenotype result
Generated value Annotated value
Plant height 1.56 m 1.62 m
Leaf angle
Leaf length 0.691 m 0.732 m
Ear position 0.619 m 0.588 m
5 Conclusion
In this paper, we present a corn key point detection model that achieves an
impressive precision rate of 99.4%. With the corresponding RGB-D data,
this model enables the rapid acquisition of three-dimensional corn
phenotypic data. Compared to traditional manual calculations, our proposed
method offers convenience and speed. In contrast to point cloud-based
methods, our approach demonstrates superior robustness and applicability
in outdoor environments. Additionally, we have addressed the issue of
excessive parameters to accommodate the miniaturization of the model by
incorporating the GS-ConV model, resulting in a 2% reduction in the
parameter count while maintaining accuracy. However, Our work is
ongoing, more work is under working to obtain more accurate 3D
phenotypes and be adapted to more scenarios.
References
1. Zermas D et al (2020) 3D model processing for high throughput phenotype extraction. The case
of corn. Comput Electron Agric 172(2020):105047
2.
Toshev A, Szegedy C et al (2014) Deeppose: human pose estimation via deep neural networks.
In: 2014 IEEE conference on computer vision and pattern recognition. Institute of Electrical and
Electronic Engineers. IEEE, Columbus, US, pp 1653–1660
3.
Krizhevsky A, Sutskever I, Hinton GE et al (2017) ImageNet classification with deep
convolutional neural networks. Commun ACM 60(6):84–90
[Crossref]
4.
Wei SE, Ramakrishna V, Kanade T et al (2016) Convolutional pose machines. In: Proceedings of
the 2016 IEEE conference on computer vision and pattern recognition. Institute of Electrical and
Electronics Engineers. IEEE, Washington, pp 4724–4732
5.
Newell A, Yang KY, Deng J et al (2016) Stacked hourglass networks for human pose estimation.
In: Proceedings of the 14th European conference on computer vision. Springer, Amsterdam,
Berlin, pp 483–499
6.
Lifshitz I, Fetaya E, Ullman S et al (2016) Human pose estimation using deep consensus voting.
In: European conference on computer vision Amsterdam. ECCV, Berlin, Germany, pp 246–260
7.
Zhang F et al (2019) Fast human pose estimation. In: Proceedings of the 2019 IEEE/CVF
conference on computer vision and pattern recognition. Institute of Electrical and Electronics
Engineers. IEEE Computer Soc, USA, pp 3517–3526
8.
Wu Wentong et al (2021) Application of local fully convolutional neural network combined with
YOLO v5 algorithm in small target detection of remote sensing image. PloS One
16(10):e0259283
[Crossref]
9.
Wang C-Y, Bochkovskiy A, Liao H-YM (2022) YOLOv7: trainable bag-of-freebies sets new
state-of-the-art for real-time object detectors. In: Proceedings of the IEEE/CVF conference on
computer vision and pattern recognition, 2023
10.
Li H et al (2022) Slim-neck by GSConv: a better design paradigm of detector architectures for
autonomous vehicles. arXiv:2206.02424
11.
Liu K, Cao J, Yu K et al (2019) Wheat TaSPL8 modulates leaf angle through auxin and
brassinosteroid signaling. Plant Physiol 181(1)
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed
Computing , Studies in Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_7
Xiaozhong Xue
Email: [email protected]
Linni Wang
Email: [email protected]
Ayaka Ehiro
Email: [email protected]
Yahui Peng
Email: [email protected]
Abstract
Glaucoma is one of the main causes of blindness, it is characterized by an
increase in the area ratio of optic cup (OC) and optic disc (OD). Therefore,
OC segmentation is an important basis for computer-aided diagnosis of
glaucoma. However, the accuracy of OC segmentation still needs to be
improved due to the low contrast, occlusion by blood vessels, and different
scales of OC in fundus images. This study applies swin-unet to do OC
segmentation. The swin-unet is composed of U-shaped structure and swin-
transformer block. U-shaped structure can extract features from different
scales and fuse a large number of features to solve the difficulties of low
contrast and different scales. In addition, swin-transformer block can extract
the features of the relationship between each patches, which can solve the
difficulties of occlusion by blood vessels. In this study, swin-unet obtained
mean intersection of union (IoU) of and mean DICE of ,
which is better than the segmentation results of U-Net and swin-
transformer. Therefore, the effectiveness of swin-unet in OC segmentation
is proved in this study.
1 Introduction
Fundus image is the photo of retina. In clinical, ophthalmologists generally
use the fundus image to diagnose various fundus diseases, such as
glaucoma [1]. The Fig. 1 is an example of fundus image, various lesions
and organs of retina can be observed from this image. Optic disc (OD) is a
highly bright oval-shaped yellowish region in fundus images, optic cup
(OC) is a brighter part in OD. Physiologically, OD and OC are regions
where blood vessels and optic nerves pass through the retina [10]. Blood
vessels are red lines scattered throughout the retina. Fovea appears as a
darker area on the retina, which is the most sensitive area of vision [11].
Fig. 1 Fundus image
Glaucoma is one of the main causes of loss of vision in the world. Due
to the destruction that leads to blindness is in incremental and insidious,
glaucoma is also called as “silent thief of vision” [20]. However, if
glaucoma can be found and treated in time, it could effectively prevent
blindness. One symptom of glaucoma is an increase in intraocular pressure
[2]. Due to the fact that OC is a concave structure on the retina, the degree
of OC concave increases when intraocular pressure increases. Therefore, as
shown in Fig. 2, in the fundus image of glaucoma patient, the area ratio of
OC and OD will increase [23], which is also an important basis for
computer-aided diagnosis of glaucoma. Furthermore, the segmentation of
OD and OC is the prerequisite for calculating their area ratio.
Fig. 2 The comparison of normal and glaucoma fundus images, the blue curve and red curve are
boundaries of OD and OC, respectively (a is an example of normal, b is an example of glaucoma)
Fig. 4 The different scales of OC (The red curves are boundaries of OC)
The main contributions of this paper are as follows:
Swin-unet is applied to OC segmentation and its effectiveness in OC
segmentation is demonstrated.
The three difficulties in OC segmentation: low contrast, covered by blood
vessels, and multiple scales, have been solved with targeted solutions.
This paper is composed as follows: Sect. 2 briefly introduces the related
works of OC segmentation; Sect. 3 explains the swin-unet that is applied to
OC segmentation in detail; Sect. 4 describes the experimental process and
shows the experimental results; Sect. 5 draws some conclusions.
2 Related Work
At present, the algorithms of OC segmentation are mainly divided into four
categories: vessels kinking—[6, 24], active contour model—[14], clustering
—[5], and deep learning-based methods [12, 22].
OD and OC are the depression parts on the retina, and the slope of OC
depression is stronger than OD. A knot will appear when the blood vessels
pass through the boundary of OD and OC. Damon et al. [24] and Wong et
al. [6] judge the OC boundaries by these knots. This method can solve the
difficulty of OC being covered by blood vessels. However, due to the
different angle when taking fundus images, not every fundus image has
knots. In addition, it is necessary to segment blood vessels before detecting
the knots, this pre-processing will bring more errors. Mishra et al. [14] uses
the active contour model-based method and Chandrika and Nirmala [5] uses
clustering-based method. However, these methods are failed in
segmentation of OC with low contrast. Furthermore, these methods need
some post-processing, such as ellipse fitting, to compensate for under-
segmentation of OC caused by vessel occlusion. These post-processing will
also reduce the OC segmentation accuracy.
With the rapid development of deep learning, it has also been widely
used in OC segmentation. Among various deep learning models, the models
based on U-Net [12] and GAN [22] are most commonly used. Due to its U-
shaped structure, U-Net can extract the features of different scales, and the
skip connection mechanism can improve the amount of extracted features.
However, it cannot solve the difficulty that OC is covered by blood vessels.
The GAN model using semi-supervised learning may help solve the
difficulty of vascular occlusion, but it does not perform well in fundus
images with low contrast.
In summary, the three difficulties in OC segmentation: low contrast,
multiple scales, and covered by blood vessels are still not solved.
Fig. 6 Two types of self-attention mechanisms (a is conventional self attention, while b is window
self attention from swin-transformer. Red squares: the range of implementing self attention)
Self-Attention: Self-attention is first proposed in the field of natural
language processing [21], and later applied in the field of computer vision
[7]. In natural language processing, the calculation of self-attention is often
based on a word, while in computer vision, it is usually based on a small
patch. The calculation process of self-attention can be simply summarized
into the following 3 steps:
Encoding a word or patch as a vector, then a sentence or image is
encoded as a matrix.
Multiply the encoded matrix with three weight matrices , , and
to obtain matrices Q, K, and V, where , , and are three
trainable matrices.
Finally, the new feature map can be calculated using
.
(2)
(3)
(4)
4 Experiment
4.1 Dataset
Five public fundus image datasets: DRISHTI-GS, ORIGA, RIM-ONE R3,
REFUGE, and G1020 are used in this study.
The DRISHTI-GS [19] dataset contains 101 fundus images. All images
were token centered on OD with a field-of-view (FOV) of 30-degrees and
resolutions of pixels. The ground-truth of OD and OC
segmentation results are annotated by 4 experts.
The ORIGA [26] dataset contains 650 fundus images. The resolution of
each image is about . All images are tagged with manually
segmented OD and OC by trained professionals from Singapore Eye
Research Institute.
Unlike other datasets, there are 159 stereo fundus images in RIM-ONE
R3 [9] dataset. The ground-truth of OD and OC segmentation results are
annotated by 2 experts.
The REFUGE [8] dataset is from Retina Fundus Glaucoma Challenge,
which contains 1200 fundus images. These images are divided into training
set, testing set, and validation set. For each image, both OD and OC
segmentation results are manually marked.
The G1020 [3] dataset contains 1020 fundus image. All images were
token with a FOV of 45-degrees. Same as the above datasets, the ground-
truth of OD and OC segmentation results are annotated.
The detail information of each dataset are listed in Table 1.
Table 1 The detail of each dataset used in this study
4.2 Pre-processing
The pre-processing applied in this study is the extraction of regions of
interest (ROI). OC only occupies a very small part in fundus image, so it is
necessary to extract the ROI. This can not only reduce the calculation
amount, but also improve the segmentation accuracy. Otherwise, as
described in Sect. 4.1, the fundus images in different datasets have different
resolution and FOV. It is necessary to normalize the scale of OC. As shown
in Fig. 10, this study directly uses ground-truth to find the center and radius
of OD, which are utilized to crop the ROI image. The process of ROI
extraction is mainly divided into two steps: center and radius detection and
removal of invalid image.
Center and radius detection: the ground-truth of OD is a binary image,
the center of OD can be detected by the connected area in this binary
image. Then looking for the radius of OD in , , , and
four directions. The maximum value on these four directions is treated as
the radius of OD. As shown in Fig. 11, the red circle and green line are
the detected center and radius, respectively.
Removal of invalid image: as shown in Fig. 12, in this process, some
images with incomplete OD are discarded. Some images could not find
the center of OD, which will also be discarded. Finally, the ROI image
can be cropped according to the center and radius of OD. The side length
of ROI image is 4 times the radius of OD.
Fig. 10 Flowchart of pre-processing
Fig. 11 The detected center and radius of OD (Red circle: center, blue lines: candidate radius, green
line: selected radius)
4.4 Results
In this study, 2778 fundus images are divided into training sets and testing
sets at the ratio of 9 : 1, 7 : 3, and 5 : 5. For each ratio, the data are
randomly divided 10 times, and the final result can be obtained by the
average of each experiment result. IoU and DICE are used to measure the
results of OC segmentation. The formula of IoU and DICE are expressed as
follows:
(5)
(6)
where,
GT and Result are ground-truth and segmentation result, respectively.
The segmentation results of each model are shown in Table 3, and the
examples of segmentation results by swin-unet are shown in Fig. 16.
Table 3 OC segmentation results
Fig. 16 The OC segmentation results by swin-unet (Red curve: ground-truth, green curve:
segmentation result)
4.5 Discussion
As shown in Table 3, compared with U-Net and swin-transformer, swin-
unet achieves the highest mean IoU and mean DICE in OC segmentation. In
different ratios of training sets and testing sets, swin-unet obtained mean
IoU of 77.39, 77.60, and mean DICE of 86.68, 86.79, ,
respectively. This experimental results prove the effectiveness of swin-unet
in OC segmentation.
5 Conclusion
In this study, swin-unet is applied to OC segmentation for the first time, and
a mean IoU of and a mean DICE of are obtained. This
segmentation accuracy is improved compared with U-Net and swin-
transformer. The experimental results show that swin-unet is effective in
OC segmentation and can effectively solve the difficulties in OC
segmentation: 1. the low contrast between OC and other areas; 2. the OC is
covered by blood vessels; 3. the different OC has different scale.
References
1. Ahmed A, Ritambhar B, Kaamran R, Vasudevan L (2015) Optic disc and optic cup segmentation
methodologies for glaucoma image detection: a survey. J Ophthalmol 2015:1–28
[Crossref]
2.
Alward W, Feldman F, Cashwell L, Wilensky J, Geijssen H, Greeve E, Quigley H, Skuta G,
Lichter P, Blondeau P, Collaborative Normal-Tension Glaucoma Study Group (1998) The
effectiveness of intraocular pressure reduction in the treatment of normal-tension glaucoma. Am
J Ophthalmol 126(4):498–505
3.
Bajwa MN, Singh GAP, Neumeier W, Malik MI, Dengel A, Ahmed S (2020) G1020: a
benchmark retinal fundus image dataset for computer-aided glaucoma detection. In: 2020
International joint conference on neural networks (IJCNN). IEEE, pp 1–7
4.
Cao H, Wang Y, Chen J, Jiang D, Zhang X, Tian Q, Wang M (2023) Swin-unet: Unet-like pure
transformer for medical image segmentation. In: Computer vision–ECCV 2022 workshops: Tel
Aviv, Israel, October 23–27, 2022, proceedings, Part III. Springer, pp 205–218
5.
Chandrika SM, Nirmala K (2013) Analysis of CDR detection for glaucoma diagnosis
6.
Damon WWK, Liu J, Meng TN, Fengshou Y, Yin WT (2012) Automatic detection of the optic
cup using vessel kinking in digital retinal fundus images. In: 2012 9th IEEE international
symposium on biomedical imaging (ISBI). IEEE, pp 1647–1650
7.
Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M,
Minderer M, Heigold G, Gelly S et al (2020) An image is worth 16x16 words: transformers for
image recognition at scale. arXiv:2010.11929
8.
Fu H, Li F, Orlando JI, Bogunović H, Sun X, Liao J, Xu Y, Zhang S, Zhang X (2019) Refuge:
retinal fundus glaucoma challenge. https://doi.org/10.21227/tz6e-r977
9.
Fumero F, Sigut J, Alayón S, González-Hernández M, González de la Rosa M (2015) Interactive
tool and database for optic disc and cup segmentation of stereo and monocular retinal fundus
images
10.
Hayreh SS (1969) Blood supply of the optic nerve head and its role in optic atrophy, glaucoma,
and oedema of the optic disc. Br J Ophthalmol 53(11):721. https://doi.org/10.1136/bjo.53.11.721
[Crossref]
11.
He H, Lin L, Cai Z, Tang X (2022) Joined: prior guided multi-task learning for joint optic
disc/cup segmentation and fovea detection. In: International conference on medical imaging with
deep learning. PMLR, pp 477–492
12.
Kuruvilla J, Sukumaran D, Sankar A, Joy SP (2016) A review on image processing and image
segmentation. In: 2016 International conference on data mining and advanced computing
(SAPIENCE), pp 198–203
13.
Lu S, Zhao H, Liu H, Li H, Wang N (2023) Pkrt-net: prior knowledge-based relation transformer
network for optic cup and disc segmentation. Neurocomputing 538:126,183
14.
Mishra M, Nath MK, Dandapat S (2011) Glaucoma detection from color fundus images. Int J
Comput Commun Technol (IJCCT) 2(6):7–10
15.
Pachade S, Porwal P, Kokare M, Giancardo L, Meriaudeau F (2021) NENet: Nested efficientNet
and adversarial learning for joint optic disc and cup segmentation. Med Image Anal 74:102,253
16.
Prastyo PH, Sumi AS, Nuraini A (2020) Optic cup segmentation using u-net architecture on
retinal fundus image. JITCE (J Inf Technol Comput Eng) 4(02):105–109
[Crossref]
17.
Raj PK, Kumar JH, Jois S, Harsha S, Seelamantula CS (2019) A structure tensor based Voronoi
decomposition technique for optic cup segmentation. In: 2019 IEEE international conference on
image processing (ICIP). IEEE, pp 829–833
18.
Shankaranarayana SM, Ram K, Mitra K, Sivaprakasam M (2019) Fully convolutional networks
for monocular retinal depth estimation and optic disc-cup segmentation. IEEE J Biomed Health
Inform 23(4):1417–1426
[Crossref]
19.
Sivaswamy J, Krishnadas S, Joshi GD, Jain M, Tabish AUS (2014) Drishti-gs: retinal image
dataset for optic nerve head (onh) segmentation. In: 2014 IEEE 11th international symposium on
biomedical imaging (ISBI). IEEE, pp 53–56
20.
Turkoski BB (2012) Glaucoma and glaucoma medications. Orthopaed Nurs 31(1):37–41
[Crossref]
21.
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I
(2017) Attention is all you need. In: Advances in neural information processing systems, vol 30
22.
Wang S, Yu L, Yang X, Fu CW, Heng PA (2019) Patch-based output space adversarial learning
for joint optic disc and cup segmentation. IEEE Trans Med Imaging 38(11):2485–2495
[Crossref]
23.
Weinreb RN, Aung T, Medeiros FA (2014) The pathophysiology and treatment of glaucoma: a
review. JAMA 311(18):1901–1911
[Crossref]
24.
Wong D, Liu J, Lim J, Li H, Wong T (2009) Automated detection of kinks from blood vessels for
optic cup segmentation in retinal images. In: Medical imaging 2009: computer-aided diagnosis,
vol 7260. SPIE, pp 459–466
25.
Xue X, Wang L, Du W, Fujiwara Y, Peng Y (2022) Multiple preprocessing hybrid level set model
for optic disc segmentation in fundus images. Sensors 22(18):6899
[Crossref]
26.
Zhang Z, Yin FS, Liu J, Wong WK, Tan NM, Lee BH, Cheng J, Wong TY (2010) Origa-light: an
online retinal fundus image database for glaucoma analysis and research. In: 2010 Annual
international conference of the IEEE engineering in medicine and biology, pp 3065–3068.
https://doi.org/10.1109/IEMBS.2010.5626137
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed
Computing , Studies in Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_8
Manan Padsala
Email: [email protected]
Devesh Jani
Email: [email protected]
Tanmay Bisen
Email: [email protected]
Aastha Shayla
Email: [email protected]
Susham Biswas
Email: [email protected]
Abstract
Urban landscapes, emblematic of modernization and growth, are
increasingly faced with the intricate challenge of visual pollution. This
nuanced form of pollution, often overshadowed environmental discussions,
profoundly influences the aesthetic harmony and mental well-being of
urban inhabitants. In this research, we present an innovative methodology
to detect visual pollution using drone-captured imagery. Our distinctive
dataset captures a spectrum of visual pollutants, from graffiti, faded
signage, and potholes to more complex issues like cluttered sidewalks and
unkempt facades. Leveraging this dataset, we fine-tuned pre-trained object
detection models, specifically YOLOv6, achieving remarkable accuracy in
detecting these visual pollutants from images. Central to our study is the
introduction of the Visual Pollution Index (VPI), a metric formulated
through the multiplicative integration of the Counting Categories Ratio
(CCR) and the Severity-Weighted Score (SWS). To provide a spatial
representation of visual pollution levels, we further introduce heatmap
visualizations. These heatmaps, overlaid on urban maps, offer a vivid
depiction of pollution hotspots, enabling city planners and stakeholders to
pinpoint areas of concern. Grounded in real-world perceptions, our
approach offers a comprehensive lens to assess, visualize, and address
visual pollution in urban environments.
1 Introduction
Urban environments, as the epicenters of human activity and innovation,
have witnessed unprecedented growth over the past few decades. While this
growth has brought about numerous advancements and opportunities, it has
also introduced a myriad of challenges, one of which is visual pollution.
Visual pollution, a term that encompasses unsightly and out-of-place man-
made objects within public and private spaces, has become a growing
concern for urban planners, environmentalists, and city dwellers alike [1].
The concept of visual pollution is not new; however, its significance has
grown in tandem with rapid urbanization. Visual disturbances, ranging from
graffiti, faded signage, and potholes to cluttered sidewalks and unkempt
facades, can degrade the aesthetic appeal of urban areas, impacting not only
the visual harmony but also the psychological well-being of residents [2].
Such disturbances can lead to decreased property values, reduced tourist
interest, and even adverse health effects due to stress and mental fatigue [3].
With the advent of technology, particularly in the realms of drone
imagery and machine learning, there exists an opportunity to address this
issue in a more systematic and data-driven manner. Drones, with their
ability to capture high-resolution images from vantage points previously
inaccessible, offer a unique perspective on urban landscapes [4]. When
combined with advanced object detection algorithms, such as YOLOv6,
these images can be analyzed to detect and quantify visual pollutants with
remarkable accuracy [5].
This paper introduces a novel approach to quantify visual pollution
using the Visual Pollution Index (VPI), a metric derived from drone-
captured imagery and object detection techniques. Furthermore, we present
heatmap visualizations to spatially represent visual pollution levels,
providing a tool for urban planners and stakeholders to make informed
decisions.
2 Literature Survey
Urban environments are increasingly being scrutinized for their aesthetic
appeal, given the rapid urbanization and the subsequent challenges it brings.
Visual pollution, an often-overlooked aspect, plays a crucial role in
determining the aesthetic quality of a place. The term “visual pollution”
refers to the entire set of unsightly and visually unpleasing elements in an
environment. This can range from graffiti, billboards, overhead power lines,
and even the architectural design of buildings.
For instance, a 2016 study by Chmielewski et al. delves into the
commercialization of public space by outdoor advertising and its potential
negative impact on the quality of life and enjoyment of public spaces [6].
The research illustrates that visual pollution can be quantified by correlating
public opinion with the number of visible advertisements. Using a 2.5D
outdoor advertisement dataset from a busy urban street in Lublin, Poland,
the study translates visibility into visual pollution. The findings suggest that
streetscape views with more than seven visible advertisements result in
visual pollution in the studied context. Our study extends this methodology
by incorporating a more comprehensive set of visual pollutants and
applying advanced deep learning techniques for detection.
Building on this, a 2019 study by Ahmed et al. delves deeper into the
realm of visual pollution [7]. The researchers propose a novel approach to
detect visual pollutants using deep learning techniques. More importantly,
they suggest the potential of creating a “Visual Pollution Index” (VPI) in
the future. This index would serve as a tool for urban planners and
professionals in urban environmental management, allowing them to
evaluate and compare the visual aestheticism of different geographic
regions. While they only proposed the idea, our research has taken the
initiative to design and implement such an index.
A distinguishing feature of our research is the utilization of a unique and
novel high-quality dataset, which stands in contrast to the commonly used
publicly available datasets of lower quality. This innovative dataset
enhances the accuracy, detection capabilities, and Intersection over Union
(IoU) metrics in our study. The superior quality of our dataset ensures more
reliable results, setting our research apart from previous studies in the
domain.
Further, a 2019 study by Wakil et al. titled “A Hybrid Tool for Visual
Pollution Assessment in Urban Environments” provides a systematic
approach for the development of a robust Visual Pollution Assessment
(VPA) tool [8]. The research introduces a methodology that integrates both
expert and public perspectives to rank Visual Pollution Objects (VPOs).
Using empirical decision-making techniques, the VPA tool produces a
point-based visual pollution scorecard. After extensive testing in Pakistan,
the tool offers regulators a consistent method for visual pollution
assessment and equips policy makers with a foundation for evidence-based
strategies. Our research builds on this, refining methods to evaluate visual
pollutants across urban settings.
Another 2022 study by Alharbi and Rangel-Buitrago focuses on the
visible deterioration and negative aesthetic quality of the landscape in
coastal environments [9]. Factors such as erosion, marine wrack, litter,
sewage, and beach driving are identified as contributors to visual pollution,
particularly in the Rabigh coastal area of the Kingdom of Saudi Arabia. The
research employs the Coastal Scenery Evaluation System (CSES) to assess
the scenic quality of 31 coastal sites.
In addition, a 2016 study by Madleňák and Hudák discusses the concept
of “visual smog,” which has emerged as a social concern in recent decades
[10]. This refers to the contamination of public spaces by aggressive and
often illegally placed advertisements that are not proportionate in size. The
study aims to measure the level of visual smog on selected road
communications, taking into account the number of ads and billboards near
roads, the distance between billboards and roads, and the density of these
billboards. The research also incorporates an analysis of traffic accidents on
the chosen road communications.
Lastly, a study by Yilmaz titled “In the Context of Visual Pollution:
Effects to Trabzon City Center Silhouette” examines the influence of visual
pollution on city silhouettes [11]. Silhouettes, indicative of a city’s history
and structure, are becoming markers of visual pollution. The research
underscores that many cities now feature buildings that lack harmony and
environmental consideration, leading to uniform concrete façades that
eclipse their historical essence. Using Trabzon as a case study, a coastal city
renowned for its historical richness, the study contrasts old and new city
images to assess the aesthetic shifts, highlighting the interventions that have
reshaped its distinctive silhouette.
In conclusion, the aforementioned studies collectively highlight the
growing importance of addressing visual pollution in urban settings. Our
research builds on these foundational studies, offering a more
comprehensive and practical solution to the problem. By designing the
Visual Pollution Index, we aim to provide urban planners and
environmentalists with a robust tool to assess and mitigate visual pollution
in our cities.
3 Problem Statement
As urban environments undergo rapid transformation, the issue of visual
pollution has become increasingly prominent, impacting the aesthetic and
overall quality of life in cities. While various studies have attempted to
address this concern, there is a distinct lack of a standardized, universally
applicable metric like the Visual Pollution Index (VPI) for quantifying and
addressing visual pollutants comprehensively. The development and
refinement of such an index, backed by advanced technological
methodologies, is crucial for enabling urban planners and environmentalists
to make informed decisions and implement effective strategies to combat
visual pollution.
4 Methodology
The entire methodology adopted for this study, including data collection,
processing, and analysis, is depicted in a comprehensive flowchart
presented in Fig. 1.
4.1.2 Timing
For the best lighting and image clarity, drone operations were planned
between 10 AM and 4 PM. This slot was selected after initial assessments
highlighted the difficulties of capturing under variable light conditions.
Here, the Severity Value corresponds to the value of severity against which
a particular visual pollutant is mapped, e.g., 0.2 for Very Low Severity. It’s
worth noting that this severity value can be replaced with custom weights
for each category based on specific user needs, allowing for a more tailored
assessment of visual pollution in different contexts.
(2) Object Detection and VPI Computation: Each image within the
group undergoes processing by the object detection model to identify
and classify visual pollutants. Leveraging the detected pollutants and
their respective severities, the VPI for the group of images is
determined. This VPI value, in conjunction with the group’s
coordinates, is archived for the subsequent heatmap creation.
Heatmap Coloring: Each VPI data point is depicted on the map using a
color spectrum, transitioning from light green (indicating a VPI close to 0)
to deep maroon (indicating a VPI nearing 1). The gradation in color
intensity offers a visual cue about the severity of visual pollution in
different regions. When the VPI is visualized for a limited set of images
covering a smaller area, the resultant heatmap might appear more localized
and detailed as depicted in Fig. 3. Conversely, with a denser array of data
points covering a broader region, the heatmap manifests as a more
continuous and expansive visual representation as depicted in Fig. 4.
Fig. 3 Pixelated heatmap of VPI for a limited dataset (Color bar represents VPI values)
Fig. 4 Artist’s illustrations of potential VPI heatmaps for India and Bengaluru
where
– (P ∩ G) represents the overlapping area between the predicted and
ground truth bounding boxes.
– (P ∪ G) represents the combined area of both bounding boxes.
Elevated IoU scores suggest superior model accuracy. A score of 1
denotes an ideal match, while a score of 0 signifies no correspondence. We
have achieved IoU values of 0.83 for Graffiti, 0.78 for Faded Signage,
0.88 for Potholes, 0.92 for Garbage, 0.82 for Construction Road, 0.71 for
Broken Signage, 0.79 for Bad streetlight, 0.77 for Bad Billboard, 0.89
for Sand on Road, 0.78 for Cluttered Sidewalk, and 0.79 for Unkempt
Facade. Overall, our models posted an average IoU rating of 0.81, which
suggests a praiseworthy degree of precision in identifying building
materials from the visuals.
Fig. 5 Aerial images of an elevated pollution zone chosen for physical survey
Fig. 6 Aerial images of some minimal pollution zones chosen for physical survey
6 Comparative Analysis
The quantification and visualization of visual pollution in urban landscapes
using aerial imagery and computational methods have garnered attention in
the urban planning and environmental aesthetics domain. Several studies
have delved into understanding the impact of visual pollutants on urban
aesthetics and the well-being of residents. Here, we present a comparative
analysis of our work with notable contributions in the literature:
1. Chmielewski, S. (2020): Chmielewski’s research titled “Chaos in
Motion: Measuring Visual Pollution with Tangential View Landscape
Metrics” delved into the concept of visual pollution (VP) in the form of
outdoor advertisements (OA) as a threat to landscape physiognomy.
The study proposed a methodological framework for measuring VP
using tangential view landscape metrics, backed by statistically
significant proofs. The research utilized raster products derived from
aerial laser scanning data to characterize areas in Lublin, East Poland.
The study highlighted the lack of consensus on the definition of VP and
the need for a quantified approach to address this challenge [27].
2. Zaeimdar, M., Khalilnezhad Sarab, F., and Rafati, M. (2019): This
study titled “Investigation of the relation between visual pollution and
citizenry health in the city of Tehran” focused on the impact of visual
contamination on the health of citizens in two urban areas of Tehran.
The research revealed a significant relationship between visual
contamination and various health indicators of citizens, including
physical signs, social function, anxiety, insomnia, and depression [28].
Our Contribution
7 Use Cases
1. Urban Aesthetics and Revitalization:
– Heatmap Insights: Urban designers can harness the power of the
generated heatmaps to discern visual pollution intensity across
varied locales. This empowers them to pinpoint areas needing
aesthetic enhancements or rejuvenation.
– Beautification Initiatives: Recognizing regions with pronounced
visual pollution can guide city planners in orchestrating targeted
beautification drives, green initiatives, and public art installations.
8 Conclusion
In this research, we have unveiled a pioneering methodology to assess
visual pollution in urban environments by harnessing aerial imagery and
cutting-edge object detection mechanisms. Our Visual Pollution Index
(VPI) emerges as a holistic metric, encapsulating both the variety and
gravity of visual pollutants. With the support of our distinctive dataset and
severity categorization, our results highlight the transformative potential of
technology in reshaping our understanding of urban aesthetics. The
congruence between our system’s evaluations and real-world perceptions
attests to the precision and dependability of our methodology. As cities
worldwide grapple with the challenges of urbanization, instruments like
ours will be instrumental in guiding efforts towards creating visually
harmonious urban landscapes. The horizon looks promising, with
opportunities for future research to refine, enhance, and broaden the scope
of our approach, weaving in real-time data streams and more intricate
analytical tools.
9 Future Scope
Integration with Smart City Infrastructure: As urban landscapes evolve
into smart cities, there’s potential to integrate the VPI with sensors and
cameras placed throughout the city. This would facilitate real-time
monitoring of visual pollution, enabling swift interventions and
continuous urban beautification efforts.
Predictive Analysis for Urban Aesthetics: By harnessing the power of
artificial intelligence and machine learning, future versions of this system
could anticipate areas prone to visual pollution. This predictive capability
could be based on urban growth patterns, historical data, and socio-
economic factors, allowing for proactive measures.
Virtual Reality and Augmented Reality Enhancements: The next frontier
could involve the use of Virtual Reality (VR) to simulate the visual
experience of different urban areas based on their VPI scores.
Additionally, Augmented Reality (AR) can be employed to superimpose
potential solutions or improvements on existing urban landscapes,
providing stakeholders with a futuristic vision of possible enhancements.
Adaptive Urban Design Frameworks: With the insights derived from
VPI, urban designers can develop adaptive design frameworks. These
would be dynamic urban design strategies that evolve based on real-time
VPI data, ensuring cities remain visually appealing amidst rapid
urbanization.
Acknowledgements
We extend our heartfelt gratitude to Kesowa Infinite Ventures Pvt Ltd for
their diligent efforts in collecting the invaluable dataset used in this study.
Their commitment to quality and precision has been instrumental in the
success of our research. We are deeply appreciative of their generosity in
granting us exclusive access to this data and allowing its use for our study.
References
1. Lynch K (1984) The image of the city. MIT Press
2.
Nasar JL (1994) Urban design aesthetics: the evaluative qualities of building exteriors. Environ
Behav 26(3):377–401
[Crossref]
3.
Ulrich RS (1984) View through a window may influence recovery from surgery. Science
224(4647):420–421
[Crossref]
4.
Anderson K, Gaston KJ (2013) Lightweight unmanned aerial vehicles will revolutionize spatial
ecology. Front Ecol Environ 11(3):138–146
[Crossref]
5.
Li C, Li L, Jiang H, Weng K, Geng Y, Li L, Ke Z, Li Q, Cheng M, Nie W, Li Y, Zhang B, Liang
Y, Zhou L, Xu X, Chu X, Wei X, Wei X (2022) YOLOv6: a single-stage object detection
framework for industrial applications. arXiv:2209.02976
6.
Chmielewski S, Lee DJ, Tompalski P, Chmielewski TJ, Wężyk P (2016) Measuring visual
pollution by outdoor advertisements in an urban street using intervisibilty analysis and public
surveys. Int J Geogr Inf Sci 30(4):801–818
[Crossref]
7.
Ahmed N, Islam MN, Tuba AS, Mahdy MRC, Sujauddin M (2019) Solving visual pollution with
deep learning: A new nexus in environmental management. J Environ Manag 248
8.
Wakil K, Naeem MA, Anjum GA, Waheed A, Thaheem MJ, Hussnain MQ, Nawaz R (2019) A
hybrid tool for visual pollution assessment in urban environments. Sustainability 11
9.
Alharbi OA, Rangel-Buitrago N (eds) Scenery evaluation as a tool for the determination of
visual pollution in coastal environments: the Rabigh coastline, Kingdom of Saudi Arabia as a
study case. Mar Pollut Bull 181
10.
Madleňák R, Hudák M (2016) The Research of Visual Pollution of Road Infrastructure in
Slovakia. In: Mikulski J (ed) Challenge of transport telematics. TST 2016. Communications in
computer and information science, vol 640. Springer, Cham
11.
Yilmaz D, Sagsöz A (2011) In the context of visual pollution: effects to Trabzon city Center
Silhoutte. Asian Soc Sci 7(5):98
[Crossref]
12.
Mitra D, Banerji S (2018) Urbanisation and changing waterscapes: a case study of New Town,
Kolkata, West Bengal, India. Appl Geogr 97:109–118
[Crossref]
13.
https://www.qgis.org/en/site/about/index.html
14.
Shorten C, Khoshgoftaar TM (2019) A survey on image data augmentation for deep learning. J
Big Data 6(1):60
[Crossref]
15.
Armstrong JS (2004) The Graffiti problem. General Economics and Teaching, University Library
of Munich, Germany
16.
Wardhana N, Ellisa E (2023) Youth tactics of urban space appropriation: case study of
skateboarding and graffiti. J Asian Arch Build Eng
17.
Stabnikov V, Chu J, Myo AN et al (2013) Immobilization of sand dust and associated pollutants
using bioaggregation. Water Air Soil Pollut 224
18.
Oladumiye EB (2013) Urban environmental graphics: impact, problems and visual pollution of
signs and billboards in Nigeria. Int J Educ Res 1:89
19.
Sarkar S (2003) Qualitative evaluation of comfort needs in urban walkways in major activity
centers. Transp Q 57(4):39–59
20.
Ściężor T (2021) Effect of street lighting on the urban and rural night-time radiance and the
brightness of the night sky. Remote Sens 13(9):1654. https://doi.org/10.3390/rs13091654
[Crossref]
21.
Andjarsari S, Subadyo AT, Bonifacius N (2022) Safe construction and visual pollution of
billboards along main street. In: IOP conference series: earth and environmental science, vol 999,
no 1, p 012015
22.
Lafta ZQ, Al-Shamry MMH (2022) The spatial variation of the aspects of environmental
pollution in Al-Hilla city and its environmental effects. Baltic J Law Polit 15(2)
23.
Naveen N, Mallesh Yadav S, Sontosh Kumar A (2018) A study on potholes and its effects on
vehicular traffic. Int J Creat Res Thoughts 6(1):2320–2882
24.
Latypova A, Lenkevich A, Kolesnikova D, Ocheretyany K (2022) Study of visual garbage as
visual ecology perspective. gmd 4(2):153–172
25.
https://geopandas.org/en/stable/
26.
Everingham M et al (2010) The PASCAL visual object classes (VOC) challenge. Int J Comput
Vision 88(2):303–338
[Crossref]
27.
Chmielewski S (2020) Chaos in motion: measuring visual pollution with tangential view
landscape metrics. Land 9(12):515. https://doi.org/10.3390/land9120515
[Crossref]
28.
Zaeimdar M, Khalilnezhad Sarab F, Rafati M (2019) Investigation of the relation between visual
pollution and citizenry health in the city of Tehran (case study: municipality districts No.1 & 12
of Tehran). Anthropog Pollut 3(1):1–10
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed
Computing , Studies in Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_9
Sirirat Wacharawichanant
Email: [email protected]
Abstract
Proteins are essential structural and functional components of human cells.
Understanding and identifying proteins can provide valuable insights into
their structure, function and role in human body. Subcellular proteins
provide the expression that characterizes the many proteins and their
conditions across cell types. This work proposed a classification model for
subcellular protein patterns using XGBoost with transfer learning of CNN
as the feature extractor. In the model training process, we used ResNet50,
VGG16, Xception, and MobileNet as the pre-trained models based on the
transfer learning technique to extract different features. The proposed model
was used to classify subcellular proteins into 28 patterns. The XGBoost
with ResNet50, VGG16, MobileNet, and Xception model achieved an
accuracy level of 92.20% 92.77%, 91.63%, and 91.44%, respectively. The
XGBoost with ResNet50, VGG16, MobileNet, and Xception model
obtained an F1 score of 0.9179, 0.9233, 0.9131, and 0.9152, respectively.
Considering the F1 score, All XGBoost with transfer learning of CNN
models gave a high score. Therefore, all evaluation parameters clearly
demonstrate the high performance of the subcellular protein pattern
classification model.
1 Introduction
Proteins are essential structural and functional components of human cells.
Understanding and identifying proteins can provide valuable insights into
their structure, function and role in human body. The human protein atlas
(HPA) is a large source of proteomic data that need to annotate and
characterize the many proteins and their conditions across cell types. The
subcellular protein is section of the HPA provides the expression and
spatiotemporal distribution of proteins encoded. It is made up of a protein
database of proteomic microscopy images in various different cell lines and
tissues.
Since the monstrous amounts of image data, it is difficult for the
machine to identify the data of these cells. Therefore, a powerful tools are
needed to annotate and characterize the proteins and their conditions. For
instance, Newberg et al. [1] presented an automated method for processing
and classifying major subcellular patterns in the Atlas images. They used
support vector machine and random forest as the classification frameworks
at determining subcellular locations. In a similar way, Shwetha et al. [2]
used a conventional approach to extract the various features the images.
Then the features are fed into a classifier. Moreover, they used a
convolutional neural network (CNN) to extract features and classify the
images into 15 classes. Although the human protein classification problem
has been successfully solved by deep learning methods, the extraction of
discriminant features from images is a very hard and time-consuming
process. Extreme Gradient Boosting(XGBoost) is a highly efficient
algorithm which pushes the limit of computational resources. A interesting
study on XGBoost is proposed by Sugiharti el al. [3]. They combined the
CNN with XGBoost for enhancement the accuracy of early detection of
breast cancer on mammogram images. In a like manner, Punuri et al. [4]
used CNN in conjunction with XGBoost technique in facial emotion
recognition. Additionally, Jiang et al. [5] used transfer learning with
XGBoost method to predict protein-coupled receptor interaction. They used
XGBoost as a weak classifier and using the TrAdaBoost algorithm based on
Jensen-Shannon divergence. After that, the transfer learning of CNN was
used for model training. Khan et al. [6] proposed CNN-XGBoost method
for recognizing three dimensions of emotion. They applied several feature
extraction techniques such as fast fourier transformation, and discrete
cosine transformation.
In summary, the combination technique of XGBoost and transfer
learning of CNN is the challenging and most effective way to classify
subcellular protein pattern images. Another challenge is creating a
classification model in the multiclass and imbalance of data. In this work,
we will be used transfer learning of CNN to train the model on the HPA
images dataset to retrieve the features. Then we applied XGBoost classifier
to the extracted features from the transfer learning technique to classify the
subcellular protein into 28 patterns. Furthermore, we have performed the
data pre-processing and augmentation in a dataset. The hybrid model of
transfer learning of CNN and XGBoost will be fine-tuned with the optimal
parameters to provide the best performance.
The paper is organized as follows. Section 2 describes the research
methods, techniques, and evaluation metrics used in this study. Section 3
describes the experimental processes, including data preparation, model
creation, and parameter tuning. Section 4 presents the results of the
experiments and their evaluation. Finally, Sect. 5 concludes the paper.
2 Methodology
In this section, we will describe the research methodology and techniques
used to confirm that the processed classification method is the most
accurate. These techniques include convolutional neural networks, transfer
learning, extreme gradient boosting, and evaluation metrics.
2.1 Transfer Learning of Convolutional Neural Network
A convolutional neural network (CNN) is a powerful type of artificial
neural network that can be used for image recognition and processing.
CNNs have a basic architecture that consists of multiple layers, each of
which is responsible for recognizing and translating different types of
information from images. Transfer learning is a technique that uses an
existing model as a starting point for training a new model. This can save
time and improve performance, as the new model can learn from the
knowledge that the existing model has already learned. In this work, we
used the ResNet50, VGG16, Xception, and MobileNet as pre-trained
models. These models were selected because they are lightweight,
frequently used, low-resource, and high-efficiency models.
Residual Network (ResNet) is a popular deep learning model for
computer vision applications. It was first introduced in 2015 by He et
al. [7]. ResNets have different layer-based architectures, depending on their
variants. The original ResNet architecture has 34 layers and uses skip
connections to connect different layers. These skip connections allow the
network to learn residual mappings, which are functions that map the input
to the output. ResNets are made by stacking residual blocks together. Each
residual block consists of two 3 3 convolutional layers with a ReLU
activation function. In this study, we used ResNet50, which has 50 layers.
ResNet50 is a popular choice for computer vision tasks because it is
accurate and efficient.
VGG [8] is a classic CNN architecture that improves on AlexNet by
using multiple 3 3 convolutional filters instead of larger filters. The VGG
architecture consists of the following layers. Input layer receives a 224
224 image. Convolutional layers use multiple 3 3 convolutional filters.
Rectified linear unit activation function (ReLU) use for reducing training
time. Hidden layers use ReLU instead of Local Response Normalization to
improve overall accuracy. Pooling layers use to reduce the dimensionality
and number of parameters of the feature maps. VGG uses three fully
connected layers. In this work, we used VGG16, which has 16 layers.
Xception [9] is improved from Inception architecture(GoogLeNet) with
a modified depthwise separable convolution. The architecture of Xception
consists of depth-wise separable convolution layers and their residual
connections. The depth-wise separable convolution is an Inception module
with a large number of layers. This modification gives more efficiency in
terms of computation time. The depth-wise separable convolutions consist
of two major processes include depth-wise convolution, and point-wise
convolution. The Xception model is composed of 36 convolutional layers
forming the feature extraction, and more than 22 million parameters.
MobileNet [10] is a mobile and embedded vision application of CNN.
MobileNet is widely used in object detection, fine-grained classifications,
face attributes, etc. MobileNet is the most efficient and lightweight neural
network that uses depth-wise separable convolutions to construct light-
weight deep convolutional neural networks. The archtecture of MobileNet
is based on depth-wise separable filters which are comprised of two layers,
depth-wise convolution filters, and point convolution filters. Depth-wise
convolution filters are applied to each input channel independently, unlike
standard convolution filters which are applied to all input channels at once.
The point convolution filter is used to combine the output of depth-wise
convolution with 1 1 convolutions to create a new feature.
(1)
(2)
Equation (2) implies the scoring function which is used to measure the
quality of the tree structure q. XGBoost performs with speed and accurate
prediction by merging both weak and strong learners iteratively to meet the
highest classification accuracy rate[12]. In general, XGBoost employs the
weak predictor principle of the ensemble decision tree. A performance
achieves by optimizing the loss function value in boosting process. The
XGBoost model for classification problems is called XGBClassifier. We
have used XGBoost in the classification step. After applying CNN to
extract the features, the retrieved features are used by XGBoost to classify
all of the subcellular protein patterns. The advantage of XGBoost is its
scalability, fast speed, and high accuracy.
Fig. 1 An architecture of the proposed XGBoost model based on transfer learning as feature
extractor
(3)
where Accuracy is the percentage of data points that are correctly classified.
(4)
where Precision is the percentage of predicted positive data points that are
actually positive.
(5)
(6)
3 Experiments
The subcellular protein pattern classification processes are designed for
multiple steps which are illustrated in Fig. 1.
Fig. 3 Number of images in each class
Fig. 4 An example of images in dataset
5 Conclusion
This work proposed a classification model for subcellular protein patterns
using XGBoost with transfer learning of CNN as the feature extractor. In
the model training process, we used ResNet50, VGG16, MobileNet and
Xception as the pre-trained models based on the transfer learning technique
to extract different features. The proposed model was used to classify
subcellular proteins into 28 patterns. The XGBoost with ResNet50, VGG16,
MobileNet, and Xception model achieved an accuracy level of 92.20%
92.77%, 91.63%, and 91.44%, respectively. The XGBoost with ResNet50,
VGG16, MobileNet, and Xception model obtained an F1 score of 0.9179,
0.9233, 0.9131, 0.9152, respectively. Considering the F1 score, All
XGBoost with transfer learning of CNN models gave a high score.
Therefore, all evaluation parameters clearly demonstrate the high
performance of the subcellular protein pattern classification model.
References
1. Li J, Newberg JY, Uhlén M, Lundberg E, Murphy RF (2012) Automated analysis and
reannotation of subcellular locations in confocal images from the human protein atlas. PLOS
ONE 7(11):1–10, 11
2.
Shwetha TR, Thomas SA, Kamath V, Niranjana KB (2019) Hybrid xception model for human
protein atlas image classification. In: 2019 IEEE 16th India council international conference
(INDICON), pp 1–4
3.
Sugiharti E, Arifudin R, Wiyanti DT, Susilo AB (2021) Convolutional neural network-xgboost
for accuracy enhancement of breast cancer detection. J Phys Conf Ser 1918(4):042016
4.
Punuri SB, Kuanar SK, Kolhar M, Mishra TK, Alameen A, Mohapatra H, Mishra SR (2023)
Efficient net-xgboost: an implementation for facial emotion recognition using transfer learning.
Mathematics 11(3)
5.
Tengsheng J, Yuhui C, Shixuan G, Hu Z, Lu W, Fu Q, Yijie D, Haiou L, Wu H (2022) G protein-
coupled receptor interaction prediction based on deep transfer learning. IEEE/ACM Trans
Comput Biol Bioinform 19(6):3126–3134
6.
Khan MS, Salsabil N, Alam MGR, Dewan MAA, Uddin MZ (2022) Cnn-xgboost fusion-based
affective state recognition using EEG spectrogram image analysis. Sci Rep 12(1):14122
7.
He K, Zhang X, Ren S, Sun J (2015) Deep residual learning for image recognition. CoRR
abs/1512.03385
8.
Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image
recognition. In: Bengio Y, LeCun Y (eds) 3rd international conference on learning
representations, ICLR 2015, San Diego, CA, USA, May 7–9, 2015, conference track proceedings
9.
Chollet F (2017) Xception: deep learning with depthwise separable convolutions. In: 2017 IEEE
conference on computer vision and pattern recognition (CVPR). IEEE Computer Society, Los
Alamitos, CA, USA, pp 1800–1807
10.
Andrew G, Howard, Menglong Z, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M,
Adam H (2017) Efficient convolutional neural networks for mobile vision applications.
Mobilenets
11.
Chen T, Guestrin C (2016) Xgboost: a scalable tree boosting system. In: Proceedings of the
22Nd ACM SIGKDD international conference on knowledge discovery and data mining, KDD
’16. ACM, New York, NY, USA, pp 785–794
12.
Friedman JH (2000) Greedy function approximation: a gradient boosting machine. Ann Stat
29(5):1189–1232
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed
Computing , Studies in Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_10
Yoshihiro Nishi
Email: [email protected]
Abstract
Financial sentiment analysis can help investors make more efficient
decisions. Finan-cial sentiment analysis involves analyzing text data to
extract sentiments and assists market participants in making informed
investment decisions. In recent years, there have been many efforts to
improve analysis accuracy using machine learning and deep learning.
Particularly FinBERT, which is pre-trained on financial datasets and excels
in analyzing financial documents. However, the output may require more
information to make decisions. It has the disadvantage of being difficult to
use for critical decision-making because of the need for help explaining the
output results. This study applies the BERTopic and SHAP to financial
sentiment analysis by FinBERT. It proposes a Shapley FinBERTopic
system, which contributes to better investor decision-making by explicitly
showing the impact of news article topics and individual words on stock
prices. The results of the experiments with this system showed an increased
interpretability of the results compared to using FinBERT alone. A more
detailed analysis will be conducted in the future.
Keywords Decision support system – Text analysis – Financial markets –
Explainable AI – SHAP – BERTopic – FinBERT
1 Introduction
Sentiment analysis using natural language processing is gaining attention in
the financial sector for analyzing information about stock prices and market
trends [1–3]. Sentiment analysis can analyze text data to extract positive,
negative, or neutral information. Market participants can make more
accurate investment decisions by understanding sentiment analysis results.
The adoption of deep learning models to sentiment analysis in the financial
sector is a topic that has received much attention in recent research.
Previous research has used deep learning models such as LSTM (Long
Short-Term Memory) [4], CNN (Convolutional Neural Network), and
BERT (Bidirectional Encoder Representations from Transformers) [3, 5].
Among them, FinBERT has been pre-trained on a financial-specific dataset
and is highly accurate in the analysis of financial documents [6]. These
methods have the potential to provide higher forecasting accuracy.
However, challenges remain in the explanatory and interpretability of deep
learning models. The deep learning-based analysis is becoming increasingly
valuable, but it has the disadvantage of being difficult to use for critical
decision-making because of the need for help explaining the output results.
Topic Modeling is a method used in text mining and natural language
processing to discover abstract topics in a collection of documents. The
primary aim is to identify hidden thematic structures within the text body. A
common technique employed for this purpose is Latent Dirichlet Allocation
(LDA), which bases the discovery of topics on word frequency and co-
occurrence [7]. This method has been widely used for various purposes like
document clustering, organizing large blocks of textual data, information
retrieval, and feature selection [8]. BERTopic is a modern take on topic
modeling that utilizes BERT embeddings. Unlike traditional methods like
LDA, BERTopic explores contextual word embeddings to apprehend
semantic relationships among words and documents [9]. One of the features
of BERTopic is its semantic understanding, which is achieved through
BERT's contextual embeddings. This feature enables BERTopic to generate
more coherent and interpretable topics than traditional methods. Another
feature is the hierarchical topic reduction, which provides a more granular
understanding of topics. BERTopic also offers built-in visualization support
to help understand and interpret the topic distribution across the corpus. It is
customizable to a reasonable extent, supporting different transformer
models and offering parameters to tweak the topic modeling process as per
requirements. Additionally, BERTopic is known for creating stable topics
even with small datasets, addressing a common issue with traditional topic
modeling techniques.
Researchers are working to develop XAI (Explainable Artificial
Intelligence), as the importance of transparency and accountability in AI has
been increasing recently [10]. XAI technology makes decision-making
more understandable to humans [11, 12]. SHAP (Shapley Additive
Explanations) have received particular attention among XAI techniques.
SHAP can evaluate the importance of each feature to the predictions of a
machine learning model [13]. SHAP is a technique for improving the
explanatory power of predictions in machine learning models and applies to
natural language processing analysis. SHAP allows one to explain how the
model is making predictions, thereby improving the explanatory nature of
the model [14].
This study applies the BERTopic and SHAP to financial sentiment
analysis by FinBERT. It proposes a Shapley FinBERTopic system, which
contributes to better investor decision-making by explicitly showing the
impact of individual news article topics and individual words on stock
prices. A Shapley FinBERTopic uses investors as a representative example
to analyze model-based decision-making by contributing to a more efficient
understanding of the model's features by those involved in the decision-
making process, thereby supporting better decision-making. In addition, the
effectiveness of the proposed method is also experimented in conjunction
with the proposed method.
2 Related Works
The financial industry is actively using natural language processing
techniques for analysis [15–19]. Also, Financial sentiment analysis using
natural language processing techniques plays an essential role in the
financial sector and has been the subject of many studies [1–3]. Financial
natural language processing sentiment analysis was used in one study to
examine the relationship between news articles and stock prices, and the
results showed well against existing quant funds [20]. Another study used
natural language processing sentiment analysis to predict stock prices by
analyzing [21]. As described above, financial natural language processing
sentiment analysis plays an essential role in the financial field, and its
application is expected to expand.
FinBERT is a BERT model specific to the financial sector and is used
for sentiment analysis of stock markets and financial instruments [22]. In a
study that evaluated the construction and performance of FinBERT,
researchers showed that FinBERT is more suitable for text mining in the
financial field than the general BERT model [6]. FinBERT is expected to be
a useful deep-learning model for sentiment analysis in financial markets.
However, the default output from FinBERT sentiment analysis is only a
single label and score. The interpretability of the model's output is limited,
as with many deep learning models.
Topic Modeling embodies a fundamental approach within text mining
and natural language processing, facilitating the unraveling of abstract
topics dispersed across a corpus of documents. The principal objective is to
discern concealed thematic architectures embedded within the textual body.
A frequently employed technique for realizing this objective is the LDA,
which predicates discussing topics on word frequency and co-occurrence
metrics [8]. This method has garnered widespread application across
diverse realms, such as document clustering, orchestration of voluminous
textual data, information retrieval, and feature selection, attesting to its
utility and robustness [23–25]. BERTopic emerges as a nuanced rendition of
topic modeling, harnessing the prowess of BERT embeddings to fathom the
semantic landscapes within texts [9]. Also, BERTopic is gradually being
used in the financial domain, and there are reports that it is more flexible
than LDA and LSA (Latent Semantic Analysis), can capture semantic
relationships that LDA and LSA cannot, and provides meaningful and
diverse topics [26]. In this study, the operational essence of BERTopic is to
ameliorate the interpretability lacuna that encumbers the sentiment analysis
utilization from FinBERT, hence rendering these outputs more intelligible
and actionable. Incorporating BERTopic engenders an auxiliary stratum of
topic-centric interpretability, situating itself as a vital instrument to amplify
the comprehension of sentiment trajectories in financial textual data. This
augmentation harbors the potential to catalyze more enlightened decision-
making paradigms within the financial sector, underscoring the substantive
value of augmenting interpretability in machine-driven sentiment analysis.
The symbiotic amalgamation of BERTopic and FinBERT not merely
endows the system with an enhanced cognizance of the predominant
sentiment in financial dialogues but also enriches the understanding of the
contextual themes intricately woven within the sentiment indices.
Further, efforts to improve the interpretability of analytical models are
progressing in the field of XAI. Among them, SHAP is a method for
explaining the predictions of machine learning models. This method
evaluates the extent to which individual features contribute to predictions.
Specifically, it evaluates how the predictions change when individual
features are removed from the model and calculates the importance of the
features based on the degree of change. SHAP uses the Shapley value, a
method used in game theory to quantify how multiple players distribute the
benefits of cooperation. The method accurately assesses how much an
individual feature contributes to the model's predictions. SHAP can be
applied to a wide variety of machine learning models. In addition,
visualization of SHAP values provides an intuitive understanding of the
importance of features [13], and numerous studies have shown their
effectiveness. For example, SHAP has been proposed as an interpretation
method for tree-based models from local explanations to global
understanding [27]. There are other examples where SHAP has been
applied to make the explanatory properties of machine learning models
more robust and easier to interpret [28]. Some usefulness of SHAP has also
been reported in efforts to apply SHAP to classification and analysis models
in the financial field [29–31]. There has also been an attempt to explain
agent behavior in financial stock trading by employing SHAP in Deep Q
Network (DQN), a deep reinforcement learning architecture [32]. There are
also active efforts to apply SHAP to natural language processing, and some
have applied SHAP to specific models [14]. These studies show that SHAP
is a promising method for explaining predictions in machine learning and
deep learning models using natural language processing.
As described above, natural language processing financial sentiment
analysis is a promising field that is expected to develop in the future.
Pretrained models such as FinBERT, specialized in the financial field, have
been released, making natural language processing financial sentiment
analysis via FinBERT possible. However, the interpretability of the model's
default output is still an issue, and there is room for progress. Incorporating
BERTopic alongside FinBERT could potentially address the interpretability
issue by uncovering thematic structures within financial texts, thereby
providing a layer of topic-based interpretability to the sentiment analysis
output generated by FinBERT, thus contributing to the ongoing
advancements in Explainable AI (XAI) within the domain of financial
sentiment analysis. Further, efforts to improve the interpretability of
analytical models in the field of XAI are progressing, and models such as
SHAP have emerged. SHAP has been reported as a promising method for
explaining predicates in machine learning and deep learning models using
natural language processing. Combining them to build new analytical
models has the potential to improve the interpretability of the models
efficiently.
4 Experiments
In this study, 337,430 news articles distributed by Thomson Reuters in the
financial markets in 2017 were used as data for analysis. Figures 3 and 4
show the results of tokenizing the obtained news articles by finbert-tone and
analyzing them by FinBERT. The vertical axis indicates the number of news
articles, and the horizontal axis shows the FinBERT score. The horizontal
axis means the certainly with which the target news article can be the label
of the output result. 37,803 news articles were classified as Positive,
297,521 news articles were classified as Neutral, and 2,860 news articles
were classified as Negative. News articles classified as Negative were
2,106. High scores of 0.9–1.0 predicted a high percentage of labels for all
labels.
Acknowledgements
This research was partially supported by Telecommunications Advancement
Foundation, JSPS KAKENHI Grant Number JP20K01751and Keio
University Academic Development Funds.
References
1. Kearney C, Liu S (2013) Textual sentiment analysis in finance: a survey of methods and models.
SSRN Electron J.
2.
Kearney C, Liu S (2014) Textual sentiment in finance: a survey of methods and models. Int Rev
Financ Anal 33:171–185
[Crossref]
3.
Man X, Luo T, Lin J (2019) Financial sentiment analysis (FSA): a survey. In: 2019 IEEE
international conference on industrial cyber physical systems (ICPS). pp 617–622. https://doi.
org/10.1109/ICPHYS.2019.8780312
4.
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
[Crossref]
5.
Devlin J, Chang M-W, Lee K, Toutanova K (2018) BERT: pre-training of deep bidirectional
transformers for language understanding. arXiv:1810.04805
6.
Araci D (2019) FinBERT: financial sentiment analysis with pre-trained language models. arXiv:
1908.10063
7.
Blei DM, Ng AY, Jordan MI (2003) Latent Dirichlet allocation. J Mach Learn Res 3:993–1022
8.
Blei DM, Lafferty JD (2007) A correlated topic model of science. Ann Appl Stat 1(1):17–35.
https://doi.org/10.1214/07-AOAS114
[MathSciNet][Crossref]
9.
Grootendorst M (2022) Bertopic: neural topic modeling with a class-based TF-IDF procedure.
arXiv:2203.05794
10.
Adadi A, Berrada M (2018) Peeking inside the black-box: a survey on explainable artificial
intelligence (XAI). IEEE Access 6:52138–52160
[Crossref]
11.
Molnar C (2022) Interpretable machine learning. Github.io. https://christophm.github.io/
interpretable-ml-book/
12.
Doshi-Velez F, Kim B (2018) Towards a rigorous science of interpretable machine learning.
arXiv:1702.08608
13.
Lundberg SM, Lee S (2017) A unified approach to interpreting model predictions. In:
Proceedings of the 31st international conference on neural information processing systems
(NIPS’17). pp 4768–4777
14.
Mosca E, Szigeti F, Tragianni S, Gallagher D, Groh G (2022) SHAP-based explanation methods:
a review for NLP interpretability. In: Proceedings of the 29th international conference on
computational linguistics. pp 4593–4603
15.
Xing FZ, Cambria E, Welsch RE (2017) Natural language based financial forecasting: a survey.
Artif Intell Rev 50:49–73
[Crossref]
16.
Nishi Y, Suge A, Takahashi H (2021) Construction of a news article evaluation model utilizing
high-frequency data and a large-scale language generation model. SN Bus & Econ 1
17.
Nishi Y, Suge A, Takahashi H (2020) News articles evaluation analysis in automotive industry
using GPT-2 and co-occurrence network. In: New frontiers in artificial intelligence. pp 103–114
18.
Nishi Y, Suge A, Takahashi H (2020) Construction of news article evaluation system using
language generation model. Agents Multi-Agent Syst: Technol Appl 2020:313–320
19.
Nishi Y, Suge A, Takahashi H (2019) Text analysis on the stock market in the automotive
industry through fake news generated by GPT-2. In: Proceedings of the artificial intelligence of
and for business
20.
Schumaker RP, Chen H (2010) A discrete stock price prediction engine based on financial news.
Computer 43(1):51–56. https://doi.org/10.1109/mc.2010.2
[Crossref]
21.
Attanasio G, Cagliero L, Garza P, Baralis E (2019) Combining news sentiment and technical
analysis to predict stock trend reversal. In: Proceedings of the 2019 International Conference on
Data Mining Workshops (ICDMW). https://doi.org/10.1109/icdmw.2019.00079.
22.
Huang AH, Wang H, Yang Y (2022) FinBERT: a large language model for extracting information
from financial text. Contemp Account Res
23.
Mahajan A, Dey L, Haque SM (2008) Mining financial news for major events and their impacts
on the market. In: 2008 IEEE/WIC/ACM international conference on web intelligence and
intelligent agent technology. https://doi.org/10.1109/wiiat.2008.309
24.
Hagen L (2018) Content analysis of e-petitions with topic modeling: how to train and evaluate
LDA models? Inf Process Manag 54(6):1292–1307. https://doi.org/10.1016/j.ipm.2018.05.006
[Crossref]
25.
Bastani K, Namavari H, Shaffer J (2019) Latent dirichlet allocation (LDA) for topic modeling of
the CFPB consumer complaints. Expert Syst with Appl 127:256–271. https://doi.org/10.1016/j.
eswa.2019.03.001
[Crossref]
26.
Raju SV, Bolla BK, Nayak DK, Kh J (2022) Topic modelling on consumer financial protection
bureau data: an approach using BERT based embeddings. In: 2022 IEEE 7th international
conference on for convergence in technology (I2CT). pp 1–6 (2022)
27.
Lundberg SM et al (2020) From local explanations to global understanding with explainable AI
for trees. Nat Mach Intell 2(1):56–67. https://doi.org/10.1038/s42256-019-0138-9
[Crossref]
28.
Parsa AB, Movahedi A, Taghipour H, Derrible S, Mohammadian A (2020) (Kouros): toward
safer highways, application of XGBoost and SHAP for realtime accident detection and feature
analysis. Accident Analysis & Prev 136:105405. https://doi.org/10.1016/j.aap.2019.105405
[Crossref]
29.
Ohana JJ, Ohana S, Benhamou E, Saltiel D, Guez B (2021) Explainable AI (XAI) models
applied to the multi-agent environment of financial markets. In: Calvaresi D, Najjar A, Winikoff
M, Främling K (eds) Explainable and transparent AI and multi-agent systems. EXTRAAMAS
2021. Lecture notes in computer science, vol 12688. Springer, Cham. https://doi.org/10.1007/
978-3-030-82017-6_12
30.
Xia X, Zhang X, Wang Y (2019) A Comparison of feature selection methodology for solving
classification problems in finance. J Phys: Conf Ser 1284(1):012026. https://doi.org/10.1088/
1742-6596/1284/1/012026
[Crossref]
31.
Xiaomao X, Xudong Z, Yuanfang W (2019) A comparison of feature selection methodology for
solving classification problems in finance. J Phys: Conf Ser 1284:012026. https://doi.org/10.
1088/1742-6596/1284/1/012026.
32.
Kumar S, Vishal M, Ravi V (2022) Explainable reinforcement learning on financial stock trading
using SHAP. arXiv:2208.08790
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed
Computing , Studies in Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_11
Yuji Tanaka
Email: [email protected]
Masaaki Kunigami
Email: [email protected]
Hiroshi Takahashi
Email: [email protected]
Takao Terano
Email: [email protected]
Abstract
This paper describes new applications of a Large Language Model (LLM)
for business domains. So far, we have conducted research on agent-based
simulation models to uncover complex socio-technical systems. However,
to let ordinary business people understand the models and their
consequences, conventional validation or visualization methods are not
enough. We must explain the plausible results through cases with natural
languages. In our previous studies, we have reported a method for
describing simulation results in natural language and grounding them with
actual business case. Based on the results, we utilize a Large Language
Model for the generation. From this study we have achieved the following
results: (1) simulation results are comprehensively analyzed and
systematically classified, and (2) the classification results are used as
prompts for with a LLM or ChatGPT, and (3) the LLM generates plausible
business cases with natural language. We have confirmed that the generated
cases are coincide with previous manual generated explanations and easy to
understand for ordinary business people.
1 Introduction
A socio-technical system is a large-scale complex system that serves as a
social infrastructure, such as transportation, information and
communication, finance, and electric power [1]. These systems comprise
technical subsystems, as well as social subsystems that stem from the
individuals and organizations overseeing their functioning, along with their
intricate interactions. To attack the systems, we must consider technical,
social, and communication issues [2].
In the realm of designing and implementing socio-technical systems, the
need for clear boundaries becomes even more pronounced as the scope and
complexity of these systems increase. However, a challenge arises due to
the shifting positions and perspectives of stakeholders, which can change in
response to fluctuations in system boundaries (Fig. 1) [3].
2 Related Work
2.1 Understanding the Content of the Agent Model
In the literature, they have proposed various methods for analyzing and
understanding the contents of agent models and their simulation results. In
previous studies, (1) a methodology for describing simulation results in
natural language [6] (Sect. 2.3), (2) a methodology for describing models in
a description language such as Unified Modeling Language (UML) [7], and
(3) a methodology for describing simulation results in a formal language [8,
9] were proposed. While formal language description has an advantage in
terms of comparability of results, natural language description is easier to
understand in terms of ease of stakeholder feedbacks. The two
methodologies are complementary to each other. In this paper, we focus on
examining method with a LLM described in natural language) since our
goal is to effectively engage with various stakeholders.
Fig. 2 Methodology for grounding agent simulation results with actual business cases [6]
3 Proposed Method
3.1 Outline
In this paper, we propose a method to enhance the utility of describing the
results of agent simulations in natural language to create business cases.
The concrete procedure is as follows (Fig. 3):
4 Experiment
In this chapter, we apply the proposed method to an agent model of
financial chain failures in our previous study. We then generate virtual
business cases that represent the propagation of the crisis among financial
institutions based on the simulation results.
This paper utilizes the GPT-4 API proposed by OpenAI [24] and was run in
September 2023. The hyperparameters were set to temperature = 1.0 or 0.9,
and the maximum length was set to be large enough to ensure uninterrupted
output.
(2) Prompts addressed in this paper
Two examples of outputs for the inputs described in (2) are depicted in
Fig. 8a, b. The outputs are displayed in correspondence with the results of
the simulation log analysis described in Sect. 4.3. The simulation settings
and the results/facts of the simulation logs are presented in the figures and
are indicated by <italics>. Appropriate assumptions and conclusions based
on the log facts are <underlined>, while items that cannot be directly read
from the logs or assumptions are <shaded>.
Fig. 8 a Example of business case output. b Examples of business case output and related questions
and answers (questions and answers are extracted and aggregated from multiple output results)
Both virtual cases generally describe the facts of the logs accurately; for
example, from the second paragraph to the fifth paragraph in Fig. 8a and
from the second paragraph to the fourth paragraph in Fig. 8b. Furthermore,
appropriate conclusions and summaries are generated under proper
inferences; for instance, in the final paragraph in Fig. 8a and in the first and
fifth paragraphs in Fig. 8b.
In contrast, the logs also describe matters and inferences that are not
directly readable from the logs themselves. These are the statements in the
last paragraph of Fig. 8a and “Question 3” in Fig. 8b regarding the bank’s
risk management and the role of the supervisory authority. As described in
Sect. 4.1, the supervisory authority is not included in this model or regarded
as an agent in the simulations in this paper. They are also absent from the
text provided as prompts. This point will be further discussed in Sect. 5.2.
5 Discussion
5.1 Potential Utilization in the Case Method
The results in Sect. 4.4 (3) not only describe the contents of the simulation
logs in a factual manner, but also generate conclusions and questions and
answers based on appropriate assumptions. Particularly, in the case depicted
in Fig. 8b, the case description in the first part is followed by a brief
summary and related questions and answers. This case study can be used as
a review document or as a starting point for facilitation in the case method
[12–14] described in Sect. 2.2.
Fig. 9 Methodology for amalgamate agent simulation and gaming simulation [3]
6 Concluding Remarks
In this paper, we propose a new application of a LLM for describing
simulation results in natural language and grounding them in actual
business cases. Specifically, (1) simulation results are comprehensively
analyzed and systematically classified, and (2) the classified results are
written down in natural language using a large language model. From the
results, we state that the answer of the title is ‘YES’. The results indicate
that it is possible to identify “possible” logs based on an overview of the
overall structure of the simulation trials. We also discovered that could-be
and plausible business cases using a LLM without the templates in our
previous studies.
The contents of the generated cases were generally consistent with that
of the simulation results. In addition to various summaries and conclusions,
simulation-based questions and answers were also understandable. They
could also be used as reference information when upgrading the simulation
model.
In the future, we would like to consider a framework that further
integrates and deepens both the constitutive and participatory approaches
described in Sect. 5.3, fostering a sense of conviction among various
stakeholders in socio-technical systems and facilitating the acquisition of
their approval.
Acknowledgements
The authors would like to thank Enago (www.enago.jp) for the English
language proofing. The authors also would like to thank DeepL and
ChatGPT for the English translation.
Notes All claims expressed in this article are solely those of the authors
and do not necessarily represent those of their affiliated organizations, or
those of the publisher, the editors, and the reviewers.
References
1. Hollnagel E, Woods DD, Leveson N (2006) Resilience engineering: concepts and precepts. CRC
Press
2.
Abbas R, Michael K (2023) Socio-technical theory: a review. In: Papagiannidis S (eds) Theory
hub book, http://open.ncl.ac.uk/. ISBN: 9781739604400
3.
Kunigami M, Terano T (2021) Amalgamating agent and gaming simulation to understand social-
technical systems. In: Kaneda T, Hamada R, Kumazawa T (eds) Simulation and gaming for
social design. Translational systems sciences, vol 25. Springer, Singapore. https://doi.org/10.
1007/978-981-16-2011-9_11
4.
Terano T (2010) Why agent-based modeling in social system analysis? Oukan (J Transdiscipl
Fed Sci Technol) 4(2):56–62 (in Japanese) https://doi.org/10.11487/trafst.4.2_56
5.
Terano T (2013) State-of-the-art of social system research 1—world and Japanese research—
cutting-edge social simulation works between computer and social sciences. J Soc Instrum
Control Eng 52(7):568–573 (in Japanese). https://doi.org/10.11499/sicejl.52.568
6.
Kobayashi T, Takahashi S, Kunigami M, Yoshikawa A, Terano T (2012) Analyzing
organizational innovation and deviation phenomena by agent based simulation and case design.
In: proceedings of 9th int. conf. innovation & management. Wuhan University of Technology
Press, pp 778–791
7.
Goto Y, Sugimoto A, Takizawa Y, Takahashi S (2014) Methodology for facilitating
understandings of complex agent-based models by gaming. Trans Inst Syst, Control, Inf Eng
27(7):290–298 (2014) (in Japanese)
8.
Kunigami M, Kikuchi T, Terano T (2022) A formal model for the business innovation case
description. J Syst, Cybern Inform 20(1):296–318. https://doi.org/10.54808/JSCI.20.01.296
9.
Kikuchi T, Kunigami M, Terano T (2023) Agent modeling, gaming simulation, and their formal
description. In: Kaihara T, Kita H, Takahashi S, Funabashi M (eds) Innovative systems approach
for facilitating smarter world. Design Science and Innovation, Springer, Singapore. https://doi.
org/10.1007/978-981-19-7776-3_9
10.
Kikuchi T, Kunigami M, Takahashi H, Toriyama M, Terano T (2019) Description of decision
making process in actual business cases and virtual cases from organizational agent model using
managerial decision-making description model. J Inf Process 60(10):1704–1718 (in Japanese).
https://cir.nii.ac.jp/crid/1050282813791883520
11.
George AL, Bennett A (2005) Case studies and theory development in the social sciences. In:
BCSIA studies in international security, Cambridge, Mass., MIT Press
12.
Barnes LB, Christensen CR, Hansen AJ (1994) Teaching and the case method: text, cases, and
readings. Harvard Business Review Press
13.
Takeuchi S, Takagi H (2010) Introduction to case method teaching. Keio University Press (in
Japanese)
14.
Gill TG (2011) Informing with the case method: a guide to case method research. Informing
Science Press, Writing and Facilitation
15.
Tanaka Y, Kunigami M, Terano T (2018) What can be learned from the systematic analysis of the
log cluster of agent simulation. Stud Simul Gaming 27(1):31–41 (in Japanese). https://doi.org/10.
32165/jasag.27.1_31
16.
Kikuchi T, Kunigami M, Takahashi H, Toriyama M, Terano T (2020) Classification of simulation
results for formal description of business case. J Jpn Soc Manag Inf Res Note 29(3):199–214
(2020) (in Japanese). https://doi.org/10.11497/jjasmin.29.3_199
17.
Goto Y (2021) Hierarchical classification and visualization method of social simulation logs
reflecting multiple analytical interests. Trans Soc Instrum Control Eng 56(10):463–474 (in
Japanese)
18.
Chung JJY, Kim W, Yoo KM, Lee H, Adar E, Chang M (2022) TaleBrush: sketching stories with
generative pretrained language models. In: Proceedings of the 2022 CHI conference on human
factors in computing systems (CHI ’22) (2022). https://doi.org/10.1145/3491102.3501819
19.
Yuan A, Coenen A, Reif E, Ippolito D (2022) Wordcraft: story writing with large language
models. In: 27th international conference on intelligent user interfaces, ACM, pp 841–852
20.
Kikuchi T, Kunigami M, Yamada T, Takahashi H, Terano T (2016) Agent-based simulation
analysis on investment behaviors of financial firms related to bankruptcy propagations and
financial regulations. transactions of the Japanese Society for Artificial Intelligence 31(6), AG-
G_1-11 (2016) (in Japanese). https://doi.org/10.1527/tjsai.AG-G
21.
Kikuchi T, Kunigami M, Yamada T, Takahashi H, Terano T (2016) Analysis of the influences of
central bank financing on operative collapses of financial institutions using agent-based
simulation. In: Proceedings of the 40th annual computer software and applications conference
(COMPSAC), 2, The Institute of Electrical and Electronics Engineers, Inc. (IEEE), pp 95–104.
https://ieeexplore.ieee.org/document/7552186
22.
Alaeddini M, Madiès P, Reaidy PJ, Dugdale J (2022) Interbank money market concerns and
actors’ strategies – a systematic review of 21 century literature. J Econ Surv:1–81. https://doi.
org/10.1111/joes.12495
23.
Tanaka Y, Kikuchi T, Kunigami M, Yamada T, Takahashi H, Terano T (2017) Classification of
simulation results using log clusters in agent simulation. JSAI Special Interest Group on
Business Informatics (SIG-BI #7) (in Japanese)
24.
OpenAI, GPT-4 Technical Report (2023) https://arxiv.org/pdf/2303.08774
25.
Wolfram S (2023) What is ChatGPT doing and why does it work? Wolfram Media Inc
26.
Ji Z, Lee N, Frieske R, Yu T, Su D, Xu Y, Ishii E, Bang YJ, Madotto A, Fung P (2023) Survey of
hallucination in natural language generation. ACM Comput Surv 55(12):1–38
27.
Axelrod R (1997) The complexity of cooperation agent-based models of competition and
collaboration. Princeton University Press
28.
Terano T (2013) Beyond KISS principle: agent-based modeling toward complex adaptive
systems. J Jpn Soc Artif Intell (Spec Issue: Complex Syst Collect Intell) 18(6):710–715 (in
Japanese)
29.
Kikuchi T, Kunigami M, Yamada T, Takahashi H, Terano T (2018) Agent-based simulation of
financial institution investment strategy under easing monetary policy for operative collapses. J
Adv Comput Intell Intell Inform 22(7):1026–1036 (in Japanese). https://doi.org/10.20965/jaciii.
2018.p1026
OceanofPDF.com
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
R. Lee (ed.), Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed
Computing , Studies in Computational Intelligence 1153
https://doi.org/10.1007/978-3-031-56388-1_12
Haruaki Tamada
Email: [email protected]
Abstract
The Git platforms, such as GitHub, are big data providers that store various
development histories. Then, many researchers analyze data from such
platforms from various aspects. Recently, AI-based systems have been used
to construct themselves. However, there are no studies to measure the
quality of the projects, and the ideal images of the projects need to be
defined. This paper aims to find the ideal images of the projects in OSS
(Open Source Software). For this, we extract the time-series project metrics
from the famous OSS projects to categorize for detecting the patterns of
projects’ growth. Our approach tries to explain the patterns and to give the
decision for the patterns, good or not. The time-series metrics from projects
include the number of stargazers, forks, commits, etc. The number of
stargazers should increase as time passes, and the number of forks tends to
decrease. The stargazer pattern indicates that many developers watch the
repository since it was managed well. We conducted the case study to
analyze the time series data from 10 repositories in GitHub. As a result, we
found that the transitions of the number of issues typically form the
sawtooth wave. The sawtooth wave pattern suggests that new issues are
reported continuously and inventoried periodically. Therefore, the projects
with the patterns indicate they are managing well, and many developers pay
attention.
1 Introduction
Recently, there are widly spreaded to distributed pull-based developments
with the platform like GitHub [9]. Many project managers struggle to
manage their projects based on agile practice [8] and scrum.1 Unfortunately,
no best practices for project management have been established.
Our research group proposed a method to characterize projects based on
the time-series data of the project metrics [12, 13]. Also, we introduced the
concept of Project as a City (PaaC) by aligning the obtained features with
the smart city. In PaaC, we assume that the current state of the city (project)
is measured, the future state of the city is defined, and actions are taken to
fill the difference to approach the ideal image. Then, we developed a tool to
analyze the time-series data of the project metrics based on the given
GraphQL [5]. However, the tool, named Argo, is just fetch the specified
time-series data of a repository in GitHub through GitHub GraphQL API.2
It is not enough to find the ideal image of the project.
This paper aims to find the ideal image of the project by analyzing the
time-series data of the project metrics. If the ideal image is clear, the
developers can work toward the goal, evaluate the progress, and corrects the
way of the project goes. For this, we find and categorize the patterns of the
time-series data from the projects. Then, we give the decision for the
patterns to be the best practice or anti-pattern from the professionals’ view.
This paper organizes as follows. Section 2 describes the proposed
method. Next, Sect. 3 shows case studies of our proposed method to find
the ideal image of the projects. Then, we introduce related works in Sect. 5.
Finally, we summarize the paper and discuss future work in Sect. 6.
2 Proposed Method
2.1 Time-Series Data
The purpose of this paper is to find the ideal image of the projects from the
projects’ time-series data. First, we extract time-series metrics from
projects. Then, we depicts the line-chart of the metrics to identify the
characteristics of the data. Next, professionals invited us investigated the
charts to give the description for the patterns. Finally, we and professional
try to find the ideal image of the projects from the patterns through the
discussion.
The ideal image of the projects are generally different by projects’ scale,
characteristics, developers and some aspects. Therefore, we assume that the
ideal image is not unique. However, if the proposed method find the similar
patterns from the most projects, the patterns might be effective for the
projects. Then, the similar patterns may become the ideal image of the
projects.
We use the time-series data from the projects to identify the patterns.
The time-series data reflect the developers’ behaviors in the projects.
Therefore, we obtain the certain time-series data of the project repository.
The example of time-series data is the number of stargazers, forks,
commits, issues, and pull requests. Also, we do not use qualitative data such
as the source codes, the content of issues and pull requests. Those
qualitative data also reflect the project behaviors, however, it is difficult to
identify the patterns. Hence, our approach employes only the quantitative
data and their transitions.
2.2 Analysis
To analyze the projects p from the obtained time-series data, we extract the
data in a period and divide so that n. In other words, we compute the
frequencies of the data in the period. Then, we convert the time-series data
from certain method f into . The datum
represents a sum of metric value of a part of period. Now, a moving average
is computed for each of the m items from to obtain the
.
Besides is a function to calculate the average value of the given data.
Next, we extract 3-grams from and obtain
. By
focusing , we categorize the patterns into the following five types:
Rising: .
Falling: .
V-shaped:
Inverted V-shaped:
Flatten:
Figure 1 depicts above patterns. Dots in each pattern shows
. Vertical locations of each dot indicates the
values of . Note that, the values in Fig. 1 are just for examples. By
applying strictly the above categorizing rules, since the rules of flatten is
quite sensitive, no patterns would be categoized into it. Therefore, we
introduce the buffer for accepting gaps to the rules of flatten as
.
2.3.1 Rising
This pattern indicates that some developers post new issues to the project.
The most of issues were posted, the metrics values sharply rise.
2.3.2 Falling
If the metrics values acts this falling pattern, we inferred the developers in
the project tried to solve the issues. For the popular project, many
developers and users posts issues one after another. Therefore, it is
important to clean up unnecessary issues, repeatedly. At this time, the
metrics values forms a sharp falling pattern.
2.3.3 ZigZag
This pattern indicates that the metrics values repeat the V-shaped and
inverted V-shaped patterns. That pattern was guessed the project works
well. Since some problems were solved for by the efforts of the project
team, and the product was used to enough to find the new problems.
2.3.4 Flatten
This pattern indicates that the metrics values are almost constant and do not
change. The causes of this pattern are two reasons. One is that the project
activities are stagnated. The other is that the number of posted and closed
are almost equals, unexpectedly. Therefore, we should consider the commit
counts and opened/closed issue counts in the period of this pattern in order
to distinguish the reasons.
Fig. 2 The chart of remained issue count per month of babylonjs/babylon.js
3 Case Studies
We demonstrate the proposed method through three case studies. The first
case study is the preliminary study to understand the four patterns. In this
study, we chose one project and illustrated the graph of the the issue count
transitions. The second case study is to identify the typical transition
patterns from the actual OSS projects. We investigated the issue count
transitions of 514 repositories and find the ratio of each pattern. The third
case study is the evaluation of the four patterns from 514 repositories.
Note that we ran Argo for fetching data from 514 repositories in GitHub
on iMac Retina 4K, Intel Core i5 3GHz 6-core 2019 with macOS Ventura
13.4.1, 16GB RAM. Moreover, the total time and data length were
summarized in Table 2.
Table 2 Fetched data size and fetching time
From the Fig. 3a, the project has started at 2013 in Flatten pattern and
the pattern has changed to Rising from Oct. 2017. Then, it changed to
ZigZag pattern from Jan. 2018 via Falling pattern after that.
Similarly, Fig. 3b indicates that the project has started in ZigZag and
Rising patterns. After that, the patterns has changed to Falling and ZigZag
patterns until Apl. 2018. Finally, the patterns in the project turned to Flatten
pattern. From Fig. 3a and b, what this means is that a project may be active
or stagnant at different times.
The repositories golang/go and apple/swift are both prominant
projects. Figure 3c and d are the charts from both project. From figures,
although golang/go has one great falling, both project increase gradually
with ZigZag patterns. The repository apple/swift looks like to resolve
unimportant issues at the beginning of the year, a slightly large falling
appeared on that time.
4 Discussion
4.1 Ideal Pattern
This section discusses the ideal patterns of issues in OSS projects based on
the results of case studies. The ideal patterns of issues in OSS projects are
considered to be different depending on the project’s situation and goals.
The huge project which attracts the attentions from the begining such as
microsoft/vscode acts the sawtooth pattern of large amplitude and the
Rising pattern. Figure 5 shows the issue count transition of
microsoft/vscode. Actually, projects with more stars tend to have
more issues, since the correlation value was calculated as
between the latest active issue count and latest stargazer count ( ).
The projects with a high number of stargazers indicate that many
developers pay attention to them. Thus, such projects receives a lot of
feedback and issue reports. The projects are considered to be stably
growing, and regularly resolving issues and pull requests.
5 Related Works
Gunnar et al. surveyed the transitions of code velocity from activities of
issues for famous repositories [6]. Their work found the code velocity was
not slowed down in the target repositories. It encourages our work since
their results might become an anti-pattern by the falling code velocity for
some reason.
Niranjan presented a tool, GitRank, to rank the repositories in GitHub
using the quality of source codes, popularity, and maintainability [4]. Their
work and ours differ in the exploited metrics and the purpose of the results.
Our work focuses on repository management, therefore, did not care on
source code quality. The developer will use both results for more effective
repository management and development.
There are many other works mining the repositories in GitHub.
Steinmacher et al. conducted the quantitative and qualitative analysis of
pull-request from non-core members [10]. Constantiou et al. proposed the
method for presenting the summary of the Ruby projects on the timeline by
the number of commits, lines, and projects [2]. Pinto et al. analyzed the
behavior of contributors caused by the development language in GitHub
[7]. Borges et al. analyzed the characteristics of the projects that earned
many stars in GitHub [1]. Tian et al. tried to find developers who have geek
talent from activities across GitHub and Stackoverflow [11]. Our work is
related to the above works analyzing the projects in GitHub. However, we
focus on the factors contributing to the spontaneous software evolution for
the ideal project management, and introduce on the new concept of the
project as a city.
6 Conclusion
The goal of this paper is to investigate the ideal management method for the
project based on the histories from OSS projects hosted on GitHub. For
this, we focus on the number of issues in the project. We collected the issue
count transitions of 514 active repositories and classified the transition
patterns. As a result, the famous projects showed a large amplitude
sawtooth pattern. The projects started with a few developers act a small
amplitude sawtooth pattern. However, the amplitude of the sawtooth pattern
gradually increased as the project gained attention.
The following tasks remain for our future works. The first is to analyze
more repositories. Specifically, our analysis could benefit from an increased
sample size, encompassing a broader range of repositories. The second is to
analyze other metrics, for example issue comments, pull requests, etc. The
third is to clarify the specific actions of developers as best practices and
anti-patterns. Lastly, delving deeper into the underlying behaviors of
contributors to identify specific actions that could be categorized as best
practices or anti-patterns would offer actionable insights for effective
project management.
References
1. Borges H, Hora A, Valente MT (2016) Understanding the factors that impact the popularity of
github repositories. In: 2016 IEEE international conference on software maintenance and
evolution (ICSME), pp 334–344. https://doi.org/10.1109/ICSME.2016.31
2.
Constantinou E, Mens T (2017) Socio-technical evolution of the ruby ecosystem in github. In:
2017 IEEE 24th international conference on software analysis, evolution and reengineering
(SANER), pp 34–44 (2017). https://doi.org/10.1109/SANER.2017.7884607
3.
Dabic O, Aghajani E, Bavota G (2021) Sampling projects in github for msr studies. In:
Proceedings of the 2021 IEEE/ACM 18th international conference on mining software
repositories (MSR), pp 560–564. https://doi.org/10.1109/MSR52588.2021.00074
4.
Hasabnis N (2022) GitRank: A framework to rank github repositories. In: Proceedings of the
2022 IEEE/ACM 19th international conference on mining software repositories (MSR), pp 729–
731. https://doi.org/10.1145/3524842.3528519
5.
Kaide K, Tamada H (2022) Argo: projects’ time-series data fetching and visualizing tool for
github. In: Proceedings of the 23rd ACIS international conference on software engineering,
artificial intelligence, networking and parallel/distributed computing (SNPD 2022 Summer)
6.
Kudrjavets G, Nagappan N, Rastogi A (2023) Are we speeding up or slowing down? on temporal
aspects of code velocity. In: Proceedings of the 2023 IEEE/ACM 20th international conference
on mining software repositories (MSR), pp 267–271. https://doi.org/10.1109/MSR59073.2023.
00046
7.
Pinto G, Steinmacher I, Gerosa MA (2016) More common than you think: an in-depth study of
casual contributors. In: 2016 IEEE 23rd international conference on software analysis, evolution,
and reengineering (SANER), pp 112–123. https://doi.org/10.1109/SANER.2016.68
8.
Project Management Institute: Agile Practice Guide (2017) Project management institute
9.
Saito Y, Fujiwara K, Igaki H, Yoshida N, Iida H (2016) How do github users feel with pull-based
development? In: 2016 7th international workshop on empirical software engineering in practice
(IWESEP), pp 7–11. https://doi.org/10.1109/IWESEP.2016.19
10.
Steinmacher I, Pinto G, Wiese IS, Gerosa MA (2018) Almost there: a study on quasi-contributors
in open-source software projects. In: 2018 IEEE/ACM 40th international conference on software
engineering (ICSE), pp 256–266. https://doi.org/10.1145/3180155.3180208
11.
Tian Y, Ng W, Cao J, McIntosh S (2019) Geek talents: Who are the top experts on github and
stack overflow? Comput Mater Continua 61(2):465–479
[Crossref]
12.
Toda K, Tamada H, Nakamura M, Matsumoto K (2019) Characterizing project evolution on a
social coding platform. In: Proceeding 20th ACIS international conference on software
engineering, artificial intelligence, networking and parallel/distributed computing (SNPD 2019),
pp 525–532
13.
Toda K, Tamada H, Nakamura M, Matsumoto K (2020) Capturing spotaneous software
evolution in a social coding platform with project-a s-a-city concept. Int J Softw Innov 8(3):35–
50
[Crossref]
Footnotes
1 https://www.scrum.org/resources/what-scrum-module.
2 https://docs.github.com/graphql.
3 https://seart-ghs.si.usi.ch.
4 https://github.com/tamadalab/argo.
OceanofPDF.com