Intelligent Computing and Internet of Things
Intelligent Computing and Internet of Things
Intelligent Computing
and Internet of Things
First International Conference on Intelligent Manufacturing
and Internet of Things and 5th International Conference on Computing
for Sustainable Energy and Environment, IMIOT and ICSEE 2018
Chongqing, China, September 21–23, 2018
Proceedings, Part II
123
Communications
in Computer and Information Science 924
Commenced Publication in 2007
Founding and Former Series Editors:
Phoebe Chen, Alfredo Cuzzocrea, Xiaoyong Du, Orhun Kara, Ting Liu,
Dominik Ślęzak, and Xiaokang Yang
Editorial Board
Simone Diniz Junqueira Barbosa
Pontifical Catholic University of Rio de Janeiro (PUC-Rio),
Rio de Janeiro, Brazil
Joaquim Filipe
Polytechnic Institute of Setúbal, Setúbal, Portugal
Igor Kotenko
St. Petersburg Institute for Informatics and Automation of the Russian
Academy of Sciences, St. Petersburg, Russia
Krishna M. Sivalingam
Indian Institute of Technology Madras, Chennai, India
Takashi Washio
Osaka University, Osaka, Japan
Junsong Yuan
University at Buffalo, The State University of New York, Buffalo, USA
Lizhu Zhou
Tsinghua University, Beijing, China
More information about this series at http://www.springer.com/series/7899
Kang Li Minrui Fei
•
Intelligent Computing
and Internet of Things
First International Conference on Intelligent Manufacturing
and Internet of Things and 5th International Conference on Computing
for Sustainable Energy and Environment, IMIOT and ICSEE 2018
Chongqing, China, September 21–23, 2018
Proceedings, Part II
123
Editors
Kang Li Zhile Yang
The University of Leeds Shenzhen Institute of Advanced Technology
Leeds Chinese Academy of Sciences
UK Shenzhen
China
Minrui Fei
Shanghai University Dongsheng Yang
Shanghai Northeastern University
China Shenyang
China
Dajun Du
Shanghai University
Shanghai
China
This Springer imprint is published by the registered company Springer Nature Singapore Pte Ltd.
The registered company address is: 152 Beach Road, #21-01/04 Gateway East, Singapore 189721,
Singapore
Preface
This book constitutes the proceedings of the 2018 International Conference on Intel-
ligent Manufacturing and Internet of Things (IMIOT 2018) and International Confer-
ence on Intelligent Computing for Sustainable Energy and Environment (ICSEE 2018),
which were held during September 21–23, in Chongqing, China. These two interna-
tional conference series aim to bring together international researchers and practitioners
in the fields of advanced methods for intelligent manufacturing and Internet of Things
as well as advanced theory and methodologies of intelligent computing and their
engineering applications in sustainable energy and environment. The new conference
series IMIOT is jointly organized with the well-established ICSEE conference series,
under the auspices of the newly formed UK-China University Consortium in Engi-
neering Education and Research, with an initial focus on intelligent manufacturing and
sustainable energy.
At IMIOT 2018 and ICSEE 2018, technical exchanges within the research com-
munity took the form of keynote speeches, panel discussions, as well as oral and poster
presentations. In particular, two workshops series, namely the Workshop on Smart
Energy Systems and Electric Vehicles and the Workshop on Communication and
Control for Distributed Networked Systems, were held again in parallel with
IMIOT 2018 and ICSEE 2018, focusing on the two recent hot topics of the integration
of electric vehicles with the smart grid, and distributed networked systems for the
Internet of Things.
The IMIOT 2018 and ICSEE 2018 conferences received 386 submissions from over
50 different universities, research institutions, and companies from both China and UK.
All papers went through a rigorous peer review procedure and each paper received at
least three review reports. Based on the review reports, the Program Committee finally
selected 135 high-quality papers for presentation at the IMIOT 2018 and ICSEE 2018.
These papers cover 22 topics and are included in three volumes of the CCIS series,
published by Springer. This volume of CCIS includes 52 papers covering 9 relevant
topics.
Located at the upstream Yangtze basin, Chongqing constitutes the most important
metropolitan area in the southwest of China. It has a glorious history and culture and
serves as a major manufacturing center and transportation hub. Chongqing is also
well-known for its spicy food and hotpot, attracting tourists and gourmets from around
the world. In addition to academic exchanges, participants were treated to a series of
social events, including receptions and networking sessions, which served to build new
connections, foster friendships, and forge collaborations. The organizers of
IMIOT 2018 and ICSEE 2018 would like to acknowledge the enormous contribution
of the Advisory Committee, who provided guidance and advice, the Program Com-
mittee and the numerous referees for their efforts in reviewing and soliciting the papers,
and the Publication Committee for their editorial work. We would also like to thank the
editorial team from Springer for their support and guidance. Particular thanks are of
VI Preface
course due to all the authors, as without their high-quality submissions and presenta-
tions the conferences would not have been successful.
Finally, we would like to express our gratitude to our sponsors and organizers, listed
on the following pages.
Honorary Chairs
Fusheng Pan Chongqing Science and Technology Society/Chongqing
University, China
Shilong Wang Chongqing University, China
Mark Price Queen’s University Belfast, UK
General Chairs
Ming Kim Lim Chongqing University, China
Kang Li Queen’s University Belfast, UK
Local Chairs
Xuda Qin Tianjin University, China
Fuji Wang Dalian University of Technology, China
Yingguang Li Nanjing University of Aeronautics and Astronautics, China
Adam Clare University of Nottingham, UK
Weidong Chen Shanghai Jiaotong University, China
Rui Xiao Southeast University, China
Furong Li Bath University, UK
Min-Sen Chiu National University of Singapore, Singapore
Petros Aristidou University of Leeds, UK
Jinliang Ding Northeastern University, China
Bing Liu University of Birmingham, UK
Shan Gao Southeast University, China
Mingcong Deng Tokyo University of Agriculture and Technology, Japan
Zhengtao Ding The University of Manchester, UK
Shiji Song Tsinghua University, China
Donglian Qi Zhejiang University, China
Wanquan Liu Curtin University, Australia
Patrick Luk Cranfield University, UK
Guido Maione Technical University of Bari, Italy
Chen Peng Shanghai University, China
Tong Sun City University London, UK
Yuchu Tian Queensland University of Technology, Australia
Xiaojun Zeng The University of Manchester, UK
Huaguang Zhang Northeastern University, China
Shumei Cui Harbin Institute of Technology, China
Hongjie Jia Tianjin University, China
Youmin Zhang Concordia University, USA
Organization IX
Organization Committee
Chairs
Congbo Li Chongqing University, China
Minyou Chen Chongqing University, China
Adrian Murphy Queen’s University Belfast, UK
Sean McLoone Queen’s University Belfast, UK
Publication Chairs
Zhile Yang Chinese Academy of Sciences, China
Jianhua Zhang North China Electric Power University, China
Hongjian Sun Durham University, UK
Trevor Robinson Queen’s University Belfast, UK
Publicity Chairs
Qingxuan Gao Chongqing University, China
Junjie Chen Southeast University, China
Brian Falzon Queen’s University Belfast, UK
Ben Chong University of Leeds, UK
Secretary-General
Yan Ran Chongqing University, China
Dajun Du Shanghai University, China
Rao Fu Queen’s University Belfast, UK
Yanxia Wang Queen’s University Belfast, UK
Registration Chairs
Guijian Xiao Chongqing University, China
Shaojun Gan Queen’s University Belfast, UK
X Organization
Sponsors
Organizers
Co-organizers
The Electric Field Analysis and Test Experiments of Split Type Insulator
Detection Robot . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 205
Pengxiang Yin, Xiao Hong, Lei Zheng, Biwu Yan, and Hao Luo
Contents – Part II XIX
IoT Systems
Image Processing
1 Introduction
The multi-objective flexible shop scheduling problems (MOFJSP) evolved from flexible
shop scheduling problems (FJSP) which has been proven to be NP-hard problem [1]. It
can comprehensively consider multiple targets and better meet modern production
requirements, but it is more complicated combinatorial optimization problem than FJSP
because of its increased machine selection, extended search scope as well as multiple
choices of optimization objectives. Therefore, this issue has attracted more and more
attention from academia and industry.
Two optimization approaches are ususlly developed for multi-objective flexible job
shop scheduling problem. The first method is weighted sum approach, namely, the multi-
objective optimization problems is transformed into a single objective optimization
problem by assigning different weights to each objective, and then solved by some meta-
heuristic approaches. Neverthless, this method only obtains a single solution, thus it is
difficult to find Pareto-optimal solutions for decision maker’s choice. The second
approach is called Pareto-based multi-objective evolutionary algorithms which is
n m
∑ ∑
Min Wmax = max {Wk } = max Pi,j,k Xi,j,k (2)
1≤k≤m 1≤k≤m
i=1 j=1
n ni m
∑ ∑ ∑
Min Lmax = Pi,j,k Xi,j,k (3)
i=1 j=1 k=1
Where Tmax is the maximum completion time of all jobs; Wmax is the maximum
workload of all machines; Lmax is the total processing time of all jobs.
Parameter initialization
Initialize phase
N
Is the stop criteria satisfied?
Y
Output Pareto non-inferior solutions
The SFLA is a novel meta-heuristic algorithm, which makes use of the cultural envi‐
ronment information named as “meme” among the population to perform its evolution.
The searching process is performed using the instruction of memes generated by the
cooperation and competition among the population. In addition, the redistribution of the
sub-population makes searching process less possible to be trapped in local minima. The
traditional meme evolution operator are shown below.
( )
ΔX = r and () ∗ Xg − Xw (6)
( )
newXw = Xw + ΔX ΔXmax ≥ ΔX ≥ ΔXmin (7)
Where, in Eq. (4) ΔX denotes the moving step of frog individual, rand() represents
the random number uniformly distributed between 0 and 1, Xb and Xw represents the
best and worst frog in each meme group respectively, Eq. (5) denotes the update of the
location of the worst frog individuals in the meme group, if the frog could produce a
better position after the above update, then replace the original position with a new
position, otherwise replace Xg with Xb, and update with (6) (7) formula above. A new
solution is randomly generated instead of the worst frog if the above method still fails
to produce better frog individuals.
In this paper, the fly strategy of particles in PSO is embedded into the local update
operator of SFLA to guide the evolution of frog individuals in the meme group, the local
position update operator is adjusted as follows,
( k ) ( )
Vk+1
id
= wV k
id
+ c 1 r1 Pid
− Xk
id
+ c2 r2 Pk
gd
− Xk
gd (8)
k+1
Xid = Xkid + Vk+1
id (9)
( )
Vk+1
id
= wVkid + c3 r3 Pkgd − Xkid (10)
k+1
Xid = Xkid + Vk+1
id (11)
Where, in Eqs. (8), (9), (10) and (11) the parameters such as c1 , c2 , c3 , r1 , r2 , r3 and w
are the same meaning as those in the standard PSO algorithm. But the difference is that
Xkid represents the worse frog individual in the meme group, Pkid and Pkgd denotes the
optimal position of the meme group and all frog individuals respectively. According to
Eqs. (8) and (9),the frog individual’ speed and displacement is updated, The objective
function (Objv’) of the updated individual is calculated and compared with objective
function (Objv) of frog individual before update in Pareto dominance relation,
(a) if Objv’≻ Objv, then the speed and displacement of the individual is replaced with
the updated frog individual’s.
(b) if Objv ≻ Objv’, then update the speed and displacement according to Eqs. (10) and
(11), a new individual is randomly generated instead of the original individual if
the frog is still dominated by the original individual.
(c) if Objv’∼ Objv, then accept the speed and displacement of the updated individual
at a probability of 50%.
The Pareto relation can be described as followed, ( ) for multi-objective optimization
problem, min[fi (x), i = 1, 2, ⋯ , n], if fi (Xa ) ≤ fi Xb and a strict inequality exists at
least, then Xb is dominated by Xa, symbolized as Xa ≻ Xb, if the two vectors do not
dominate each other, it can be symbolized as Xa ∼ Xb.
8 M. Gou et al.
The outside elite documents stored non-dominated individuals in the iteration which
overcome the defects of NSGA II algorithm without using external elite population to
preserve the historical non-dominated individuals which caused some better solutions
to be lost, falling into premature convergence. However, the external elite document is
only used to store the non- dominated individuals searched in the iteration, and these
individuals do not react to the evolutionary population, resulting in reducing the conver‐
gence speed of the algorithm and decreasing the diversity of Pareto solutions. Therefore,
this paper proposes the local optimization strategy including selection strategy and local
optimization to optimize the external Pareto solution set. The selection strategy decides
how to choose the important search area to avoid blind search and the local optimization
is to optimize the important search regions. The selection strategy is as follows.
(1) Strategy 1, In order to avoid the repeated search for a region, a non-dominated
individuals on the current stored in the external elite document is designated the
largest number of local search, once more than the search times, the search area is
discarded.
(2) Strategy 2, A sphere radius (R) is specified in advance, aiming at guiding the search
toward promising regions. For non-dominated individuals that do not reach the
search upper limit (L), the number of non-dominated individuals within the radius
Improved Shuffled Frog Leaping Algorithm 9
of the hyper-sphere is counted as the center, the smaller amount means the less non-
dominated individuals around it, indicating the new non-dominated individuals are
discovered with the high probability. If the minimum number exceeds one, the
average euclidean distance is calculated within the radius of the hyper-sphere, The
larger distance means the more sparse distribution of solutions, resulting in more
likely to find new non-dominated solutions. The local optimization algorithm for
external elite documents is given in Fig. 2.
the descending order based on crowding-distance after the first 20% as a global optimal
solution to guide the meme within the group of the worst individual updates.
Xm[ L] 3 4 2 4 3 3 2 2 1
encoding
Xp[ L] 1 3 2 2 1 3 3 1 2
Xp[ L] O11 O31 O21 O22 O12 O32 O33 O13 O23
decoding
Xm[ L] M3 M4 M2 M4 M3 M3 M2 M2 M1
Where Oij represents the j-th operation of the i-th job, when decoding, the procedure
firstly reads the operation coding successively and then randomly select a machine in a
given optional machine table of operation. The operations processed on each machine
after decoding are {M1 : O23 }, {M2 : O21 , O33 , O13 }, {M3 : O11 , O12 , O32 }, {M4 : O31 , O22 }.
The velocity vector of the worst individual in the meme group is updated according to
the Eqs. (8) and (10), and the position vector is updated according to the Eqs. (9) and
(11). Since FJSP belongs to integer encoding, there may be a decimal fraction after
updating, so it is necessary to adjust the infeasible solution. The feasible operation is
obtained according to the reordered new displacement in ascending order. The adjust‐
ment results are shown in the following Table 2. At the same time, the process also has
machine constraint, hence there is a need to reselect the machine after the adjustment
operation, according to the procedure to find the process of optional machine set in the
optional machine table, and then randomly select a processing machine.
optional machine set. The objective function is calculated. If the mutated individual
dominates the original one, the original individual is replaced.
5 Computational Results
5.1 Problem 8 × 8
5.2 Problem 10 × 10
This is an instance of 10 × 10 with 30 operations taken from Kacem et al. [12]. The
comparison of the SFLA-PSO algorithm with others is showed in Table 5, in which Si
denote i th solution. The comparison result shows that the proposed SFLA-PSO algo‐
rithm obtains more and better Pareto solutions than other algorithms.
5.3 Problem 15 × 10
This is an instance of 15 × 10 with 56 operations taken from Du et al. (2008). The
comparison of the SFLA-PSO algorithm with others is showed in Table 6, in which Si
denote i th solution. The comparison result shows that the proposed SFLA-PSO algo‐
rithm is not bad.
6 Conclusions
This paper studies a multi-objective flexible job shop scheduling problem with the goal
of minimum the maximum completion time, maximum workload of all machines and
the total processing time of all job, and establishes a multi-objective optimization model.
Since it is NP-hard, an improved shuffled frog leaping algorithm were proposed to
heuristically solve them. The key contribution of this paper is that The particle swarm
flight strategy is embedded in the SFLA, and the dynamic crowding density sorting
method is used to update the non-dominant population external elite files, to improve
the performance of algorithm. Moreover,A new local optimization strategy is adopted
to achieve the high development ability of the new algorithm. The comparison results
with other typical algorithms demonstrated the effectiveness of the proposed algorithm.
The improved shuffled frog leaping algorithm proposed in this paper improves the
search performance of the shuffled frog leaping algorithm and extends the applicability
of the algorithm. How to improve the SFLA algorithm in solving more complex prob‐
lems and improve the convergence precision of the algorithm, simulated annealing
algorithm, the chaotic theory and immune algorithm with strong local search ability is
an important research direction in the future of the authors.
References
1. Garey, M.R., Johnson, D.S., Sethi, R.: The complexity of flowshop and jobshop scheduling.
Math. Oper. Res. 1(2), 117–129 (1976)
2. Deb, K.: A fast and elitist multiobjective genetic algorithm. NSGA-II. IEEE Trans. Evol.
Comput. 6(2), 182–197 (2002)
3. Jadaan, O.A., Rajamani, L., Rao, C.R.: Non-dominated ranked genetic algorithm for solving
multi-objective optimization problems, NRGA. J. Theor. Appl. Inf. Technol. 1, 60–67 (2008)
4. Shao, X.Y., Liu, W.Q., Liu, Q.: Hybrid discrete particle swarm optimization for multi-
objective flexible job-shop scheduling problem. Int. J. Adv. Manuf. Technol. 67(9–12), 2885–
2901 (2013)
14 M. Gou et al.
5. Zhong, Y., Yang, H., Mo, R., Sun, H.: Optimization method of flexible job-shop scheduling
problem based on niching and particle swarm optimization algorithms. Comput. Integr.
Manuf. Syst. 12(21), 3231–3238 (2015)
6. Lei, D., Li, M., Wang, L.: A two-phase meta-heuristic for multiobjective flexible job shop
scheduling problem with total energy consumption threshold. IEEE Trans. Cybern. PP(99),
1–13 (2018)
7. Eusuff, M., Lansey, K.: Optimization of water distribution network design using the shuffled
frog leaping algorithm. J. Water Resour. Plann. Manage. 129(2003), 210–225 (2003)
8. Luo, X.H., Yang, Y., Li, X.: Modified shuffled frog-leaping algorithm to solve traveling
salesman problem. J. Commun. 30(7), 130–135 (2009)
9. Ren, W.L., Zhao, C.W.: A localization algorithm based on SFLA and PSO for wireless sensor
network. Inf. Technol. J. 12(3), 502–505 (2012)
10. Xu, Y., Wang, L., Liu, M., Wang, S.Y.: An effective shuffled frog-leaping algorithm for
hybrid flow-shop scheduling with multiprocessor tasks. Int. J. Adv. Manuf. Technol. 68(5–
8), 1529–1537 (2013)
11. Lei, D., Zheng, Y., Guo, X.: A shuffled frog-leaping algorithm for flexible job shop scheduling
with the consideration of energy consumption. Int. J. Prod. Res. 55(11), 3126–3140 (2017)
12. Kacem, I., Hammadi, S., Borne, P.: Approach by localization and multiobjective evolutionary
optimization for flexible job-shop scheduling problems. IEEE Trans. Syst. Man Cybern. Part
C Appl. Rev. 32(1), 1–13 (2002)
13. Xia, W.J., Wu, Z.M.: An effective hybrid optimization approach for multi-objective flexible
job-shop scheduling problems. Comput. Ind. Eng. 48(2005), 409–425 (2005)
14. Zhang, G.H., Shao, X.Y., Li, P.G., Gao, L.: An effective hybrid particle swarm optimization
algorithm for multi-objective flexible job-shop scheduling problem. Comput. Ind. Eng.
56(2009), 1309–1318 (2009)
15. Xing, L.N., Chen, Y.W., Yang, K.W.: Multi-objective flexible job shop schedule, design and
evaluation by simulation modeling. Appl. Soft Comput. 9(2009), 362–376 (2009)
Hysteretic Model of a Rotary
Magnetorheological Damper in Helical Flow
Mode
1 Introduction
The configuration of RMRD in helical flow mode is shown in Fig. 1. It mainly consists
of outer cylinder, inner cylinder, piston, coil and rod. The inner cylinder rotates with
the rod while piston and outer cylinder keep still. Spiral groove in the rod is filled with
fluid. MR fluid flows in duct 1 and duct 2. The details of helical flow mode can be seen
in Ref. [17].
Total torque derived from three ducts can be calculated by Eqs. (1)–(4).
Tmax ¼ T1 þ T2 þ T3 ð1Þ
Z l
DP
T1 ¼ xhdx cos hðrd 0:5rh Þ ð2Þ
0 l
w
T2 ¼ 2pr23 Lg þ 2pr22 la s2 ð3Þ
d
Hysteretic Model of a Rotary Magnetorheological Damper 17
Inner Outer
Rod Coil Piston
cylinder cylinder
w
T3 ¼ 2pr33 Lg þ 2pr32 la s3 ð4Þ
d
here, T1 , T2 and T3 are torque in duct 1, duct 2 and duct 3, respectively. T1 is derived
from the effects of helical flow mode. T2 and T3 are both calculated based on working
principle of shear mode. In our prior work [17], a rotary damper is designed and tested.
In this study, the thickness of duct 2 and duct 3 is changed from 1.5 mm and 2 mm to
1 mm and 1 mm, respectively. Figure 2 shows the comparison of the magnetic field
intensity between the initial damper and novel one. The novel one shows higher
magnetic field intensity than the initial one both under 0.2 A and 2.0 A. The higher
magnetic field intensity is beneficial to improve coulomb damping force.
The components and assembly of the novel damper can be seen in Fig. 3. The
volume of MR fluid in RMRD is 385 106 m3 which is also smaller than the initial
damper. Test apparatus of the RMRD is shown in Fig. 4. Rack-and-pinion is applied to
convert the linear motion to rotational motion. Characteristics of torque-angle and
torque-velocity loops are displayed in Figs. 5, 6 and 7. The torque can be improved by
18 J. Yu et al.
Rotary MR
damper
Rack and
Pinion
1 ebw
z¼ ð5Þ
1 þ ebw
Even though these type curves are enough to capture torque-angle or force-
displacement, the nonlinear hysteretic characteristics cannot be captured accurately. To
describe the hysteresis, hysteresis factor in Eq. (6) is introduced to replace the velocity
term in Eq. (5). To realize the reversibility of model, normalization method is used to
20 J. Yu et al.
separate hysteretic part and current-dependent part. The two parts follow the multi-
plication rule which is shown in Eq. (7).
w ! n1 w þ n2 sgn / € or w ! n1 w þ n2 sgnð/Þ þ n3 /
€ þ n3 / ð6Þ
Td ¼ TI G ð7Þ
€ and / are angular acceleration and angle, respectively. TI and G are current-
here, /
dependent part and hysteretic part, respectively. The maximum torque vs. velocity
under different current is shown in Fig. 9. The maximum torque is mainly controlled by
current while the velocity has little effect on maximum torque. The mathematical
expression of current-dependent part is shown in Eq. (8).
2
þ c3 I
T1 ¼ c 1 e c 2 I þ c4 ð8Þ
X
N 2
J¼ T^t Ts ð12Þ
k¼1
here, N is the number of data. T^t and Ts are the model torque and test data, respectively.
The optimization results can be seen in Table 1.
Experimental and theoretical results are compared in Fig. 10. The hysteresis
characteristics can be captured by the model in the visual. To further evaluate the
accuracy of the novel model, the arctangent model and hyperbolic tangent model are
identified and compared. Their equations are shown in Eqs. (13) and (14). The mean
deviation, root-mean-square error (RMSE) and coefficient of determinations R2 [20] are
set as evaluation coefficients. Values of the arctangent model, the hyperbolic tangent
model and the novel model about mean deviation are 1.6638, 1.6744 and 1.3808,
respectively. The values about RMSE are 96.7687, 96.5282 and 86.1562, respectively.
The lower values of mean absolute error and RMSE show that the novel model has
more accuracy than the other two models. Besides, the values about R2 are 0.9658,
0.9660 and 0.9729, respectively. The higher determination coefficient verifies the
trustable ability of the novel model. The differences between experimental results and
22 J. Yu et al.
theoretical results are also calculated and shown in Fig. 11. It is noted that the novel
model owing better accuracy.
a1 2 a4 a8 sgnð/Þ
Td ¼ /_ þ a tan½a7 ð/_ þ Þ ð13Þ
1 þ a2 e a 3 I p 1 þ ea5 ða6 þ I Þ 1 þ a9 ea10 I
Td ¼ ða1 I þ a2 Þ/_ þ ða3 I þ a4 Þ/ þ a5 I 2 þ a6 I þ a7 tanh½a8 /_ þ a9 sgnð/Þ ð14Þ
(a) Torque vs. angle /0.6 Hz (b) Torque vs. angle /0.8 Hz (c) Torque vs. angle /1.0
Hz
(d) Torque vs. velocity /0.6 Hz (e) Torque vs. velocity /0.8 Hz (f) Torque vs. velocity
/1.0 Hz
4 Conclusion
To capture the hysteretic characteristics of the RMRD in helical flow mode, a novel
model is proposed and analyzed based on the experimental results. The hysteretic
characteristics are captured by a shape function and hysteresis factor. The current-
dependent part and the hysteretic part are separated to ensure the reversibility. The
model are identified by CFPSO algorithm and compared with two typical models. The
mean absolute error, RMSE and coefficient of determinations are introduced to evaluate
accuracy of the models. Results show that the novel model can describe the nonlinear
hysteretic characteristics with feasibility.
Acknowledgments. We would like to thank the authors of the references for their enlighten-
ment. This research is also supported financially by the National Natural Science Foundation of
People’s Republic of China (Project No. 51675063), the Program for New Century Excellent
Talents in University (No. NCET-13-0630) and the State Scholarship Fund of China Scholarship
Council (No. 201706050094). These supports are gratefully acknowledged.
References
1. Wang, D.H., Liao, W.H.: Magnetorheological fluid dampers: a review of parametric
modelling. Smart Mater. Struct. 20, 1–34 (2011)
2. Chen, P., Bai, X.-X., Qian, L.-J., Choi, S.-B.: A new hysteresis model based on force–
displacement characteristics of magnetorheological fluid actuators subjected to squeeze
mode operation. Smart Mater. Struct. 26, 1–10 (2017)
3. Wereley, N.M., Panng, L., Kamath, G.M.: Idealized hysteresis modeling of electrorheolog-
ical and magnetorheological dampers. J. Intell. Mater. Syst. Struct. 9, 642–649 (1998)
4. Li, W.H., Yao, G.Z., Chen, G., Yeo, S.H., Yap, F.F.: Testing and steady state modeling of a
linear MR damper under sinusoidal loading. Smart Mater. Struct. 9, 95–102 (2000)
5. Ismail, M., Ikhouane, F., Rodellar, J.: The hysteresis Bouc-Wen model, a survey. Arch.
Comput. Methods Eng. 16, 161–188 (2009)
6. Dahl, P.R.: Solid friction damping of mechanical vibrations. AIAA J. 14, 1675–1682 (1976)
7. Şahin, İ., Engin, T., Çeşmeci, Ş.: Comparison of some existing parametric models for
magnetorheological fluid dampers. Smart Mater. Struct. 19, 1–11 (2010)
8. Ma, X.Q., Wang, E.R., Rakheja, S., Su, C.Y.: Modeling hysteretic characteristics of MR-
fluid damper and model validation. In: Proceedings of the IEEE Conference on Decision and
Control, vol. 2, pp. 1675–1680 (2002)
9. Choi, S.-B., Lee, S.-K., Park, Y.-P.: A hysteresis model for the field-dependent damping
force of a magnetorheological damper. J. Sound Vib. 2, 375–383 (2001)
10. Yu, J., Dong, X., Zhang, Z.: A novel model of magnetorheological damper with hysteresis
division. Smart Mater. Struct. 26, 1–15 (2017)
11. Chen, P., Bai, X.-X., Qian, L.-J., Choi, S.B.: An approach for hysteresis modeling based on
shape function and memory mechanism. IEEE/ASME Trans. Mech. 23, 1270–1278 (2018)
12. Pawlus, W., Karimi, H.R.: A comparative study of phenomenological models of MR brake
based on neural networks approach. Int. J. Wavelets Multiresolut. Inf. Process. 11, 1–30
(2013)
13. Miah, M.S., Chatzi, E.N., Dertimanis, V.K., Weber, F.: Nonlinear modeling of a rotational
MR damper via an enhanced Bouc-Wen model. Smart Mater. Struct. 24, 1–14 (2015)
24 J. Yu et al.
14. Tse, T., Chang, C.: Shear-mode rotary magnetorheological damper for small-scale structural
control experiments. J. struct. Eng. ASCE 130, 904–911 (2004)
15. Boston, C., Weber, F., Guzzella, L.: Modeling of a disc-type magnetorheological damper.
Smart Mater. Struct. 19, 1–12 (2010)
16. Imaduddin, F., Mazlan, S.A., Zamzuri, H.: A design and modelling review of rotary
magnetorheological damper. Mater. Des. 51, 575–591 (2013)
17. Yu, J., Dong, X., Wang, W.: Prototype and test of a novel rotary magnetorheological damper
based on helical flow. Smart Mater. Struct. 25, 1–15 (2016)
18. Sun, S.S., Ning, D.H., Yang, J., Du, H., Zhang, S.W., Li, W.H.: A seat suspension with a
rotary magnetorheological damper for heavy duty vehicles. Smart Mater. Struct. 25, 1–10
(2016)
19. Yu, J., Dong, X., Zhang, Z., Chen, P.: A novel scissor-type magnetorheological seat
suspension system with self-sustainability. J. Intell. Mater. Syst. Struct. 29, 1–12 (2018)
20. Dong, X., Yu, J., Wang, W., Zhang, Z.: Robust design of magneto-rheological (MR) shock
absorber considering temperature effects. Int. J. Adv. Manuf. Tech. 90, 1735–1747 (2017)
Dynamic Production Scheduling Modeling
and Multi-objective Optimization
for Automobile Mixed-Model Production
1 Introduction
Due to end customers’ diversified demands, more companies have adopted production
by order strategy, resulting in multiple types of products are collinearly produced on
mixed-model production lines. In consequence, wide variety of parts need to be
assembled which results in higher inventory cost. When formulating a production plan,
the amount of parts inventory needs to be known, so production sequence can be
reasonably arranged. Since production flow sequence determines parts usage sequence,
Foundation items: This work was supported in part by the Key Technology Research and System
Integration of Discrete Intelligent Manufacturing Workshop, China (No. cstc2016zdcy-
ztzx60001), the Fundamental Research Funds for the Central Universities of China under Grant
2018CDXYJX0019, and the National Nature Science Foundation of China under Grant
51805053 as well as Grant 51575069.
optimizing production queue can help to balanced parts consumption ratio. The pur-
pose of equalization is to reduce fluctuations in consumption rate. This not only
increases the robustness of production line, it is also possible to predict the con-
sumption rate of parts, which is the key to keeping inventory at a low level. Through
the strategy of milk run [1], stocks will eventually reach a dynamic balance, and it will
ultimately achieve the goal to reduce inventory costs.
Prior to the use of intelligent algorithms, most of production scheduling was done
manually. Shop scheduling problem is a typical NP-hard problem with exponential
explosion characteristics [2], and it is difficult to find the optimal solution relies on
manual scheduling. However, the emergence of intelligent algorithms solves this
problem. The results filtered by tens of millions of iterations are often much better than
the results of manual scheduling [3].
Therefore, this article introduces a dynamic scheduling model that covers the
general assembly shop, the painting shop, and painted body storage (PBS) which is a
buffer zone connects the two shops. Parts inventory in the general assembly shop will
dynamically change as time goes. The optimization goal of the general assembly shop
is to equalize the consumption rate of various parts as much as possible while ensuring
each car can be assembled on time; the painting shop is to reduce the number of color
switching in production queue and lower spraying cost; and PBS is to rearrange the
production queues coming out of the painting shop, and to reduce waiting time of
vehicles in the buffer zone, under the premise of meeting the general assembly shop’s
requirements. The NSGA-2 has been selected to achieve these multiple optimization
goals. This algorithm has a reliable performance in solving multi-objective optimiza-
tion problems and is widely used [4].
2 Mathematical Model
Many car companies are currently running on order production mode. An order infor-
mation represents a specific model of car. In the workshop, an order i corresponds to a
list of parts Mi , so the optimization of production queue is essentially the optimization of
fMi g. Pre-definition is mandatory for each model of car. Considering extensibility of the
mathematical model, 42 different models are predefined, shown in Table 1.
Among them, T1 are customized parts that can be used only for the specified model
of the specified configuration, such as seats and engines. T2 are common parts in
different car models that can be used only for the specified configuration, like in various
electronic auxiliary systems. T3 are common parts in different configurations that can
be used only for specified models, like interior trimming panels and wheels. T4 are
common parts for all vehicles, like different fasteners.
For the general assembly shop, the parts consumption speed needs to be more
balanced. The theoretical consumption rate of each part is
J ¼ P N1
EVj ¼ V ð1Þ
N1
i¼1 Mi;j
Deviation between actual consumption rate and theoretical consumption rate is the
optimization goal of the general assembly shop.
P N2
J Þ2
ðVj V
j¼1
min f1 ¼ ð2Þ
N1
For the painting shop, reducing the frequency of color switching can reduce coating
costs. The number of color switching can be expressed as
8
< 0; i¼1
Si ¼ 0; colorðiÞ ¼ colorði 1Þ ð3Þ
:
1; colorðiÞ ¼
6 colorði 1Þ
There are many kinds of buffers for connecting the general assembly shop and the
painting shop, linear, circular, backward and so on [5]. In general, waiting buffer time
of car body is always proportional to the optimization effect of buffer. Therefore, linear
buffer has been selected for our model.
For PBS, the less time car bodies are waiting in PBS, the more efficient PBS will
be. The waiting time can be expressed as
28 Z. Shen et al.
8
< iði0 Þ, i0 ¼ 1
Wi ¼ iði0 Þ, iði 1Þ\iði0 Þ
0
ð5Þ
: 0
iði 1Þ, iði0 1Þ iði0 Þ
Three optimization objectives have strong correlations and involve many variables.
Changes in any workshops’ production queue will affect another workshop queue and
buffer. Therefore, a multi-objective optimization algorithm is necessary. The typical
multi-objective evolutionary algorithms are NSGA-2, PESA-2 and SPEA-2. Each of
these three algorithms has its advantages and disadvantages. The advantage of NSGA-2
is that it has high operational efficiency and good distribution of solution sets in low-
dimensional problems; its disadvantage is that the diversity of solution sets is not ideal
in high-dimensional problems. The advantage of PESA-2 is that the convergence of its
solution is very good; but the disadvantage is that the selection operation can only
select one individual at a time, the time consumption is very large, and the class
diversity is not good. The advantage of SPEA-2 is that it can obtain a well-distributed
solution set, but its clustering process takes a long time to maintain diversity, and the
operating efficiency is not high.
In order to balance the running time with the quality of reconciliation, NSGA-2 was
eventually selected. Besides, we have designed a heuristic algorithm that can directly
calculate the downstream shop queue within constraints of the linear buffer, based on
the queue of upstream shop and the optimization goal of downstream shop which are
embed it in the NSGA-2.
The heuristic algorithms are divided into inbound rules and outbound rules. The
inbound rules are as follows:
1. If the last car in a lane has the same color of the waiting car, enter the lane.
2. Enter a lane with least car.
Its outbound rules are as follows:
1. In the waiting outbound car, select the same color as the last outbound car.
2. Unless the number of waiting cars in PBS is greater than M, otherwise there is no
car outbound.
3. When the number of waiting cars in PBS is greater than M, select the lane with the
highest number of cars, and outbound.
It should be pointed out that the general assembly shop is set to be the upstream
workshop and the painting shop as the downstream workshop. The reason is that the
optimization of the painting shop is simpler than the general assembly shop. A simple
Dynamic Production Scheduling Modeling and Multi-Objective Optimization 29
4 Simulation Examples
The simulated data is based on the production plan of 1,000 cars in a car factory within
one day. In addition, a traditional algorithm is designed to compare with the above
algorithm. This traditional algorithm is actually applied to the factory, rely on PLC
control and will determine the inbound and outbound sequence of body in PBS
according to a fixed priority. Both that embedded heuristic algorithm based on NSGA-2
and traditional algorithms are programmed by Matlab.
30 Z. Shen et al.
Due to computational complexity, the production plan has been divided into 10
equal proportions smaller plans, so only one portion (100 cars) production queue needs
to be calculated. Then connect them end to end in a cyclic manner to get the total
production queue, so this total production queue is “equalized”. Inventory changes of
parts should conform to the rule in Fig. 2.
Two example was run 50 times. The operating environment of the program is Intel
Xeon CPU E3-1240 3.50 GHz, and the RAM is 32 GB. The average time of operation
of embedded heuristic algorithm was 278.8 s, which is sufficient for a one-day pro-
duction plan of a car factory. And traditional algorithm uses the same production
sequence of painting shop with embedded heuristic algorithm.
The results of the two algorithms are shown in Tables 3 and 4 respectively.
affect the inventory of the assembly shop, but the waiting time will not be. In order to
study what negative impact the waiting time would have on the production line, some
inspection work was done.
When observing the results of the calculation, an unrealistic situation was dis-
covered. The results show that when waiting time is bigger, a kind of defect is more
likely to appear. This defect will result vacancies at the production queue of the general
assembly shop. The reason for this defect is that there are fewer bodies available in the
PBS when the queue is first started. If there is no body that meets the outbound rules,
there will be a vacancy in the assembly shop queue.
To avoid this situation, the production speed of the painting shop needs to be
greater than the general assembly shop, and the M value increases over time for a short
period of time when the queue is just starting to run. When the available car body in the
PBS is sufficient, the M value no longer increases.
Actually, as long as the car body can meet the constraints of continuous outbound,
they are all equivalent regardless of the waiting time. If the embedded heuristic
algorithm can ensure continuous outbound car body in PBS. The traditional algorithms
do not have obvious advantages in Optimization goal 3. Instead, there is a clear
disadvantage in optimizing goal 1. It can be considered that the embedded heuristic
algorithm has a comparative advantage over the traditional algorithms.
32 Z. Shen et al.
5 Conclusion
References
1. Sadjadi, S.J., Jafari, M., Amini, T.: A new mathematical modeling and a genetic algorithm
search for milk run problem. Int. J. Adv. Manuf. Technol. 44(1–2), 194–200 (2009)
2. Jalilvand-Nejad, A., Fattahi, P.: A mathematical model and genetic algorithm to cyclic
flexible job shop scheduling problem. J. Intell. Manuf. 26(6), 1085–1098 (2015)
3. Soleimani, H., Kannan, G.: A hybrid particle swarm optimization and genetic algorithm for
closed-loop supply chain network design in large-scale networks. Appl. Math. Model. 39(14),
3990–4012 (2015)
4. Deb, K.: A fast elitist non-dominated sorting genetic algorithm for multi-objective
optimization: NSGA-2. Lect. Notes Comput. Sci. 1917, 849–858 (2000)
5. Chen, GY: Research on buffer design and resequence in automobile production line.
Huazhong University of Science and Technology (2007)
Tandem Workshop Scheduling Based
on Sectional Coding and Varying Length
Crossover Genetic Algorithm
1 Introduction
optimization problem. Bai et al. [5] proposed a tandem workshop control model, which
reduced the complexity of AGVs scheduling and improved the efficiency of the system.
Tang et al. [6] solved the problem of task scheduling and coordination control in
tandem AGV systems by neuro-endocrine coordination mechanisms. Hou et al. [7]
developed a collaborative optimization method to solve the problem of machine
allocation and loop layout for tandem AGV systems. Reza et al. [8] putted forward a
method based on tabu search and genetic algorithm to avoid infeasible solutions from
crossover operation.
The paper will be organized as follows: in Sect. 2, a tandem workshop scheduling
system model was defined. And we established the mathematical model in Sect. 3.
Next, we explained the proposed sectional coding and the varying length crossover
genetic algorithm (VGA) as detailed in Sect. 4. In the Sect. 5, we compared the result
of the proposed algorithm with the fixed length crossover genetic algorithm (FGA) to
verity the effectiveness of the algorithm. At last, we made a conclusion.
2 Problem Description
In a tandem workshop scheduling system, the distribution of the machines are given,
the AGVs move in a bi-directional path, and only one AGV is worked in different areas
to avoid collisions. In different areas, the location of a machine is designated as a
transfer station, which is used to handle transportation tasks between different areas,
and all transfer stations constitute a transfer center. AGV has different operating speeds
in different areas, which can save transportation time, increase production efficiency,
and save costs. The tandem workshop scheduling model is shown in Fig. 1.
The optimization for tandem AGV workshop aims to provide the excellent scheme
of path planning, so that the material distribution time of tandem workshop scheduling
system is reduced effectively.
36 H. Sun and X. Zheng
3 Mathematical Model
R X
X Mj X
N X Mj
Min dijrj sj Tijrj sj : ð1Þ
i¼1 j¼1 rj ¼1 sj ¼1
Constraints:
X Mj X
N X Mj
dijrj sj qjrj sj Q 8i 2 f1; 2; . . .; Rg: ð2Þ
j¼1 rj ¼1 sj ¼1
dijrj sj
tijrj sj ¼ 8i 2 f1; 2; . . .; Rg 8j 2 f1; 2; . . .; N g: ð3Þ
v1
0
0
dijr j sj
tijr j sj
¼ 8j 2 f1; 2; . . .; N g 8i 2 f1; 2; . . .; Rg: ð4Þ
v2
Mj X
X N Mj X
X N
0
Tijrj sj ¼ dijrj sj tijrj sj þ dijrj sj tijr j sj
8i 2 f1; 2; . . .; Rg: ð5Þ
rj ;sj ¼1 j¼1 rj ;sj ¼1 j¼1
Mj
X
dij1sj ¼ 1 8i 2 f1; 2; . . .; Rg 8j 2 f1; 2; . . .; N g: ð6Þ
sj ¼1
Mj
X
dijrj 1 ¼ 1 8i 2 f1; 2; . . .; Rg 8j 2 f1; 2; . . .; N g: ð7Þ
rj ¼1
Tandem Workshop Scheduling Based on Sectional Coding 37
Mj
X Mj
X
dijrj h dijhsj ¼ 0 8h 2 f1; 2; . . .; M g 8i 2 f1; 2; . . .; Rg
rj ¼1 sj ¼1
ð8Þ
8j 2 f1; 2; . . .; N g:
Where:
Formula (1) shows the total time for AGV to perform all transportation tasks in
tandem workshop.
Formula (2) indicates that the load of a AGV cannot exceed the load-carrying quota.
Formula (3) indicates that in the machine area, the operating time of AGV transport
materials between two machines.
Formula (4) indicates that in the transfer center, the operating time of AGV
transport materials between two machines.
Formula (5) represents the total time of different areas to accomplish tasks.
Formula (6)–(8) means that different AGV will return to the starting position after
completing all the transportation tasks from the transfer station.
Notes:
Tijrj sj : In the area j, the time for the i-th AGV to complete the transportation tasks
from the machine rj to the machine sj .
qjrj sj : In the machine area j, transport the weight of the material from machine rj to
machine sj .
0
dijrj sj : In the transfer center j, the distance of the i-th AGV from the machine rj to the
machine sj .
dijrj sj : In the transfer center j, the distance of the i-th AGV from the machine rj to the
machine sssssj .
8
< ¼ 1 If the i-th AGV is arranged for material transportation from
dijrj sj machine rj to machine sj :
:
¼ 0 Other
In this paper, a tandem workshop AGV scheduling model is established based on the
minimum material transportation time as the objective function. Compared with the
traditional genetic algorithm, the coding method of chromosomes is redesigned, which
could generate chromosomes to meet the needs of the optimization problem. The
objective function is used as the fitness function of the algorithm. With improved
crossover and mutation operation, the certain cross rate and mutation rate are set, so
that the superior individuals of the previous generation are not lost, and this can ensure
the variety.
38 H. Sun and X. Zheng
The following Figs. 3, 4 and 5 show the specific operation of the crossover
operation:
As shown in Figs. 6 and 7, the specific operation for the mutation operation.
40 H. Sun and X. Zheng
5 Experimental Verification
materials required to be transported between the machines in the workshop are shown
in Table 3 below.
The problem is solved by applying our algorithm. The detailed parameters are as
follows: The population size is N = 100, the crossover rate is 0.6, and the maximum
iteration number of the algorithm is 10000 times. The algorithm of DE, the fixed length
crossover genetic algorithm and the genetic algorithm of this paper were performed 50
times respectively. The data results are shown in Table 4 below:
The VGA proposed in this paper adopts the sectional coding rules, which can avoid
illegal solutions in the crossover and mutation operation. The crossover operation with
varying length crossover enables the population to have more potential, retain the fine
quality of the individual and improve the computing capability of the algorithm. At the
same time, the mutation operation combining insertion mutation and reverse sequence
mutation can maintains the diversity of the population and increases the local search
ability of the algorithm to avoid premature phenomena.
Figure 8 shows the average fitness value curve of different algorithms when
reaching the termination condition. Since the VGA that we proposed employs the
mutation operation combining two mutation methods, the DE algorithm with different
mutation rates is used for comparison. For the DE, the optimal value decreases as the
mutation rate increases. In the case of high mutation rate, the optimal value of DE is
obviously larger than that of FGA and VGA. According to the experimental results,
sectional coding and varying length crossover genetic algorithm have faster conver-
gence rate and better value of the objective function than the traditional algorithm.
6 Conclusion
An optimization model is proposed to optimize the tandem AGV workshop with the
goal of minimizing the total distribution time. Only one AGV is responsible for
material transportation in different areas, the coding rules, crossover and mutation
operation of traditional genetic algorithms are improved, and the proposed algorithm is
Tandem Workshop Scheduling Based on Sectional Coding 43
Acknowledgements. This work was supported by the Guidance Program for Natural Science
Foundation of Liaoning (No. 20170540138).
References
1. Pen C.T., Du Z.J.: Design of AGV based on AT89S52 MCU. Wireless Internet Technology
13 (2017)
2. Bozer, Y.A., Srinivasan, M.M.: Tandem configuration for automated guided vehicle systems
and the analysis of single vehicle loops. IIE Trans. 23, 72–82 (1991)
3. Zhang, H.: Design and implementation of a hybrid scheduling model for tandem workshop
resources in cloud computing environment. CIT 5, 8–11 (2017)
4. Zhou, Q., Liu, J., Wei, F.L.: Single-field tandem workshop scheduling optimization based on
genetic taboo search algorithm. I J. Chang. Univ. Sci. Technol. (Nat. Sci.) 4, 32–38 (2014)
5. Bai, S.F., Tang, D.B., Gu, W.B., Zheng, K.: Research of multiple AGV systems based on
tandem workshop control module. CNEU 3, 8–12 (2012)
6. Tang, D.B., Lu, X.C., Zheng, K.: Research on tandem AGV scheduling based on neuro-
endocrine coordination mechanism. Mach. Build. Autom. 4, 112–115 (2015)
7. Hou, L.Y., Liu, Z.C., Shi, Y.J., Zheng, X.J.: Optimizing machine allocation and loop layout in
tandem AGV workshop by the collaborative optimization method. Neural Comput. Appl. 4,
959–974 (2016)
8. Rezapour, S., Zanjirani-Farahani, R., Miandoabchi, E.: A Machine-to-loop assignment and
layout design methodology for tandem AGV systems with single-load vehicles. Int. J. Prod.
Res. 49, 3605–3633 (2011)
9. Chen, Z.T.: Research and application on job shop scheduling problem based on improved
genetic algorithm, p. 32. Dalian University of Technology, Dalian
Robust Bi-level Routing Problem
for the Last Mile Delivery Under Demand
and Travel Time Uncertainty
Abstract. Designing the last mile delivery system in a lean way has
become an important part of serving customers efficiently and econom-
ically. However, in practice, the uncertainty in customer demand and
travel times often means vehicles capacity may be exceeded along the
planed route and vehicles miss theses time windows, increasing the cost,
reducing efficiency and decreasing the customer satisfaction. Previous
studies have lacked an uncertainty-based view, and few studies have
discussed how to develop an uncertain model. To address this issue,
the bi-level routing problem for the last mile delivery is formulated
as a robust vehicle routing problem with uncertain customer demand
and travel times. In addition, a modified simulated annealing algo-
rithm is proposed and tested in computational experiments. The results
show that the proposed model has good performance for uncertainty
processing.
1 Introduction
The bi-level multisized terminal location-routing problem (BL-MSTLRP ) for
the last mile delivery, as a classical combinatorial optimization problem, first is
introduced by Zhou [8] under the e-commence environment, that aims to find the
optimal set of terminals and the optimal set of routes for a fleet of vehicle deliv-
ering products or service to a given set of customers groups. It is worthy that the
Supported by the National Social Science Fund of China (Grant No.
18BJY066), Fundamental Research Funds for the Central Universities (Grant No.
106112016CDJXZ338825), Chongqing key industrial generic key technological inno-
vation projects (Grant No. cstc2015zdcy-ztzx60009), Chongqing Science and Tech-
nology Research Program (Grant No. cstc2015yykfC60002).
c Springer Nature Singapore Pte Ltd. 2018
K. Li et al. (Eds.): ICSEE 2018/IMIOT 2018, CCIS 924, pp. 44–54, 2018.
https://doi.org/10.1007/978-981-13-2384-3_5
Robust Bi-level Routing Problem for the Last Mile Delivery 45
BL-MSTRP, as the classical optimization for last mile delivery, assumes that all
the input data, such as customer demands and travels time, are both determin-
istic and are known in advance. However, the solutions derived by deterministic
model are often infeasible when applied to the real-world situation, particularly
in the era of on-time delivery [7]. To tackle problem, this paper proposes robust
bi-level routing problem (RBLRP ) for the last mile delivery under demand and
travel time uncertainty, an extended version of the BL-MSTRP for the last mile
delivery. Note that robust optimization is advantageous due to its computational
tractability and its ability to deal with the practical applications, in which only
partial information or a small amount of historical data about the uncertain
parameters is available [1–4]. In this study, the uncertainty sets of customer
demand and uncertainty sets of travel times of bi-level were derived with partial
information about the distributions of the uncertain parameters and a modified
SA algorithm was proposed to tackle this issue.
The reminder of this paper is organized as follows. Section 2 describes the
considered RBLRP and the process of uncertainty transformation. In addition,
the corresponding robust mathematical model is also presented. In Sect. 3, a
modified SA algorithm is designed for the RBLRP. Section 4 presents a compu-
tational experiment using the modified versions of Zhou ’s instances [8]. Finally,
conclusions are drawn and presented in Sect. 5.
d 23
q5 - q5 , q5 + q5
q5
d12
Fig. 1. The problem description of RBLRP with uncertainty for last mile delivery with
HD and CP service programming.
The customer demand uncertainty set Uqki and the travel distance uncertainty
set Udk is defined as follows based on the robust optimization approach [4].
with ⎧
⎨
= q̃i ∈ R|Nl | |q̃i = q̄i + αi q̂i ,
k
Uqki |αi | ≤ Γkqi ,
⎩ (2)
i∈Nlk
k
|αi | ≤ 1, Γqi = θqi NC , ∀i ∈ NC
k
Equation (1) reflects the overall demand uncertainty set uqi of customer groups
i ∈ NC is the Cartesian product of the demand uncertainty set for each vehi-
cle, which consists of CP demand and HD demand. Equation (2) denotes the
uncertain demand of customer groups with HD demand; q̄i represents the nom-
inal value of uncertain demand; and q̂i denotes the maximum deviation from
the nominal value for each vertex i ∈ NC . αi is the auxiliary variable, and
Γkqi is the uncertainty budget that controls the level of uncertain demand of
customer groups. Since the uncertainty experienced by each vehicle is related
k
to the number customer group visited, Γqi change by at
of customers in each
k
most θqi NC , where k
θqi NC represents the smallest integer that is larger
than or equal to θqi NCk . And θqi is the demand uncertainty budget coefficient
of customer groups, and takes value in the interval [0, 1]. If θqi = 0, Γkqi = 0
and q̃i = q̄i – the influence of the demand uncertainty
of customer groups can
be ignored completely. If θqi = 1, Γkqi = NCk – each customer demand q̃i of
customer groups can take any value in the interval [q̄i − q̂i , q̄i + q̂i ].
Robust Bi-level Routing Problem for the Last Mile Delivery 47
For the distance within the customer groups, because of the spatial aggregation
feature of customers locations for the last mile delivery, the overall delivery
demand of a customer group keeps constant while individual demand constantly
changes, which is difficult to calculate the tour travel distance for all customers.
Here, the approximate continuous model for customer group (modified from
Zhou et al. [8]) was introduced to model the nominal travel distance L̄ (n) of the
tour in customer group with n customers, which is given by
with
⎧
⎨ k
k
UL(n) = L̃ (n) ∈ R|NC | L̃ (n) = L̄ (n) + λi L̂ (n) , |λi | ≤ ΓkL(n) ,
⎩ k (9)
i∈NC
|λi | ≤ 1, ΓkL(n) = θL(n) NCk , ∀i ∈ NCk
Equation (9) shows the overall distance uncertainty set uL(n) of the tour in cus-
tomer groups is the Cartesian product of the travel time uncertainty set UdkA of
the tour in a customer group for each vehicle. NCk denotes the set of customer
groups on the route of the vehicle. Similarly, the parameters in the Eq. (10) has
the resemble meanings to those in Eq. (2). NCk represents the nominal value of
uncertain demand L̃ (n) and L̂ (n) denotes the maximum deviation from the
nominal value for each customer group ∀i ∈ NCk . λi is the auxiliary variable,
and ΓkL(n) is the uncertainty budget that controls the level of uncertain travel
distance of the tour in the customer groups and ΓkL(n) is defined as equaling
k
θL(n) NCk , where θL(n)
k NC represents the smallest integer that is larger
than or equal to θL(n) NC . And θL(n) is the travel time uncertainty budget
coefficient of the tour in a customer group, and takes value of between 0 and 1.
For the distance between terminals in level 1 and between customer groups in
level 2, without loss of generality, the travel time uncertainty set udA between
terminals and customer groups is given by
with
⎧
⎨ k
UdkA = d˜ ∈ R|A | d˜ij = d¯ij + βij dˆij , |βij | ≤ ΓkdA ,
⎩ (11)
(i,j)∈Ak
|βij | ≤ 1, ΓkdA = θdA Ak , ∀ (i, j) ∈ Ak
Equation (11) shows the overall distance uncertainty set udA is the Cartesian
product of the travel time uncertainty set UdkA for each vehicle. Similarly, in
Eq. (12), Ak denotes the set of arcs on the route of the vehicle k, which consists
of the set between customer groups and the set between terminals. d¯ij repre-
sents the nominal value of uncertain distance d˜ij and d˜ij denotes the maximum
deviation from the nominal value for each arc ∀ (i, j) ∈ Ak . βij is the auxiliary
variable, and ΓkdA is the uncertainty budget that controls the level of uncertain
k
travel time
of customer and terminals and ΓdA is defined as equaling
groups
θdA A – where
k represents the smallest integer that is larger than
k θdA A
k
or equal to θdA A . And θdA is the distance uncertainty budget coefficient, and
takes value of between 0 and 1.
Robust Bi-level Routing Problem for the Last Mile Delivery 49
xjik − xijk = 0, ∀i ∈ ND , k ∈ K (16)
j∈N1 j∈N1
U0uk = q̃jC xijk , ∀u ∈ ND , k ∈ K (17)
i∈N1 j∈ND
Ujik − Uijk = q̃iC , ∀i ∈ ND , k ∈ K (18)
j∈N1 j∈N1
Constraints (13)–(22) are used for limiting the vehicle routing of level 1. Con-
straints (14) guarantee that each terminal is served at most once. Constraints
(15)–(17) are the flow conservation constraints, which ensure that vehicle starts
from DC and ends at DC. Constraints (18) ensure calculate the load of vehicles.
Constraints (19) are the flow constraints for demand. Constraints (20) make sure
that remaining demand of a vehicle should not exceed its capacity. Constraints
(21) stipulate that vehicle’s maximum working hours should not be exceeded.
Constraints (22) ensure that the remaining demand of a vehicle is zero after
50 X. Huang et al.
serving the final vertex. Note that constraints (20)–(22) are both related to both
level 1 and level 2.
xijk = 1, ∀i ∈ N2 , k ∈ K (22)
j∈NC
zim = 1, ∀i ∈ NC (23)
m∈ND
x0jk = 1, ∀k ∈ K (24)
j∈N2
xi0k = 1, ∀k ∈ K (25)
i∈N2
xjik − xijk = 0, ∀i ∈ NC , k ∈ K (26)
j∈N2 j∈N2
U0uk = q̃jH xijk , ∀u ∈ NC , k ∈ K (27)
i∈N2 j∈NC
Ujik − Uijk = q̃iH , ∀i ∈ NC , k ∈ K (28)
j∈N2 j∈N2
(30)
q̃iC zim ≤ bm ym , ∀k ∈ K, m ∈ ND (31)
i∈NC
Constraints (23)–(32) are used for limiting the level 2. Constraints (23) guaran-
tee that each customer group is served exactly once by a level 2. Constraints (24)
ensure that each customer group is assigned to a terminal exactly once. Con-
straints (25)–(31) have a similar meaning with constraints (15)–(21). Constraints
(32) make sure that the terminal’s capacity cannot be exceeded by customers
assigned.
In this work, permutation representation is used and the splitting procedure [5,6]
is integrated into the encoding strategy. As far as the chromosome of level 1 and
level 2, the DCs, denoted as 0 between the head and tail of the chromosome,
are generated at random, aiming to split the giant tour into some small routes
without violating the vehicle capability and working time constraints.
Robust Bi-level Routing Problem for the Last Mile Delivery 51
j1 j j+1 j j1 j1 j j+1
4 Computational Experiments
The computational experiment instance was derived from the latest research
instance of Zhou [8] designed for the BL-MSTLRP with deterministic customer
demands and travel times. To simplify the calculation, the datasets of terminal
size information and vehicle type information were ignored. The speed of vehicles
is 30 km/h between nodes and 20 km/h in customer groups with working time
about 8 h. In addition, the customer demand transfer coefficient is subject to
distribution as follows.
⎧
⎪ 0, 0 ≤ dij ≤ 500
⎪
⎨
pijCH = f (dij ) = 0.3, 500 < dij ≤ 800 (32)
⎪
⎪ 0.7, 800 < dij ≤ 1000
⎩
1, dij > 1000
For the robust purpose, the nominal value q̄i (i ∈ NC ) and t̄ij ((i, j) ∈ A) was
assumed to equal to the corresponding customer demand and travel time in
case set data from the Zhou, respectively. Moreover, we assumed that the max-
imal demand deviation q̂i was 0.2q̄i , and the maximal travel time deviation t̂ij
was 0.2t̄ij and θqi = θL(n) = θdA was 0.6. For the SA, The terminal tempera-
ture Tmin = 0.001, maximal evolutional generation Iiter = 100 and temperature
update factor r = 0.99. Finally, the computational experiments were imple-
mented using MATLAB R2014a, and the experiments on a computer with an
AMD A8–7650K Radeon R7 CPU at 3.30 GHz and 8 GB RAM under Windows
X64 system.
Robust Bi-level Routing Problem for the Last Mile Delivery 53
1600
1400
1200
1000
Y-coordinate
800
1
600
400
Terminal
0
0 500 1000 1500 2000 2500 3000
X-coordinate
5 Conclusion
The last mile delivery problem has gained widespread attention in recent years,
but the previous studies lack enough analysis of the uncertain phenomenon. This
paper focus on the research on the uncertainty optimization of the last mile. We
found that robust optimization performed outstandingly on uncertainties and
also had good performance in dealing with large-scale network uncertainties.
54 X. Huang et al.
References
1. Ben-Tal, A., Ghaoui, L., Nemirovski, A.: Robust Optimization (2009)
2. Bertsimas, D., Sim, M.: The price of robustness. Oper. Res. 52(1), 35–53 (2004).
https://doi.org/10.1287/opre.1030.0065
3. Han, J., Lee, C., Park, S.: A robust scenario approach for the vehicle routing problem
with uncertain travel times. Transp. Sci. 48(3), 373–390 (2014). https://doi.org/10.
1287/trsc.2013.0476
4. Hu, C., Lu, J., Liu, X., Zhang, G.: Robust vehicle routing problem with hard time
windows under demand and travel time uncertainty. Comput. Oper. Res. 94, 139–
153 (2018). https://doi.org/10.1016/j.cor.2018.02.006
5. Kloimüllner, C., Papazek, P., Hu, B., Raidl, G.R.: A cluster-first route-second
approach for balancing bicycle sharing systems. In: Moreno-Dı́az, R., Pichler, F.,
Quesada-Arencibia, A. (eds.) EUROCAST 2015. LNCS, vol. 9520, pp. 439–446.
Springer, Cham (2015). https://doi.org/10.1007/978-3-319-27340-2 55
6. López-Santana, E., Rodrı́guez-Vásquez, W., Méndez-Giraldo, G.: A hybrid expert
system, clustering and ant colony optimization approach for scheduling and routing
problem in courier services. Int. J. Ind. Eng. Comput. 9(3), 369–396 (2018). https://
doi.org/10.5267/j.ijiec.2017.8.001
7. Taş, D., Dellaert, N., Van Woensel, T., De Kok, T.: Vehicle routing problem with
stochastic travel times including soft time windows and service costs. Comput. Oper.
Res. 40(1), 214–224 (2013). https://doi.org/10.1016/j.cor.2012.06.008
8. Zhou, L., Lin, Y., Wang, X., Zhou, F.: Model and algorithm for bilevel multisized
terminal location-routing problem for the last mile delivery. Int. Trans. Oper. Res.
(2017). https://doi.org/10.1111/itor.12399
A Comprehensive Fault Diagnosis System
and Quality Evaluation Model
for Electromechanical Products by Using
Rough Set Theory
1 Introduction
Reliability and quality performance for electromechanical products will directly affect
the stability of the operational stability. Thus, every company should make the most of
its sources and regulating the cost from time to time to product quality [1]. It is
introduced that during the manufacturing process of electromechanical products, the
fault diagnosis system and quality evaluation of their key procedures must be
J. Pang—This work was supported by the National Natural Science Foundation of China
(No. 71671130, No. 71301120, No. 51705048).
strengthened to ensure the qualification and stability to materials and workmanship [2].
The fault diagnosis system is a key part of product quality evaluation, and product
working performance is affected by the processing quality in through their manufac-
turing process [3]. Moreover, the quality evaluation goals can be translated into a set of
preference or utility functions that represent the value of alternative sets of conse-
quences by decision theorists [4]. Through introducing the reliability specifications and
quality test methods specified for electromechanical products, the basic fault diagnosis
and quality control requirements can be understood by the application [5]. This paper
presents a fault diagnosis system for alarm processing and fault diagnosis in product
manufacturing process.
On the other hand, a comprehensive fault diagnosis system and quality evaluation
model for electromechanical products is the basis of scientific decision-making. Fault
diagnosis system is an important part for quality evaluation model, a comprehensive
evaluation with rough set theory can improve the electromechanical products quality
[6]. Furthermore, combination with rough sets and decision-making theory has enor-
mous development potentials in the field of fault diagnosis system and quality evalu-
ation for electromechanical products [7]. To ensure the accuracy and stability of quality
evaluation, this comprehensive fault diagnosis system sets up fault recognition
mechanism before acting [8]. Domestic and foreign scholars have done a lot of research
work on comprehensive quality evaluation methods. In the next, under the framework
of rough set theory, this study presents a multi-stage innovation quality evaluation
model with the theory of rough set analysis. A quality evaluation model for identifi-
cation, diagnosis, estimation and forecast of fault diagnosis system using the data
fusion technique and the rough set theory was established [9]. This paper describes how
to apply rough set theory in quality evaluation model validity of fault diagnosis sys-
tems. Objective weighting evaluation method based on actual data source to make
quantitative evaluation of indicators, the weight is determined according to the cor-
relation or variation coefficient of each index.
Considering the above aspects, a new method for quality evaluation and its
application in ball valve fault diagnosis with rough set theory is proposed in this study.
This study will be arranged as follows. The implement process comprehensive fault
diagnosis system and quality evaluation model for electromechanical products by using
rough set theory in the following section. In the case study, a testing attribute evalu-
ation is set up, the attributes of t quality evaluation model are calculated, and the
experimental data are analyzed to verify or identify the model’s key attributes for a
high fault diagnosis system. In the last section, some conclusions on the approved
diagnosis system and evaluation model are given.
Rough set theory is an effective tool to deal with vagueness and uncertainty. It provides
effective means for people to understand things correctly and make scientific decisions.
The rough set theory was proposed by Poland scientist in 1982, and after the contin-
uous development of scientific researchers on rough set theory [10]. Rough set theory is
a new mathematical tool with strong knowledge obtaining ability after probability
A Comprehensive Fault Diagnosis System and Quality Evaluation Model 57
theory, fuzzy sets, mathematical theory of evidence [11]. At present, rough set theory
has been applied to various fields, such as machine learning, fault diagnosis, control
algorithm acquisition, process control and knowledge acquisition in relational database,
and has achieved great success [12].
In brief, the rough set theory and method is quite an efficient means to deal with
complex system, which is applied to a lot of fields. An algorithm of attribute reduction
based on attribute importance of rough set was used to fault diagnosis system [13]. An
extraction method of decision rules for fault diagnosis by using rough set theory is
presented to extract simple and effective diagnostic rales for quality evaluation in this
paper. The implement process of fault diagnosis system and quality evaluation model
for electromechanical products by using rough set theory is shown in Fig. 1.
jPOSc ðDÞj
cc ðDÞ ¼ ð4Þ
jUj
X
n
jXi j jXi j 1 X n
IðPÞ ¼ j1 ¼1 jXi2 j ð6Þ
i¼1
jUj jUj 2
jUj i¼1
Where X represents the cardinal number of the set X, Xi/U represents the relative
cardinality of the equivalent class Xi in U.
And the importance of attribute Aa is defined as:
X
m
xðai Þ ¼ SGFAfai g ðai Þn SGFAfai g ðai Þ ð8Þ
i¼1
60 J. Pang et al.
3 A Case Study
A method of fault diagnosis of ball valve by using rough set theory is proposed in the
case study. A ball valve is widely used in the pneumatic and hydraulic control system,
which controls the circulation of the medium in the pipe through opening and closing
of the ball valve. The existing problems of manufacturing process of ball valve were
analyzed, including unstable machining quality, imperfect technics, incongruous pro-
cess design and low assembly quality. The comprehensive evaluation method consists
of objective weighting method and subjective weighting method. Rough set theory was
introduced to solve this problem, and further the quality evaluation model for fault
diagnosis ability was proposed.
In this paper, the comprehensive evaluation method is used to evaluate the index of
the quality of the ball valve. With the use of rough set theory, the operational quality of
ball valve was improved and optimized. In the process of work, if the valve leakage or
failure cannot start normally, it will cause the system fail to achieve the prescribed
workflow and affect the environmental pollution. Typical application of the ball valve
of quality improving and enhancement of processing capacity were introduced in this
case study. Rough set theory is applied to evaluate the quality of ball valve. The index
that affects the quality of ball valve is determined through comprehensive index weight
values, and the key part is controlled during the production process of ball valve
quality. To realize the complementary advantages of uncertainty, the final compre-
hensive evaluation result is more reasonable.
To determine the reliability correction factor, it is necessary to comprehensively
consider the composition structure, manufacture and assembly, and use of the product.
However, the above indicators are too general and difficult to quantify in practice.
Therefore, it is necessary to divide the above several large segments into easily
quantified indicators, establish a more comprehensive index system to evaluate the
degree of difference.
In view of the quality characteristics and consulting related manuals and large
number of failure data, the ball valve can be obtained the factors which affecting the
quality of life cycle. Then, the comprehensive evaluation hierarchy model of numerical
control machine tool subsystem quality correction factor is established, which is shown
in Fig. 2.
The layered model consists of three layers, the top layer is the target layer, the
middle layer is the criterion layer and the bottom layer of the index layer. Rule layer
contains leakage of packing, leakage of main seal, side snapping or bite motion,
abnormal pneumatic actuator. And each factor contains number of factors for the
specific comparative indicators, which is used to build the hierarchical model of index
layer. Then, the simplification of rough set theory to the index system is described
below. Firstly, the indexes of two level in Fig. 2 are set as the attributes of the
information system, that is, A ¼ fX1 ; X2 ; . . .; X11 g. Ball valve can be evaluated as a
collection of objects in the system, for the convenience of processing, there are eight
ball valves, namely U = {I, II, …, VIII}.
In the evaluation process, the corresponding evaluation results of each secondary
index are measured in four grades, which is shown in Table 1.
A Comprehensive Fault Diagnosis System and Quality Evaluation Model 61
In the next, the information system S = (U, A, V, f) of rough set is simplified, and a
simple set is R1 = {X1, X2, X5, X7, X9, X11}. And the Table 1 corresponding to the
information system S can be further reduced to Table 2.
From Table 2, comprehensive assessment standards are namely “excellent”,
“good”, “general” and “poor”, and can be respective by 4, 3, 2, 1, which are used to
build an information system S of rough set. And the corresponding level one and two
levels index sets are {F1, F2, F3, F4} = {{X1, X2}, {X5}, {X7}, {X9, X11}}.
Then, the computing process of secondary classes is simplified with the framework
among U = {I, II, III, IV, V, VI, VII, VIII} and A = {X1, X2, X5, X7, X9, X11}.
62 J. Pang et al.
Let F1 = {X1, X2}, then we can get U/IND(F1) = {{I}, {II, V}, {III, VI}, {IV, VII},
{VIII}} through the Formula (6):
12 þ 22 þ 22 þ 22 þ 12 50
IðF1Þ ¼ 1 ¼ :
82 64
So we have a hypothesis that U/IND(F1 − {X1}) = {{I, III, IV, VI, VII}, {II, V,
VIII}} is calculated with Formula (6):
52 þ 32 30
IðF1 fX1gÞ ¼ 1 ¼ :
82 64
Next, the calculating results can be get through the Formula (7):
50 30 20
SGFF1fX1g ðX1) = I(F1) I(F1 fX1gÞ¼ ¼ :
64 64
50 48 2
SGFF1fX2g ðX2) = I(F1) I(F1 fX2gÞ¼ ¼ :
64 64
X
2
20
xðX1Þ ¼ SGFF1fX1g ðX1Þn SGFF1fXig ðXiÞ ¼ 0:9091:
i¼1
20 þ 2
In much the same way, the other weights can be obtained as:
After that, depending on the results, we can meet with the comprehensive fault
diagnosis system and quality evaluation model of ball valve to discuss further action.
Data collection module is the base of the condition monitoring and quality evaluation
model in the manufacturing process. Next, we point out the merits and main problems
of the ball valve quality evaluation model. Development of the multiparameter con-
dition monitoring and quality evaluation model for ball valve. The experimental results
demonstrate that the proposed fault diagnosis system and quality evaluation model is
correct and effective.
4 Conclusions
This paper studies the decision rules acquisition methodology by using rough set theory
for fault diagnosis and quality evaluation. The application of comprehensive quality
evaluation model of ball valve in a manufacturing firm was discussed. This study offers
one new approach for automatic diagnosis rules knowledge in intelligent quality
evaluation. Rough set theory not only can enrich intelligent decision theory, but also
effectively provides the reference to the research of relative areas. Moreover, a fault
diagnosis system and quality evaluation model for electromechanical products is
developed based on the information fusion technology and rough set theory. Examples
show that the evaluation is feasible effective and provides a new way for the evaluation
of the quality of the ball valve. The method of fault diagnosis and quality evaluation
model can be also prospected for other electromechanical products by using rough set
theory.
Acknowledgments. This work was supported by the National Natural Science Foundation,
China (No. 71671130, No. 71301120, No. 51705048), the postdoctoral program of Zhejiang
University & Zhejiang Linuo Fluid Control Technology Company of China (No. 174102), the
key project of teaching reform of Wenzhou University (No. 15jg07), engineering practice edu-
cation center of Wenzhou University & Zhejiang Linuo Fluid Control Technology Company of
China.
References
1. Bassi, F., Clerici, R., Aquario, D.: Evaluating quality of the didactics at university: the
opportunities offered by latent class modeling. TQM J. 30(2), 168–180 (2018)
2. Huang, H.B., Li, R.X., Yang, M.L., Lim, T.C., Ding, W.P.: Evaluation of vehicle interior
sound quality using a continuous restricted Boltzmann machine-based DBN. Mech. Syst.
Sig. Process. 84, 245–267 (2017)
3. Oliveira, J.C.M., Pontes, K.V., Sartori, I., Embirucu, M.: Fault detection and diagnosis in
dynamic systems using weightless neural networks. Expert Syst. Appl. 84, 200–219 (2017)
64 J. Pang et al.
4. Vidal Legaz, B., Maia De Souza, D., Teixeira, R.F.M., Anton, A., Putman, B., Sala, S.: Soil
quality, properties, and functions in life cycle assessment: an evaluation of models. J. Clean.
Prod. 140, 502–515 (2017)
5. Ma, C., et al.: Sound quality evaluation of the interior noise of pure electric vehicle based on
neural network model. IEEE Trans. Ind. Electron. 64(12), 9442–9450 (2017)
6. Gao, Z., Chin, C.S., Chiew, J.H.K., Jia, J., Zhang, C.: Design and implementation of a smart
lithium-ion battery system with real-Time fault diagnosis capability for electric vehicles.
Energies 10(10), 1503 (2017)
7. Zhao, R., Li, C., Tian, X.: A novel industrial multimedia: rough set based fault diagnosis
system used in CNC grinding machine. Multimed. Tools Appl. 76(19), 19913–19926 (2017)
8. Moayyedian, M., Abhary, K., Marian, R.: Optimization of injection molding process based
on fuzzy quality evaluation and Taguchi experimental design. CIRP J. Manuf. Sci. Technol.
21, 150–160 (2018)
9. Nazari, R., Seron, M.M., De Dona, J.A.: Actuator fault tolerant control of systems with
polytopic uncertainties using set-based diagnosis and virtual-actuator-based reconfiguration.
Automatica 75, 182–190 (2017)
10. Chiaselotti, G., Ciucci, D., Gentile, T., Infusino, F.: Rough set theory and digraphs. Fundam.
Inform. 153(4), 291–325 (2017)
11. Gardiner, E.J., Gillet, V.J.: Perspectives on knowledge discovery algorithms recently
introduced in chemoinformatics: rough set theory, association rule mining, emerging
patterns, and formal concept analysis. J. Chem. Inf. Model. 55(9), 1781–1803 (2015)
12. Hu, D., Yu, X., Wang, J.: Statistical inference in rough set theory based on Kolmogorov-
Smirnov goodness-of-fit test. IEEE Trans. Fuzzy Syst. 25(4), 799–812 (2017)
13. Huang, C.-C., Tseng, T.-L., Tang, C.-Y.: Feature extraction using rough set theory in service
sector application from incremental perspective. Comput. Ind. Eng. 91, 30–41 (2016)
14. Khan, M.A.: A probabilistic approach to rough set theory with modal logic perspective. Inf.
Sci. 406–407, 170–184 (2017)
15. Yao, Y., Zhang, X.: Class-specific attribute reducts in rough set theory. Inf. Sci. 418–419,
601–618 (2017)
The Research of Improved Wolf Pack
Algorithm Based on Differential Evolution
1 Introduction
1.1 A Subsection Sample
Swarm intelligence algorithms such as Particle Swarm Optimization (PSO) [1–3], Fish
Swarm Optimization (FSA) [4, 5], Ant Colony Optimization (ACO) [6], Differential
Evolution Algorithm (DE) [7], are widely used to solve optimization problems and
have their own merits in solving different problems. The wolf pack search (WPS) was
proposed by Yang et al. [8] in 2007. The Wu et al. [9] proposed a new Wolf Pack
algorithm (WPA) on the basis of analyzing the behavior of hunting and prey among
wolves in 2013. The WPA simulates the characteristics of wolves and groups to
collaboratively hunt. In addition, the abstract behavior of walking, summoning, raids,
siege, “winner is king” head wolf generation mechanism and the “strong survival”
group update their mechanism to solve the optimization problems. Although the
algorithm has a short time of introduction due to its good performance in multi-peak,
high-dimension functions. In addition, it has been widely used in three-dimensional
underwater track planning [10], hydropower reservoir optimization scheduling [11] and
other human production activities. However, there are inevitably some shortcomings:
easy to fall into local optimal solution, large computational resource cost and low
robustness. Authors in [12] proposed an improved wolf pack algorithm, which pro-
posed the wolf updating rules and introduced the phase factor according to the basic
idea of the traditional wolf pack algorithm. At the same time, it optimized the types of
step lengths of the traditional wolves and designed a new wolves. The location update
formula is validated by the test function simulation. Literature [13] proposed a cultural
wolf pack algorithm combined with cultural algorithm. This algorithm can effectively
solve the blindness problem of artificial wolf search. Through testing and analyzing
three complex functions, the validity of the culture wolf pack algorithm is verified. In
Literature [14], the idea of solving the current local optimum in the PSO algorithm was
introduced into the walking and summoning behavior of the wolf pack algorithm, and
the suboptimal solution obtained by the chaos method is optimized. The improved
algorithm greatly improves the accuracy of the search and avoids falling into local
optimal. The effectiveness of the algorithm is verified by simulation.
Aiming at the deficiencies of WPA, this paper proposes an improved wolf pack
algorithm based on differential evolution (DIWPA). Based on the intelligent behavior
improvement of WPA, DIWPA introduces crossover, mutation and selection in dif-
ferential evolution, which greatly improves the optimization performance and robust-
ness of the algorithm.
This paper introduces the basic principle of DIWPA firstly. Then, DIWPA is used
to conduct optimization test on 12 benchmark functions and compare to 3 typical
optimization algorithms. The test results show that DIWPA has good performance for
all kinds of test function optimization, and can avoid premature algorithm.
The DIWPA inherits all the behavior of the WPA and improves it. The main
improvements are as follows:
① Due to the lack of guidance in wandering in WPA, the algorithm is unable to fully
search the solution space and easily fall into a local optimum. The DIWPA
introduces the search factor to the search wolves to make the algorithm traverse the
entire space.
② In the face of high-dimensional function optimization, it may inevitably consume a lot
of computing resources and may be difficult to jump out of the optimization process
and reduce the robustness in WPA. The DIWPA introduces the maximum number of
raid wolves to improve computational efficiency and algorithm robustness.
③ All the artificial wolves will siege at the same determined step size in siege
behavior in WPA, so the adaptive siege step length is introduced in the siege
behavior and the differential evolution strategies is introduced to improve the
global search capabilities in DIWPA.
The Research of Improved Wolf Pack Algorithm Based on Differential Evolution 67
where Xit represents the ith artificial wolf in the tth generation; X L and X U are the lower
and upper bounds of the variable X ¼ ðx1 ; x2 ; ; xD Þ respectively; D is the variable
dimension.
(2) The head wolf produces rules
In the initialized population, the fitness value of each artificial wolf is calculated by
Eq. (2). The artificial wolf with minimal fitness value will be elected to be the head
wolf in the iterative process and the fitness value of head wolf is Qlead. During the
process, the artificial wolf with the optimal fitness value after each evolution need to
compare with the head wolf in the previous generation. If the current fitness value of
the artificial wolf is better, the artificial wolf will replaces the head wolf and to be a
new head wolf. For the minimum problem, Qi is considered better than Qlead when
Qi < Qlead.
Q ¼ f ðX Þ ð2Þ
where stepa is the walking distance of the search wolf, c is the search factor of search
wolf, and is a random number within [−1, 1].
At that time, the fitness value of the position where the search wolf is located is Qip.
The direction in which the fitness value is the smallest and smaller than the current
position fitness value Qio is selected to be further advanced, and the search wolf
information Xi is updated. Comparing the artificial wolf with the minimal fitness value
Qmin with the head wolf Qlead after the end of one-time walking behaviour. If
Qmin < Qlead, the wolf with the minimal fitness value becomes the new head wolf and
initiate the summoning; otherwise, repeat the walking behaviour until a certain wolf i
with the fitness value Qi < Qlead, or the number of walks T1 reaches the maximum
number of walks T1max. It should be pointed out that each search wolf has different
fitness values so the search directions h of each search wolf is different, it will be taken
within [hmin, hmax].
(4) Raid behaviour
The head wolf initiates the summoning action by howling, calling the raid wolves to
approach the head wolf quickly. The position of the raid wolf i in the dth (d=1, 2, …,
D) dimensional space at the (k + 1)th evolution is shown in Eq. (4):
xkidþ 1 ¼ xkid þ stepdb gkd xkid =gkd xkid ð4Þ
where stepb is the step size of the raid wolf, and gkd is the position of the head wolf of
the kth generation in the dth (d = 1, 2, …, D) dimensional space.
In the process of wolves rushing, the wolf with the minimal fitness value Qmin is
compared with the head wolf Qlead after each raid. If Qmin < Qlead, the wolf with the
minimal fitness value becomes a new head wolf, the raid behaviour ends; otherwise, the
raid continues until the fitness value of a certain wolf i is smaller than the head wolf
fitness value or the raid reaches the maximum raid T2max.
(5) Siege behaviour
The raid wolves will unite all the search wolves and move to the position where the
head wolf is after the raid behaviour. For all the wolves of kth generation, assuming
that the position of the head wolf in the dth (d = 1, 2, …, D) dimensional space is Gkd ,
the siege behaviour of wolves is shown in Eq. (5):
xid ¼ xkid þ k stepdi ðGkd xkid Þ=Gkd xkid ð5Þ
where k is a random number distributed uniformly between [−1, 1]; stepi is the
adaptive siege step size of wolf i. Taking the difference between the fitness difference in
the search space upper and lower bound and the search space and the optimal value as
The Research of Improved Wolf Pack Algorithm Based on Differential Evolution 69
the reference value, a positive correlation relationship between the adaptive siege step
length and the difference between the fitness value of wolf i and the fitness value of
head wolf is established. Adaptive siege step size stepi is shown in Eq. (6):
XU XL
stepi ¼ ðQi Qlead Þ ð6Þ
maxQ0 minQ0
where maxQ0 and minQ0 are the worst fitness value and optimal fitness value of the
artificial wolf produced by the algorithm initialization.
If the fitness value of the position of the artificial wolf after the siege is smaller than
the fitness value of the original position, the position information of the artificial wolf
will be updated; otherwise, the position of the artificial wolf is not changed.
In particular: the stepa and stepb have the relationship is shown in Eq. (7):
stepa ¼ stepb =2 ¼ X U X L =S ð7Þ
where S is the step size factor, it is the physical quantity that describes the artificial wolf
to search for fine degree in the space to be searched for.
(6) Variation behaviour
After the siege behaviour, two different artificial wolf (kth generation) X1k , X2k are
randomly selected in the group, and the vector difference is assigned to the weight and
then the third randomly selected artificial wolf X3k is added. Artificial wolf variation
vector vki þ 1 for any artificial wolf target vector Xik has expressed in Eq. (8):
vki þ 1 ¼ X3k þ F X1k X2k ð8Þ
where F is a scaling factor, it’s a value between [0, 2], indicating the degree of scaling
of the vector difference. The larger of F, the greater influence of the difference vector on
vki þ 1 , which is beneficial to keeping the diversity of wolves; on the contrary, F can
promote the local refined search.
(7) Crossover behaviour
After the variation behaviour, the artificial wolf variation vector vki þ 1 and the artificial
wolf target vector Xik are mixed according to the Eq. (9) to generate the artificial wolf
test vector uki þ 1 :
(
vkij þ 1 randðjÞ CRorj ¼ randnðiÞ
ukij þ 1 ¼ ð9Þ
xkij randðjÞ [ CRorj 6¼ randnðiÞ
where j represents the jth variable of the artificial wolf, randnðiÞ 2 ½1; 2; ; D is the
identification of the randomly selected dimension variable; CR is the crossover prob-
ability factor, generally in the value between [0, 1]. The larger the CR is, the more
70 Y. Wang et al.
favourable of the algorithm performs local search and also to speed up the convergence
rate; the smaller the CR is, the more conducive to maintaining the diversity of wolves
and global search.
(8) Choice behaviour
After the crossover behaviour, the artificial wolf test vector uki þ 1 and the artificial wolf
target vector Xik are obtained. If the fitness value of the test vector is smaller than the
target vector, uki þ 1 is replaced by Xik to become the (k + 1)th generation and the
artificial wolf position is updated. Otherwise, use Xik directly as the (k + 1)th gener-
ation and update. It is shown in Eq. (10):
uki þ 1 Qðuki þ 1 Þ \ QðXik Þ
Xik þ 1 ¼ ð10Þ
Xik Qðuki þ 1 Þ QðXik Þ
For the artificial wolf vector Xik þ 1 , the artificial wolf with the minimal fitness value
Qmin should be found and compared with the head wolf fitness Qlead. If Qmin < Qlead,
the artificial wolf with the minimal fitness value is selected as the head wolf.
(9) Wolves update mechanism
The prey hunted by the wolves is not evenly distributed to every wolf and will be
distributed according to the principle of “weak meat and strong food”. This may lead to
the weak wolves to be starved to death. This principle will be simulated by removing R
wolves with worst fitness and produced R artificial wolves according to formula (1).
R is a random integer between ½n=ð2 bÞ; n=b and b is the updates scale factor of
wolves.
Through the above actions, the wolves constantly update and evolve until the
algorithm reaches the maximum number of iterations Gmax or when the optimal value
obtained by the algorithm reaches the preset precision.
range of the variable value, and the index is the dimension D of the function. The higher
the test function dimension D, the greater the computational complexity of the algorithm
and the higher the performance requirement of the algorithm.
The standard deviation, average value, optimal value, calculation success rate,
worst value, and average consuming time of the four algorithms on the 12 functions are
shown in Table 3. The average convergence curves of the functions are shown in
Fig. 1.
The Research of Improved Wolf Pack Algorithm Based on Differential Evolution 73
50 0 10
10 10 10
WPA DIWPA PSO DE WPA DIWPA PSO DE WPA DIWPA PSO DE
0 -50 0
10 10 10
10 -100 -10
10 10
-100
10
-150 -20
10 10
-150
10
-200 -30
10 10
-200
10
-250 -40
10 -250 10 10
0 500 1000 1500 2000 0 500 1000 1500 2000 0 500 1000 1500 2000
Number of iterations Number of iterations Number of iterations
(a) The average iteration (b) The average iteration (c) The average iteration
curve of Sphere curve of Matyas curve of Booth
20
50 500 10
10 WPA DIWPA PSO DE WPA DIWPA PSO DE
WPA DIWPA PSO DE
400 0
0 10
10
The optimal value/log
-20
-50 10
10
200
-100 -40
10 10
100
-150 -60
10 10
0
-200 -80
10 -100 10
0 500 1000 1500 2000 0 500 1000 1500 2000 0 500 1000 1500 2000
Number of iterations Number of iterations Number of iterations
(d) The average iteration (e) The average iteration (f) The average iteration
curve of Eggcrate curve of Trid6 curve of Sumsquares
5 10
10 5 10
10
WPA DIWPA PSO DE WPA DIWPA PSO DE
WPA DIWPA PSO DE
The optimal value/log
0 0 5
10 10 10
The optimal value/log
-5 -5
10 10 0
10
-10 -10
10 10
-5
10
-15 -15
10 10
0 500 1000 1500 2000 0 500 1000 1500 2000 0 500 1000 1500 2000
Number of iterations Number of iterations Number of iterations
(g) The average iteration (h) The average iteration (i) The average iteration
curve of Rastrigin curve of Griewank curve of Quadric
10 5
10 5 10
WPA DIWPA PSO DE 10
WPA DIWPA PSO DE
WPA DIWPA PSO DE
5 0
10 0 10
10
The optimal value/log
The optimal value/log
0 -5
10 -5 10
10
-5 -10
10 -10 10
10
-10 -15
-15
10 10 10
(j) The average iteration (k) The average iteration (l) The average iteration
curve of Rosenbrock curve of Ackley-1 curve of Ackley-2
4 Conclusion
In order to improve the optimization performance of traditional wolf pack algorithm, this
paper proposes an improved wolf pack algorithm based on differential evolution
(DIWPA). The main improvements are as follows: (1) The DIWPA introduces the search
factor of search wolf in the walking behaviour, which makes the algorithm optimization
can traverse the entire space; (2) The DIWPA introduces the maximum number of raid
wolves to improves computational efficiency and algorithm robustness; (3) the adaptive
siege step size and the differential evolution strategies is introduced to improve the global
search capabilities in DIWPA. The DIWPA is used to conduct optimization test on 12
benchmark functions and compare to 3 typical optimization algorithms. The test results
show that DIWPA has great robustness and global search ability, especially has excellent
performance in multi-peak, high-dimension, indivisible functions.
References
1. Kennedy, J., Eberhart, R.: Particle swarm optimization. In: 1995 Proceedings of IEEE
International Conference on Neural Networks, vol. 4, pp. 1942–1948. IEEE (2002)
2. Tehsin, S., Rehman, S., Saeed, M.O.B., et al.: Self-organizing hierarchical particle swarm
optimization of correlation filters for object recognition. IEEE Access PP(99), 1 (2017)
3. Wei, L.X., Li, X., Fan, R., et al.: A hybrid multi-objective particle swarm optimization
algorithm based on R2 indicator. IEEE Access PP(99), 1 (2018)
4. Singhal, P.K., Naresh, R., Sharma, V.: Binary fish swarm algorithm for profit-based unit
commitment problem in competitive electricity market with ramp rate constraints. Gener.
Transm. Distrib. IET 9(13), 1697–1707 (2015)
5. Li, X.L., Shao, Z.J., Qian, J.X.: An optimizing method based on autonomous animate: fish
swarm algorithm. Syst. Eng.-Theory Pract. 22, 32–38 (2002)
6. Li, Y., Zhang, C., Yang, Q., et al.: Improved ant colony algorithm for adaptive frequency-
tracking control in WPT system. IET Microwaves Antennas Propag. 12, 23–28 (2017)
7. Wu, L.H.: Differential Evolution Algorithm and Its Application. Hunan University (2007)
8. Yang, C., Tu, X., Chen, J.: Algorithm of marriage in honey bees optimization based on the
wolf pack search. In: International Conference on Intelligent Pervasive Computing, pp. 462–
467. IEEE Computer Society (2007)
9. Wu, H.S., Zhang, F.G., Wu, L.H.: A new swarm intelligence algorithm—wolf pack
algorithm. Syst. Eng. Electron. 35(11), 2430–2438 (2013)
10. Zhang, L., Zhang, L., Liu, S., et al.: Three-dimensional underwater path planning based on
modified wolf pack algorithm. IEEE Access 5(99), 22783–22795 (2017)
11. Wang, J.Q., Jia, Y.Y., Xiao, Q.Y.: Wolf pack algorithm in the optimal operation of
hydropower station reservoirs. Adv. Sci. Technol. Water Resour. 35(3), 1–4 (2015)
12. Hui, X.B., Guo, Q., Wu, P.P., et al.: An improved wolf pack algorithm. Control Decis. 32(7),
1163–1172 (2017)
13. Qian, R.X.: A wolf algorithm based on cultural mechanism. Inf. Technol. 2015(12), 98–102
(2015)
14. Cao, S., Jiancheng, A.: Otsu image segmentation method for improved wolf group
optimization algorithm. Microelectron. Comput. 34(10), 16–21 (2017)
15. Solano-Aragón, C., Castillo, O.: Optimization of benchmark mathematical functions using
the firefly algorithm with dynamic parameters. In: Castillo, O., Melin, P. (eds.) Fuzzy Logic
Augmentation of Nature-Inspired Optimization Metaheuristics. SCI, vol. 574, pp. 81–89.
Springer, Cham (2015). https://doi.org/10.1007/978-3-319-10960-2_5
Multi-objective Optimization Genetic Algorithm
for Multimodal Transportation
1
School of International Business, Sichuan International Studies University,
Chongqing 400031, China
[email protected]
2
College of Mechanical Engineering, Chongqing University, Chongqing 400044, China
1 Introduction
In order to facilitate modeling and analysis, the virtual network is defined as follows.
Let G = (V, E, M) be the network graph under consideration. It is the abstraction of
logistics nodes and traffic connections in the geographical map under consideration. V
is a set of the nodes in the actual logistics transportation network correspond. E is a set
of edges, corresponding to all possible transportation paths. M is a collection of trans‐
portation modes. To facilitate problem solving, the following assumptions are made.
Hypothesis 1 the same shipment cannot be transported separately;
Hypothesis 2 there are multiple modes of transport between two adjacent transfer
stations to choose from, but only one mode of transport can be selected;
Symbol description:
{ }
∑∑ ∑∑∑
Z2 = min dijm ∕vm + max(Di , qijm ttikl ) (1)
ij m ij i kl
S.T.
⎧Q if i = O
∑ ∑ ⎪
qijm − qjim = ⎨ −Q if i = D ∀ i ∈ V, m ∈ M (2)
j i ⎪0 otherwise
⎩
80 X. Guiwu and X. Dong
li ≤ Ai ≤ ui , ∀ i ∈ V (4)
∑
xijm ≤ 1 ∀ (i, j) ∈ E, m ∈ M, (5)
m
{
1 k≠l
𝜔kli = , (6)
0 k=l
The first optimization objective function in Eq. (1) is to reduce the total transportation
cost, and the second optimization goal is to reduce the total transportation time. Equa‐
tion (2) is the traffic balance condition in the network flow. Formula (3) ensures the
tolerance of time requirements. Equation (4) represents that the node should meet the
predetermined time window constraints. Formula (5) indicates that there is at most one
transport mode for each link. The values of variables (6) and (7) are limited.
In order to make the multimodal transportation model closer to the real scene, we studied
the customized time window, the transportation distances of different transportation
modes, different transportation speeds, the conversion costs and conversion time of
different nodes, which is different from the previous issue of multimodal transportation
routes. There are multiple edges between some nodes. Moreover, once the mode of
transportation has changed at certain nodes, additional conversion costs and time are
still included. The problem has become very complicated, and the key is to find it difficult
to find a best multimodal transport network. Therefore, a two-layer multi-objective
hybrid Taguchi algorithm is proposed to find the optimal transport path.
The dual-layer multi-target hybrid Taguchi algorithm consists of two layers. The
upper layer aims to find the overall global optimal Pareto frontier and transportation
mode, while the lower layer searches for the local optimal Pareto frontier of the fixed
path and transport mode. The lower level optimization method determines the upper
feasible solution space. The multimodal transport path problem model can be reduced
to the following two-level multi-objective optimization problem.
min Z = {Z1 , Z2 )
(xijm ,𝜔kli )
S.T.
⎧Q if i = O
∑ ∑ ⎪
qijm − qjim = ⎨ −Q if i = D ∀ i ∈ V, m ∈ M, (9)
j i ⎪0 otherwise
⎩
li ≤ Ai ≤ ui , ∀ i ∈ V, (11)
{
1 k≠l
𝜔kli = , (12)
0 k=l
∑
xijm ≤ 1 ∀ (i, j) ∈ E, m ∈ M, (13)
m
Since the position of the Pareto front before the optimization is not known, it is hoped
that the chromosomes of the initial population can be evenly dispersed throughout the
feasible solution space, and there is a potential advantage of using the Taguchi experi‐
mental method to generate these chromosomes.
Algorithm 3. Generation of the initial population of transport modes
Step 1: Read the path provided by the upper layer
82 X. Guiwu and X. Dong
Step 2: All the edges in the path have the same transport method available? If appli‐
cable, apply to select chromosomes and go to step 4. If not, go to step 3
Step 3: Divide the path into several sections based on the same available transport
method. Sides with the same available transport methods are connected together.
Each section chooses its own orthogonal table.
Step 4: Calculate the shipping cost and time for each chromosome.
Step 5: Sort the chromosomes according to the fast non-dominated sorting method.
Step 6: Select the chromosome based on chromosome rankings.
The Taguchi experiment method also incorporates crossovers in order to inherit good
genes from the father. This article uses a two-level orthogonal array with l factors, where
l is the length of the feasible path. Two levels correspond to the parents of two chro‐
mosomes.
Algorithm 4. Crossing based on taguchi experimental method
Step 1: Randomly select the parents of the two paired pools;
Step 2: Select the orthogonal table according to the length of the chromosome;
Step 3: Use orthogonal experiments to generate progeny;
Step 4: Sort children according to the fast non-dominated sorting method;
Step 5: Select two chromosomes based on the chromosomes.
End
Step 5: Select: Intersect each chromosome with crossover probability. Continue to
choose until the selected number of chromosomes reach to half of it
Step 6: Cross: Crosses the selected chromosome. If the crossed path is feasible, then
go to Step 8
Step 7: Execute Algorithm 2
Step 8: Variation: The probability that each chromosome will mutate. If the crossover
path is feasible, then go to Step 10.
Step 9: Execute Algorithm 2
Step 10: Use the fast non-dominated sorting method to sort all chromosomes
Step 11: Preserve the next chromosome of the next generation
End
4 Numerical Simulation
In order to verify the models and algorithms constructed, based on the logistics network
in Eastern China, it is assumed that there are a batch of containers that need to be trans‐
ported. The transport can be carried out in 35 city nodes. The transportation methods
include roads, railways and waterways, but not all cities have three links. An example
was constructed with 35 nodes and 136 edges. Figure 1 shows the map of logistics and
transportation in eastern China. Many major cities along the Yangtze River rely on
waterways, railroads and trucks for transportation. There is a simple and effective model
in the multimodal network environment. The model can provide some good trade-off
solutions. These nodes next to the Yangtze River can be connected to each other by
barges, such as to nodes 1, 3, 9, …, 93. The following parameters
G1 = 220, pc1 = 0.1, pm1 = 0.1, G2 = 50, pc2 = 0.2 and pm2 = 0.2 are used in the
simulation. The stop condition is simply to reach the highest iteration of the predeter‐
mined upper and lower levels. The two predetermined maximum iterations are 20
generations.
There are three common modes of transportation, namely roads, railways and water‐
ways. The total transportation costs include transportation costs and transportation costs.
84 X. Guiwu and X. Dong
The mode of transportation determines the cost of transportation. Transit costs include
labor costs and equipment costs during transportation. The transfer container is a
standard container, 8 feet high, 8 feet wide and 20 to 55 feet long. Containers can be
transported by trucks, railroads and barges. In general, containers can hold 15 tons.
Because there is no universal standard to describe the shipping rate, transshipment rate,
transportation speed and transit time. This article uses the relevant data provided by [14]
as the basis for calculation, as shown in Tables 1, 2 and 3. The transport rates and speeds
given in Table 1 are shown in Tables 2 and 3, respectively.
Consider one task is 1000 containers from node 1 (Shenzhen) to node 93 (Dandong).
The goods are more focused on the quality of transportation. Assuming that the trans‐
portation continues until it reaches the destination, simulations are performed. The
results are shown in Fig. 2. The Pareto solution and other non-dominated optimal solu‐
tions are given in the figure. These solutions become alternatives to the carrier’s decision.
The carrier needs to select according to the shipper’s requirements (Table 4).
Multi-objective Optimization Genetic Algorithm 85
5 Conclusion
References
1. Lozano, A., Storchi, G.: Shortest viable path algorithm in multimodal networks. Transp. Res.
Part A 35(3), 225–241 (2001)
2. Boussedjra, M., Bloch, C., EI Moudni, A.: An exact method to find the intermodal shortest
path. In: Proceedings of the 2004 IEEE International Conference on Networking, Sensing &
Control, Taiwan, China (2004)
3. Brands, T., Wismans, L.J.J., Berkum, E.C.V.: Multi-objective optimization of multimodal
passenger transportation networks: Coping with demand uncertainty. In: Papadrakakis, M.,
Karlaftis, M.G., Lagaros, N.D. (eds.) An International Conference on Engineering and
Applied Sciences Optimization, Kos Island, Greece, 4–6 June 2014, pp. 547–561 (2014)
4. Brands, T., Berkum, E.C.V.: Performance of a genetic algorithm for solving the multi-objective,
multimodal transportation network design problem. Int. J. Transp. 2(1), 1–20 (2014)
5. Zhang, Y., et al.: A bi-objective model for uncertain multi-modal shortest path problems. J.
Uncertainty Anal. Appl. 3(8), 1–17 (2015)
6. Osuji, G.A., Okoli Cecilia, N., Opara, J.: Solution of multi-objective transportation problem
via fuzzy programming algorithm. Sci. J. Appl. Math. Stat. 2(4), 71–77 (2014)
7. Yunhe, Z., Dong, L.B.L., Hongyan, G.: Research on a generalized shortest path method of
optimizing intermodal transportation problems. J. China Railway Soc. 28(4), 22–26 (2006)
8. Wei, H., Li, J., Liu, N.Z.: An algorithm for shortest path with multimodal in time - varying
network. Chin. J. Manag. Sci. 14(4), 56 (2006)
9. Jianyong, Z., Yaohuang, G.: A multimodal transportation network assignment model. J. China
Railway Soc. 24(4), 114–116 (2002)
10. Tao, W., Gang, W.: A combined optimization model for transportation modes of multimodal
transport. Eng. Sci. 7(10), 46–50 (2005)
11. Zhong, W., Jinsheng, S., Ailing, H., et al.: Research on model of the shortest time path and
transport cost in multimodal transportation. Eng. Sci. 8(8), 61–64 (2006)
12. Wendong, Y., Wenfang, W.: Analyzing and modeling of multimodal transportation with time
window. J. Nanjing Unv. Aeronaut. Astronaut. 41(1), 111–115 (2009)
13. Guiwu, X., Yong, W.: Optimization algorithm of multimodal transportation with time
window and job integration of multiagent. J. Syst. Eng. 26(3), 379–387 (2011)
14. Manuel, D., Rossetti, H.N.: WebShipCost - Intermodal Transportation Linkage Cost
Assessment Via the WWW. The Mack-Blackwell Transportation Center (2005)
Big Data Analytics
The Research of 3D Power Grid Based on Big
Data
Abstract. The visual graphics can reflect the hiding relationships between
different data by color, location, and the topological structure. This paper pro-
poses a 3D power grid framework by integrating big data and 3D visualization
technologies to achieve flexible, multidimensional, dynamic, and panoramic
display of power data. The proposed framework consists of data acquisition,
data analysis and mining and 3D visualization module. We detailed the 3D
method from modeling, management and visualization techniques, and analyzed
the modeling elements from geospatial data, electric graphic data, electric
attribute data and data mining result dataset. Finally, we give two scenarios of
3D power system. Most of tools used in this paper are open source software.
This can ensure the system stability, flexible and easy to use for the management
of large distributed data resources and avoid duplication system development.
1 Introduction
As the power system becomes more and more complex and complicated, the control of
the safe and stable operation of power grid becomes very complicated. In order to
improve the operation level of power grid, the power system imposes higher require-
ments on high-performance computing and visualization technologies. But, due to the
heterogeneous nature of power systems, it’s difficult to achieve interoperation between
various power databases. This results in the phenomenon of information islands [1].
However, a new trend in IT service is managing, analyzing and visualizing the multi-
source power big data by high-performance computer clusters to construct smart power
grid [2–5]. Big data has 4 V characteristics, focusing on the fusion, analysis and mining
of multi-source and multi-type data. It has high data throughput speed, distributed
computing, and multi-source data association mining analysis capabilities. Therefore,
the characteristics of big data technology fit the needs of massive power data analysis
and calculation.
Currently, 2D visualization technology has been widely used in the power industry.
The power data is displayed by two-dimensional lines, charts, scatter plots and images
[6, 7], or by geographic information system (GIS) [8–10]. However, big data tech-
nology and GIS technology have been decentralized in the power sector. Also, 2D GIS
is based on two-dimensional abstract graphical symbols, which have limited ability to
present information to users. And it’s difficult to accurately deal with many 3D
problems in 2D GIS. Therefore, if the high-performance computing capability of big
data technology can be effectively combined with geographical 3D visualization
technology, the real-time power grid operation and control will become clear, intuitive
and effective. Therefore, this paper represents a power grid framework based on big
data and 3D technologies to achieve flexible, dynamic visualization of power data.
This section introduces a framework based on big data technologies and 3D techniques
to manage and visualize large amounts of power data. This framework consists of data
acquisition and conversion modules, big data platform, power data analysis and mining
modules, and 3D visualization modules.
(1) Data acquisition and conversion modules: in the proposed framework, power data
is mainly extracted from the internal system [11], for example, wide area mea-
surement system, energy management system, device detection and monitoring
system, etc. External data is collected from socio-economic, urban population,
geographic information system. The two kinds of data are imported into big data
platform by three ways. Traditional relational data is converted by Apache Sqoop
[12]. Streaming data or timing data is processed by Apache Storm [13].
Unstructured data could be imported by Apache Flume [14]. Specifically, the data
which will be managed and processed in big data platform includes power
topology elements, 3D models, related attribute data (such as electric equipment
information) and geospatial data (e.g. streets, buildings, roads, water, etc.)
(2) Big data platform: We use Apache Hadoop [15] to build big data platform. The
underlying storage form is HDFS which will be analyzed directly. In order to
facilitate the scheduling and management of cluster resources, we adopt YARN
[16] as the manager. Above data management module is data processing services,
including off-line processing, machine learning, memory computing. Offline pro-
cessed data can be stored by HBASE [17]. HIVE [18] generates SQL query
statement and submits to other analysis and processing modules. Offline processing
is mainly used to meet non-real-time requirements. But for time-critical computing,
we use memory computing, such as Impala [19] or Spark [20].
(3) Data analysis and mining modules: Multi-source heterogeneous data often implies
valuable relationship. For this purpose, machine learning library (such as Spark R
and Spark Mllib) is employed. Except for providing simple data query and general
analysis capabilities, this framework includes the 3D modeling method for domain
application, For example, the data mining results of equipment state evaluation,
risk assessment, load capacity assessment and fault prediction. The modeling
results will be submit to visualization client (such as web browser, desktop 3D
viewer) (Fig. 1).
(4) 3D visualization of analysis modules: 3D visualization module is the core of this
framework. It consists of two parts. One part is power data spatialization module,
the other is the power data visualization module. Spatialization module access to
The Research of 3D Power Grid Based on Big Data 91
JDBC ODBC
Big Data
Platform Cluster Resource Scheduling Management (YARN)
power big data platform through JDBC, ODBC or Hive. The spatialized objects
include data mining and analysis results and static attribute data. Then, the spa-
tialized results are transferred into three-dimensional symbols (such as points, lines,
and surfaces) and annotations. Finally, three-dimensional client utilizes digital
terrain models, remote sensing images, power equipment, 3D symbols and anno-
tations to construct a panoramic virtual scene.
3 3D Techniques
Comparing with the 2D power GIS, 3D modelling process is complex and requires a
combination of different software to construct power and geospatial objects. This
section will detail the visualization techniques for 3D power system (Fig. 2).
• SketchUp [21] is an open source software for 3D modelling. It has been used in a
wide range of applications such as architectural and civil design. Power tower and
electric equipment can be easily modelled in SketchUp. However, due to lack of
geographical projection information, 3D models created by SketchUp can’t be
directly import into 3D power system. In order to ensure that the model can be
92 B. He et al.
Internet
HTML JavaScript Java Applet
Explorer
Mobile Client Desktop Client
Visualization Techniques
WebGL OpenGL
OpenLayers
Map,GML CityGML
Google SketchUp
3D Geodatabase Skyline TerraExplorer
GeoServer 3D Models
(3DCityDB)
ESRI CityEngine
Autodesk Revit
PostGIS
Modeling Techniques
displayed properly, the usual practice is firstly defining the local coordinates of the
area to be modelled, and then convert it to system’s geographic coordinate.
• Skyline TerraExplorer Pro [22], ESRI CityEngine [23] are two commercial 3D GIS
desktop viewer and creator. They provide a wide range of powerful terrain analysis
tools to improve decision making. They can manage and visualize geospatial fea-
tures, imageries, elevation raster, urban models and dynamic objects. These data
could be published and utilized by web or desktop client. Secondary development
power 3D system could have rich functionality based on this two software, and the
development process is easier.
• 3D GIS has been used to the acquisition, management, analysis in macroscopic
geographical environment, and show buildings generalization appearance and
location. It’s unable to obtain further constructional details (such as indoor pattern,
various pipelines and facilities configuration and materials, etc.). The inner com-
ponents of 3D objects haven’t explicit spatial relationship. This results in today’s
power 3D GIS staying in viewing three-dimensional appearance. For the city power
3D map based on big data technologies, the detailed power information is very
helpful for power management staff to know the panorama operation status of
power network. Autodesk Revit [24] software is a such tool to build Building
Information Modelling (BIM). In order to achieve three-dimensional multi-scale
power map visualization, it needs BIM models integrated into 3D GIS.
3D models sharing. Currently model data management often adopts relational database
such as SQL Server, Oracle, and associates model files with database to achieve read
and display. For the model retrieval, each model will be classified, labelled and added
property, and then establish text index database to achieve quick search. This method is
inefficient. It requires an efficient way for large data management in big data platform.
The 3D City Database (3DCityDB) is a free Open Source package consisting of a
database schema and a set of software tools to import, manage, analyse, and visualize
virtual 3D city models according to the CityGML standard. 3D power models can be
transformed to CityGML. The geometry, topology, semantics between the different
classes of the power objects are kept in the CityGML file. Then, 3DCityDB import tool
could be used to save the CityGML file into relational database, and tile the large,
complex 3D power objects to easily visualize in virtual globes like Google Earth or
Cesium WebGlobe.
We3D Visualization
WebGL OpenGL
(2) Electric graphic data: this type of data includes geographic wiring diagram for
power facilities such as (overhead lines, cable lines), equipment (substations,
switching stations, towers, substations, transformers, switches, etc.). The other
three types of data could be associated with electric graphic data to build a 3D
panoramic visualization environment (Fig. 5).
(3) Electric attribute data: include monitoring data, power sales data, social statistics,
and etc.
(4) Data Analysis and Mining Results: Power system management process involves a
lot of data analysis. This type of data sets comes from load capacity assessment,
equipment state evaluation and fault prediction, equipment risk assessment, etc.
3D Modeling Method
3D Symbols 3D Spatialization
will be displayed by setting colors, geometries and symbol size by analysed data value.
Dynamic spatialization is used to visualize the data analysis and mining results. The
obtained graphics data sets will be directly graded and classified for visualization. The
spatialization method includes inverse distance weighted method, ordinary kriging,
cubic splines, and etc. The visualization method includes categories (unique value,
unique value with many Fields), quantities (graduated colors, graduated symbols,
proportional symbols), charts (pie, bar, stacked), and multiple attributes (Fig. 6).
5 Scenarios
This section will describe the using of 3D power visualization framework in the fol-
lowing two scenarios:
uses Storm to process it. According to the purpose of the application, the visualization
method is divided into three types: abnormal state detecting, real-time operation state
visualization, post hoc analysis. During the abnormal status detecting process, the
abnormal value combine with the devices coordinate will be transmit to visualization
client and symbolized by abnormal types to achieve equipment malfunction status
alerts. Real-time status could be visualized by time interval and associated with the
observed value of the related equipment and power lines. The status value need to the
space-time transformation to convert the indicator data or time data into three-
dimensional space, and the format index axis is the Z-axis three-dimensional graphics,
and can also be used hierarchical, categorical displayed. The post hoc analysis is
focused on the need for large amounts of data analysis requirements. According to the
specified interval, the analysis results will be symbolized, colored and loaded into 3D
scene.
6 Conclusions
Acknowledgment. The work described in this paper is supported by the National High Tech-
nology Research and Development Program of China (863 Program) (2015AA050201), and the
Chinese National Natural Science Foundation (41701167) and (6433012). We gratefully
acknowledge the valuable cooperation and help of Guangzhou Power Supply Co. Ltd.
References
1. Liu, Q.: Data fusion and application of power equipment based on big data on Panorama.
Informatization Constr. (9), 63–65 (2015)
2. Zhou, J.Z., Hu, R.Q.Y., Qian, Y.: Scalable distributed communication architectures to
support advanced metering infrastructure in smart grid. IEEE Trans. Parallel Distrib. Syst. 23
(9), 1632–1642 (2012)
3. Fang, X., Yang, D.J., Xue, G.L.: Evolving smart grid information management cloudward: a
cloud optimization perspective. IEEE Trans. Smart Grid 4(1), 111–119 (2013)
4. Vu, H., Liu, J.K., Huang, X.: A secure cloud computing based framework for big data
information management of smart grid. IEEE Trans. Cloud Comput. 3(2), 233–244 (2014)
5. Maheshwari, K., Birman, K., Wozniak, J., Zandt, D.V.: Evaluating cloud computing
techniques for smart power grid design using parallel scripting. In: Proceedings of 2013 13th
IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGrid),
Persistent Link, pp. 319–326 (2013)
6. Ravikumar, G., Pradeep, Y., Khaparde, S.A.: Graphics model for power systems using
layouts and relative coordinates in CIM framework. IEEE Trans. Power Syst. 28(4), 3906–
3915 (2013)
7. W.Zhangjin,H.Xingwen,D.Ying,Z.Rui,Z.Qingguo, “A CGI + AJAX + SVG based moni-
toring method for distributed and embedded system,” in Proc. 2008 IEEE 1st Int. Conf. UBI-
Media Computing,2008,pp. 144–148
8. Tang, K., Zhong, Y.J., Wu, H.F., He, P.Y., Xu, L.J., Qi, W.Q.: Design and implementation
of GIS basic data quality management tools for power network. Electr. Power Inf. Commun.
Technol. 14(2), 98–101 (2016)
9. Bai, Y.Q., Nie, J., Chen, Y., Li, Z., Chen, T.: Construction of GIS platform for Chongqing
grid. J. Geomatics 37(5), 73–75 (2012)
98 B. He et al.
10. Wei, W., Liu, B.H., Dai, J.Z., Xia, H., Li, Z.: Design and implementation of GIS-based
Wuhan Grid management systems. China Electr. Power 37(1), 161–163 (2009)
11. Zhang, D.X., Miao, X., Liu, L.P., Zhang, X., Liu, K.Y.: Research on development strategy
for smart grid big data. Proc. CSEE 35(1), 2–12 (2015)
12. Apache Sqoop. http://sqoop.apache.org/
13. Apache Storm. http://storm.apache.org/
14. Apache Flume. http://flume.apache.org/
15. Apache Hadoop. http://hadoop.apache.org/
16. YARN. http://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/YARN.html
17. HBASE. http://hbase.apache.org/
18. HIVE. http://hive.apache.org/
19. Impala. http://impala.io/
20. Spark. http://spark.apache.org/
21. SketchUp. http://www.sketchup.com/
22. Skyline TerraExplorer Pro. http://www.skylinefrance.com/en/home
23. ESRI CityEngine. http://www.esri.com/software/cityengine/
24. Autodesk Revit. http://www.autodesk.com/products/revit-family/ overview
25. 3D City Database. http://www.3dcitydb.org/3dcitydb/3dcitydbhomepage/
26. Cesium. https://cesiumjs.org/
27. WebGL Earth. http://www.webglearth.com/
28. World Wind. https://goworldwind.org/
29. ossimPlanet. https://trac.osgeo.org/ossim/
Identifying Service Gaps from Public Patient
Opinions Through Text Mining
1
School of Management, Chongqing Technology and Business University,
Chongqing 400067, China
[email protected], [email protected]
2
School of Engineering, Cardiff University, Cardiff, CF24 3AA, UK
1 Introduction
At the present stage, it is widely accepted that collecting, responding and processing the
comments from customers is extremely important to the organizations. The feedback
from customers has been a crucial source to evaluate the quality of service and products.
With the development of the technology, reporting their experience and public opinions
on the Internet has become a popular way for patients and customers [1]. In the area of
the service industry, large organizations have performed a huge number of studies to
discover the reasonable approaches to process the comments from consumers. Sentiment
analysis and text mining based on the natural language processing and machine learning
algorithms play a critical role in these studies [2]. Obviously, it could be meaningful if
sentiment analysis can be conducted in the field of healthcare. Through this, the infor‐
mation of sentiment and emotion can be captured from the textual data of patients’
experience. Academic research in this area has largely focused on establishing the
models with higher accuracy for ‘automatically predicting the sentiment embedded
within the text’. Additionally, based on the research of the National Health Service
(NHS) (2013) in England, the sentiment analysis witnessed it could become the novel
resource for patients when they were making choice of hospitals. Based on this, the test
for patients’ satisficing to the service of hospitals will be conducted by analysis a vast
amount of textual data on the Patient Opinion website in this project. Three questions
will be focused on, which are whether this hospital could be proposed by the specific
patient, whether the attitude of the staff is appropriate and whether the waiting time is
reasonable.
Textual comments have attracted considerable attention for the past few years. Literature
highlights that marketing department had used manual and machine learning methods
to process the information. The deeper understanding of the comments can be gained if
the analysis is performed with the manual manners. However, the review for the huge
quantity of text might be inconsistent without ‘procedural models’ [4].
In terms of the medical service, online health communities and the patient-authored text
they contained, accompanied by the potential value, have attracted more and more atten‐
tion. Graebner et al. (2012) suggested that the major direction of consumer comments anal‐
ysis with text mining has basically focused on establishing more accurate models for fore‐
casting the sentiment embedded within the textual recommendations. Previous studies
showed that the linguistic text mining approaches had better performance than that of
manual methods in the categorization of the comments because they performed better in the
precision of prediction. In the field of medicine, enterprises like Google Flu [5] and
HealthMap [6] have already demonstrated that the PAT (patient-authored text) is a depend‐
able source for illness tracking and medical research. Additionally, websites such as Cure‐
Together and PatientLikeMe have put great effort into the research for effect exerted by the
patient opinions on the healthcare service.
A large number of works focused on the sentiment analysis, which was also regarded
as opinion mining. It involved the study of the analysis of person’s opinion, sentiments
to the product, services etc., and the factors influence them. Machine learning means the
Identifying Service Gaps from Public Patient Opinions 101
model is trained on the structure of the data. In terms of sentiment analysis, the literature
displays several key aspects of it. The bag-of-words approach can be used to process
the text with unigrams, bigrams and trigrams in the analysis. With consideration of
potential emotional comments, the sentiment could be determined in several ways by
the algorithm. As can be seen in the research of the feedback of the healthcare experience
on Twitter in the US, monitoring social media can obtain the meaningful, unsolicited
and real-time data, while the traditional survey system cannot capture it. One of the most
important reasons is social media can provide the opinions from patients that cannot be
obtained from other routes with the cautious observation [7].
Besides those manners of classification, automatic topic detection based on docu‐
ment clustering is an alternative approach for extracting health-related hot topics in
online communities [8]. Patterson (2011) illustrated that document clustering is a valid
manner to clearly differentiate the groups of clinical narratives. Recently, many studies
applied clustering manners to this user-generated medical text to discover the topics that
interest patients who post their experience online. The study of Denecke focused on
blogs related to healthcare and separated the topics of these blogs into two groups:
informative and affective [9]. Brody (2010) processed the texts based on Latent Dirichlet
Allocation (LDA) topic models to explore the salient directions of online evaluations of
healthcare systems. Therefore, it is valuable to perform the clustering approaches in text
mining.
In summary, the main mission of the service industry, such as the healthcare systems,
is offering the satisfactory service to meet the requirements of the recipients. Addition‐
ally, since people rely increasingly upon the internet and social networks, the amount
of the text present online continues to rise sharply, which provides organizations with
a source to exploit business intelligence. Moreover, collecting and analyzing consumer
comments witness their importance have a continuous increase. Hence improving the
service quality based on the advice from the Internet and social media has drawn signif‐
icant attention because of its vast potential value, outstanding convenience and high
efficiency. Previous literature highlights several research limitations, such as the manual
preprocess, size of the dataset and the variety of specialities, which also point out the
directions of further research.
2 Methodology
The data analysis techniques were conducted to the textual comments from Patient
Opinion about the service experience of cancer patients. The objective was to test
whether the prediction for patients’ views on several topics could be obtained automat‐
ically based on their feedback. The machine learning classification approach was chosen
and trained from the labelled data, which was regarded as the training group. And the
comments can be classified into various categories based on the given examples, with
the help of the open-source data mining software—Weka, which had been widely
adopted in the previous researches of text mining in many domains. To scale the relia‐
bility of the experiments, the results of this research were compared with the quantitative
rating provided by the Patient Opinion. There were two domains of the service of
102 M. Tang et al.
healthcare predicted, which were waiting for time and attitude of the staff. In terms of
clustering, after converting the textual data into the valid format of Weka, the expectation
maximization (EM) clustering was conducted.
Sentiment analysis, which is also named opinion mining, is to process people’s opinions
and extract the sentiments towards the products and services [10]. Machine learning is
about learning from the structure of predefined data and discover the principles and
functions can process the unlabelled information [11]. There are two main types of
machine learning algorithms, which is supervised and unsupervised respectively. Senti‐
ment analysis or text classification should be involved by supervised fields, which means
that the training data should have already had the class labels by the start of the operation.
The fundamental process of supervised machine learning method for sentiment analysis
was shown in Fig. 1, the labelled data should be provided as the training set and the
algorithm can learn and output the trained model. The reliability of the model should be
tested with the unlabelled data.
Identifying Service Gaps from Public Patient Opinions 103
Labelled Data
(Training set) Unlabelled Data
Machine Learning
Algorithm Model Trained Model
Prediction
Fig. 1. The fundamental process of supervised machine learning method for sentiment analysis
high scores were ‘ranked and combined with expert opinions’, the topics of the clusters
were identified.
N
f (w, ci) ∗ log (1)
|{cj|f (w, cj) ≥ f (w, ci), j = 1, 2, … , N|}|
To reduce the processing time of classification, the number of words in the learning
process was limited to no more than 10,000 based on the frequency of each word. After
the experiment, the prediction performance of every method was compared with that of
others. The results outputted by Naïve Bayes Multinomial, J48 Decision Trees and
Support Vector Machine of 3 domains of healthcare service (overall evaluation to the
service, whether the communication is good and whether the waiting time is reasonable)
were shown as follow (Tables 2 and 3).
Identifying Service Gaps from Public Patient Opinions 105
Table 2. The results of the overall evaluation, waiting time, communication and the overall
evaluation without ‘Stop Word List’
Overall rating Waiting time Communication Overall rating without ‘Stop
Word List’
NBM J48 SVM NBM J48 SVM NBM J48 SVM NBM J48 SVM
Accuracy 86.60% 76.30% 78.60% 80.20% 80.30% 81.60% 83.60% 81.60% 80.70% 85.90% 72.60% 72.90%
F-measure 0.87 0.81 0.82 0.83 0.85 0.82 0.85 0.8 0.81 0.85 0.81 0.8
Recall 0.87 0.81 0.82 0.83 0.85 0.82 0.85 0.8 0.81 0.85 0.81 0.8
Time (s) 0.08 226 316 0.03 36 61 0.03 92 106 0.09 252 332
Based on the results, the NBM algorithm observed that it had the best performance
in the overall assessment. Then, obviously, the accuracy of 3 approaches changed in the
experiment for ‘waiting time’ and ‘communication’. On the one hand, the J48 Decision
Trees and Support Vector Machine performed better, which increased roughly 3%. The
comparison also suggests that the NBM could perform better with the large database.
Conversely, the J48 DT and SVM algorithms could outperform the NBM if the size of
the dataset is not large enough. In terms of the influence of ‘StopWordList’, it displayed
that the list had a helpful effect on the work of the algorithms. All 3 methods became
more accurate with using the Weka’s list of stop words. Moreover, the F-measure and
Recall index also increased slightly with the time taken to complete the process
becoming less. For the final prediction of the test set (Table 4), there was 77% comments
showed the service satisfied the patients, which was fundamentally in a similar range
with the quantitative survey and the statistical analysis of Patient Opinion.
The results indicate that the output of the sentiment analysis of the comments can
be efficient and have reasonable accuracy [15]. And it can accomplish the effective
extraction and identification of the critical issues of the review. It can also prove that
there are dramatic meaningful information in the unstructured feedback of patients and
it could significantly benefit the organizations if processed appropriately. With the help
of the sentiment analysis, in both overall view and certain domains, an aggregate eval‐
uation of objectives can be provided.
In the ‘staff’ group, basically, the patients’ views to doctors and staffs were satis‐
factory. There was one word deserved to be mentioned - ‘professional’, which was
mentioned more times in writers’ praises than other words. It suggested that the patients
may pay more attention to attitudes and skills in the professional field instead of char‐
acters and personalities. On the negative side, patients and reporters often criticized the
topic of ‘shortage’, which showed that the shortage of staffs was still the critical problem
in the healthcare system. Comparing the topics around ‘staff’, the second group labelled
‘communication’ was comprised of features much more specific. ‘Respect’ and ‘effi‐
cient’ are two domains with which the criticisms were most concerned. There were also
4 words displayed the drawbacks of the communication, which was ‘poor’, ‘ignored’,
‘lack of’ and ‘no’, and showed the ignorance, might partly as a result of overworked
status, of the staffs and nurses is the most serious problem. In terms of ‘care’ label, it
involved the key features of the care service itself and the results observed the similar
situation. The praises from patients were almost full of the texts about care and
‘patience’, ‘empathy’, ‘dignity’ and ‘compassion’ are the aspects with the most concern
of recipients. The number of comments about nurses was considerably more than that
about other workers in the healthcare systems, which made the nurse become the
Identifying Service Gaps from Public Patient Opinions 107
essential part of the hospitals. Among all of these words, ‘patience’ played the most
important role in service provided by nurses. There was the other special label is ‘end
of life’, which represented the status of cancer patients in their last time. These comments
were often written by their relatives and families. The praise also occupied most of the
texts and the most important word in this field was ‘respect’, which suggested, to the
patients in their final time, respect was crucial to them.
To discover the drawbacks of the healthcare service, the comments with negative
sentiment in the fields of ‘waiting time’ and ‘communication’ were selected as our
objectives. Obviously, it can be observed that patients with emergency and appointments
were more likely to complain about the waiting time. Since the words involve ‘hour’
were much more the phrases have ‘minute’. Therefore, the further study of the limit
needs to be performed here. In the area of ‘communication’, which the ignorance and
rude manners of staffs mentioned most times in the comments.
Above all, professional skills of staff, nurses’ patience and the manners of commu‐
nication have drawn the most attention from patients. The shortage of staff, overworked
schedules and ignorance for an effective communication are the problems that need to
be improved immediately.
4 Conclusion
This work demonstrates that it is possible for organizations to predict the evaluation
embedded with the huge amount of comments through sentiment analysis and explore
the hot medical related topics through clustering analysis. By conduct these approaches
based on text mining, several valuable conclusions can be drawn from the obtained
results. Most of the reporters from Patient Opinion have a basically satisfactory expe‐
rience but the comments mention the ‘waiting time’ observe that the negative feedback
occupies the larger part than the positive group, which is mostly because of the appoint‐
ment and emergency. Then, in the field of communication, more than 70% of the
comments tagged with positive sentiment and efficiency and manners of staff play the
crucial roles in communication between patients and hospitals. The criticism for the
communication mainly focuses on the ignorance of the help workers. In addition, for
nurses, patients and their families, they pay more attention to their patience while doctors
and staffs are praised mainly because of their professional skills. Finally, respect plays
the most important role in the healthcare for cancer patients in their last days.
References
1. Greaves, F., Millett, C.: Consistently increasing numbers of online ratings of healthcare in
England. J. Med. Internet Res. 14(3), e94 (2012)
2. Tumasjan, A.: Predicting elections with Twitter: what 140 characters reveal about political
sentiment. In: Fourth International AAAI Conference on Weblogs and Social Media,
Washington DC, pp. 178–185 (2010)
3. Zimlichman, E., Levin-Scherz, J.: The coming golden age of disruptive innovation in health
care. J. Gen. Intern. Med. 28, 865–867 (2013)
108 M. Tang et al.
4. Ziegler, C., Skubacz, M., Viermetz, M.: Mining and exploring unstructured customer
feedback data using language models and treemap visualizations. In: IEEE/WIC/ACM
International Conference on Web Intelligence and Intelligent Agent Technology, pp. 932–
937. IEEE, Sydney (2008)
5. Ginsberg, J.: Detecting influenza epidemics using search engine query data. Nature 457,
1012–1014 (2008)
6. Freifeld, C.C.: HealthMap: global infectious disease monitoring through automated
classification and visualization of internet media reports. J. Med. Res. 15, 150–157 (2008)
7. Greaves, F., et al.: Use of sentiment analysis for capture patient experience from free-text
comments posted online. J. Med. Internet Res. 15(11), e239 (2014)
8. Lin, Y., et al.: A document clustering and ranking system for exploring MEDLINE citations.
J. Am. Med. Inform. Assoc. 14, 651–661 (2007)
9. Denecke, K., Nejdl, W.: How valuable is medical social media data? Content analysis of the
medical web. Inf. Sci. 179, 1870–1880 (2009)
10. Pang, B., Lee, L.: Opinion mining and sentiment analysis found. Trends Inf. Retr. 2(1–2), 1–
138 (2008)
11. Ivanciue, O.: Weka machine learning for predicting the phospholipidosis including potential.
Curr. Top. Med. Chem. 8(18), 1691–1709 (2008)
12. Witten, I.H., Frank, E.: Data Mining: Practical Machine Learning Tools and Techniques.
Morgan Kaufmann Publishers, San Francisco (2005)
13. Frank, E., et al.: Data mining in bioinformatics using Weka. Bioinformatics 20(15), 2479–
2481 (2004)
14. Li, J., et al.: Discovery of significant rules for classifying cancer diagnosis data.
Bioinformatics 19(Suppl. 2), 1193–2103 (2003)
15. Alemi, F., et al.: Feasibility of real-time satisfaction surveys through automated analysis of
patients’ unstructured comments and sentiments. Qual. Manag. Health Care 21(1), 9–19
(2012)
16. Abegaz, T., Dillon, E., Gilbert, J.E.: Exploring affective reaction during user interaction with
colors and shapes. Proc. Manuf. 3(Suppl. C), 5253–5260 (2015)
17. Dong, A., Lovallo, D., Mounarath, R.: The effect of abductive reasoning on concept selection
decisions. Des. Stud. 37(Suppl. C), 37–58 (2015)
18. Evans, P.: From deconstruction to big data: how technology is reshaping the corporation.
MIT Technol. Rev. (2015). Stanford, California
19. Hsu, F.-C., Lin, Y.-H., Chen, C.-N.: Applying cluster analysis for consumer’s affective
responses toward product forms. J. Interdiscip. Math. 18(6), 657–666 (2015)
20. Chen, R., Xu, W.: The determinants of online customer ratings: a combined domain ontology
and topic text analytics approach. Electron. Commer. Res. 17(1), 31–50 (2017)
21. Holy, V., Sokol, O., Cerny, M.: Clustering retail products based on customer behaviour. Appl.
Soft Comput. 60(Suppl. C), 752–762 (2017)
How to Verify Users via Web Behavior
Features: Based on the Human Behavioral
Theory
Jiajia Li, Qian Yi(&), Shuping Yi, Shuping Xiong, and Su Yang
Abstract. Nowadays, the Internet has penetrated into people’s daily life in
many ways, and people usually use the form of web account to carry out web
activities. However, the problem of user account being illegally occupied by
others has become increasingly prominent, and the security of user account is
not guaranteed. On the other side, the network’s advantage of recording data
easily provides a valuable opportunity to comprehensively record a tremendous
amount of human behaviors. Based upon the status and problems, we present an
approach to distinguish genuine account owners from intruders. This study
constructs a series of web behavior features based on web usage logs, and uses
machine learning to identify users based on Bayesian theorem and structural risk
minimization criterion. Experiments show that web behavior analysis is a fea-
sible way for user verification.
1 Introduction
With the rapid development of Internet technology, people are becoming more and
more inseparable from the Internet. According to the 39th report of CNNIC (China
Internet Information Center) “By the end of 2016, the number of netizens in China has
reached 731 million, and the total number of new netizens has increased 42 million 990
thousand over the whole year” [1]. The development of mobile Internet has brought
great convenience to people’s life. It also provides support platform for information
sharing, which further reduces transaction costs and facilitates communication. In
general, it improves the efficiency of the overall resource utilization. But at the same
time, there are also various security problems of network. So far, the common risk we
faced is that the user’s account number is embezzled. Under the condition of account
password leakage, there are some shortcomings in traditional protection methods,
which makes user verification become one of the key issues in account security. In
order to solve these problems, we decide to use the real time updated web log and adopt
the web behavior features which based on the behavior analysis to carry on the con-
tinuous security monitoring to the user.
The field of human web behavior, elucidation of human behavior in the web
environment is quickly gaining eminent position [2]. Guo et al. establish a dynamics
model of human interest attenuation based on the data of blog releasing. The results
show that blog arrival time interval distribution is a mixed distribution with exponential
and power-law feature. It demonstrates the variety of human web behavior dynamics
[3]. Neelima et al. dig the frequency of users accessing each page, and find users’
sessions to analyze users’ time spent on a specific page [4]. Zhong et al. calculate the
features for every page based on analysis of association rules, then build up a per-
sonalized user browsing behavior model [5]. Via click stream analysis, Leung et al.
contrast the different behavioral patterns of consumers which during weekdays and
weekends [6]. Cui et al. conducts a pioneering research in how to use model or method
to represent human behaviors. They mention that forecasting and predicting are major
goals in modeling human behaviors, but it also faces various challenges [7]. In most of
existing research, some of which discuss web behavior from the time dimension, and
others focus on the analysis of the connection between users and browsing pages. But
unlike those studies, this paper maps the behavior theory of the real world into the
virtual world, so as to explore the user web behavior from different dimensions.
On Classification and Prediction, Lin et al. use three integrated feature sets
(EEG + Modeling, EEG + Task, and EEG + Task + Condition) training a linear dis-
criminant analysis classifier, then integrating human behavior modeling and data
mining to predict human errors [8]. By leveraging observable information of the user’s
Twitter feedback, Pennacchiotti and Popescu use Gradient Boosted Decision Trees-
GBDT to build user classification model. Finally, automatically infer the values of user
attributes, such as political orientation or ethnicity [9]. According to the characteristics
of user interaction with the mouse, Feher et al. introduce a method to verifies users
[10]. Pao et al. propose a user verification approach based on user trajectories. Such as
on-line game traces, mouse traces, handwritten characters, and traces of the movements
of animals in their natural environments [11]. These studies not considered user
behavior related to physiological and psychological needs comprehensively, therefore,
the current methods do not fully reflect the individual differences of users, nor can they
guarantee enough effectiveness. Based on the web behavior which is refined by web
log analysis, we propose a general user verification approach, which does not require
any extra sensing equipment.
Based on the characteristics of individual differences in human web behavior, this
paper is no longer just a study of the relationship between users and web pages but
make use of Web logs to construct a series of behavior features which are related to
user habits, and then, it draws support from two kinds of machine learning methods to
verify users identity.
2 Feature Construction
patterns of action, besides, people’s attitude will change their behavior accordingly
when they are in a state of discomfort [12]. In this study, web behavior refers to the
action that takes place on human-computer interface when people surf the Internet.
Compared with the traditional entity scene, the existence of the web log to make it
possible for accurately and comprehensively collection of human behaviors in the
absence of other additional equipment. The internal drive of web behavior includes
psychological and physiological demands together with the way of thinking, which
same as the offline behavior. Moreover, it combined with the constraints on actual
conditions can decide the possibility and trend of behavior. In the same way, web
behavior also shows five major characteristics as bellows: Adaptability; combined with
the work and rest time, user always choose the most convenient time to surf on Internet,
besides, the size of browser window is also depend on their vision and operating
environment. Relevance; in fact, most of the time user care more about what they are
interested in, and the occurrence of web behavior is related to it, meanwhile, it also
affects the persistence and frequency of behavior. Diversity; user can perform web
browsing, mouse clicking, input text and other multiple forms of action on the human-
computer interaction interface. Dynamic and Controllability; If the external environ-
ment or user’s motivation changes, user can decide whether to change the current
action and how to change it.
2.2 Definitions
Based on the above theory, we analyze the web logs of AML (Advanced Manufac-
turing Technology) to get 13 features of user behavior. The definitions and descriptions
of these web behavior features used in this study are given below.
(1) Week: It indicates that the behavior is happened in which day of the week. It
transforms by the ‘timestamp’ which is recorded on web log. First, read the data
of ‘timestamp’ from web log, such as 2017-05-25T19:47:19.123Z. Next, the
information of year, month and date was selected. Finally, convert it to ‘week’.
For example: ‘2017-05-25T19:47:19.123Z’ is transformed into ‘Thur’.
(2) Time: It indicates that the behavior is happened in which time of the day. It
transforms by the ‘timestamp’ which is recorded on web logs. First of all, we
divide the 24 h of a day into 6 parts, and the partition rule is shown in Table 1.
After that, we extract the information in the ‘timestamp’ which represents hours,
and transform it into the Time’s value in the end.
(4) P-Web: The ratio of web browsing height to the total height of the web page. It
should be noted that the height of user’s mouse sliding on a Web page is called
web browsing height. Generally, users will browse more content when the page is
more consistent with his interests. The height of web browsing and the total height
of web page are directly given by web logs.
(5) Event: The type of operation event that user does on web pages. There are 8 types
involved in this study, as illustrated in Table 2.
(6) Language: The language used by the user for typing. In our study, this feature’s
value is mostly English or Chinese, which reflects users’ habitual language and is
related to their living environment.
How to Verify Users via Web Behavior Features 113
3 Methodologies
In this section, we describe the process of user verification and explain the two machine
learning methods that we use: Naïve Bayes and Support Vector Machine (SVM). The
flow chart of the process is shown in Fig. 3. The process is implemented in five steps.
The first step is to obtain the data from the data source, then select the high frequency
operation users as the research objects. The second step is to deal with the data
according to the principle of feature construction in the Sect. 2. The third step is to
further process the data so that it meets the input requirements of machine learning. The
fourth step is to use the classifier called Naïve Bayes and SVM to determine whether
114 J. Li et al.
the web behavior belongs to the account owner. The fifth step is to discuss the per-
formance of the classifiers.
C : C ¼ fy1 ; y2 ; . . .; yn ; . . .g
The set C represents the set of all categories, yn represents the category n.
How to Verify Users via Web Behavior Features 115
I : I ¼ fx1 ; x2 ; . . .; xm ; . . .g
It makes any xðxi 2 IÞ has one and only one y ðyj 2 f ðxi ÞÞ corresponded to it.
Among them, f is called classifier, and the task of classification algorithm is to con-
struct classifier f.
A. Naïve Bayes
Classification ideology of Naïve Bayes is to solve the probability of occurrence of
category yj ðj ¼ 1; 2. . .; n; . . .Þ when element xi appears. After that, the probable
maximum category y will be selected as the category of element xi . The classification
process of Naïve Bayes is divided into three phases, and the specific content
arrangements are as follows:
(1) Preparation; above all, the feature attributes are determined and partitioned
according to the practical matter, next a part of the elements is selected and
classified though known data, in the last step, we take it as a training set. In this
phase, the input is all the data to be classified, and the output includes both the
feature attributes and the training set. The quality of the features and training
samples plays a decisive role in the design of classifier, therefore, this is a crucial
phase.
(2) Classifier Design; in this phase, a classifier will be generated after input feature
attributes and training dataset. The main work of this phase is not only to calculate
the probability of each category in the training set, but also to solve the conditional
probability, that is to calculate the probability of occurrence of all categories when
the feature iði ¼ 1; 2; . . .; nÞ occurs. Ultimately, the results are recorded to design a
classifier.
(3) Application; the test data is used to verify the classification models which build in
phase (2).
Naïve Bayes classification is a typical generative learning method which has the
advantages of simple logic and easy implementation. Bayes faster than other complex
methods in learning and classification, indeed, it has excellent performance in text
classification and spam filtering.
B. Support Vector Machine
Support vector machine (SVM) is a binary model by Computing support vector
which has strong generalization capability. By the way, it is widely applied into fields
such as face recognition and speech recognition. This algorithm maximizes the interval
between different types of samples in the feature space by learning a linearity classi-
fication facet, and then transforms it into a convex two order programming problem.
That means maximizing the interval of two classes of sampling points under constraint
conditions. Similar to Naïve Bayes, the classification process of SVM has been three
116 J. Li et al.
phases: Preparation, Classifier Design and Application. In the second phases, the main
task is to find a hyperplane of optimal decision for classification.
P ¼ TP=ðTP þ FPÞ
Precision can measure the correct rate of the model in predicting positive samples.
(2) Recall (R):
R ¼ TP=ðTP þ FN Þ
Recall can measure the reliability of the model in predicting positive samples
How to Verify Users via Web Behavior Features 117
F1 ¼ 2TP=ð2TP þ FP þ FN Þ
4 Experimental Result
Firstly, we collected the web usage log of AML which in the period of May 2017 to
March 2018. Secondly, according to the user’s operation frequency, the top ten user’s
data set has been retained. And then, the data are preprocessed by the principle of
feature construction in part 2 of this paper. Thirdly, since most of the machine learning
methods can only accept the input of quantitative features, we use the encoding method
which called One-Hot to convert the non-numeric features into numerical. After that, in
order to gain as much effective information as possible from the data set, we use cross
validation to partition the training set and test set, and the proportion of the two is 7:3.
At the same time, this can also help us evaluate the predictive performance of the
model. After the data preparation is completed, we use Naïve Bayes and SVM algo-
rithms to implement user identity recognition based on web behavior. This process is
implemented in Python Programming.
Figure 4 below shows the prediction results of SVM for the test set, and the
prediction accuracy is 65%. According to the predicted results by SVM, there are 812
and 1212 web behaviors that belong to user 2 and user 7, of which has 801 and 1196
are correctly classified to their owns, and only a few are misjudged as other users. But
for other users, the prediction is not satisfactory. Among them, 203 web behaviors
belong to user 4 are wrongly awarded to users 7. Moreover, only 18 of the 130 web
behaviors that belong to user 8 are correctly classified. On the whole, many behaviors
that are not user 7 are wrongly classified to him.
The accuracy of Naïve Bayes is 64%, as shown in Fig. 5. The prediction result for
user 2 is the most ideal, and 800 of web behavior records which belongs to him are
correctly identified. But from the point of view of the user 1, the prediction results are
not good. There are 333 web behavior records in the test set which belongs to the user
1. Among it has 146 records are classified to user 8, and there is also having 59 and 53
records are misjudged to user 4 and user 7. It is disappointing that only has 46 records
are correctly identified.
Table 4 shows the overall performance of SVM and Naïve Bayes. The value of
prediction, recall and F1 -score of each user is presented in this report, and the last row
express the model performance with regard to the whole test set. From this one report
we can see that SVM is better than Naïve Bayes in precision and recall, but for F1 -
score, Bayes is slightly higher than SVM.
As the result of analysis, SVM has higher accuracy but less reliable for positive
samples. But for Bayes, it is more balanced for the distribution of accuracy and
reliability.
5 Conclusions
We present 13 web behavior features which associated with the physiological, psy-
chological and living environment of users to verify their identity. It is proved that this
is a feasible task under the support of machine learning method, although the results
vary with the choice of the model. It is a valuable study to verify user identity with the
help of web behavior features, which is because the user’s equipment information and
network environment are very easy to fake, but the behavior of someone is hard to copy
owing to his uniqueness, so that we could guarantee the security of accounts by the user
web behaviors, without using other additional devices.
The future direction of work includes the following points: (1) We will try diverse
classification and prediction methods to improve the accuracy and reliability of user
identification. (2) On the basis of the existing research, we will further explore more
features of human web behavior. (3) The importance and contribution of every web
behavior feature under different approaches will be discussed in the next step, and the
web user is presented in a comprehensive way from the perspective of human
behavioral science.
120 J. Li et al.
Acknowledgment. This work was supported by Fundamental Research Funds for the Central
Universities NO. 106112016CDJXY110003, 2016.1-2017.12 and the National Natural Science
Foundation of China under Grant No. 71671020.
References
1. CNNIC: The thirty-ninth statistical report on the development of Internet in China (2017).
http://www.cnnic.net.cn/hlwfzyj/hlwxzbg/hlwtjbg/201701/t20170122_66437.htm
2. Géczy, P., Izumi, N., Akaho, S., et al.: Human behavior on web: what is known. In:
International Proceedings of Economics Development & Research (2011)
3. Guo, J.L., Fan, C., Guo, Z.H.: Weblog patterns and human dynamics with decreasing
interest. Eur. Phys. J. B 81(3), 341–344 (2011)
4. Neelima, G., Rodda, S.: Predicting user behavior through sessions using the web log mining.
In: International Conference on Advances in Human Machine Interaction, pp. 1–5. IEEE
(2016)
5. Zhong, J., Yan, C., Yu, W., et al.: A kind of identity authentication method based on
browsing behaviors. In: Seventh International Symposium on Computational Intelligence
and Design, pp. 279–284. IEEE Computer Society (2014)
6. Leung, R., Law, R., Masiero, L., et al.: Behavior of online visitors to hotel ICON: a
weekday-weekend analysis. e-Rev. Tour. Res. 7 (2016)
7. Cui, P., Liu, H., Aggarwal, C., et al.: Uncovering and predicting human behaviors. IEEE
Intell. Syst. 31(2), 77–88 (2016)
8. Lin, C.J., Wu, C., Chaovalitwongse, W.A.: Integrating human behavior modeling and data
mining techniques to predict human errors in numerical typing. IEEE Trans. Hum.-Mach.
Syst. 45(1), 39–50 (2017)
9. Pennacchiotti, M., Popescu, A.M.: A machine learning approach to Twitter user
classification. In: International Conference on Weblogs and Social Media, Barcelona,
Catalonia, Spain. DBLP, July 2011
10. Feher, C.: User identity verification via mouse dynamics. Inf. Sci. 201(19), 19–36 (2012)
11. Pao, H.K., Fadlil, J., Lin, H.Y., et al.: Trajectory analysis for user verification and
recognition. Knowl.-Based Syst. 34(5), 81–90 (2012)
12. Von Mises, L., Greaves, B.B.: Human Action: The Treatise on Economics. Liberty Fund,
Carmel (2007)
Authentication Using Users’ Mouse Behavior
in Uncontrolled Surroundings
Fan Mo, Shiquan Xiong ✉ , Shuping Yi, Qian Yi, and Anchuan Zhang
( )
1 Introduction
Human’s behavior is continuous in both real world or online. Before using the Internet,
they already develop behavioral patterns on many things from using a pen to pressing
a button. After they get online, their behaviors continue with offline behavioral patterns,
develop into online patterns, and become stable gradually. Based on that, analysis of
their online behavior is meaningful and feasible, since each people’s behavior online
probably could only match with himself/herself. Studies on trusted interaction mecha‐
nism are extremely important nowadays, because of the rapid development of the
Internet and security needs of persons and institutions online.
Authentication using behavioral biometrics is a very important issue waiting to be
tackled nowadays. People use access control mechanisms, like username, password,
token, fingerprints, or signature, to protect against another person’s unauthorized access.
This means that a user needs to give his/her identity’s proof when logging in a web.
Another drawback of user verification methods is that they require dedicated hardware
devices such as fingerprint sensors are expensive and are not always available. Although
2 Literature Review
were asked to install a monitoring system and ran the software for nine weeks. These
fusion methods were able to achieve high accuracy [21]. However, the time for collecting
data was too long, and their methods cannot be used on websites.
Overall, authentication for websites generally needs more varied than that on a
traditional computer. Studies that used behavioral biometrics designed experiment to
collect data from users who used the same computer, and their variable control is strict.
This means that the generalization of these authentications models is not so easily into
the daily usage of web development. First, one person may have several devices to access
the websites. The mouse or computer used by a person may be different. Second, on the
condition of the Internet, invaders are highly possible to use different computers and
measures to remote control users’ data. To deal with this problem, this research considers
the complex environment of using computers and tries to build models to verify mouse
behavior of individuals.
3.1 Participants
There were more than twenty participants in this experiment. The participants were
recruited from the Department of Industry Engineering, Chongqing University. They
were asked to use online websites to search information, to read texts and pictures, and
to download files if they want. There were no specific tasks they are asked to finish, and
there is no target amount of tasks, so the participants can choose to use the web a lot,
only use it a little, or even not use it. In that conditions, only data from eight participants
(two are females and six are males) were chosen to be analyzed in next analysis because
of their data were from the period more than 3 months. The average age of participants
was 31.25 (SD = 12.848). Six of them were postgraduate students, and two were profes‐
sors. All the participants had more than 10 years’ experience of using internet and mouse.
Participants used mouse normally because there is no target in this experiment. The data
were collected when they used websites to browse webs, so it was a normal mouse
surfing.
3.2 Procedure
A behavioral biometric system fusing user data from mouse, which is based on front-
end web development, is presented in this research. There are mainly four steps (the last
three steps are as shown Fig. 1.
124 F. Mo et al.
First, users log into the websites, and use it of his/her own free will; meanwhile, the
system records users mouse data. An example JavaScript code can be seen as follows.
The second phase is features acquisition. The gathered data are processed by R, and the
specific actions (e.g. click, or move) are extracted. Then, features of the action are
calculated. These features are saved as feature data sets. The third phase is modeling
and the fourth phase is verification. More details about the last three parts are described
in Sect. 4.
A data collection system was developed. Javascript and Java EE in front-end web are
used to develop the system which is official websites of the research group (http://
www.cquieaml.com/) as shown in Fig. 2. The websites include several parts, such as
academic research, business exchanges and cooperation, blog article, about us, and the
management system. Javascript codes recorded users mouse data, like click and move,
and then transformed the data into back-end server and saved them. The data that are
used in following analyses can be downloaded in the management system.
Two personal computers (Dell Inspiron 3268-R1308) are used for controlling the
management system. Participants access the online websites using their own computers,
including MacBook Air (2015 early), Lenovo (ThinkPad X1 CAR), Acer (4750G), and
Acer (TMP238-M-P6J9). The sampling rate of the mouse data is not consistent, from
10 Hz to 200 Hz. By resampling, the rate is fixed at 100 Hz, so the interval time between
two data records (e.g. two movement segments) is 0.01 s.
Eight users’ behavior of this system is recorded for more than 3 months, from April
2017 to July 2017. All of them are members of the research group and have their own
accounts on the websites. Although the users know the websites can record their
behavior, they do not pay attention to it since they generally do some basic work, such
as downloading papers or posting a message on the forum, and the recording of their
behavior does influence their feelings, so it can be regarded as uncontrolled surround‐
ings.
4 Data Processing
Data on mouse behavior are cleaned, classified, and featured through R programming.
User characteristics that relate to the way that a particular user interacts with the websites
are collected by the system. This is done by monitoring mouse movements. Features are
calculated on these actions. After data collection, the raw data is processed to create
features for active authentication testing and identification. The calculated features are
selected from prior works. These features are used in many related studies associated
with authentication, since they have relatively high reliable.
The original data only involved five values, including type of the mouse event, x-
coordinate of the mouse pointer (x), y-coordinate of the mouse pointer (y), time of the
mouse event (t), and users ID. R codes are developed to classify the type of mouse event
and extract the features. The main mouse events in webs are left-click (LC) and the
movement sequence (MS). Unlike the mouse behaviors when interacting with files
systems or games, right-click (LC) is not commonly used on websites. The right click
motion on pages is to open the webpage ‘tab’, which helps users to have advanced
functions with the websites. These functions are usually not used when surfing the web,
so the right click is not going to be analyzing in follow analysis.
Therefore, the research focused on the two behavior and extracted features of them.
Features used in this research are from existing studies. The data format changes as
shown in Fig. 3.
126 F. Mo et al.
The basic events of mouse behavior include pressing the mouse button (including
left button and right button), lifting the mouse button (also including left button and right
button), and the movement of the mouse which is recognized by a photoelectric detector.
( )
⎧ EventPR = PRi , xi , yi , ti , IDuser
⎪ ( )
⎪ EventLL = LLi , xi , yi , ti , IDuser
⎪ ( )
⎨ EventLR = LRi , xi , yi , ti , IDuser (1)
⎪ ( )
⎪ EventM = (Mi , xi , yi , ti , IDuser
⎪ Event = MS , x , y , t , ID )
⎩ MS i 1:n 1:n 1:n user
Left click usually is composed of three sequential basic events: pressing the mouse
button (PL), small movement (M), and lifting the mouse button (LL). Users may not
keep still when pressing the button, which is the reason why there is a small movement.
Features of left click involved interval time between pressing and lifting the mouse
button and the slide distance between them.
{( ) ( ) ( ) }
EventLC = LCi , [(PLi , xi , yi , ti , MSi , x1:n , y1:n , t1:n , LLi , xi , yi , ti ], IDuser (2)
Interval time in LC equals that ti of LLi minus ti of PLi, and that is, interval time in
LC = tLLi − tPLi. Similarly, slide distance in LC is:
√
( )2 ( )2
xLL − xPL + yLL − yPL
i i i i
(3)
Each movement has values of all above features, so the minimum, maximum, range,
mean, and standard deviation of them are used to reflect the features of the movement
sequence. In this part, 45 features are used to measure the movement sequences.
There are also four features that are able to use to reflect the movement, including
duration
√( of movement tn, traveled distance Sn, jitter S′ ∕Sn, straight distance
) 2 ( )2
S′ = x1 − xn + y1 − yn . Therefore, there are 49 features in total.
4.2 Learning
R codes are used to build the machine learning methods. Classifiers are constructed for
both left-click (LC) and movement sequence (MS). Specifically, the two actions instance
forms a row whose columns contain the features that are associated with the action, and
its label is given by the ID of the user who performed the action, so it is the supervised
learning. Support Vector Machine, Random Forest, and K-Mean Clustering are tested,
and then the method of random forest is used to establish models, because of its relatively
high accuracy and lesser time consumption. This research not aims to get the best results
of machine learning methods; it just shows a good learning results can be gained in
uncontrolled surroundings. The mouse data is organized using the techniques of classi‐
fication and regression trees. An algorithm in the random forest runs, and then randomly
selects the predictor and repeats the process. Hundreds of regression trees are generated.
The random selection of the predictive factor is averaged as a result of the classification.
The data from feature data sets are split into two parts: training sets (60%) and testing
sets (40%), using random sampling methods. This method picks sample randomly and
put it into a set. Each training sample is used to construct a decision tree. ‘randomForest’
library of R is used. The number of trees to grow is 500. The importance of predictors
is set as needed to be assessed, and the proximity measure among the rows is set as
128 F. Mo et al.
4.3 Verification
In the end, authentication is conducted based on the two behavior sub-models of LC and
MS. The authors comprise an inducer used to build verification model for the user by
training on past behavior, often given by samples. During verification, the induced model
is used to classify new samples that are acquired from the user. There are two main kinds
of models which are models of click and models of movement. Invaders’ mouse behavior
is recorded by using the account of members in the group, the reliability of his/her click
behavior and movement behavior are calculated respectively.
To give the final reliability of users’ mouse behavior, the two reliabilities are mixing.
To reasonably and simply mix the two authentication results from the models, below
equation is used.
α ≥ A&𝛽 ≥ B (4)
Where α is the reliability of models of left click, and β is the reliability of models of
the movement sequence. A is the toleration of reliability of left click, and B is the toler‐
ation of reliability of movement sequence. When both reliabilities from the two models
reach the threshold, the individual’s behavior can be considered as credible.
Data of mouse behavior are gathered from March 29th, 2017 to August 25th, 2017, and
there are more than 1.6 million records of mouse events from eight people. Users’ click
behavior and movement behavior are modeling using Random Forest based on the
training data sets, respectively. There are 500 trees for both of the two models. After
that, testing data sets are used to verify the accuracy of the models. As shown in Fig. 4,
the accuracy is not very high. In terms of the event of left click, the only accuracy of
three participants’ behavior is up to 50%, including MF (72%), YS (55%), and ZHJ
(79%). The results are not good even worse than random guessing. Several predictions
are very bad. For instance, YJ’s left clicks are all classified into XHQ (55%) and ZAC
(45%). None of LJJ left clicks are grouped to other. With regards to the event of move‐
ment sequence, there are five participants who have accuracy up to 50%, including LJJ
(66%), MF (64%), XSQ (64%), ZAC (52%), and ZSJ (51%). These results are better than
Authentication Using Users’ Mouse Behavior 129
the results of left click but still very poor. The reasons for it is the unbalanced sample
size of different participants, because the data are collected from uncontrolled experi‐
mental surroundings.
Fig. 4. Left figure: Pie figure of classifying events of LC. Pie figure of classifying events of MS.
In order to improve the prediction accuracy, the sample size is balanced by multiplying
the size of a small sample group, and random sampling the large sample group. Then,
the models are rebuilt using these new data sets. As shown in Tables 2 and 3, the accuracy
significantly increases. The last column is the error rates of grouping. As to the left click,
the error rates of MF, SHY, and XSQ are 0%, which means that all the predictions are
absolutely right. The error rates of LJJ are the highest, up to 15.58%. The overall error
rate is 3.36%.
As to the accuracy of the movement sequence, the overall error rate is 4.21%, which
is also low. All the predictions of participants’ movement sequence behavior are rela‐
tively high, except for the predictions of ZAC. The error rate of predicting ZAC’s
movement sequence is 62.96%. It should be noticed that ZAC is the most active partic‐
ipants in the data collection system, so the data of ZAC is maximum; nevertheless, in
the results of the accuracy of movement sequence, the sample size of ZAC is the
minimum. That is why the result is not reasonable.
130 F. Mo et al.
Data on mouse behavior are cleaned, classified, and featured through R program‐
ming. User characteristics that relate to the way that a particular user interacts with the
websites are collected by the system. This is done by monitoring mouse movements.
Features are calculated on these actions. After data collection, the raw data is processed
to create features for active authentication testing and identification. The original data
only involved five values, including the type of the mouse event, x-coordinate of the
mouse pointer (x), y-coordinate of the mouse pointer (y), time of the mouse event (t),
and users ID. R codes are developed to classify the type of mouse event and extract the
features. The main mouse events in webs are left click (LC) and the movement sequence
(MS), so the researchers focused on the two behavior and extracted features of them.
Features used in this research are from existing studies. The data format changes as
shown in Fig. 3. Interviews of one participant showed that he used both mouse and the
MacBook touchpad. As we can see, behavior of the two interaction ways must be very
different, but results indicated that all mouse behavior from one person can be learned,
and the classifiers can be strong. That is, although participants used several different
devices to log in the system, it seems that the models could handle this problem. There‐
fore, in some extent, this finding may be extended to other studies, so researchers do not
need to control the experiment strictly, which can reduce the cost and also make models
fit with real conditions better.
However, the results are high enough, so a test of unauthorized behavior from an invader
is conducted. There is an invader whose code is ZSJ. ZSJ used MF’s account to login
to the websites and tried to simulate MF’s mouse behavior. After about ten minutes’
use, 35 left click events and 42 movement events are recorded. No matter left click events
and movement events, none of ZSJ’s mouse behavior is recognized as MF’s, and the
accuracy of authentication is 100%.
Prediction results in Sect. 5.2 are hard to reach a really high accuracy (i.e.
100%), but when the system knows the behavior from the identical account, the
accuracy of authentication could be much higher. This situation conforms to the real
environment that an invader knows others’ account by stealing or password breaker.
The uncontrolled mouse behavior models provide the second protection to prevent
Authentication Using Users’ Mouse Behavior 131
invader’s unauthorized behavior. However, the system needs to collect enough data
in seconds in order to extract them into mouse action, so it may cost several minutes
to ensure the accuracy of recognition.
6 Conclusion
This research uses data from websites which can be considered as an uncontrolled
experiment. Data are continuously gathered from websites, which means the data are
unbalanced and quite conform to the real situation. The data for the uncontrolled experi‐
ment is further used to build mouse behavior models through the machine learning
method (i.e., random forest). Only left click and movement sequence of mouse events
are considered, but error rates are still very high. In terms of left click, error rates lower
than 3.36%, and in terms of movement sequence, the error rates lower than 4.21%. In
the end, a case study of unauthorized behavior from an invader is conducted. The results
institute that the authentication using users’ mouse behavior is quite high. This research
provides a rapid and general approach to authentic user behavior on the network, which
is meaningful for the development of authentication using behavioral biometrics in the
uncontrolled environment.
Further work should focus on conducting this experiment on larger sample size. Real
Environment usually involves a large group of users on websites, and the identification
of potential invaders need to base on all these users’ behavior. Besides, mass data might
lead to more processing time, so how to improve the classifiers is vital. In addition to
that, the influence of users’ psychological characteristics, as well as their age and expe‐
rience of using smart devices, on their mouse behavior may deserve further research.
This could explore more about how online behavior related to the specific users who
take the behavior.
Acknowledgment. This work was supported by Fundamental Research Funds for the Central
Universities NO. 106112016CDJXY110003, 2016.1-2017.12 and the National Natural Science
Foundation of China under Grant No. 71671020.
References
6. Zheng, N., Paloski, A., Wang, H.: An efficient user verification system via mouse movements.
In: Proceedings of the 18th ACM Conference on Computer and Communications Security,
CCS 2011, pp. 139–150. ACM, New York (2011). http://doi.acm.org/
10.1145/2046707.2046725. Accessed 25 Aug 2017
7. Gamboa, H., Fred, A.: A behavioral biometric system based on human-computer interaction,
pp. 381–392 (2004). http://adsabs.harvard.edu/abs/2004SPIE.5404..381G. Accessed 25 Aug
2017
8. Mondal, S., Bours, P.: Combining keystroke and mouse dynamics for continuous user
authentication and identification. In: 2016 IEEE International Conference on Identity,
Security and Behavior Analysis, ISBA, pp. 1–8 (2016)
9. Park, J., Han, S.H., Kim, H.K., Moon, H., Park, J.: Developing and verifying a questionnaire
for evaluating user value of a mobile device. Hum. Factors Ergon. Manuf. Serv. Ind. 25(6),
724–739 (2015)
10. Gaines, R.S., Lisowski, W., Press, S.J., Shapiro, N.: Authentication by keystroke timing: some
preliminary results. Report No. RAND-R-2526-NSF. Rand Corp Santa Monica CA, May
1980. http://www.dtic.mil/docs/citations/ADA484022. Accessed 25 Aug 2017
11. Brown, M., Rogers, S.J.: User identification via keystroke characteristics of typed names
using neural networks. Int. J. Man-Mach. Stud. 39(6), 999–1014 (1993)
12. Haider, S., Abbas, A., Zaidi, A.K.: A multi-technique approach for user identification through
keystroke dynamics. In: 2000 IEEE International Conference on Systems, Man, and
Cybernetics, vol. 2, pp. 1336–1341 (2000)
13. Han, S.J., Kim, S.-U.: Placement of a touchpad and click-buttons to relieve arm fatigue and
discomfort in a laptop PC. Hum. Factors Ergon. Manuf. Serv. Ind. 27(3), 131–137 (2017)
14. Monrose, F., Rubin, A.: Authentication via keystroke dynamics. In: Proceedings of the 4th
ACM Conference on Computer and Communications Security, CCS 1997, pp. 48–56. ACM,
New York (1997). http://doi.acm.org/10.1145/266420.266434. Accessed 25 Aug 2017
15. Pusara, M.: An examination of user behavior for user re-authentication (2007). https://
search.proquest.com/openview/1bd296a5f59df64c30768646f674eeab/1?pq-origsite=gsch
olar&cbl=18750&diss=y. Accessed 25 Aug 2017
16. Bergadano, F., Gunetti, D., Picardi, C.: User authentication through keystroke dynamics.
ACM Trans. Inf. Syst. Secur. 5(4), 367–397 (2002)
17. Pusara, M., Brodley, C.E.: User re-authentication via mouse movements. In: Proceedings of
the 2004 ACM Workshop on Visualization and Data Mining for Computer Security, VizSEC/
DMSEC 2004, pp. 1–8. ACM, New York (2004). http://doi.acm.org/
10.1145/1029208.1029210. Accessed 26 Apr 2018
18. Everitt, R.A.J., McOwan, P.W.: Java-based internet biometric authentication system. IEEE
Trans. Pattern Anal. Mach. Intell. 25(9), 1166–1172 (2003)
19. Shen, C., Cai, Z., Guan, X., Du, Y., Maxion, R.A.: User authentication through mouse
dynamics. IEEE Trans. Inf. Foren. Secur. 8(1), 16–30 (2013)
20. Ahmed, A.A.E., Traore, I.: Anomaly intrusion detection based on biometrics. In: Proceedings
from the Sixth Annual IEEE SMC Information Assurance Workshop, pp. 452–453 (2005)
21. Bailey, K.O., Okolica, J.S., Peterson, G.L.: User identification and authentication using multi-
modal behavioral biometrics. Comput. Secur. 43, 77–89 (2014)
Construction of Network User Behavior
Spectrum in Big Data Environment
Mengyao Xu, Fangfei Yan, Biao Wang, Shuping Yi(&), Qian Yi,
and Shiquan Xiong
1 Introduction
With the development of Internet technology, people’s lives have become more con-
venient. At the same time, information security issues have also received more and
more attention. Identification of user identities can be achieved by studying the
behavior patterns of network users, thus a new method can be offered for the estab-
lishment of information security protection mechanism. At present, many scholars have
explored how to reflect the behavior patterns of network users. Frhan [1] developed a
visualization model named as Social Pattern Clustering WebClickviz
(SPCWebClickviz) to visualize the social networking data based on user activities and
then clusters them into specified groups. Qiang and Ji-min [2] analyzed Web logs to
build user behavior patterns. Zhu et al. [3] proposed a method for defining user roles
based on behavior patterns and clusters the behavior of telecommunication network
users during socialization. Katerina and Nicolaos [4] investigated the potential corre-
lation between mouse behavioral patterns or keystroke dynamics and a set of End-User
Development (EUD) behavioral attributes. These studies are currently only from a
single point of view to establish the user behavior patterns, either only use the front-end
data (mouse or keyboard related data), or only use the back-end data (log data),
research is too one-sided, therefore, the user’s behavior patterns should be described in
a more comprehensive and three-dimensional manner by establishing a network user
behavior spectrum.
The concept of behavior spectrum was first used by Makkink’s in 1936 in the study
of European pied Avocet. The behavior spectrum is widely seen in animal behavior.
Nishida et al. [5] used behavior spectrum to list and describe all behavior patterns
recorded by chimpanzees in the Mahale Mountains National Park, Hall and Heleski [6]
demonstrated that the horse’s behavior spectrum plays an important role in assessing
training methods; Ivanov and Krupina [7] studied Contagion behavior cause changes in
the behavior spectrum in Rats. There are few literatures that human behavior is studied
by using behavioral spectrum, only one paper can be found so far. The behavior
spectrum was defined by Qin et al. [8] as an effective user network access behavior
monitoring and measurement method. However, the data that reflect the physiological
behavior is not collected, that is, the data collected in the article does not fully represent
the behavior of the user.
All the behaviors of the network user are recorded by using the behavior spectrum
of the network user in this paper, and then the behavior pattern of the network user are
constructed. Perceived behavioral features and physiological behavioral features can be
used to construct a network user’s behavioral spectrum. Perception [9] is the organi-
zation, identification, and interpretation of sensory information in order to represent and
understand the presented information, or the environment. In this paper, perception
refers to the direct reaction of an objective thing through the sensory brain in the human
brain. After the brain perceives the object, user will be guided to make an action. For
example, when the operating system’s own language is unfamiliar to the user, the
language will immediately be transformed into his familiar language. Therefore, the
user’s perceived behavior is defined that the influence of the external environment,
personality traits, and the habits formed during the growth process on the user’s net-
work behavior. Physiology [10] is the scientific study of normal mechanisms, and their
interactions, which works within a living system. In this paper, physiology refers to the
life activities of living organisms and the functions of various organs. Therefore, the
user’s physiology is different and the user’s behavior is also different. For example:
The elderly move the mouse generally slower than young people. Therefore, the user’s
actions using the input tools (mouse, keyboard), including movement direction and
movement speed, movement trajectory, movement time, tapping rate, tapping time are
defined as the user’s physiological behavior.
The remainder of this paper is organized as follows. The data collection is
described in Sect. 2. The method for constructing a network user behavior spectrum is
described in Sect. 3 and an example is given in Sect. 4. Finally, the concluding remarks
are addressed in Sect. 5.
2 Data Collection
Data collection is an essential step before data analysis. In this paper, the data is
analyzed from the third person point of view to get the user’s behavior features and a
one-to-one mapping with the actual behavior is achieved, so different types of data are
Construction of Network User Behavior Spectrum 135
needed to collect. At present, data is obtained through the self-built website and the
enterprise management system, and data from the system’s background log and using
embedding code technology is collected.
The enterprise management system data is obtained from the background log of the
cooperative enterprise. The operations of all employees in the system are recorded,
these employees belong to different departments and have different responsibilities, but
each user has a clear operation task. As shown in Fig. 1, the collected features are
divided into four basic categories (time attributes, operational attributes, individual
basic attributes and company related attributes) and used to analyze user behaviors. As
shown in Fig. 2, the self-built website data is obtained from the background log and
embedding code technology of the academic exchange website established by the
laboratory.
From the Sect. 2, the user behavior features are divided into perceptual state and
physiological state. The network user behavior spectrum is constructed by analyzing
the network behavior of the perceptual state and physiological state. Next, the enter-
prise data is used to establish a network behavior analysis method based on the per-
ception state and the mouse data of a self-built website is used to establish a network
behavior analysis method based on physiological state.
(a) (b)
enterbutton enterfuction
ennull
Make a list
save
cancellation
delete
modify Sales Invoice
increase
invalid
rate per unit time on the X-axis during the Kth mouse moving; ⑤ Vx(k) = Sy(k)/T(k),
k = 1, 2.., n, The moving rate per unit time on the X-axis during the Kth mouse
moving.
From the eigenvalues, not only the user’s range of moving distance and moving
rate can be gotten, but also the standard deviation of the sample eigenvalues can be
calculated, from which you can see the difference in user behavior.
4 Case Study
Based on the analysis method of the Sect. 3 for network user’s perceptual state
behavior and physiological state behavior, an example is used to illustrate the user’s
network behavior spectrum building process. 150 data from No. 46 are used for
experimental analysis.
In terms of perceptual behavior, according to the four attributes, five features are
selected: user name, date, IP, head and browser; In terms of physiological behavior,
Construction of Network User Behavior Spectrum 139
(a) 40 33
Horizontal moving distance
1050
700
350
0
145
103
109
115
121
127
133
139
151
157
163
169
175
181
187
193
199
1
7
31
13
19
25
37
43
49
55
61
67
73
79
85
91
97
-350
-700
-1050
(b) 40 33
Vertical moving distance
150
-50 103
1
7
109
115
121
127
133
139
145
151
157
163
169
175
181
187
193
199
49
13
19
25
31
37
43
55
61
67
73
79
85
91
97
-250
Fig. 4. (a) The moving distance of the mouse in the horizontal direction of the number 33 and
40. (b) The moving distance of the mouse in the vertical direction of the number 33 and 40
(a) 40 33
Horizontal moving rate
10
0
113
106
120
127
134
141
148
155
162
169
176
183
190
197
15
22
29
36
43
50
57
64
71
78
85
92
99
1
8
-5
(b) 40 33
3
Vertical moving rate
2
1
0
127
106
113
120
134
141
148
155
162
169
176
183
190
197
15
22
29
36
43
50
57
64
71
78
85
92
99
1
8
-1
-2
-3
-4
-5
Fig. 5. (a) The moving rate of the mouse in the horizontal direction of the number 33 and 40.
(b) The moving rate of the mouse in the vertical direction of the number 33 and 40
140 M. Xu et al.
according to the user’s mouse usage, four features are selected: coordinates of the
mouse on the X axis, coordinates of the mouse on the Y axis, time, user name.
Statistical analysis and correlation analysis are performed on the features. The results
are shown in Figs. 6, 7, 8 and 9 and Table 2.
head Control
center
Homepage
Account
management
IP
106.91.50.77
14.106.175.192
14.110.128.84
222.178.10.230
222.178.10.232
222.178.10.238
222.178.10.239
101
106
111
116
121
126
131
136
141
146
-50
-100
-150
-200
-250
0
11
16
21
26
31
36
41
46
51
56
61
66
71
76
81
86
91
96
1
6
101
106
111
116
121
126
131
136
141
146
-5
-10
-15
5 Conclusions
In this article, a method of representing user behavior patterns based on the user
behavior spectrum is proposed by us. Firstly, the related definitions of perceptual
behavior and physiological behavior are given. Then, the feature values of the user’s
perceptual behavior and physiological behavior are used for statistical analysis and
correlation analysis. Thus, analytical methods based on perceptual behavior and
physiological behavior are established. Finally, these analysis methods are used to
establish one user’s behavior spectrum, and the results showed that the combination of
two behaviors to construct the user’s behavior pattern is more comprehensive and
three-dimensional than analyzing only a single behavior. The purpose of describing the
behavior can be achieved by establishing the behavioral spectrum. Only after the
description of the behavior, problems such as the behavior’s development, mechanism,
function and evolution can be proposed. Moreover, the establishment of user behavior
spectrum is significant for the identification of user identities.
In the future, there is still a lot of work that needs to be done by us. The analysis of
the combination of perception and physiological behavior needs to be more thorough,
the mining of behavioral data needs to be deeper, and the construction of the user’s
behavioral spectrum needs to be more complete. In addition, only the data from self-
built websites is used by us to establish a behavioral spectrum. However, self-
constructed websites are short-lived and have less data. In the future, more data from
cooperative companies will be collected and the establishment of user behavior
spectrum for employees in the company will be completed. Also, corporate data will be
used to study the evolution of user behavior.
Construction of Network User Behavior Spectrum 143
Acknowledgment. This work was supported by Fundamental Research Funds for the Central
Universities NO. 106112016CDJXY110003, 2016.1-2017.12 and the National Natural Science
Foundation of China under Grant No. 71671020.
References
1. Frhan, A.J.: Visualization and analysis of user behaviour patterns for multimedia content
view in social networks. In: 5th ISEEE International Symposium on Electrical and
Electronics Engineering, pp. 1–7. IEEE Press, Galati, Romania (2017)
2. Qiang, W., Ji-min, L., Wan-hai, Y., et al.: User identification in the preprocessing of web log
mining. Comput. Sci. 29(4), 64–66 (2002)
3. Zhu, T., Wang, B., Wu, B., et al.: Role defining using behavior-based clustering in
telecommunication network. Expert Syst. Appl. 38(4), 3902–3908 (2011)
4. Katerina, T., Nicolaos, P.: Mouse behavioral patterns and keystroke dynamics in end-user
development: what can they tell us about users’ behavioral attributes? Comput. Hum. Behav.
83, 288–305 (2018)
5. Nishida, T., Kano, T., Goodall, J., et al.: Ethogram and ethnography of Mahale chimpanzees.
Anthropol. Sci. 107(2), 141–168 (1999)
6. Hall, C., Heleski, C.: The Role of the ethogram in equitation science. Appl. Anim. Behav.
Sci. (2017)
7. Ivanov, D.G., Krupina, N.A.: Changes in the ethogram in rats due to contagion behavior.
Neurosci. Behav. Physiol. 47(8), 987–993 (2017)
8. Qin, T., et al.: Behavior spectrum: an effective method for user’s web access behavior
monitoring and measurement. In: IEEE Global Communications Conference, pp. 961–966
(2013)
9. Schacter, D.L.: Psychology. Worth Publishers, New York (2011)
10. Prosser, C.L.: Comparative Animal Physiology, Environmental and Metabolic Animal
Physiology, 4th edn, pp. 1–12. Wiley-Liss, Hoboken (1991). ISBN 0-471-85767-X
Relation Analysis of Heating Surface’s Steam
Temperature Difference and Fouling Degree
Based on the Combination of Thermodynamic
Mechanism and Production Data Mining
Abstract. This paper aims to monitor heating surface fouling degree of boiler
without fume temperature data recorded. Based on the thermodynamics, this
paper analyzes the heat balance mechanism of the heat transfer characteristics of
the convection heating surface of the boiler and the endothermic characteristics
of the steam and obtains the relationship between the heat transfer coefficient
and the temperature difference and it is proved that the temperature difference
between the heating surface and heat transfer coefficient has a strong correlation
through the gray correlation analysis method. This paper analyzes the actual
production data by correlation analysis and regression analysis: the amount of
heat released from the combustion of the boiler is certain when the load is stable,
and based on the model for the temperature difference of the working fluid to
represent the degree of dust accumulation is established. This paper validates the
result that the model is able to reflect the degree of dust deposition on the
convective heating surface well by simulation and field data verification
methods.
1 Introduction
It is a common phenomenon that ash particles produced during solid fuel combustion
adhere to surfaces of heat transfer tubes inside boilers. How to estimate fouling degree
is one of the most urgent puzzles for electric engineers in industry. As for now, the soot
blowing of large coal-fired power station boilers generally employs a fixed quantitative
method to purge each heated surface every certain time [1]. This blind and centralized
soot-blowing method is likely to result in insufficient soot-blowing or excessive soot-
blowing, which will not only damage boiler tube but also affect the stability of the
operation, and reduce the generator unit’s operating efficiency [2]. Therefore, it is very
necessary to transform the “timed and quantitative” soot-blowing method into “on-
demand distribution” soot-blowing methods so as to improve the safety and economy
of boiler operation.
Researchers at home and abroad have optimized the boiler soot-blowing system
through various methods [3–5]: Prof. Wei of North China Electric Power University
proposed the concept of clean factor [3], the ratio of real heat absorption and maximum
transferable heat of ideal heat exchanger [4], that is, the ratio of heat transfer effective-
ness to characterize the degree of dust accumulation; the method of exporting the
temperature of flue gas as a whole reflects the total degree of ash pollution in the
furnace and its effect. And as for specific heated surface, there is no way to know its
fouling degree.
This paper takes a 300 MW coal-fired power plant boiler (DC furnace) as the
research object, and analyzes the gray correlation between the temperature difference
between the inlet and outlet working surface of the convective heating surface and the
cleaning factor calculated based on the thermodynamic mechanism based on the
excavation of a large amount of actual production data. The characteristics prove that
the temperature difference on the side of the heating surface can reflect the degree of
dust accumulation. This method can characterize the degree of dust accumulation based
on limited measurement points of heated temperature without calculating the
endothermic enthalpy value. This method is proved to be reliable through field test.
Q1 ¼ KHDt=Bj : ð1Þ
In the formula, H is heat surface area, m2; Dt is heat transfer temperature and
pressure, C; Bj Calculated fuel consumption, kg/s; K is heat transfer coefficient,
kW=ðm2 CÞ.
For the heated surface of the fouling external surface of the pipe, the heat transfer
coefficient K, which does not take into account the thermal resistance and scaling of the
metal pipe, is calculated as
146 J. Wang et al.
1
K¼ ð2Þ
1=a1 þ e þ 1=a2
In the above formula, a1 and a2 are the exothermic coefficient of the flue gas
against the wall and the wall of the working medium, kW=ðm2 CÞ; e is the coefficient
of ash pollution, ðm2 CÞ=kW. As the degree of dust accumulation on the heated
surface increases, the fouling coefficient e will gradually increase, and the value of heat
transfer coefficient K will gradually decrease.
700 800
1MPa 11MPa
700
600 12MPa
2MPa
3MPa 13MPa
600
500 14MPa
15MPa
500
16MPa
400
enthalpy difference
enthalpy difference
400
300
300
200
200
100 100
0 0
0 50 100 150 200 250 300 0 20 40 60 80 100 120 140 160 180 200
temperature difference temperature difference
From the above two figures, it can be found that the enthalpy difference is
approximately linear with a certain pressure and temperature difference Fig. 2.
Dh ¼ aDT: ð4Þ
Thus, the convective heat absorption can also be expressed by the following
formula
Q2 ¼ DaDT=Bj ð5Þ
Qualitative Analysis of Relevance. For a specific convective heating surface, the heat
Q1 transferred by the convection and the heat Q2 absorbed by the working fluid should
be in balance [7]. The conclusion can be made from (1) and (5),
KHDt DaDT
¼ : ð6Þ
Bj Bj
DaDT
K¼ : ð7Þ
HDt
From the formula (7), the heat transfer temperature and pressure can be calculated
through the inlet and outlet temperature parameters on the steam side and the flue gas
side, however, it will inevitably be limited in practical applications. Based on the above
formula, it can be implied that there is a certain correlation between heat transfer
coefficient K and the temperature difference ΔT.
Correlation is defined as the measure of the size of systems’ association with time or
different objects [8]. In the process of system development, if the trend of heat transfer
coefficient K is consistent with that of temperature difference ΔT of the working fluid, it
means the degree of synchronization changes is relatively high, vice versa.
Quantitative Analysis of Relevance. The degree of correlation is essentially the
degree of difference between the geometric shapes of the curves. The correlation
coefficient for the reference sequence K and the comparison sequence ΔT at each time
can be calculated by the following formula:
Dmin þ qDmax
n¼ : ð8Þ
DðtÞ þ qDmax
Among them, the resolution coefficient q is generally between 0 and 1; DðtÞ is the
absolute difference between each point on the K curve and the ΔT curve, Dmin is
minimum of the absolute difference at all time, and Dmax is the maximum one.
148 J. Wang et al.
The correlation r is a number that represents the correlation degree which is cal-
culated by merging the correlation coefficients at each time (i.e., each point in the
curve) into one value. And take this value as the index to evaluate the relationship
between ΔT and K. The formula is as follows:
1X N
r¼ nðtÞ: ð9Þ
N t¼1
In this paper, the heat transfer coefficient K model is established using the low-
temperature reheater with a wide range of working surface heat and temperature
measurement points in the research object, and the correlation between heat transfer
coefficient and the temperature difference at the working side is analyzed by the gray
correlation analysis method to verify that the temperature difference in the working side
reflects the degree of dust accumulation.
Take the above 300 MW coal-fired power plant boiler (DC furnace) as an example. In
the load range of [160 MW, 300 MW], the load stability section was selected to
analyze the soot blowing system.
280 0.9
0.8
eat transfer coefficient
260
0.7
load/MW
0.6
240
0.5
0.3
200
5:47 8:47 11:47 14:47 0.2
time/h 5:47 8:47 11:47 14:47
time/h
From Fig. 4, it can be seen that the per-unit value of heat transfer coefficient (K) of
the heating surface in the low-temperature reheater increases from 0.3 to 0.9 after the
start of soot blowing. As the time go by, the degree of dust accumulation in the low
reheating area gradually increases, and the K value gradually decreases.
108
0.9
106
0.8
temperature difference/
104
102 0.7
100
0.6
98
0.5
96
94 0.4
92
5:47 8:47 11:47 14:47
time/h 0 100 200 300 400 500 600
time/min
Through the analysis of the field data, it is proved that under certain load condi-
tions, the temperature difference ΔT on the working side is positively correlated with
the heat transfer coefficient K, so the temperature difference on the working side can
reflect the degree of dust accumulation on the heated surface.
the independent variable, and the coal supply B and the wind volume W were used as
the dependent variables to perform the linear regression analysis.
The relationship between the amount of actual coal B on one day and the amount of
^ is shown in Fig. 7. And the relationship between the amount of
coal fitted by the load B
actual wind W and the amount of wind fitted by the load is shown in Fig. 8.
Fig. 7. Linear fitting of coal quantity to load. Fig. 8. Linear fitting of wind W to load
From the results above, it can be concluded that there is a good linear relationship
between the coal feed amount B, the total air flow rate W, and the load P. That is, when
the load is constant, the amount of heat released from the combustion of the boiler is also
constant, and the magnitude of the load is corresponding. Therefore, the difference in
heat absorption per unit working medium on the convective heating surface under the
same load can reflect the degree of fouling degree on the convective heating surface.
In the above formula, DT is the actual value of the temperature difference between
the inlet and the inlet of the convective heating surface whose unit is °C. From formula
Relation Analysis of Heating Surface’s Steam Temperature Difference 151
(10), it can be inferred that the fouling degree coefficient b is between 0 and 1, and b
approaching 0 means that the heated area is severely covered by ash; while b
approaching 1 means that the heated surface tends to be clean.
According to the boiler’s 696,960 samples in 484 days of operation, 287 samples
with stable load in the range of [160 MW, 300 MW] are analyzed. Since for different
heat receiving surfaces under different loads, the working medium heat absorption
interval values differs as well, and the temperature difference between the import and
export of the lower limit value of the TL, the upper limit value TH and the load
relationship is shown in Table 2:
Table 2. Upper and lower limit of inlet and outlet heater surface temperature.
Heating surface Lower limit TL Upper limit TH
Screen superheater −0.0136 P+86.1 −0.155 P+149.7
High temperature superheater −0.017 P+37.66 −0.066 P+58.09
Low temperature superheater −0.08 P+55.26 −0.23 P+115.34
High temperature reheater 0.028 P+74 −0.13 P+139.5
Low temperature reheater −0.09 P+169.73 −0.3 P+207.33
That is, the temperature difference of the working surface of the heating surface
indicates the degree of dust accumulation, which needs to be analysed and determined
in conjunction with the load at that time. The larger the load, the greater the temper-
ature difference of the working fluid will be correspondingly larger, and the difference
in the dust level represented by the temperature difference of the working fluid under
different loads is different.
Fig. 9. Low-temperature reheater heat excha- Fig. 10. High-temperature superheater heat
nge model. exchange model.
flue gas side, the working side of the low temperature superheater, and the flue gas side
of the rear screen superheater. The results are shown respectively in Table 3 and
Table 4.
According to the output results in Tables 3 and 4, it can be concluded that whether
on the surface of the low temperature reheater or on that of high temperature
superheater, when the temperature difference between the inlet and outlet working
fluids decreases, the heat transfer coefficient will also decrease.
40 0.6
39 0.5
temperature difference/
38 0.4
37 0.3
36 0.2
35 0.1
34 0
11:05 13:05 15:05 17:05 19:05 21:05 11:05 13:05 15:05 17:05 19:05 21:05
time/h time/h
Fig. 11. Temperature difference curve of work- Fig. 12. Fouling degree coefficient b curve
ing fluid in high temperature superheater of high-temperature superheater.
From the above figure, it can be seen that the temperature difference of the working
surface of the high-temperature superheater heating surface after soot blowing rises
from 35 °C to 40 °C, and the fouling degree coefficient b increases from 0.1 to 0.5. As
time went by, the high reheating area gradually accumulated ash, and the temperature
difference and coefficient of dust accumulation gradually decreased.
6 Conclusion
The temperature difference varies between the inlet and outlet side on different heating
surfaces under different load. However, it is not difficult to find the relationship
between the temperature difference limit of each heating surface and the load through
the production data. And to some extent, the real-time temperature difference data can
effectively reflect the degree of fouling degree. The results show that:
(1) The degree of correlation between the temperature difference on the heating
surface side of the working surface and the heat transfer coefficient is strong.
(2) The temperature difference on the heating surface of the heating surface can
effectively reflect the degree of dust accumulation.
154 J. Wang et al.
(3) In the case of a smokeless temperature measurement point of the boiler, it can
provide guidance to the working personnel and help achieve on-demand blowing.
This study only focused on stable conditions, the unit parameters themselves have a
certain degree of volatility, and further processing was done using data filtering
methods. Under varying conditions, the parameters of the unit have a wide range of
changes and needs to be further explored in the future.
References
1. Shi, Y.H., Yun, F., Wei, T.Z.: Research on ash fouling monitoring of heating surface in coal
fired utility boiler. J. Power Syst. Eng. 26(06), 49–50 + 53 (2010)
2. Qian, H., Song, L., Chen, Q.Q.: Research on the believe rule base of soot blowing demand
based on production data mining. J. Therm. Power Gener. 46(06), 113–118 (2017)
3. Yan, W.P., Liang, X.J., Zhou, J., Ye, X.M., Zheng, Z.: Ash monitoring and sootblowing
optimization of a 300 MW coal-fired utility boiler. Proc. CSEE 20(01), 58–61 + 129 (2007)
4. Wu, S.Y.: Effect of fouling on heat transfer performance of heat exchanger. J. Petro-Chem.
Equip. 29(1), 10–12 (2000)
5. Xu, X.H., Zhou, K.Y., Wei, H.Q., Xu, J.Q.: Increment of furnace exit gas temperature for
monitoring ash fouling in the furnace of coal-fired boilers. J. Proc. CSEE 29, 21–26 (2011)
6. Zhou, Q.T.: Principle of Boilers, 2nd edn, pp. 183–185. China Electric Power Press, Beijing
(2009)
7. Yan, J., Yu, X., Wang, Y.: Water and Steam Thermodynamic Properties Chart, 3rd edn,
pp. 13–53. Higher Education Press, Beijing (2015)
8. Sun, C.Z., Song, Y.T.: Theory discussion about grey interconnect degree model. J. Global
Geol. 19(03), 248–252 + 270 (2000)
Variable Selection Methods in Dredger
Production Model
1 Introduction
Dredgers are designed and manufactured for underwater earth and rock works. They
can be used to excavate canals, widen waterways, remove siltation from reservoirs, dig
trenches, collect ore and so on. The TSHD is more popular than other kinds of dredgers
because of its ability to complete the work of digging, loading, transporting, dis-
charging (blowing) independently and change the site itself. Nowadays, the dredging
equipment and auxiliary equipment of TSHDs are becoming gradually perfect. How-
ever, the operation and management level of TSHDs is generally not high, which
affects the dredging efficiency of TSHDs directly.
The optimization of the dredging parameters of TSHD is to obtain the relevant
construction parameters, so that the dredger can be in good operation condition and
achieve the purpose of high efficiency, high quality and low consumption when
dredging under the guidance of these parameters. However, giving proposed con-
struction parameters are not simple, the main difficulties are as follows: 1. It is unre-
alistic to establish an accurate input-output mechanism model for a complex nonlinear
multivariable time-varying system such as dredging operation system from a basic
principle. 2. The characteristics of the actual dredging equipment are greatly influenced
by the construction environment and operation conditions. 3. The dredging process
must ensure the safety of the slurry transport system and unreasonable parameter
adjustment will cause abnormal situations such as blockage, rupture and forced
shutdown.
Holland IHC company is one of the most important dredger manufacturing and
research units in the world. They also studied the optimization of dredging operations
and put forward new concepts such as “Efficient Dredging” and “Accurate Dredging”
[1]. But the results of their dredging optimization are not detailed, testing results and
key technologies not transparent, usually only a concept. In China, the research reports
on the optimization of dredging work are rare and floating on the surface, so it is
difficult to study their essence and specific performance.
Modeling and optimization of dredging operations often use model predictive
control [2], and the relationship between production and influencing factors must be
understood before modeling. In this paper, we focus on maximizing the production of
dredging. According to the experience of actual operator and some simulation exper-
iments, the factors that affect the yield can be summarized as follows: speed to ground,
vacuum degree of dredge pump, depth and pressure of drag head, soil characteristics [3,
4], pump speed, high pressure flushing system, overflow using [5] and so on.
Accordingly, there are many parameters related to production. If taking all the
parameters as the inputs of TDS model, it is easy to have problems such as over fitting,
low precision and long modeling time. The purpose of variable selection is to select a
certain feature set from production-related parameters as inputs to establish an effective
TDS model, which can improve the accuracy while removing redundant information,
reducing the complexity, improving the generalization performance and simplifying the
model. Three methods are applied to realize the screening of production-related
parameters, namely, least absolute shrinkage and selection operator (LASSO), mean
impact value (MIV) and genetic algorithm (GA).
2 Method Description
Taking the magnitude and dimensional difference of measured data of dredgers into
account, the data is first normalized in the early stage of variable screening. This paper
uses deviation normalization to process data and the formula is as follows:
xk xmin
xk ¼ ð1Þ
xmax xmin
where xmin is the minimum number of data series and xmax is the largest number.
mt ðtÞ
vt ðtÞ qw mt ðtÞ qw vt ðtÞ
TDSðtÞ ¼ vt ðtÞqq ¼ qq ð2Þ
qq qw q q qw
Variable Selection Methods in Dredger Production Model 157
where qw (1.024 ton/m3) and qq (2.65 ton/m3) denote the density of water and quartz;
mt and vt are loading quality and capacity.
When establishing the yield model, it is assumed to be a linear model. The
production-related parameters are recorded as x1 ; x2 ; . . .. . .xp as independent variables,
and the TDS is calculated according to the real data and Eq. (2) and recorded as y as the
dependent variable. Supposing that n groups measurement data of TSHD have been
standardized as follows: y1 ; x11 ; x12 ; . . .x1p ; y2 ; x21 ; x22 ; . . .x2p …… yn ; xn1 ; xn2 ; . . .xnp .
Create a linear model of production and production-related parameters:
y ¼ b1 x1 þ b2 x2 þ . . .. . .bp xp þ e ð3Þ
01 0 1 0
1 0 1
y1 x11 ... x1p b1 e1
B .. C B .. .. .. C ¼ ð x B .. C B .. C
where y ¼ @ . A x ¼ @ . . . A 1 xp Þ b ¼ @ . A e ¼ @ . A
yn xx1 xnp bp en
ð4Þ
y ¼ xb þ e ð5Þ
X
P
^ arg min ðy xbÞT ðy xbÞ s:t:
b¼ jbi j t ð6Þ
b2RP i¼1
X
P
^ arg min ðy xbÞT ðy xbÞ þ k
b¼ jbi j ð7Þ
b2RP i¼1
3. Move β j from 0 towards its least-squares coefficient x j , r ,until some other competi-
tor xk has as much correlation with the current residual as does x j .
4. Move β j and β k in the direction defined by their joint least squares coefficient of the cur-
rent residual on x j , xk ,until some other competitor xl has as much correlation with the
current residual.
5. Continue in this way until all p predictors have been entered. After min ( n − 1, p )steps,
algorithm stops.
1 1 1
f ðXÞ ¼ ¼ ¼P ð8Þ
^
SE sseðT TÞ n
ð^t t Þ2 i i
i¼1
^ f^t1 ; ^t2 ; . . .^tn g is the TDS of test set; T ¼ ft1 ; t2 ; . . .tn g is the real
In the formula, T¼
TDS of test set; n is the number of samples for the test set.
To avoid the influence of randomness of the initial weights and thresholds of BP
network on the fitness function calculation, the weights and thresholds of BP neural
network are optimized by GA for each individual in calculating the fitness function
value, that is, the improved BP network using GA mentioned above.
160 Y. Zhang et al.
3 Experimental Result
3.1 Sources of Data
Modern TSHDs are equipped with advanced equipment, such as DGPS high-precision
positioning systems, draft gauge indicators, drag arm position indicators, concentration
meters, flowrate indicators and output calculating devices. So various parameters in the
process of dredging can be measured.
Variable Selection Methods in Dredger Production Model 161
The experimental data in this paper are taken from the measured dredging data of a
TSHD in Xiamen. According to the actual operator’s experience and some simulation
results, the primary selection of production-related parameters is listed in Table 1.
screening method, their mean value being 0. But these three items must not be con-
sidered as having no effect on TDS, because their functions are obvious in the sea area
where sediment characteristics and wind waves change significantly. Especially in the
maritime space where soil is dense and influence of wave flow is great, it is necessary
to open the high-pressure flushing system to break the soil and increase the pressure of
the wave compensator and the opening of the water diversion window.
Accordingly, it can be concluded that the results of these three variable selection
methods above conform to actual measurement data. However, the results cannot be
used as input for all TDS models, because the soil characteristics and construction
environments encountered in each working condition are different.
4 Conclusion
The screening results of three methods above are not entirely consistent, but the
variables that should be kept are roughly the same. The main difference is reflected in
the ranking of importance, which is caused by the algorithm principle of LASSO and
MIV. LASSO pays attention to the linear fitting degree of production-related param-
eters to TDS, while MIV method takes the sensitivity of parameters as the importance.
LASSO has a greater impact on the penalty item. When the penalty parameter k is
larger, more coefficients tend to be 0 and less production-parameters are persisted. In
addition, it assumes that the TDS model is a linear model. Although it is not in line
with the actual situation, it has obvious advantages in solving the collinearity problem.
MIV takes the sensitivity of production-related parameters as importance, leading
to some important but constant parameters missing. But for the specific working
environment, if the soil or environmental factors in dredging process have not changed
significantly, these important control variables can remain unchanged.
Variable Selection Methods in Dredger Production Model 165
GA is the best way to establish effective TDS model, but it only works for specific
conditions, and has no universality like MIV. What’s more, MIV and GA are based on
the BP network model, and there are inherent defects in the BP network, such as the
randomness of initial weight and threshold value, local minimization, slow conver-
gence speed, different choice of network structure and sample dependence. Although
the BP network is optimized by GA, it cannot make up for all the defects of the BP
network, at the same time, the application of GA algorithm increasing the running time
of the program.
Concluding, these three methods above have their own advantages and disadvan-
tages, and GA is the best in terms of modeling effect. They also have one thing in
common: the final input variables are not universal, but they are in line with the actual
situation of the specific dredging operation. Therefore, for the different sea areas, these
three methods above should be reused to determine the best input parameters.
Acknowledgments. This work was financially supported by the Science and Technology
Commission of Shanghai Municipality of China under Grant (No. 17511107002).
References
1. Increased accuracy in data for Dutch dredging operations. World Pumps, 2017(26), 28–29
(2017)
2. Braaksma, J., Babuska, R., Klaassens, J.B., et al.: Model predictive control for optimization
the overall dredging performance of a trailing suction hopper dredger. In: Proceedings of the
18th World Dredging Congress (2007)
3. Stano, P.M., Tilton, A.K., Babuska, R.: Estimation of the soil-dependent time-varying
parameters of the hopper sedimentation model: The FPF versus the BPF. Control Eng. Pract.
24, 67–78 (2014)
4. Chen, X., Miedema, S.A., van Rhee, C.: Numerical modeling of excavation process in
dredging engineering. Procedia Eng. 102, 804–814 (2015)
5. Lendek, Z., Babuska, R., Braaksma, J., de Keizer, C.: Decentralized estimation of overflow
losses in a hopper dredger. Control Eng. Pract. 16, 382–406 (2008)
6. Tibshirani, R.: Regression shrinkage and selection via the LASSO: a retrospective. R. Statist.
Soc. B 73(3), 273–282 (2011)
7. Zhang, L., Li, K.: Forward and backward least angle regression for nonlinear system.
Automatic 53, 94–102 (2015)
8. Moradi, M.H., Abedini, M.: A combination of genetic algorithm and particle swarm
optimization for optimal DG location and sizing in distribution systems. Int. J. Electr. Power
Energy Syst. 1, 66–74 (2011)
9. Zhang, Z.: An estimation of coal density distributions by weight based on image analysis and
MIV-SVM. In: 2015 IEEE Advanced Information Technology Electronic and Automation
Control Conference (2015)
10. Chen, D., Hu, B., Shao, X., Su, Q.: Variable selection by modified IPW (iterative predictor
weighting)-PLS (partial least squares) in continuous wavelet regression models. Analyst 129
(7), 664–669 (2004)
11. Duan, F., Xiao, F., Jiang, J.J., Ma, L.: Automatic variable selection method and a
comparison for quantitative analysis in laser-induced breakdown spectroscopy. Spectrochim.
Acta, Part B 143, 12–17 (2018)
Fault Diagnosis and Maintenance
A New Data Analytics Framework
Emphasising Pre-processing in Learning
AI Models for Complex Manufacturing
Systems
1 Introduction
Fig. 1. Visualisation of proposed data analysis design flow for manufacturing data
Fig. 2. Detail of the EDA and pre-processing steps required in the framework
number of correctly recognised as not in the class (true negative), count of sam-
ples that were incorrectly assigned a class (false positive) and those who were
not recognised as being in the correct class (false negatives), each denoted by
tp, tn, f p, f n respectively [12]. These are used to construct confusion matrices
which provide values that are used for calculating the common performance
measures to evaluate classification models, for this paper, binary classification
(Table 1).
True Value
Positive Negative Total
Predictive Positive tp fp tp + f n
Negative f n tn f n + tn
Total tp + f n f p + tn N
– Precision is a measure of a class agreement of the data labels with the clas-
tp
sifiers output labels, calculated by tp+f p;
precision.sensitivity
– F-measure is calculated by 2 precision+sensitivity and is more robust to imbal-
anced data [12].
dataset as this is the value that our model aims to predict. The quality of the
products is extremely high as only 0.58% of products fail at the final testing
stage, thus introducing a major class imbalance issue with the data. Figure 3
depicts an example of the flow of a product across the factory floor, highlighting
the numerous stations associated with different lines in the build.
Fig. 3. Example flow of Bosch factory floor depicting stations as circular nodes.
In the first instance, we perform EDA to identify key properties of the Bosch
dataset to identify correlation, redundant variables, underlying structure and
issues within the data.
Data Properties: The Bosch manufacturing dataset consists of over 2.4M jobs,
each of which have an associated ID and 4364 variables. These variables/features
represent either numeric, categorical or date measurements. We performed anal-
ysis to determine the proportion of missing observations per feature and also a
count of missing observations per ID. Initial investigation into the categorical
A New Data Analytics Framework Emphasising Pre-processing 175
features indicate an issue of extreme sparsity (around 99% missing) and thus
is not included in this paper as done in [6]. Our analysis has focused on the
numeric data as preliminaries found it to be most influential, therefore categor-
ical and date variables were not within the scope of this study. Table 2 provides
a summary of the dataset used for the research in this paper.
Fig. 4. Bar plot showing the number of features associated with the individual stations
of the Bosch production line.
Fig. 5. Visualisation of missing versus observed instances within the Bosch data
Figure 4 shows the count of numeric features associated with each station. Sta-
tions 24, 25, 30 and 31 contain the largest number of features, so we assume that
these stations process more products and could be more influential.
3.2 Pre-processing
Initial analysis was performed to check for outliers in the features through visual-
isations of the distributions. Correlations between features as well as the response
were calculated. This demonstrated that features from the later stages of the
build were more highly correlated than those from earlier in the process. The
class imbalance is high, therefore if this is not handled appropriately, any model
built with this data will result in a biased approach predicting that the product
to be in the majority class i.e. pass. Before implementing sampling methods to
handle class imbalance, a number of stages of preprocessing are necessary.
improving interpretability. Here we used the top 50 features indicated from the
algorithm and their associated observations to train a new XGBoost classifier
model. Table 3 shows an example of the ‘Gain’ values produced by XGBoost.
Sampling: To account for the extreme class imbalance, one must consider sam-
pling methods to rebalance the class variable, but investigation into the XGBoost
algorithm demonstrated its robustness to imbalanced data and was not per-
formed for this initial analysis. However, sampling methods must be considered
for our general data analysis framework when implementing alternative learning
algorithms (Fig. 6).
Table 3. Example of six variables from XGBoost which show the accuracy of model
gained by retaining these features.
Feature Gain
L1 S24 F1723 0.5328070
L1 S24 F1846 0.2248599
L1 S24 F1632 0.1162531
L1 S24 F1695 0.0611954
L3 S34 F3876 0.0403588
L2 S26 F3036 0.0132253
4 Conclusion
In this paper, a new framework has been presented which combines useful ana-
lytics tools into a different format from those previously implemented. Using a
widely available dataset from Bosch, an appropriate training dataset contain-
ing 50 features was produced, allowing an extreme gradient boosting tree to
be used as a classification prediction model. Using the framework (Fig. 2), data
preprocessing and exploratory analysis was used to create a reduced data size
highlighting the most influential features. This allowed us to perform an R imple-
mentation of an extreme gradient boosting (XGBoost) model [11] and employ
R’s inbuilt performance metrics to demonstrate a high accuracy and F-measure.
The research highlights the necessity to pre-process and we are currently work-
ing with a commercial partner to apply the research to their system to produce
an automated system for performing the analysis on manufacturing data.
References
1. Lee, J., Kao, H., Yang, S.: Service innovation and smart analytics for industry 4.0
and big data environment. Proc. CIRP 16, 3–8 (2014)
2. He, Q.P., Wang, J.: Statistical process monitoring as a big data analytics tool for
smart manufacturing. J. Proc. Control 67, 35–43 (2017). https://doi.org/10.1016/
j.jprocont.2017.06.012
3. Susto, G., Schirru, A., Pampuri, S., McLoone, S., Beghi, A.: Machine learning for
predictive maintenance: a multiple classifier approach. IEEE Trans. Industr. Inf.
11(3), 812–820 (2015)
4. Wuest, T., Weimer, D., Irgens, C., Thoben, K.D.: Machine learning in manufactur-
ing: advantages, challenges, and applications. J. Prod. Manufact. Res. 4(1), 23–45
(2016). https://doi.org/10.1080/21693277.2016.1192517
5. Moldovan, D., Cioara, T., Anghel, I., Salomie, I.: Machine learning for sensor-based
manufacturing processes. In: Intelligent Computer Communication and Processing
(ICPP) 2017, pp. 147–154. IEEE (2017)
6. Zhang, D., Xu, B., Wood, J.: Predict failures in production lines. In: IEEE Inter-
national Conference on Big Data, pp. 2070–2074. Washington, USA (2016)
7. Lee, K., Cheon, S., Kim, C.: A convolutional neural network for fault classification
and diagnosis in semiconductor manufacturing processes. IEEE Trans. Semicond.
Manufact. 25(5), 1167–1180 (2014)
8. Kotsiantis, S.B., Kanellopoulos, D., Pintelas, P.E.: Data preprocessing for super-
vised learning. Int. J. Comput. Sci. 1(2), 111–117 (2006)
9. Carbery, C.M., Woods, R. Marshall, A.H.: A Bayesian network based learning
system for modelling faults in large-scale manufacturing. In: IEEE International
Conference on Industrial Technology 2018, pp. 1357–1362. France (2018)
10. Chawla, N.V.: Data mining for imbalanced datasets: an overview. In: Maimon,
O., Rokach, L. (eds.) Data Mining and Knowledge Discovery Handbook. Springer,
Boston (2009). https://doi.org/10.1007/978-0-387-09823-4 45
A New Data Analytics Framework Emphasising Pre-processing 179
11. Chen, T., Guestrin, C.: Xgboost: a scalable tree boosting system. In: 22nd Inter-
national Proceedings on knowledge discovery and data mining, pp. 785–794. ACM
(2016)
12. Sokolova, M., Lapalme, G.: A systematic analysis of performance measures for
classification tasks. Inf. Proc. Manag. 45(4), 427–437 (2009)
13. Kaggle.com: Bosch production line performance (2016). https://www.kaggle.com/
c/bosch-production-line-performance. Accessed Nov 2017
Performance Assessment of Multivariate
Control System Based on Data-Driven
Covariance Historical Benchmark
1 Introduction
Modern industrial process controllers can meet design requirements in the early run-
ning, but without regular maintenance, the controller performance will decline with
time and eventually lead to a huge gap [1]. At present, for MIMO systems on site, such
as CCS, the control performance can only be judged by manual analysis, which is
difficult to adapt to the requirements of power production automation. It is also nec-
essary to realize the timely detection of performance degradation, provide guidance for
the optimization. Therefore, it is an urgent need to implement automatic and effective
online performance assessment and monitoring of a MIMO.
Research on control performance was first proposed by Harris [2], which shows the
construction of performance benchmark is the key to evaluation index. Depending on
the benchmark, performance evaluation methods can be divided into two types [3]. One
is a method based on mathematical modeling. The performance benchmark of this
method is mostly from some of the best controllers designed for theoretical goals. Such
methods as IMC [4], MVC [5, 6] and LQG [7] are relatively mature, but the monitoring
effect depends on the accuracy of the process model. The other is data-driven evalu-
ation methods, such as based on multivariate covariance and principal component
analysis statistics [8]. These methods have strong versatility and do not rely on process
models. Yu and Qin propose a statistical method for the MIMO process based on
historical data [9, 10], in industrial applications, the availability of benchmarks has
been validated [11]. However, a practical problem is that the selection of benchmark
data for this method requires a certain amount of expert knowledge [12]. Based on
data-driven technology, this paper presents a performance evaluation method based on
historical benchmarks for MIMO. According to the actual operation of the system, only
I/O operation data is needed and it is performed under the closed-loop conditions. It
does not require in-depth understanding of the system and does not hinder the normal
operation of the industrial process. In order to overcome the limitations of historical
benchmarks relying on expert experience, a user-definable indicator was proposed to
select historical data sets and perform distribution fitting to obtain a range of index
values divided by performance levels.
This paper is organized as follows. In Sect. 2, the performance method based on
data-driven is proposed. In Sect. 3, an experiment is illustrated to show the rationality
of this method. Finally, the conclusions are given in Sect. 4.
2 Proposed Method
2.1 The Framework of the Proposed Method
In this section, a flowchart of the proposed method is given in Fig. 1. Data mining
technology is used to analyze a large number of process data off-line, obtain the
benchmark value corresponding to the historical data sets, calculate the current value
Jact of the system operation online, and compare with the benchmark Jhist to obtain the
performance indexes, and further multiple performance levels are divided. Finally, a
comprehensive analysis of the performance of MIMO is performed.
Jhist
cðkÞ ¼ : ð1Þ
Jact
For a MIMO process, let the benchmark period be I and the monitored period II, to
assess the overall control performance during period II, a covariance index similar to
that proposed by Mcnabb and Qin [14] as follows:
182 H. Qian et al.
Evaluation result
jcovðyI Þj
Iv ¼ : ð2Þ
jcovðyII Þj
jcovðyI Þj Y q
Iv ¼ ¼ jKj ¼ ki : ð5Þ
jcovðyII Þj i¼1
Where KI ¼ diag kI;1 ; kI;2 ; ; kI;q and KII ¼ diag kII;1 ; kII;2 ; ; kII;q ,kI;i are
eigenvalues during period I corresponding to a controlled variate, and kII;i are eigen-
values during period II.
Comparing Eqs. (4), (7) and (8) can yield:
2k 3
I;1
kII;1
6 7
6 kI;2 7
6 kII;2 7
K¼ KI K1
II ¼6 .. 7: ð9Þ
6 7
4 . 5
kI;q
kII;q
kI;i
Ii ¼ki ¼ : ð10Þ
kII;i
Yq X
q
aðCÞ¼x0 k þ
i¼1 C;i
xi kC;i : ð11Þ
i¼1
Where C is a certain operating phase, kC;i are the eigenvalues correspondingly. The
weight coefficient x0 represents the importance of overall control performance,
x0 ; x1 ; xq represents the importance of each closed-loop control performance.
In this paper, a large amount of historical process data is used to traverse all
historical stages according to Eq. (11) for off-line calculations to obtain index values
aðCÞ for any stage and regard it as a random variate, denoted as X. Through mathe-
matical statistics analysis, we can the probability distribution of X and further fit the
distribution to f ð xÞ, take multiple data sets corresponding to multiple stages corre-
sponding to the peak value as historical data sets, in that way, the average value of the
eigenvalue values in the multiple stages corresponding to the historical data sets serves
as the evaluation benchmark, i.e.k;1 ; k;2 ; ; k;q .
After fitting, in most cases, X is possible to obey the lognormal distribution, thus
Y ¼ lnðXÞ Nðl; r2 Þ, the probability density function expression of X is
( h i
pffiffiffiffi
1
exp 2r1 2 ðln x lÞ2 x[0
f ðx; l; rÞ¼ 2prx : ð12Þ
0 x0
184 H. Qian et al.
1X M
k;i ¼ kk;i : ð13Þ
M k¼1
1X N
kl þ pr;i ¼ kpk;i : ð14Þ
N k¼1
k;i
I i ¼ ki ¼ : ð15Þ
kII;i
In practical work, Y obeys the normal distribution and has 3r criteria, so the
lognormal distribution X has
Pðelr x el þ r Þ¼68:27%
Pðel2r x el þ 2r Þ¼ 95:45% ð17Þ
Pðel3r x el þ 3r Þ¼ 99:73%:
Therefore, the range of indexes can be divided into 5 intervals in the direction of
ð0; 1Þ, corresponding to 5 grades, respectively poor, fair, good, nice and great, as
shown in Table 1. In the end the rating criteria is summarized as follows
The process output data from a 600 MW coal-fired heating unit of a thermal power
plant under normal operation is used as the data source of the experiment. The dataset
contains the set-point and output values of the two controlled variates (output power
and main steam pressure). Then, an example for monitoring the performance of CCS is
illustrated in this section.
k;N
I N ¼ kN ¼ : ð18Þ
kII;N
k;P
IP ¼ kP ¼ : ð19Þ
kII;P
k;N k;P
I v ¼ kN kP ¼ : ð20Þ
kII;N kII;P
Where IN and IP is the single closed-loop performance index of power and main
steam pressure, respectively, Iv represents the overall performance of the system.
The simulation results show that the electric power control performance is great for
more than 360 min, and is still nice for a long time, and the performance is good and
fair in a short time within 200 min to 400 min. The main steam pressure control
performance is great or nice most of the time, nice state time is longer, and performance
is good and general in a short time. The overall control performance is great or nice
most of time and it is good or fair in a very short time.
At time t1, IN ,IP and Iv are all greater than 1, control quality of the system better
than the benchmark, and is in the great state; At t2, the control performance is general,
which needs to be improved. There is a severe attenuation trend in t1 to t2, which
188 H. Qian et al.
should be taken seriously. Adjustment quality of the system is nice at t3; it is good at
t4. The situation does not occur, that is, the power, main steam pressure, and overall
control performance are all in poor, which means that the system is out of control.
Performance Assessment of Multivariate Control System Based 189
4 Conclusion
This paper proposes evaluation indexes for MIMO based on data-driven technology
combined with the actual operation of the control system. It can not only obtain the
control performance of a single closed-loop, but also get the overall performance of
MIMO. And a rule for selecting historical sets is also proposed, by which an improved
historical benchmark is obtained, and uses distribution fitting to divide the control
performance into five levels. The performance evaluation of a 600 MW unit coordi-
nated control system is simulated, and the objectivity and rationality of the method are
verified. For the characteristics and control quality requirements, this method can be
applied to a large number of multivariate control systems in the field. It can realize
online automatic evaluation to monitor the change of control quality, quickly respond
to severe performance degradation, and provide guidance for the optimization of
control system.
References
1. Desborough, L., Miller, R.: Increasing customer value of industrial control performance
monitoring-honeywell’s experience.In: AIChE Symposium Series, pp. 169–189 (2002)
2. Harris, T.J.: Assessment of control loop performance. Can. J. Chem. Eng. 67(5), 856–861
(1989)
3. Xin, Q., Yang, C., et al.: A review of control loop monitoring and diagnosis: prospects of
controller maintenance in big data era. Chin. J. Chem. Eng. 24(8), 952–962 (2016)
4. Zhang, G.M., Li, N., Li, S.Y.: A data driven performance monitoring method for predictive
controller. J. Shanghai Jiao Tong Univ. 45(8), 1113–1118 (2011)
5. Zhang, Q., Li, S.Y.: Performance monitoring and diagnosis of multivariate model predictive
control using statistical analysis. Chin. J. Chem. Eng. 14(2), 207–215 (2006)
6. Wang, X., Huang, B., Chen, T.: Multirate minimum variance control design and control
performance assessment: a data-driven subspace approach. IEEE Trans. Control Syst.
Technol. 15(1), 65–74 (2006)
7. Bernstein, D.S., Haddad, W.M.: LQG control with an H∞ performance bound: a Riccati
equation approach. In: American Control Conference, pp. 796–802. IEEE (2009)
8. Alghazzawi, A., Lennox, B.: Model predictive control monitoring using multivariate
statistics. J. Process Control 19(2), 314–327 (2009)
9. Yu, J., Qin, S.J.: Statistical MIMO controller performance monitoring, Part I: data-driven
covariance benchmark. J. Process Control 18(3–4), 277–296 (2008)
10. Yu, J., Qin, S.J.: Statistical MIMO controller performance monitoring, part II: Performance
diagnosis. J. Process Control 18(3–4), 297–316 (2008)
190 H. Qian et al.
11. Li, Q., Whiteley, J.R., Rhinehart, R.R.: A relative performance monitor for process
controllers. Int. J. Adapt. Control Sig. Process. 17(7–9), 685–708 (2010)
12. Wang, L., Li, N.: Performance monitoring of the data-driven subspace predictive control
systems based on historical objective function benchmark. Acta Autom. Sin. 39(5), 542–547
(2013)
13. Desborough, L., Harris, T.: Performance assessment measures for univariate feedback
control. Can. J. Chem. Eng. 70(6), 1186–1197 (1992)
14. Mcnabb, C.A., Qin, S.J.: Projection based MIMO control performance monitoring: I—
covariance monitoring in state space. J. Process Control 13(8), 739–757 (2003)
15. Yuan, Q., Lennox, B., Mcewan, M.: Analysis of multivariate control performance
assessment techniques. J. Process Control 19(5), 751–760 (2009)
Quantitative Safety Assessment Method
of Industrial Control System Based
on Reduction Factor
1 Introduction
With the arrival of intelligent manufacturing and the deep integration of industrial-
ization and information technology [1], in view of the increasing security threat and
attack of industrial control system, the network security risk assessment is becoming
more and more concerned as a method to recognize the risk of industrial control system
[2] in advance. It has become a new hot spot in the research of industrial control
network security [3].
Most of the industrial enterprises in our country have not established the infor-
mation security management system of industrial control system, the safety manage-
ment system is not perfect, and the establishment of the safety management system [4]
is lack of basis and guidance. With the continuous development of the Internet, the
proportion of information security risk is becoming more and more serious. Therefore,
it is necessary to develop the relevant standards for safety management, put forward the
specific process of safety management [5], and put forward the safety requirements of
the industrial control system from many aspects such as information security and
functional safety, and guide the industrial enterprises to establish the suitable condi-
tions according to their own actual conditions [6].
In the field of industrial control security, most of them are the combination of
analytic hierarchy process and normalization. In this paper, the fuzzy matrix method
and entropy method are adopted to reduce the consistency test [7]. The reduction factor
can also improve the accuracy of the final evaluation. It is a feasible method.
necessary to transform the complex multi factor problem into the contrast problem. By
comparing the factors, the fuzzy complementary [10] judgement matrix constructed by
the expert is converted into a fuzzy consistent judgment matrix [11].
The comparison of the two factors is divided into 6 grades: the same, important,
relatively important, obviously important, and absolutely important. It’s natural to use
digital 0.1-0.9 to quantify, as shown in Table 1.
The fuzzy complementary judgement matrix is calculated by line and recorded as:
Quantitative Safety Assessment Method of Industrial Control 195
X
n
ri ¼ rik ð2Þ
k
The consistency condition of the judgement matrix is not completely satisfied, and
it is objective or even impossible to eliminate in practice.
Use the test index:
kmax n
CI ¼ ð5Þ
n1
When less than 0.1, the judgement matrix conforms to the consistency standard
[13], and the result of hierarchical single ranking is acceptable. Otherwise,CIwill be
revised until it meets the consistency standard.kmax is the maximum eigenvalue of a
fuzzy judgment matrix.
The matrix RM is used as the original matrix, and linear normalization of it can be
got, so as to solve the homogenization problem of different quality indicators. More-
over, because the meaning of the positive and negative index values is different (The
higher the positive index is, the lower the negative index value is.). Therefore, different
algorithms is used to standardize the data for high and low index. The specific methods
are as follows:
For positive indicators:
fij minðfij Þ
j
lij ¼ ð8Þ
maxðfij Þ minðfij Þ
j j
X
n
hi ¼ k hij ln hij ð10Þ
j¼1
lij
hij ¼ P
n ð11Þ
hij
j¼1
1 hi X m
xi ¼ P
m ð xi ¼ 1Þ ð12Þ
m hi 1
i¼1
Bi ¼ Wi RAi ð14Þ
Quantitative Safety Assessment Method of Industrial Control 197
P
n
ti xi
j¼1
B1 ¼ Pn ð15Þ
xi
j¼1
P
n
ti ðxi Þa
j¼1
B2 ¼ n ð16Þ
P
ðxi Þa
j¼1
si ¼ aTi qi ð17Þ
ai is the unit amount of each element, qi is the reduction, If it’s not zero, ti was cut.
To ensure that the weight after reduction is not negative,si ð0; 1Þ.
The evaluation score of the upper layer is obtained by the formula 17. Parameter Bi
is the weight of the various factors of the I, Si is the evaluation score for the I factor. In
this way, the total score of the system can be finally obtained.
4 Case Analysis
The network risk assessment is carried on to a specific power system. First, the
evaluation object is analyzed, and the evaluation target is established by the analysis.
Aiming at the total risk target of the system, it is refined into multiple sub targets, and
establish a hierarchical structure model of industrial control system with multi-
objective, multi-level and multi criteria layer by improved analytic hierarchy process.
The level structure model is divided into four layers: the target layer is the security risk
198 H. Zhu et al.
of the enterprise control system; the first level index divides the factors affecting the
security risk of the control system into three aspects: assets, vulnerabilities and threats;
the second index is the identification of the three risk factors and subdivides according
to the attributes of each category of risk factors. In order to refine the mapping rela-
tionship between target level and target level, the second level index is further refined
into families, forming the underlying factor layer. Finally, the hierarchical structure
model is determined.
Specifically, the relative importance of each factor index is divided according to the
quantitative scaling method of fuzzy matrix priority relation, and the fuzzy judgment
matrix is constructed.
Referring to the assignment structure, the numerical results is given in the form of
matrix, and get the corresponding fuzzy judgement matrixRP1 ,RP2 ,RP3 ,RP4 andRP5 .
2 3 2 3
0:5 0:7 0:8 0:9 0:5 0:5 0:2 0:3
6 0:3 0:5 0:6 0:8 7 6 0:5 0:4 7
RP1 ¼6 7; RP2 ¼ 0:5 0:3
; RP3 ¼6
0:5 0:3 7
4 0:2 0:4 0:5 0:6 5 0:7 0:5 4 0:8 0:7 0:5 0:6 5
0:1 0:2 0:4 0:5 0:7 0:6 0:4 0:5
2 3
0:5 0:2 0:3 0:4 0:6
6 0:8 0:5 0:4 0:4 0:7 7
6 7
RP4 ¼6
6 0:7 0:6 0:5 0:5 0:8 7
7
4 0:6 0:6 0:5 0:5 0:6 5
0:4 0:3 0:2 0:4 0:5
2 3
0:5 0:4 0:7 0:6 0:6 0:5 0:8 0:4 0:2 0:8 0:9
6 0:6 0:5 0:7 0:8 0:7 0:6 0:9 0:3 0:2 0:5 0:6 7
6 7
6 0:3 0:3 0:5 0:4 0:3 0:4 0:6 0:2 0:3 0:7 0:8 7
6 7
6 0:4 0:2 0:3 0:5 0:4 0:4 0:7 0:5 0:5 0:6 0:7 7
6 7
6 0:4 0:3 0:7 0:6 0:5 0:6 0:8 0:5 0:5 0:7 0:8 7
6 7
RP5 ¼6
6 0:5 0:4 0:6 0:6 0:4 0:5 0:7 0:5 0:5 0:6 0:8 7
7
6 0:2 0:1 0:4 0:3 0:2 0:3 0:5 0:6 0:7 0:5 0:5 7
6 7
6 0:6 0:7 0:8 0:5 0:5 0:5 0:5 0:5 0:7 0:8 0:8 7
6 7
6 0:8 0:8 0:7 0:5 0:5 0:5 0:3 0:3 0:5 0:2 0:1 7
6 7
4 0:1 0:5 0:3 0:4 0:3 0:4 0:5 0:2 0:8 0:5 0:5 5
0:2 0:4 0:2 0:3 0:2 0:2 0:5 0:2 0:9 0:5 0:5
According to formula 8 and 9, each matrix is normalized and transform the fuzzy
complementary judgement matrix RPi into fuzzy consistent matrix RMPi .i ¼ 1; 2. . .5.
According to formula 10, the entropy weight xi of the i index is calculated. Finally, the
weight of each factor bi is calculated based on formula 11. The overall framework is
shown in Fig. 1.
Quantitative Safety Assessment Method of Industrial Control 199
xP5 ¼ ð 0:135 0:151 0:081 0:094 0:117 0:108 0:071 0:111 0:103 0:013 0:017 Þ
5 Conclusion
In this paper, an improved fuzzy analytic hierarchy process is applied to assess the risk
of information security in specific power systems. The results of the inverse model
gelatinization are improved by entropy method and improved reduction factor, effec-
tively assessing the importance of the system assets, the severity of the vulnerability
and the threats faced, and evaluating the system risk synthetically, thus verifying the
effectiveness and practicability of the proposed method. At the same time, it provides a
good theoretical basis for the risk management of the industrial control system, and
further enriches the theory of risk assessment of industrial control system. It has high
theoretical value and practical value for the establishment of the strengthening scheme
of industrial control system and the improvement of the risk management system.
In the comprehensive fuzzy evaluation of the relative importance of each element
and the factors of the evaluation layer, it is one of the key points of the future research
to design a better evaluation standard to deal with the evaluation data better and reduce
the subjective uncertainty.
Acknowledgments. This work was financially supported by the Science and Technology
Commission of Shanghai Municipality of China under Grant (No. 17511107002).
References
1. Peng, y, Jiang, C., Xie, F.: Progress in information security of industrial control systems.
J. Tsinghua Univ. (Nat. Sci. Ed.) 52(10), 1396–1408 (2012)
2. Peng, J., Liu, L.: Information security analysis of industrial control systems. Autom. Instrum.
33(12), 36–39 (2012)
3. Chin, Z.: Safety and management of industrial network control system. Meas. Control
Technol. 32(2), 87–92 (2013)
Quantitative Safety Assessment Method of Industrial Control 201
4. Xiong, Q., Peng, Y., Dai, Z.: Preliminary study on information security risk assessment of
industrial control system. Chin. Inf. Secur. 27(3), 57–59 (2012)
5. Zhu, X.: Information security risk assessment risk analysis method. Inf. Secur. Technol. 28
(8), 87–89 (2010)
6. Jia, C., Feng, D.: Safety assessment of industrial control system based on fuzzy analytic
hierarchy process. J. Zhejiang Univ. 50(4), 759–765 (2016)
7. Qin, C., Chen, X., Yang, Y.: Research and application of three level assessment on the safety
of tailing bank based on FAHP IE algorithm. Control Proj. 21(6) 995–1000 (2014)
8. Lu, H., Chen, D., Peng, Y.: Quantitative research on information security risk assessment of
industrial control system. Autom. Instrum. 35(10), 21–25 (2014)
9. Yuan, Y.: Research on risk management of construction projects based on fuzzy analytic
hierarchy process (FAHP). Chongqing University, Chongqing (2013)
10. Ma, L., Zhang, L., Yang, Y., et al.: Research on the security risk assessment of information
systems based on fuzzy neural network. J. Chin. Secur. Sci. 22(005), 164–169 (2012)
11. Li, J., Yin, X., Hutu, et al.: Research on network security assessment based on extension
theory. Comput. Eng. Appl. 48(21), 79–82 (2012)
12. Krutz, R.L.: Securing SCADA Systems. Wiley, Hoboken (2005)
13. Fan Hong, Feng Dengguo, Wu Yafei. Methods and applications of information security risk
assessment. Tsinghua University press, Beijing 2006 49–50
14. Kefeng, F., Ruikang, Z., Lin, L.: Research on information security standard system of
industrial control system. Top. Netw. Secur. Stand. 6 (4), 17–21 (2016)
15. Xiaofeng, Z., Xiuzhen, C.: Grey level information security evaluation model. Ind. Control
Syst. 1(6), 15–20 (2014)
16. Nicholson, A., Webber, S., Dyer, S., et al.: SCADA security in the light of cyber-warfare.
Comput. Secur. 31(4), 418–436 (2012)
Intelligent Computing in Robotics
The Electric Field Analysis and Test
Experiments of Split Type Insulator Detection
Robot
Abstract. The insulator strings of Extra High Voltage (EHV) and Ultra High
Voltage (UHV) transmission lines are longer. It is more convenient and effective
to detect insulators with robot. Currently, insulator detection robots mainly work
by climbing along the insulator string, the structure is bulky and complex, and
needs further improvement. Therefore, this paper proposes a split insulator
detection robot which is suitable for the detection of suspended insulator string.
The electric field distribution around the insulator robot is simulated and ana-
lyzed. The test experiment of 220 kV insulator string is carried out. The results
indicate that the maximum electric field strength around the split type robot is
1753 kV/m. Partial discharge will not be generated. The measured potential
value of the insulator detection robot is less than the original voltage distribution
value of the insulator. The actual voltage distribution of insulators can be
obtained by compensation and correction, then the detection of low and zero
insulators is carried out.
1 Introduction
insulator is small through infrared detection method, while the ultraviolet detection
performance is easily affected by the environment of the insulators [7, 8]. Both methods
have their own limitations, the data analysis methods need further studied in order to
improve the detection accuracy [9]. The data of contact detection is more reliable.
Traditional detection methods mainly rely on spark fork method or voltage distribution
method by manual work to distinguish the condition of insulators, which are very
inconvenient for long string insulators detection because the operating lever is usually
too long [10, 11]. With the rapid development of EHV and UHV transmission lines, the
insulator strings are too long to use manual detection methods [12, 13]. Insulator
detection by robot is convenient and effective for long insulator strings, and also
reduces the manpower investment in insulator detection work, reducing the operation
and maintenance costs of transmission lines [14, 15].
At present, insulator detection robot moves up and down along the suspension
insulator string mainly by climbing. Electric field distribution of each part of the
insulator detection robot is simulated and analyzed by researchers. Simulation results
show that the robot has a certain influence on electric field distribution of the insulator
string, but does not affect its insulation performance [16, 17]. The electric field strength
of the robot can be reduced through structural improvement and optimization, it can
avoid interference to the detection system and the malfunction of robot by corona
discharge and suspended potential discharge [3, 18]. However, due to the combination
of power and detection module on robot body, resulting in large volume, heavy weight
and complex structure of the robot body.
This paper designs a kind of split insulator detection robot for the suspension string
of the disc porcelain insulator in the straight-line tower, which has simple structure,
light weight and small volume. By simulating the electric field of the robot body and
the insulator string when the robot moves along insulator string, researchers study the
influence of the robot on insulator string and the largest field strength of each structure
of robot body. The test of 220 kV transmission line insulator string voltage detection
by designed insulator robot is carried out, the results of voltage distribution mea-
surement are analyzed.
The structure of split insulators low/zero live detection robot is showed in Fig. 1. It
contains two parts: power module and detection module. The power module and the
detection module are connected by a thin insulation rope, which can be dissembled and
installed flexibly. The detection module transmits instructions to the power module
through WIFI communication network, which controls the detection module to move
up and down and stop, then insulator detection according to the predetermined pro-
cedures is accomplished.
The power module includes cable drive motor, cable rack, guide pulley and fixed
plate. The driving motor is consisted of control chip, WIFI communication module,
motor driver, rotate speed sensor, grinding motor, power management unit and the
battery. The structure of the cable rack and guide pulley are simple and small in size, it
The Electric Field Analysis and Test Experiments 207
is easy to carry onto the tower. The connecting plate can be fixed on the cross arm
stably due to the structure of the cross arm, the connecting part is metal, so as to ensure
that the power module and the tower body are equipotential.
The detection module includes distance sensor, frame, stepper motor, probe, guide
stripes and WIFI communication module. All modules are installed on the frame and
are dismountable. The guide stripes can effectively prevent the insulators detection
robot from being stuck by the insulator’s shed while moving along the insulator string.
The detection module identifies the steel cap, shed and steel foot of the insulator
through the pre-supposed algorithm by using the range sensor. When the range sensor
recognizes the steel cap, the communication module sends the stop instruction to the
cable drive motor. When the detection module stops at the detection position, stepper
motor controls the two probes to touch the adjacent two insulators’ steel caps and
measure the voltage of the insulators. Then the voltage data can be transmitted to the
hand-held terminal through the communication module, and draw the distribution
voltage curve on the terminal which is used to judge low/zero value insulators.
Because the power module of the insulator detection robot is installed on the cross arm
of the tower and is equipotential with respect to the arm, the influence of the electric
field on power module is negligible. But the detection module may change the dis-
tribution of the surrounding electric field while moving, and the module body may also
produce partial discharge due to the high local field strength, which needs simulation
on the electric field for further analysis. Take the disc type porcelain insulator string on
the 220 kV single-loop cathead tower as simulation object, the electric field distribution
of the robot’s body, insulators and gold fittings when the insulator detection robot
208 P. Yin et al.
moves at different positions along the 220 kV disc type porcelain insulator string are
analyzed.
Fig. 2. The 3D model of the tower-line system and the insulator robot.
For the insulator robot, detection part including probe, stepper motor, connecting
plate, rack, guide stripes, communication module are built in the model. The probe is
divided into front metal and insulated connecting rod, the length of the front section is
120 mm, the length of the connecting rod is 130 mm, the inner radius of the rack is
140 mm, the thickness is 5 mm, the height is 20 mm, the guide stripes is 200 mm long
and 20 mm wide, and the radius of the circular arc is 20 mm; The size of the
communication module is 30 40 100 (mm), and the step motor size is
30 40 45 (mm). For the convenience of model subdivision, the model of the
detection part is simplified and the components with less influence on the electric field
simulation results are removed.
Then the whole model is subdivided, air bags are added respectively above the
insulator string and the tower in order to control their density. The insulator string and
the robot get encryption subdivision. The subdivision diagram of the insulator robot is
shown in Fig. 3 for example. At last, potential is loaded and the maximum field
pffiffi
intensity is studied. The maximum voltage Um ¼ 220 pffiffi 2 ¼ 179:629 kV is loaded in the
3
phase of the insulator robot, the rest phases are loaded with 0:5Um according to phase
The Electric Field Analysis and Test Experiments 209
sequence. The insulator is located at the high voltage end, the high potential Um is
loaded on the steel foot and wire connected to it, and 0 potentials are loaded on the low
voltage end which is connected with metal fittings, the iron cap, the pole tower, the
ground and the outsourced air boundary. The steel foot and iron cap with unknown
potential in the calculation model are suspended potential conductors. Coupled free
degree method is applied, which means the connected steel feet and steel cap are
coupled at equal potential. The probe of the insulator robot also needs to be coupled
when measuring the equipotential of the steel cap connected with measuring device.
Fig. 4. The influence of detection robot on the electric field intensity of insulator surface.
difference between two probes is different. The larger the voltage difference is, the
greater the field strength is. And this also verifies the effectiveness of the calculation.
The electric field simulation shows that the insulator robot has a certain influence
on the distribution of the electric field of the insulators, but it has little influence on the
maximum field strength of them, so this influence can be ignored. The maximum field
strength of the insulator is about 1753 kV/m. At the probe and the insulated connecting
rod, the radius of the probe wire is 0.3 cm. The halo field strength is about 37.6 kV/cm
calculated with the Peek formula, and the field strength of the probe surface is
17.53 kV/cm, which is a large margin. So the corona discharge will not occur.
The Electric Field Analysis and Test Experiments 211
The 220 kV insulator string was used for testing in the laboratory. The insulator string
has 13 insulators, the field layout is shown in Fig. 7. By using the iron tower to
simulate the tower cross arm, the iron frame is grounded through copper wire, the
lowest fittings are connected to a high-voltage generator, and the AC voltage effective
value is 127 kV. The insulator detection robot is set to measure twice of each insulator
when working.
212 P. Yin et al.
As there are some differences between the environment in the laboratory and for the
actual tower, the difference of the distribution voltage in the industry standard between
the measured value and the standard value is quite different. The simulation data with
the measured data are compared in Table 1. Firstly, the sum of voltage of 13 insulators
is 110.03 kV, which is about 17 kV less than the applied voltage. This result is caused
by the capacitance of the probe and the impact of the measurement circuit on the
distribution voltage. When the insulator robot is measuring the insulator, it is equiv-
alent to a parallel branch, resulting in the reduction of the equivalent impedance, and
the distribution voltage is reduced. When the robot moves along the insulator string, the
measured values are smaller than that of the original insulators, so the recorded total
voltage is about 17 kV less than the sum of the applied voltages.
The robot and the insulator string forms a equivalent parallel circuit structurally.
The configuration of this equivalent parallel circuit is constantly varying during
measurement, and the measured value from the detection robot has a complex non-
linear relationship with the original voltage distribution.
The whole model of insulator robot detection is established and verified by sim-
ulation. Firstly, the three-dimensional model according to the test field layout is set up,
distribution parameters are simulated according to the finite element method in Sect. 2,
and the equivalent branch of the insulator robot is added to the road model. The
insulators at high potential are added to the equivalent branch of the robot to calculate
the voltage values at both ends respectively, and the insulators without the robot branch
are calculated. Results of voltage distribution is shown in Table 1. When the robot is
close to one certain insulator, the voltage distribution of the simulation is different from
the original insulator voltage distribution, the maximum single voltage difference
The Electric Field Analysis and Test Experiments 213
appears on the first insulator, the difference is 3.45 kV, which is about 13.14%. The
overall difference of the whole series voltage is about 15.5 kV. When the robot is close
to insulators, the maximum difference between the simulated data and the measured
data is about 1.51 kV, and the difference is about 6.1%.
Because the insulator model is idealized, the material parameters are different from
data of the actual insulators, and the measurement error is considered at the same time,
the above error values can meet engineering requirements. It shows that the insulator
detection robot can effectively measure the insulator distribution voltage. However, due
to the difference between the total value of measurement and the actual voltage value, it
is not convenient for the staff to analyze. The voltage compensation method can be
studied so that the measurement results reflect the original voltage distribution of the
insulators as accurate as possible.
5 Conclusion
This paper designed a split type insulator detection robot which is applicable for
suspension insulator string detection, and carried out electric field analysis and test
study of insulator robot detection. Conclusions are as follows:
(1) The insulator robot has some influence on the local electric field distribution of
insulators, but it has little influence on the maximum field strength of insulators
which can be neglected. The maximum field strength of the insulator robot itself is
less than halo field strength of the air. The same research method can be used to
detect the low and zero insulators of the suspension porcelain insulator strings of
500 kV and 1000 kV transmission lines by further improving the insulation of the
insulators.
(2) The insulator robot can accurately and effectively measure the distribution voltage
of insulators. However, because of the influence of the probe and measuring cir-
cuit, the measurement value is different from the actual voltage distribution of the
insulator. The measured results should be compensated.
(3) The insulator robot can detect the distribution voltage effectively under the sim-
ulated environment with low and zero voltage value insulators. The distinguishing
and analysis of low and zero insulators can be carried out with the help of the
measured voltage distribution curve.
Acknowledgement. Fund Project: Technology project of State Power Grid Co., Ltd.
(No. 524625170046)
References
1. Zhibin, Q., Jiangjun, R., Daochun, H., et al.: Study on aging modes and test of transmission
line porcelain suspension insulators. High Volt. Eng. 42(4), 1259–1267 (2016)
2. Xian, Z., Richeng, L., Ming, Z., et al.: Online detection device of low or zero value
insulators in UHVDC Transmission line. Electr. Power 49(6), 90–94 (2016)
214 P. Yin et al.
3. Juan, J., Guanbin, W., Shiyou, M., et al.: The electric field analysis and optimal design on a
robot for insulator detection. Insul. Surge Arresters 2, 180–185 (2017)
4. Yuntu, J., Jun, H., Jian, D., et al.: The identification and diagnosis of self-blast defects of
glass insulators based on multi-feature fusion. Elect. Power 50(5), 52–58 (2017)
5. Feng, W., Hongcai, L., Xiang, P., et al.: Optimization research on condition-based
maintenance of overhead transmission line with online detection without power cut. Elect.
Power 49(10), 84–89 (2016)
6. Haipeng, S., Weiguo, L., Zhanzhan, Q.: Study on the on-line detection device of hand-held
insulator. Insul. Surge Arresters 1, 22–26 (2012)
7. Jiangang, Y., Ye, Z., Tangbing, L.I., et al.: Analysis of high-voltage ceramic insulators
infrared detection blind areas. High Volt. Eng. 43(9), 2903–2910 (2017)
8. Jing, C., Nan, Z., Zhong, M., et al.: Ultraviolet detection technology of 750 kV resistant to
porcelain insulator based on artificial pollution test. Elect. Power 49(9), 23–29 (2016)
9. Peng, L.: Study on Heating and Discharge Characteristics of Zero Value Insulators Based on
Infrared and Ultraviolet Imaging. North China Electric Power University, Beijing (2016)
10. Yong, F., Zongren, P., Peng, L., et al.: Voltage-sharing characteristics of porcelain insulators
for UHV AC transmission lines. High Volt. Eng. 36(1), 270–274 (2010)
11. Xiaowei, L., Linong, W., Junhua, W., et al.: Research on automatic detection technology for
faulty porcelain insulators on AC transmission lines. Electr. Meas. Instrum. 53(11), 110–115
(2016)
12. Liang, Z., Yong, L., Zhigang, R., et al.: Analysis on the influence factor of insulators
detection by robots and electric field method. Insul. Surge Arresters 2, 148–155 (2017)
13. Zhengzi, C., Weiguo, L., Wenbin, C.: Design and implementation of insulator resistance
measurement robot based on WiFi wireless remote control. Elect. Power 48(12), 23–26
(2015)
14. Rui, G., Bing, T., Lei, Z., et al.: Study on cap and pin porcelain insulator detecting
technology suitable for robots on transmission line. Insul. Surge Arresters 2, 141–147 (2017)
15. Yao, C., Dehua, Z., Jie, N.: The impact of zero and low resistance insulator on potential and
electric field distribution of long insulator strings in the 500 kV transmission line. Insul.
Surge Arresters 3, 29–34 (2015)
16. Ming, Z., Richeng, L., Jie, X., et al.: Feasibility research and develop of working detecting
robot for low/zero insulator. High Volt. Appar. 52(6), 160–166 (2016)
17. Linhua, Z., Yandong, C.: The research and design of overhead transmission line insulator
detecting robot. Electron. Des. Eng. 23(16), 164–166 (2015)
18. Tao, C., Daqing, S., Deli, Z., et al.: Research on the application of disc type porcelain
insulators detection robot in overhead transmission lines. Insul. Surge Arresters 2, 11–16
(2013)
Study on Early Fire Behavior Detection
Method for Cable Tunnel Detection Robot
1 Introduction
With the rapid development of cities, more and more cable tunnels have replaced
overhead lines. The power cables work under high voltage and high temperature
environment for a long time. Breakdown and short circuit may occur with insulation
aging. Further, the cable tunnel fire may be produced. In order to ensure the safe and
stable operation of power cables, monitoring and maintenance should be carried out.
Due to the narrow space and complex environment, it is very inconvenient to operate
and maintain the cable tunnel manually. Especially when a tunnel fire accident occurs,
it is not easy to find the hidden danger of the fire. Once lighting a fire, the fire will
develop rapidly along the cable line. Combustion produces a large amount of smoke,
toxic and harmful substances and high temperatures. These pose a threat to the personal
safety of the staff in the tunnel. The cable tunnel has a long distance, narrow pas-
sageways and few outlets. Rescuers are hard to get close to the source of fire.
Therefore, the cable tunnel inspection robot emerges as the times require [1–3].
Intelligent robots are used for inspection work, which can effectively reduce staff
burden. Furthermore, there are fixed-point detection systems in general tunnels, but
they have monitoring blind area. As a mobile monitoring system, the tunnel robot can
effectively compensate for the shortcomings of the fixed-point detection system. The
cable tunnel robot can be divided into walking and track type [4, 5]. The walking robot
is mainly a caterpillar type. The cable tunnels may have accumulated water, debris,
large slopes and complex pipelines, which are not conducive to walking. The orbital
robot walking on the track at the top of the tunnel is not affected by the above factors.
In 2009, the Shanghai Jiao Tong University developed a lightweight, small crawler
patrol robot [6]. The upper part arranged infrared cameras, gas sensors and signal
transceivers to collect and transmit the physical environment and gas environment in
the tunnel. In 2012, the Hangzhou power company produced a wheeled robot [7]. It
carries the gas sensor and the smoke sensor, can effectively perceive the potential
danger in the tunnel, and provide the data base for the fire prevention. The fire
extinguishing device can take effective measures to extinguish the fire according to the
feedback signal of the actual sensing device. However, because the robot carries too
many equipment, its weight and size exceed the standard. In 2014, the Xiamen Electric
Power Exploration Institute studied and produced a fixed track cable tunnel inspection
system [8]. The mobile mechanism marched along the predetermined track. Two
infrared camera devices were arranged to collect the images in the tunnel and an
ultrasonic sensing device to measure the obstacles. The whole mobile mechanism also
carries several gas sensors to detect the gas environment. However, the above research
focuses on the robot’s motion and data acquisition function. There are few studies on
the tunnel fire behavior judging by the collected data. If the fire source is detected in
time, the fire can be extinguishing as soon as possible.
In this article, the simulation model is established according to the actual size of the
cable tunnel. The tunnel fire is simulated under different fire power sources, and the
data of flue gas concentration and temperature in the tunnel are obtained. Consider the
dynamic sampling data of robots, the characteristic parameters are analyzed and
extracted. A fire detection method of cable tunnel robot based on SVM is proposed.
The effectiveness of the above method is verified by simulation data.
In order to study the early fire behavior detection method, the development law of the
tunnel fire needs to be obtained. The cable tunnel robot can measure temperature,
characteristic gas, particle concentration, etc. Unlike the fixed-point monitoring device,
the tunneling robot detects relevant data in the process of moving. It is necessary to
obtain data on temperature and smoke concentration measured by robots. However, the
fire tests are destructive and non-recoverable, and full-scale tests are expensive. A more
efficient method is as follows: first, the fire development rules are obtained by
numerical simulation, and then the simulation data are verified by typical tests. This
section will introduce the model building and simulation results.
Study on Early Fire Behavior Detection Method 217
According to the actual situation, there is ventilation equipment in the cable tunnel
under normal working conditions. In the early stage of the fire, the ventilation
equipment may be still working. Wind speed has a certain effect on the spread of the
fire. When the fire develops to a certain stage, trigger the linkage and close the fire
door. It is considered that the wind speed of the port is zero. It can delay the spread of
fire and reduce oxygen content. The two cases above are simulated respectively.
According to results of previous studies, the heat release rate of the cable material is
obtained. Combined with the shape and size of the cable in this article, the maximum
heat release rate of different numbers of cables is calculated. The fire power
218 B. Yan et al.
corresponding to different fire sizes is set to 1467 kW/m2, 2200 kW/m2 and
3385 kW/m2, respectively. The fire source is set as the T2 fire growth model. The
source of the fire rises to maximum power after 200 s, and then the power remains
unchanged. In order to simulate the situation of fire in different layers of cable, The fire
source height is set to 0.5 m, 0.9 m, 1.3 m, respectively. Considering the wind speed,
the power and the height of the fire source, different working conditions are set up as
shown in Table 1.
200 s, the cables in the top are all ignited. Typical fire spread situation is shown in
Fig. 2.
When the tunnel is not ventilated, the smoke spreads more evenly in the tunnel, and the
ventilation of the tunnel will intensify the burning of the cable and make the smoke
accumulate faster in the tunnel.
(3) Analysis of temperature variation
At the beginning of the combustion, the temperature distribution throughout the
tunnel is relatively uniform. During the development of the fire, the temperature grows
with the rise of smoke, which causes the high temperature to appear at the top of the
tunnel and spreads rapidly over the tunnel. The temperature declines with the distance.
This makes the suspended orbit inspection robot better able to detect changes in
temperature and smoke concentration on fire than the robot walking on the ground. The
typical tunnel temperature distribution is shown in in Fig. 4.
Take the following working conditions as an example: height of the fire source 0.9 m,
heat release power 3385 kW/m2 and wind speed 1 m/s. The variation of smoke con-
centration and temperature in cable tunnel fires is analyzed. The probe is arranged at
the height of the robot sensor. The smoke concentration and temperature measured by
the probe arranged in proper order from the fire source point are shown in Figs. 5 and
6. It can be seen from Fig. 5 that the temperature of the detection point above the fire
source point rises slowly due to the effect of ventilation. The temperature at the second
detection points in the downwind direction increased rapidly and the temperature value
is higher. For the turbulence produced by combustion has a great influence on the
temperature change. The temperature variation of several detection points near the fire
source point is obviously different from that of other detection points. The trend of
temperature of the detection point beyond the fire source 5 m is similar changing with
time. The difference in the rate of rise is small. But there are some differences in the
starting time of temperature rise at different locations. Therefore, when sampling
Study on Early Fire Behavior Detection Method 221
sequence is selected for analysis, the point with a certain distance from the fire source is
more regular. It can be seen from Fig. 6 that the smog concentration curve is rather
chaotic compared with the temperature. This is mainly due to the fact that the gener-
ation and diffusion of particles are very irregular under the action of flame turbulence.
Especially when the flame is burning fiercely, it is more difficult to make use of smoke
concentration. The smoke concentration changes more smoothly at a certain distance
from the fire source and when the concentration is low. Therefore, when selecting
smoke concentration data, a lower trigger value can be set. When the smoke concen-
tration exceeds the trigger value, a certain length of data is recorded and analyzed.
When the fire behavior is detected, the tunnel robot moves towards the fire source.
The data collected during the movement are different from those obtained directly by
simulation. It is necessary to analyze and process data with the speed of mobile robot.
In order to conduct fire source detection, it is necessary to determine the starting
position of the analysis data. Due to the random time and location of the fire, the robot
may begin to judge in different locations. For the temperature varies greatly with the
outside world, it is more reasonable to use smoke concentration as trigger criterion. It is
assumed that the analysis data begin to record when the smoke concentration exceeds
10%. Take the following working conditions as an example: height of the fire source
0.9 m and wind speed 1 m/s. The heat release power is 1467 kW/m2, 2200 kW/m2 and
3385 kW/m2, respectively. The smoke concentration and temperature detection value
under different heat release power and initial measurement position are shown in
Figs. 7 and 8. It shows that the sampling data sequence of smoke concentration and
temperature under different fire source power has certain rules. But it is difficult to
distinguish only through the curve.
222 B. Yan et al.
Fig. 6. Smoke concentration variation at different locations from the fire point
From the above analysis, it can be seen that fire detection is more accurate based on
robot detection data at the early stage of fire. In this paper, a fire detection method based
on SVM is proposed. When the smoke concentration exceeds the threshold, 10 tem-
perature values of robot sampling and 10 smoke concentrations are selected to form a
characteristic vector. It is input to SVM for classification and judgement. SVM
Study on Early Fire Behavior Detection Method 223
1 Xn
Min kW k2 þ P e
i¼1 i
ð3Þ
2
yi ðw /ðxi Þ þ bÞ 1 ei ; i ¼ 1; 2; ; n ð4Þ
Xl
f ð xÞ ¼ sgn a y KðX; Xi Þ þ b
i¼1 i i
ð5Þ
4 Conclusion
The simulation model is established according to the actual size of the cable tunnel in
this article. The tunnel fire is simulated under different fire power sources, and the data
of flue gas concentration and temperature in the tunnel are obtained. The variation law
of smoke concentration and temperature under different combustion conditions is
analyzed. Data sequence collected by the robot during a mobile process is analyzed.
The sampling data sequence of smoke concentration and temperature under different
fire source power has certain rules. The characteristic values are selected through
simulation data, and the fire detection model based on support vector machine is
trained. The accuracy of the test is 91.7%. The validity of the fire detection method is
verified.
Acknowledgments. This study was funded by the State Grid Corporation Headquarters Science
and Technology Project (524625160014).
References
1. Hui, Z., Ruiming, Q.: Mechanism design of a new kind of inspection robot for cable tunnel
and track optimization simulation. Mach. Des. Manuf. Eng. 47(3), 61–65 (2018)
2. Ingason, H., Li, Y.Z.: Model scale tunnel fire tests with longitudinal ventilation. Fire Saf. J. 45
(8), 371–384 (2010)
3. Rie, D.H., Hwang, M.W., Kim, S.J., et al.: A study of optimal vent mode for the smoke
control of subway station fire. Tunn. Undergr. Space Technol. 21(3), 300–301 (2006)
4. Victores, J.G., Martinez, S., Jardon, A., et al.: Robot-aided tunnel inspection and maintenance
system by vision and proximity sensor integration. Autom. Constr. 20(5), 629–636 (2011)
5. Li, Y.Z., Lei, B., Ingason, H.: The maximum temperature of buoyancy-driven smoke flow
beneath the ceiling in tunnel fires. Fire Saf. J. 46(4), 204–210 (2011)
6. Fu, Z., Chen, Z., Zheng, C., et al.: A cable-tunnel inspecting robot for dangerous
environment. Int. J. Adv. Robot. Syst. 5(3), 243–248 (2008)
7. Wei, H., Zhen, R.G., Ge, J., et al.: Research of patrol robot for electrical cable corridor.
Autom. Instrum. 12, 13–16 (2013)
8. Ling, F., Huaxia, Y., Dong, F.W.: Research on the moving structure of split cable tunnel
inspection robot. Process Autom. Instrum. 38(6), 46–50 (2017)
Speech Endpoint Detection Based
on Improvement Feature and S-Transform
1 Introduction
þZ1
Sðt; f Þ ¼ xðsÞgðt s; f Þej2pf s ds: ð1Þ
1
j f j ðtsÞ2 f 2
gðt s; f Þ ¼ pffiffiffiffiffiffi e 2 : ð2Þ
2p
From Eq. (1), we can see that ST can be regarded as a phase correction of a
continuous wavelet transform(CWT), it changes the corresponding scale parameters,
translation parameters of the wavelet transform, thus making up for the disadvantage of
phase information lacking in CWT. From Eq. (2), we can see that the window function
is characterized by both time and frequency characteristics, and the variance of the
Speech Endpoint Detection Based on Improvement Feature and S-Transform 227
window function is equal to the inverse of the frequency. Based on these character-
istics, satisfactory results obtained if a signal is transformed into the time-frequency
domain by ST and then processed by the method of spectral subtraction.
(2) S-transform
Transform the speech into the time-frequency joint domain.
(3) Through the Mel filter bank
Transform the signal into the Mel spectrum.
(4) Take logarithms for energy
The log energy output for each filter bank is:
!
X
N 1
2
E ðmÞ ¼ ln jXa ðkÞj Hm ðkÞ ; 0 m M: ð6Þ
k¼0
(5) Calculate discrete cosine transform (DCT) to get SMFCC cepstrum parameters:
X
N 1
pnðm 0:5Þ
smfccðnÞ ¼ E ðmÞcos : ð7Þ
m¼0
M
In Eq. (8), p refers to the order used in the SMFCC cepstrum analysis.
Which xðnÞ is a pure speech signal, d ðnÞ is a noise signal, and the s-transform is
performed on both sides of Eq. (9):
where m denotes the m-th sample point in the ST, and n denotes the n-th frame.
Ys ðm; nÞ can be obtained by direct discrete s-transform. However, the noise energy
spectrum Ds ðm; nÞ cannot be obtained accurately and has to be obtained approximately
by the way of estimation.
h The inormal estimation method in [9, 10] is to take the
statistical average E jDs ðm; nÞj2 in the absence of speech as jDs ðm; nÞj2 .
Speech Endpoint Detection Based on Improvement Feature and S-Transform 229
It is known that the length of the noise segment is IS, and the corresponding frame
number is NIS. The average energy value of the noise segment can be obtained as:
h i 1 XNIS
E jDs ðm; nÞj2 ¼ jXs ðm; nÞj2 : ð11Þ
NIS i¼1
By obtaining the power spectrum of the silent speech segment and subtracting the
power spectrum of the original speech, then the power spectrum of the relatively pure
speech signal is processed by square root, and after the inverse S-transform (IST), a
relatively pure speech signal xei ðkÞ can be obtained:
!
X 1XN1
xei ðkÞ ¼ f i2pnk
Xs ðm; nÞ e N ; 0 k N 1: ð12Þ
N m¼0
Noise
estimation
Quadratic
Framing ST
amplitude
square root
IST
of amplitude
The characteristics of noise and speech in the frequency domain are quite different. In
general, the voice energy of a voiced segment has a large change with the frequency
band, and there is a larger peak at the resonance peak, but the energy value in the noise
segment is relatively low, and more evenly distributed within the frequency band.
According to this characteristic of the speech signal, the band variance becomes the
230 L. Xunbo et al.
most commonly used characteristic parameter. The uniform subband separation fre-
quency band variance is proposed as the speech characteristic parameters of the signal
based on this characteristic [11].
p is so called uniform subband, which means each subband is equal bandwidth. The
amplitude of the m-th subband in the i-th frame is:
X
1 þ ðm1Þp þ ðp1Þ
XXi ðmÞ ¼ jXi ðkÞj: ð14Þ
k¼1 þ ðm1Þp
Assuming XXi ¼ fXXi ð1Þ; XXi ð2Þ; . . .; XXi ðqÞg, Then we can get the sub-band
mean value of the i-th frame speech signal:
1X
q
Ei;1 ¼ XXi ðk Þ: ð15Þ
q k¼1
The sub-band variance of the i-th frame speech signal can expressed as:
1 X
q
Di;1 ¼ ½jXXi ðkÞj Ei;1 2 : ð16Þ
q 1 k¼1
The most commonly method used in endpoint detection is a dual threshold method. By
extracting typical characteristic parameters of the speech, an proper threshold is
selected to determine the start and end points of speech and to detect speech frames and
non-speech frames [12]. This paper selects the improved spectral subtractive SMFCC
cepstral distance and uniform subband variance as features, adopts a dynamic threshold
estimation mechanism and uses the two-parameter double threshold method to perform
endpoint detection. The procedure of the spectral subtraction algorithm based on ST is
given as follows:
Speech Endpoint Detection Based on Improvement Feature and S-Transform 231
7 Experiment Results
The speech database used in the experiments here is 126 pure voices in the TIMIT
voice database. The noise signals used in the simulation includes 5 kinds of noise
(white, pink, volvo, babble and factory noises) of NOISEX-92 Datebase. The pure
speech signals and various noise signals are mixed at 3 different SNR(10 dB, 5 dB, and
0 dB). In order to verify the correctness and robustness of this method, we compared
the proposed algorithm with the algorithms mentioned in literature [4] and literature
[13]. All three algorithms perform well when SNR is 10 dB, the detection results with
SNR of 5 dB and 0 dB are shown in Figs. 3, 4 and 5, it can be seen that when the SNR
is 5 dB, the detection method based on the DWT-MFCC cepstrum distance missed
speech signals between 3 s and 3.5 s; when the SNR is 0 dB, the two comparison
algorithms were not able to detect the behind speech segments while the improved
232 L. Xunbo et al.
Pretreatment
No First-level
judgment Thresholds T3 T4
update dynamically
Exceeded the
threshold?
No
Yes
Second-level
judgment Exceeded the
threshold?
Yes
Speech frame
1
Amplitude
1
0 0
-1 -1
0 0.5 1 1.5 2 2.5 3 3.5 0 0.5 1 1.5 2 2.5 3 3.5
Ɵme/s Ɵme/s
Noise-added speech signal (SNR=5dB) Noise-added speech signal (SNR=0dB)
Amplitude
Amplitude
1 1
0 0
-1 -1
0 0.5 1 1.5 2 2.5 3 3.5 0 0.5 1 1.5 2 2.5 3 3.5
Ɵme/s Ɵme/s
Short-term spectrum reducƟon SMFCC cepstrum distance Short-term spectrum reducƟon SMFCC cepstrum distance
Amplitude
Amplitude
20 20
0 0
0 0.5 1 1.5 2 2.5 3 3.5 0 0.5 1 1.5 2 2.5 3 3.5
Ɵme/s Ɵme/s
Bandwidth variance for short-Ɵme uniform subband separaƟon
Amplitude
200 200
100 100
0 0
0 0.5 1 1.5 2 2.5 3 3.5 4
0 0.5 1 1.5 2 2.5 3 3.5 4 Ɵme/s
Ɵme/s
Fig. 3. The endpoint detection results of improved algorithm based on ST in SNR = 5 dB and
SNR = 0 dB.
algorithm in this paper can still detect the position of each segment of speech. The
experiment results can clearly show that the proposed algorithm has higher robustness
for speech endpoint detection.
In order to further verify the stability and effectiveness of the proposed algorithm,
the correct rate results are used to evaluate the detection results under different noise
Speech Endpoint Detection Based on Improvement Feature and S-Transform 233
Amplitude
Amplitude
0 0
-1 -1
0 0.5 1 1.5 2 2.5 3 3.5 0 0.5 1 1.5 2 2.5 3 3.5
Noise-added speech signal (SNR=5dB) Noise-added speech Signal (SNR=0dB)
1 1
Amplitude
Amplitude
0 0
-1 -1
0 0.5 1 1.5 2 2.5 3 3.5 0 0.5 1 1.5 2 2.5 3 3.5
DWT-MFCC Cepstrum distance DWT-MFCC Cepstrum distance
Amplitude
Amplitude
10
10
0
0 0.5 1 1.5 2 2.5 3 3.5 0
Ɵme/s 0 0.5 1 1.5 2 2.5 3 3.5
Ɵme/s
Fig. 4. The endpoint detection results based on DWT-MFCC cepstrum distance in SNR = 5 dB
and SNR = 0 dB.
0 0
-1 -1
0 0.5 1 1.5 2 2.5 3 3.5 0 0.5 1 1.5 2 2.5 3 3.5
Noise-added speech Signal(SNR=5dB) Noise-added speech signal(SNR=0dB)
1 1
Amplitude
Amplitude
0 0
-1 -1
0 0.5 1 1.5 2 2.5 3 3.5 0 0.5 1 1.5 2 2.5 3 3.5
Spectral subtraction and uniform sub-band spectrum variance Spectral subtraction and uniform sub-band spectrum variance
200 200
Amplitude
Amplitude
100 100
0 0
0 0.5 1 1.5 2 2.5 3 3.5 4 0 0.5 1 1.5 2 2.5 3 3.5 4
Ɵme/s Ɵme/s
Fig. 5. The endpoint detection results based on spectral subtraction and uniform sub-band
spectrum variance in SNR = 5 dB and SNR = 0 dB.
Table 1. The correct rate of different algorithms in different noises (SNR = 5 dB).
Noise Proposed DWT- Spectral Proposed algorithm
algorithm MFCC subtraction + uniform based on ST in pure
based on ST cepstrum sub-band spectrum speech
distance variance
white 92.7% 84.4% 90.5% 96.2%
pink 91.6% 82.8% 90.6% 99.1%
volvo 91.3% 80.4% 90.3% 97.6%
babble 90.4% 81.3% 89.9% 96.4%
factory 90.2% 79.7% 89.5% 96.0%
234 L. Xunbo et al.
Error rate/%
DetecƟon rate/%
Fig. 6. The detection rate and error rate of different algorithms under different SNR
environments when the SNR is 5 dB. As shown in Table 1, it can be clearly seen that
the improved algorithm in this paper has better detection accuracy.
The detection rate and error rate are used to evaluate the superiority of this algo-
rithm. It is easy to see that the improved algorithm has a higher detection rate and lower
error rate relative to the compared algorithms from Fig. 6. The error rate of the pro-
posed algorithm is relatively stable under different SNR, which illustrates that this
algorithm has more stable and reliable performance, stronger anti-noise robustness in
speech endpoint detection.
8 Conclusion
In the low SNR and non-stationary noise environment, traditional feature detection
methods will lead to a sharp drop in detection performance. This paper proposed a
feature-improved speech endpoint detection algorithm based on s-transform. This
method combines s-transform with MFCC and spectral subtraction to reduct noise. The
experimental results show that this algorithm has a high accuracy and has advantages in
the detection rate and error rate compared to the other two algorithms in Literature [4,
13]. It can effectively suppress noise and improve the quality of speech.
References
1. Stockwel, R.G., Mansinha, L., Lowe, R.P.: Localization of the complex spectrum: the S-
transform. J. IEEE Trans. Signal Procesing. 44(4), 998–1001 (1996)
2. Huang, L., Yang, C.: A novel approach to robust speech endpoint detection in car
environments. In: IEEE International Conference on, vol. 3, pp. 1751–1754. IEEE (2000)
3. Nakagawa, S., Wang, L.: Speaker identification and verification by combining MFCC and
phase information. In: IEEE International Conference on Acoustics, vol. 20, no. 4, pp. 4529–
4532 (2009)
4. Yin, R., Cheng, J.: Improved feature extraction algorithm based on DWT-MFCC. J. Modern
Electron. Technol. 40(9), 18–21 (2017)
Speech Endpoint Detection Based on Improvement Feature and S-Transform 235
5. Zhang, Z., Yao, E., Shi, Y.: Audio endpoints detection algorithm based on wavelet analysis
and MFCC. J. Electron. Meas. Technol. 39(7), 62–66 (2016)
6. Zeng, S., Jingxiang, L.: Speech endpoint detection method based on fusion of MFCC
distance and logarithmic energy parameter. J. Audio Eng. 40(9), 51–55 (2016)
7. Cao, D., Gao, X., Gao, L.: An improved endpoint detection algorithm based on MFCC
Cosine Value. J. Wirel. Pers. Commun. 95, 2073–2090 (2017)
8. Wang, H., Yu, Z.: SMFCC: a novel feature extraction method for speech signal. J. Comput.
Appl. 36(6), 1735–1740 (2016)
9. Goh, Z., Tan, T.-C., Tan, B.T.G.: Postprocesing method for suppressing musical noise
generated by spectral subtraction. IEEE Trans. Speech Audio Process. 6(3), 28–292 (1998)
10. Paliwa, K., Wojcicki, K., Schwerin, B.: Single-channel speech enhancement using spectral
subtraction in the short-time modulationdomain. J. Speech Commun. 52(5), 450–475 (2010)
11. Wang, Z.-F.: Speech endpoint detection method research based on double threshold-
frequency band variance. Electron. Des. Eng. 24(19), 86–88 (2016)
12. Sun, Y., Wu, Y., Li, P.: Research on speech endpoint detection based on the improved dual-
threshold. J. Chang. Univ. Sci. Technol. (Nat. Sci. Ed.) 39(1), 92–95 (2016)
13. Wang, W., Hu, G., Yang, L., et al.: Research of endpoint detection based on spectral
subtraction and uniform sub-band spectrum variance. Audio Eng. 40(5), 40–43 (2016)
A Localization Evaluation System
for Autonomous Vehicle
1 Introduction
However, due to the uncertainty of the mobile robot itself, it will inevitably affect
the accuracy of localization. Scholars have also conducted research on this issue. As the
researches did in [8], using entropy and mutual information metrics to estimate
quantitative uncertainty estimates and the effectiveness of information theory concepts
within a probabilistic framework. In [9], it suggests a position estimation algorithm
based upon an adaptive fading Kalman filter to overcome the inaccurate results due to
limited linear approximation, which the adaptive fading factor enables the estimator to
change the error covariance according to the real situation. And the literature [10]
associates the new observations to existing estimates in terms of the Kullback-Leibler
divergence to evaluate the proposed method to augment standard dynamic object
trackers which solve the problem that other traffic participants are often occluded from
sensor measurements by buildings or large vehicles in autonomous driving scenarios.
But at present, there are few methods to evaluate the results of the localization of
autonomous vehicles.
To solve this problem, we construct a localization evaluation system based on
Gaussian distribution model that is independent of its own localization system and is
related to the actual environmental information. The Kullback–Leibler divergence and
Fourier transform methods are used to process the localization data to complete the
evaluation of the localization effect.
1
pðxÞ ¼ detð2pRÞ2 expf ðx lÞT R1 ðx lÞg:
1
ð1Þ
2
Grid Hypothesis Model. Agrid hypothesis model using the existing prior information
of the surrounding environment and posteriori information of current laser real-time
collection of autonomous vehicle is constructed and it is the basis and focus of this
localization evaluation system. The prior information is the set of existing high-
definition maps and the array of data collected by the laser sensor in the area with a
certain accuracy. The posterior information is the array of data of the current auton-
omous vehicle collected by the laser sensor in real time. Comparing the posteriori
information with the priori information on the collected data of the corresponding
position of the autonomous vehicle, the matching degree of the two can be obtained.
We use the matching degree to evaluate the localization results of the autonomous
vehicle. Figure 2 illustrates how to get the priori data array and posterior data array.
As shown in Fig. 2, if the data are collected in the A and B directions, the resulting data
array is A = [0.3, 0.5, 0, 0.2, 0.7, 0.4, 0, 0], B = [0, 0.2, 0.7, 0.4, 0, 0, 0.3, 0.5]. The
priori data array and an array of posterior data can be obtained by collecting data by
laser. Then we construct a hypothesis grid to describe the model.
The output from the localization algorithm as a center to construct a localization
frame composed of m n grids. Each grid acquires data in l directions by laser and a
total of m n l sets of data are obtained. Figure 3 is a top view of the grid local-
ization frame.
A Localization Evaluation System for Autonomous Vehicle 239
Comparing the above m n l sets of data with the corresponding current real-
time laser acquisition data to obtain the matching degree, in order to better describe the
matching degree in each grid, we simulated a visualized image of the matching degree.
For the convenience to display, one of the l directions in each grid is taken, the red
arrow shown in Figs. 4 and 5, and the height(ordinate) of the red arrow is the matching
value of prior data and posterior data.
Finally, on the basis of constructing a real-time grid hypothesis evaluation model,
Kullback–Leibler divergence and Fourier Transform methods are used to evaluate the
current localization results.
240 Y. Yin et al.
the value of DKL ðPjjQÞ indicates the information loss when using Q to approximate P,
if it is close to 0, the two distributions have similar behavior, while a value close to 1
means two very different distributions. In the evaluation model, P denotes the data
distribution collected in the localization hypothesis grid, Q denotes the theoretical data
distribution of the localization hypothesis grid in the collection data set of the region
and DKL ðPjjQÞ denotes the relative entropy of the two distributions. By calculating the
distance between the data of the localization hypothesis grid and the priori data of the
localization hypothesis grid in the coordinate system, the confidence is obtained by
combining the DKL ðPjjQÞ parameter, thereby completing the localization evaluation.
Fourier Transform. Fourier Transform is an algorithm that samples a signal over a
period of time (or space) and divides it into its frequency components. Fourier analysis
converts a signal from its original time domain to a representation in the frequency
domain and vice versa. The following are the formulas for the discrete Fourier forward
and inverse transforms:
X
N1
i2pkn
XðkÞ ¼ xðnÞe N
n¼0
ð3Þ
X
N 1
2pkn 2pkn
¼ xðnÞ½cos i sin ; k ¼ 0; 1; . . .; N 1:
n¼0
N N
A Localization Evaluation System for Autonomous Vehicle 241
1XN 1
i2pkn
xðnÞ ¼ XðkÞe N
N k¼0
ð4Þ
1XN 1
2pkn 2pkn
¼ XðkÞ½cos þ i sin ; n ¼ 0; 1; . . .; N 1:
N k¼0 N N
where the latter expression follows from the former one by Euler’s formula:
3 Experiments
3.1 Evaluation Steps
Step 1: Using the pose6D localization model introduced to Euler angles to locate the
autonomous vehicle. A high-definition map and a collection of arrays of data col-
lected by the laser sensor in that area are ready.
Step 2: After several experiments on the localization of the autonomous vehicle, the
accuracy of the optimal localization frame is:
X dimension: 63; Y dimension: 63; theta dimension: 32; theta resolution: 0.00436;
0.16 meters per pixel.
The data in its localization frame are collected by laser sensor.
Step 3: Comparing the data obtained in the previous step with the priori data set
corresponding to the first step to obtain the heat map.
In this heat map, red is darkest than other colors, from red to background blue is from
dark to light. The darker the color, the closer the two sets of data are. The smaller
offset in the first heat map in Fig. 6 indicates that the localization is more accurate, the
offset in the second heat map is centered, and the third is larger, indicating poor
localization results.
Step 4: Using Kullback–Leibler divergence and Fourier transform respectively for
localization evaluation. We define data from grid hypothesis½i as dfgh½i and
data from grid pose½i as dfp½i. According to (2), we can obtain the relative
entropy:
242 Y. Yin et al.
dfgh½i
KLparam
¼ 100 dfgh½i ln : ð6Þ
dfp½i
Fig. 7. Evaluation confidence and localization results by using the Fourier transform
A Localization Evaluation System for Autonomous Vehicle 243
Fig. 8. Evaluation confidence and localization results by using the Fourier transform
Fig. 9. Evaluation confidence and localization results by using the Kullback–Leibler divergence
Fig. 10. Evaluation confidence and localization results by using the Kullback–Leibler
divergence
In Figs. 7, 8, 9 and 10, the y axis represents the value of evaluation confidence and
localization result that change with time. In Fig. 7, when the localization fails, the
confidence changes obviously and the localization failure can be immediately judged
based on the result, while in Fig. 9, the confidence changes slowly and the evaluation
speed is slower. In Fig. 8, when some small localization errors occur (such as the
impact of pedestrians on the road, changes in some buildings, etc.), result is judged as
244 Y. Yin et al.
localization failure, as shown in Fig. 10 that the evaluation analysis using the Kull-
back–Leibler divergence method is more accurate.
4 Conclusion
In this note, we have considered the evaluation of the localization effect of autonomous
vehicles under real environmental information using the Kullback–Leibler divergence
and Fourier transform evaluation models.
The experimental results show that the two evaluation models can evaluate the
localization effect. The Fourier transform evaluation model evaluates the localization of
the autonomous vehicle faster than the Kullback–Leibler divergence evaluation model.
However, the Kullback–Leibler divergence evaluation model is more effective than the
Fourier transform evaluation model in dealing with some unexpected conditions
encountered by autonomous vehicles under the real environmental information, and it
is more accurate in localization evaluation.
References
1. Kurashiki, K., Fukao, T., Ishiyama, K., Kamiya, T., Murakami, N.: Orchard traveling UGV
using particle filter based localization and inverse optimal control. In: The 3rd Symposium
on System Integration, pp. 31–36. IEEE Computer Society, Japan (2010)
2. Thrun, S., Burgard, W., Fox, D.: Probabilistic Robotics. The MIT Press, England (2006)
3. Shoukry, Y., Abdelfatah, W.F., Hammad, S.A.: Real-time Markov localization for
autonomous UGV. In: 4th International Design and Test Workshop. IEEE Computer
Society, Saudi Arabia (2009)
4. Tripathi, S.K., Sapre, R.M.: Robust target localization and tracking using Kalman filtering
for UGV-UAV coordinated operation. In: 2016 IEEE International Conference on Recent
Advances and Innovations in Engineering, India (2016)
5. Tsalatsanis, A., Valavanis, K.P., Kandel, A., Yalcin, A.: Multiple sensor based UGV
localization using fuzzy extended Kalman filtering. In: 2007 Mediterranean Conference on
Control and Automation. IEEE Computer Society, Greece (2007)
6. Kwon, J.-W., Park, M.-S., Kim, T.-U., Chwa, D.-K., Hong, S.-K.: Localization of outdoor
wheeled mobile robots using indirect Kalman filter based sensor fusion. J. Inst. Control. Rob.
Syst. 14, 800–808 (2008)
7. Nguyen, N., Tyagi, D., Shin, V.I.: An observation model based on polyline map for
autonomous vehicle localization. In: Proceedings of the IEEE International Conference on
Industrial Technology, India, pp. 2427–2431 (2006)
8. Messina, E., Madhavan, R.: Intelligent unmanned ground vehicle navigation via information
evaluation. Adv. Rob. 20, 1375–1400 (2006)
9. Sung, W.J., Choi, S.O., You, K.H.: TDoA based UGV localization using adaptive Kalman
filter algorithm. In: 2008 2nd International Conference on Future Generation Communica-
tion and Networking Symposia, China, vol. 4, pp. 99–103. IEEE Computer Society (2008)
A Localization Evaluation System for Autonomous Vehicle 245
10. Galceran, E., Olson, E., Eustice, R.M.: Augmented vehicle tracking under occlusions for
decision-making in autonomous driving. In: IEEE/RSJ International Conference on
Intelligent Robots and Systems, Germany, pp. 3559–3565 (2015)
11. Pfaff, P., Plagemann, C., Burgard, W.: Gaussian mixture models for probabilistic
localization. In: IEEE International Conference on Robotics and Automation, USA,
pp. 467–472 (2008)
12. Wikipedia. https://en.wikipedia.org/wiki/Kullback%E2%80%93Leibler_divergence#cite_
note-1
Cooperative Slip Detection Using
a Dual-Arm Baxter Robot
1 Introduction
The manipulation of objects by robots is typically accompanied by events such
as slippage between a robot’s fingers or, in the case of two non-dexterous end-
effectors, robotic arms. Humans are able to identify such events using a combi-
nation of superficial and deep mechanoreceptors [1], but for robots, with more
elementary mass, stiffness and friction sensing, such events can be hard to dis-
tinguish [2].
In industry many items manipulated by robots are too wide or heavy to
be practically held in the grasp of a single enclosed end-effector due to size
or payload limitations. In certain scenarios manipulators may work together in
c Springer Nature Singapore Pte Ltd. 2018
K. Li et al. (Eds.): ICSEE 2018/IMIOT 2018, CCIS 924, pp. 246–258, 2018.
https://doi.org/10.1007/978-981-13-2384-3_23
Cooperative Slip Detection Using a Dual-Arm Baxter Robot 247
pairs, where an object may be held in a friction only grasp between the end-
effectors, without additional support employed [3].
In the field of robotic slip detection and control, much work has been done
on single dexterous end-effectors. While several methods exist to detect slip,
such as the ones using accelerometers [4,5] and optical sensors [6,7] the main
method of detecting slip is via pressure signal analysis of piezoresistive force
sensing resistors (FSR) mounted on contact points [8]. These sensors encounter
a change in resistivity which is directly proportional to the change in pressure.
A typical time domain FSR sensor signal with slip occurring can be seen in
Fig. 1. Such a sensor is incorporated into a control loop to detect slippage based
on empirical mode decomposition in [2]. This approach relies on the analysis of
vibrations generated during slippage, however challenges are faced in discrim-
inating between noise and the small signals generated by slippage because of
low signal-to-noise ratio, necessitating the use of time domain filtering on any
feedback signals. In [9], a novel flexible, thin and lightweight FSR was developed
in order to detect slip. The authors encountered issues differentiating between
slip and changes in normal force, and so in [10], they investigated a method
for identifying object slip by analysing the frequency components of the output
measurements from the sensor. End-effector mounted strain gauge signals were
analysed after Fast Fourier Transformation in [11]. Seven band-pass filters with
different frequency bands were designed based on these signals and a threshold
for each defined. Once slip is detected, the parameters of a sliding mode slip
prevention controller are updated so that the grip force is increased in discrete,
predetermined amounts to prevent the object slipping further.
Fig. 1. A typical FSR time domain signal response with obvious high frequency com-
ponents while slip is occurring, from 5 to 7 s [2].
Similarly, in [12], FSR sensor slip signals were transformed using FFT. This
data was then used to train a standard artificial neural network for incipient
slip detection, which successfully detected slippage of five different objects, of
varying materials.
The optimal sensing layout for a single-end effector with two load cells mea-
suring normal and tangential force used in conjunction with an optical sensor is
investigated in [13]. Slip is induced through the modulation of grasp force applied
to an object. The paper concludes that it is preferable to monitor normal force,
248 S. Trimble et al.
2 System Description
This section discusses the robotic platform employed in this study including the
custom end-effectors and the Robot Operating System (ROS). The filter design
and test objects used in the experiments are also described.
Fig. 3. 3D printed end-effector fitted to Baxter. Note the three rubber contact points
for increased gripping.
2.2 End-Effectors
In order to promote the platform, a research version of the robot was released
in early 2013 along with an open-source Software Developement Kit (SDK),
integrated with the Robot Operating System (ROS) [21]. ROS is an open-source,
meta-operating system of libraries and tools built onto Ubuntu, which provides
a networked framework for controlling and communicating with robots through
publishing/subscribing to/from Rostopics. The Rostopics employed in this study
are summarized in Table 1.
The ROS network and robot block diagram is shown in Fig. 4. This peer-
to-peer network consists of one master and one client machine. The Rosmaster
PC is an Ubuntu machine with ROS installed. On this machine three custom
nodes are running. The torque control node is a custom script used to change the
torque values of the Rostopic .../gravity compensation torques/actual ef f ort.
It is commanded with an external hand-held controller, connected over USB. The
collision avoidance suppression node publishes a simple message to the collision
avoidance node to deactivate the subsystem. This normally-on safety feature is
constantly reset, so the custom node needs to publish at a minimum of 10 Hz.
Cooperative Slip Detection Using a Dual-Arm Baxter Robot 251
The Baxter robot communicates with the ROSmaster PC using TCP-IP pro-
tocol over ethernet, with a low latency of <1 ms. The robots S0 torque joint top-
ics are passed torque values using ROSmessages sent from the torque control
node in the ROSmaster PC. This applies normal force, F , through the end-
effectors onto the test object. F can then be decremented by the user to entice
slip.
Preliminary testing revealed a vast amount of undesirable signals and noise when
attempting to detect slip, due to end-effector motion and constant error correc-
tion. The torque value of the joint W 0 is measured in Nm at a sampling rate
of 95 Hz, similar to the 100 Hz sampling rate used in [22], but below the 400 Hz
used in [5,23].
Due to the real-time operation of the system, a low order infinite impulse
response (IIR) filter was preferable to minimise excessive computation and ensure
that the rosnode containing the filter would operate at the correct frequency of
200 Hz. This value was chosen as it is high enough to prevent aliasing (>2F s),
yet sufficiently low that the node scan-cycle does not have any latency issues.
252 S. Trimble et al.
As mentioned earlier, in order to design a slip control system for a dual arm
system, it is necessary to find an appropriate feedback signal for control purposes.
Determination of normal force, F , the force applied to the object by the end-
effectors during manipulation, is necessary to enable optimisation of the force in
order to prevent slip.
Fig. 6. Tangential and normal forces acting on an object while being manipulated by
two non-dexterous collaborating end-effectors, in blue (Color figure online)
It can be established from Fig. 6 that if one or both manipulators are equipped
with force feedback, F can be adjusted to a safe minimum in order to hold
the object securely without any additional support. By applying minimal F
the object can be held safely without the risk of crushing [13]. This can be
calculated when the tangential force acting on the object, Ft , and the interface
static coefficient of friction, µs , are known, as in Eq. 1:
Ft
F = (1)
µs
Cooperative Slip Detection Using a Dual-Arm Baxter Robot 253
Ft = m(g + a) (3)
Z-plane acceleration of grippers, a, must be taken into account when the robot
is in motion, along with g, acceleration due to gravity.
In order to apply the minimum grasping force to hold an object where µs
and m are unknown, F cannot be calculated. When Ft is greater than F , the
object will begin to move, therefore, the point at which minimum F is applied
can be assumed as the point of impending slippage [13]. TS0 , torque at shoulder
joint S0, and m is recorded for each test. Combining Eqs. 1, 2 and 3 gives the
relationship:
TS0 d(g + a)
= (4)
m µs
Therefore for an object of the same µs and d acting on the same system but
different m, it can be said that the ratio of threshold torque to mass remains
the same, that is:
TS01 TS02
= (5)
m1 m2
4 Experimental Methodology
The two end-effectors were commanded to identical Cartesian coordinates in the
X and Z planes in order to keep them at the same height and distance from
the robot. Both were also set to the same roll and yaw, with opposing pitch in
order to get the flat surface of the end-effectors to face each other to allow for a
grasping action as left and right F were increased.
Each object was held manually between the end-effectors and TS0 incre-
mented, decreasing the difference in Y until a firm grasp was achieved, in the
manner of Fig. 6. Torque was monitored to ensure a constant steady-state value
before slipping is induced.
The torque level on the right S0 joint was decremented manually using an
external hand-held controller until slipping was observed. The real-time torque
254 S. Trimble et al.
value of left joint W 0 was input into the filter discussed in Sect. 2.4 and the
output recorded, along with all left arm joint torque values. Slip quality was
monitored by the observer, watching for any sudden drops or unwanted parox-
ysmal movement, signifying a material that normal force is difficult to modulate
for. The observer description of the contact properties of the intersection between
the end-effectors and the object is recorded.
The experiment was repeated five times for each of the eight material/mass
combinations where slip could be successfully detected. The average slip thresh-
old torque values of left joint S0 and the filter output were recorded for each
repetition, and the ratio of these values to mass calculated to validate Eq. 5.
5 Results
Quantitative analysis of four of the materials where slip was successfully recorded
is provided in Table 2, which lists the type and description of materials used,
their masses, average threshold forces at which slip occurs, average peak values
of vibrations in joint W 0 during slip and ratios of threshold and vibration forces
to mass for each material. Figure 7 depicts torque (Nm) against time (samples)
for time windows, in which slip occurs, of the eight experiments where the top
rows (i) and (ii) correspond to the standard material whilst rows (iii) and (iv)
correspond to the same material but with additional mass. The plots reveal
the difference in slip threshold force at joint S0 and magnitude of vibrations
observed at joint W 0 as mass is increased and are typical of the materials where
a consistent good quality slip signal was observed. It can be seen that when the
slip threshold measured at S0 is crossed as in rows (i) and (iii) of Fig. 7, a spike
Fig. 7. Rows (i) and (ii) show the S0 force and absolute filter output respectively for
time windows, in which slip occurs, of the lighter mass experiments; rows (iii) and (iv)
for the heavier. The dashed lines mark the point of impending slip.
Cooperative Slip Detection Using a Dual-Arm Baxter Robot 255
in the absolute filtered signal value W 0 is observed (see corresponding rows (ii)
and (iv)). The magnitude of the spike varies depending on a combination of mass
and surface friction properties.
Fig. 8. Average frequency magnitude compared to the mass for four different objects.
The average magnitude of vibration against object’s mass for the four materi-
als is depicted in Fig. 8, showing that as mass increases, the frequency component
of the force in the 9–15 Hz range during slip at W 0 increases.
Additional materials including sheets of steel, cardboard, polypropylene and
a hard card box were also tested but proved to be challenging with regard to slip.
These are excluded from Table 2 as due to their slip behaviour it was not possible
to acquire a satisfactory slip signal with such measurements. These issues are
further discussed in the following section.
6 Discussion
As can be expected, for objects with the same coefficient of friction where slip
could be modulated consistently, the magnitude of vibration is directly propor-
tional to the mass due to additional tangential force. There is also a relationship
256 S. Trimble et al.
between the threshold holding torque at joint S0 and this tangential force, as
seen in Table 2. This was true in general for the experiments either with a glossy,
but sticky material such as plastics or laminated card, and materials with a high
friction surface, such as MDF. The ratios of both normal force to mass and
vibration magnitude to mass are approximately adhered to, as per Eq. 5.
As mentioned earlier, it was not possible to detect slip for all the materials
tested due to the intermittent or sudden nature of the slip observed. ‘Slip-stick’
is a complex known problem due to dynamic instabilities [4]. Materials with this
characteristic are not best suited to a noisy moving robotic platform, and in the
case of this study they made joint S0 force control difficult to modulate.
The strip of cardboard exhibited behaviour in this manner due to it being an
easily compressible material with a non-uniform thickness. The polypropylene
material exhibited similar behaviour, because, although smooth, it was very
sticky and required a large mass in order to cause slip. This created undesirable
torsion forces acting on the end-effectors which disrupted the contact shape
between the material and the robot. The steel sheet, with it’s relatively high
mass and very slippery surface slid out of the grasp suddenly on all five occasions,
preventing a distinct slip signal from being observed. The three-dimensional hard
card box experienced changes in orientation as slipping occurred, thus shifting
the angle of the contact surfaces. This caused sporadic vibrations forces acting
on W 0, making a slip signal problematic to capture.
In this paper a dual-arm Baxter collaborative robot with custom friction grip
end-effectors is introduced. Experiments were conducted with different materials
with varying surface friction properties and weight and real-time time-domain fil-
tering implemented on a joint torque signal to detect slip. Time domain analysis
of the filtered and threshold signals revealed that the approach was effective at
detecting slip for the laminated card, MDF, PVC and Perspex, as these materials
slipped in a smooth manner. The steel, cardboard and polypropylene exhibited
‘slip-stick’ behaviour, making it difficult to capture a slip signal. It would be
valuable to further investigate data in Fig. 7, by increasing data points with
additional materials and weights. In future work, improved signal conditioning
will be performed to enhance the slip detection signal. It would also be useful
to investigate the filter time delay, as the filtered torque signal is to be used as
a feedback signal for a fully automated closed-loop slip controller on the dual
manipulator system.
Cooperative Slip Detection Using a Dual-Arm Baxter Robot 257
References
1. Gerling, G.J., Rivest, I.I., Lesniak, D.R., Scanlon, J.R., Wan, L.: Validating a
population model of tactile mechanotransduction of slowly adapting type I afferents
at levels of skin mechanics, single-unit response and psychophysics. IEEE Trans.
Haptics 7, 216–228 (2014)
2. Deng, H., Zhong, G., Li, X., Nie, W.: Slippage and deformation preventive control
of bionic prosthetic hands. IEEE/ASME Trans. Mechatron. 22, 888–897 (2017)
3. Xian, Z., Lertkultanon, P., Pham, Q.C.: Closed-chain manipulation of large objects
by multi-arm robotic systems. IEEE Rob. Autom. Lett. 2, 1832–1839 (2017)
4. Won, H.-I., Chung, J.: Numerical analysis for the stick-slip vibration of a trans-
versely moving beam in contact with a frictional wall. J. Sound Vibr. 419, 42–62
(2018)
5. Khurshid, R.P., Fitter, N.T., Fedalei, E.A., Kuchenbecker, K.J.: Effects of grip-
force, contact, and acceleration feedback on a teleoperated pick-and-place task.
IEEE Trans. Haptics 10, 40–53 (2017)
6. Erhart, S., Hirche, S.: Model and analysis of the interaction dynamics in coopera-
tive manipulation tasks. IEEE Trans. Rob. 32, 672–683 (2016)
7. Saen, M., Ito, K., Osada, K.: Action-intention-based grasp control with fine finger-
force adjustment using combined optical-mechanical tactile sensor. IEEE Sens. J.
14, 4026–4033 (2014)
8. Aqilah, A., Jaffar, A., Bahari, S., Low, C.Y., Koch, T.: Resistivity characteristics
of single miniature tactile sensing element based on pressure sensitive conductive
rubber sheet. In: 2012 IEEE 8th International Colloquium on Signal Processing
and its Applications, pp. 223–227, March 2012
9. Teshigawara, S., Ishikawa, M., Shimojo, M.: Development of high speed and high
sensitivity slip sensor. In: 2008 IEEE/RSJ International Conference on Intelligent
Robots and Systems, pp. 47–52, September 2008
10. Teshigawara, S., Tadakuma, K., Ming, A., Ishikawa, M., Shimojo, M.: High sensi-
tivity initial slip sensor for dexterous grasp. In: 2010 IEEE International Conference
on Robotics and Automation, pp. 4867–4872, May 2010
11. Engeberg, E.D., Meek, S.G.: Adaptive sliding mode control for prosthetic hands
to simultaneously prevent slip and minimize deformation of grasped objects.
IEEE/ASME Trans. Mechatron. 18, 376–385 (2013)
12. Schoepfer, M., Schuermann, C., Pardowitz, M., Ritter, H.: Using a Piezo-resistive
tactile sensor for detection of incipient slippage. In: ISR 2010 (41st Interna-
tional Symposium on Robotics) and ROBOTIK 2010 (6th German Conference
on Robotics), pp. 1–7, June 2010
13. Dzitac, P., Mazid, A.M., Ibrahim, M.Y., Appuhamillage, G.K., Choudhury, T.A.:
Optimal sensing requirement for slippage prevention in robotic grasping. In: 2015
IEEE International Conference on Industrial Technology (ICIT), pp. 373–378,
March 2015
14. Li, Z., Tao, P.Y., Ge, S.S., Adams, M., Wijesoma, W.S.: Robust adaptive control
of cooperating mobile manipulators with relative motion. IEEE Trans. Syst. Man.
Cybern. Part B (Cybern.) 39, 103–116 (2009)
15. Lotfavar, A., Hasanzadeh, S., Janabi-Sharifi, F.: Cooperative continuum robots:
concept, modeling, and workspace analysis. IEEE Rob. Autom. Lett. 3, 426–433
(2018)
16. Arshad, A., Badshah, S., Soori, P.K.: Design and fabrication of smart robots. In:
2016 5th International Conference on Electronic Devices, Systems and Applications
(ICEDSA), pp. 1–4, December 2016
258 S. Trimble et al.
17. Makrini, I.E., Rodriguez-Guerrero, C., Lefeber, D., Vanderborght, B.: The variable
boundary layer sliding mode control: a safe and performant control for compliant
joint manipulators. IEEE Rob. Autom. Lett. 2, 187–192 (2017)
18. Chen, B., Wan, J., Shu, L., Li, P., Mukherjee, M., Yin, B.: Smart factory of industry
4.0: key technologies, application case, and challenges. IEEE Access 6, 6505–6519
(2018)
19. ReThink Robotics: Baxter hardware specifications, May 2014. Accessed 14 Jan
2018
20. Cremer, S., Mastromoro, L., Popa, D.O.: On the performance of the baxter research
robot. In: 2016 IEEE International Symposium on Assembly and Manufacturing
(ISAM), pp. 106–111, August 2016
21. Rethink robotics opens up baxter robot for researchers (2018). Accessed 5 Feb 2018
22. Stachowsky, M., Hummel, T., Moussa, M., Abdullah, H.A.: A slip detection and
correction strategy for precision robot grasping. IEEE/ASME Trans. Mechatron.
21, 2214–2226 (2016)
23. Damian, D.D., Newton, T.H., Pfeifer, R., Okamura, A.M.: Artificial tactile sensing
of position and slip speed by exploiting geometrical features. IEEE/ASME Trans.
Mechatron. 20, 263–274 (2015)
24. Tuononen, A.J.: Onset of frictional sliding of rubber–glass contact under dry and
lubricated conditions. Sci. Rep. 6(1) (2016)
25. Xu, F., Yoshimura, K.-I., Mizuta, H.: Experimental study on friction properties of
rubber material: influence of surface roughness on sliding friction. Procedia Eng.
68, 19–23 (2013)
Intelligent Control and Automation
Study on Swimming Curve Fitting
of Biomimetic Carangiform Robotic Fish
1 Introduction
After a long period of natural continuous choice, fabulous evolution of the fishes makes
their swimming ability meet the needs of survival. According to the different swimming
duration time the fish movement can be divided into two patterns: cruise and
maneuvering. The cruise refers to the long-term steady-state swimming over a long
period of time, while the maneuvering refers to generating a great acceleration or a
large steering torque in a short period of time. Also, the propulsion patterns can be
divided into two categories depending on the different parts of fish body that can
generate thrust: Body and/or Caudal Fin propulsion (BCF) and Media and/or Paired Fin
propulsion (MPF). The carangidae fish studied in this paper uses the BCF propulsion
pattern, which relies on the swing of the 1/3 parts of the body and the caudal fins, to
generate thrust. Based on the prototype of carangidea fish the study is implemented on
both structural appearance and athletic stance of swimming in this paper. The
propulsive efficiency of fish is more than 80%. Therefore, in order to generate high
speed with high propulsive efficiency the motion control for the biomimetic robotic fish
should be well studied.
The kinematics theory of robotic fish has been studied since 1920s. In 1960,
Lighthill introduced the “slender body theory” in aerodynamics into the hydrodynamic
analysis of fish swim. Using carangidae fish as the research object, he initially estab-
lished the carangidae fish kinematic model. Experimental studies on fish dynamics
mainly focused on kinematic parameters and various forms of fish movement. In 1933,
Gary recorded the movement of fish by camera, and the theory of undulatory propulsion
from the motion of the cross section. Domenici and Blake studied the C-shape start of
the angel fish, suggesting that the size of the fish is independent of the duration of the
escape process but linearly related to the turning radius. In 1999, Spierts gave the C-
sharp and S-sharp turning curves of crucian carp, and studied the kinematics difference
between the two starting methods respectively [11]. In 2001, Rosenberger observed the
morphological parameters and swimming patterns of the myliobatis-shaped fish, and
compared the effects of pectoral fin motion parameters on the swimming speed and
propulsion performance. At present, there are two main control algorithms for robotic
fish motion control: Central Pattern Generator (CPG) and curve fitting. CPG is a control
algorithm using the neural network, which is based on the motion mechanism of the
animal, and it is used as the bottom controller to generate the swimming posture of the
robotic fish. Since CPG has a high degree of non-linear characteristic, which leads to big
obstacles for its theoretical research and practical application. Moreover, the basic
movement of the CPG is composed of a series of rhythmic movements, which can
respond quickly for the robotic fish cruise movement, but it is difficult to achieve its
maneuvering. The curve fitting is to fit the trajectory of the fish through each joint of the
robotic fish, which is suitable for all kinds of motion forms, meanwhile the design
process is simple and convenient. In the traditional curve fitting, the end points fall on
the curve in the fitting process, which lead to larger fitting error. In this paper, the
minimum error criterion is proposed for curve fitting as a metric.
In this paper, we use the curve fitting method to control the three-joint biomimetic
carangiform robotic fish Based on the metric of minimum error criterion. Therefore, the
end point of the joint does not fall on the fitted curve as the traditional methods.
Experimental results show that the minimum error curve fitting method can make the
robotic fish movement more smoothly and improve the velocity of biomimetic robotic
fish movement compared with the traditional curve fitting method.
According to Lighthill’s theory of slender body [8], the fish wave is a traveling wave
group with a gradually increasing wave amplitude. It can be approximated by a
polynomial and a sinusoidal curve:
Through the biological studies we can see that the swing amplitude of the tail
motion joint is 0.075–0.1 times of the body length, and fish wave wavelength is usually
k 1LB , so the fish wave number is k ¼ 2p=k 2p=LB [16]. Therefore, in the absence
of special instructions, the parameters used in this paper are assumed as c1 ¼ 0:05,
c2 ¼ 0:03, k ¼ 8, x ¼ 2p=9. The values here are negative because the origin point of
the model is set the connection point of the head and tail movement joints, and then the
fish head direction is negative. The body wave motion equations of the biomimetic
robotic fish using FB ðx; tÞ can be drawn by MATLAB shown in Fig. 1. Each curve in
Fig. 1 corresponds to the movement posture of the robotic fish at a given moment, and
the fluctuation starting point of the tail movement joint is 0 point.
Due to technical limitations, robotic fish can not be infinitely flexible at present, but
instead of flexible curves it can be approximated by finite rigid joints. Because of the
limitation of the biomimetic robotic fish joint length, it is difficult to use the analytical
method to fit the equation of the multi-joint biomimetic robotic fish’s body wave
motion. Therefore, the numerical approximation method is used. The method firstly
divides the body wave motion equation of the biomimetic robotic fish into N move-
ments with time varying as FB ðx; iÞði ¼ 0. . .. . .NÞ, and the gestures of each moment are
fitted using three rigid joints as ðl1 ; l2 ; l3 Þ. In this way, the fish postures at all times can
be replaced by approximated rigid joints of robotic fish. Figure 2 shows the robotic fish
motion at a certain moment during cruise, while Fig. 3 shows the body wave motion
curve corresponding to this posture.
264 B. Lou et al.
In Fig. 3, hi;j ði ¼ 0. . .M 1; j ¼ 1. . .3Þ is the slip angle of each joint relative to the
horizontal line, /i;j ði ¼ 0. . .M 1; j ¼ 1. . .3Þ is rotation angle which the individual
joint of the biomimetic robotic fish need to be rotated, that is, the deflection angle of the
back joint relative to the front joint. For example, “/i;2 ” means the angle at which the
second joint is required to be rotated relative to the first joint at the time of the
movement posture FB ðx; iÞ. If this formula is expanded to a K-joint biomimetic robotic
fish, the result of numerical approximation for the rotation angle using the body wave
motion equation FB ðx; tÞ can be expressed as /i;j ði ¼ 0. . .M 1; j ¼ 1. . .KÞ.
In references [18, 19], a method of curve fitting is proposed. In the case of known
joint length, the starting point of the first joint coincides with the starting point of the
wave, and then the circle is rounded with the length of the joint. The intersection with the
curve is the coordinate of the other end point of the joint. The starting point of the second
joint is the end point of the first joint, and the end point coordinate of the second joint is
obtained by using the above mentioned method. Similarly, the coordinate of the third
joint is obtained, then the fitted curves of the three-joint robotic fish motion are obtained.
Study on Swimming Curve Fitting of Biomimetic Carangiform Robotic Fish 265
In the process of fitting, the end of each joint falls on the theoretical fish wave curve,
which causes large fitting errors, and the fluctuation curve of biomimetic robotic fish is
not smooth enough to generate accurate biomimetic robotic fish swimming motion.
In this paper, the minimum error fitting method is proposed to fit the body wave
motion curve of biomimetic robotic fish. Assuming that the biomimetic robotic fish has
k joints with the length of each joint is lj ðj ¼ 1. . .kÞ; the starting coordinate of each
joint is ðxb ; yb Þ, the end point coordinate is ðxe ; ye Þ, which means that the starting
coordinate of the j th joint is the end coordinate of the j 1 th joint; finally, each joint
can be represented by a linear equation y ¼ gðxÞ. Then, the fitting error equation can be
written as:
Z xe
eðxÞ ¼ ½gðxÞ FB ðxÞdx: ð2Þ
xb
Although the fitting error includes amplitude error and direction error, only the
amplitude error is considered in this paper, since the direction error is minimized to a
certain small value at the same time when the amplitude error is minimized. Formula
(6) represents only one joint fitting error, and it can be expanded to k th joint error
equation as:
Z
xei;j
ei;j ðxÞ ¼ ½gi;j ðxÞ FB ðx; iÞdx; ð3Þ
xb
i;j
where i ¼ 0. . .M 1; j ¼ 1. . .k,
From the above analysis, we can see that the coordinates of each joint are related,
so the starting coordinates of each biomimetic robotic fish joint can be described by the
following equation:
8
>
> xb ¼ 0 j¼0
< y i;j ¼ 0 j¼0
bi;j
ð5Þ
>
> x b ¼ xei;j1 j ¼ 2. . .k:
: i;j
ybi;j ¼ yei;j1 j ¼ 2. . .k
The initial coordinate of the first joint is (0, 0), which is the connection point
between the biomimetic robotic fish head and the tail movement joints. The connection
point is the starting point of the fish wave motion. And since the length of each joint of
the biomimetic robotic fish is given, the following equation can be obtained:
Therefore, the problem of curve fitting is transformed into the problem of finding
the joint end point coordinates in order to minimize the error equation ei;j ðxÞ when the
motion attitude equation FB ðx; iÞ with the starting point coordinate and length of the
joint are known at a certain moment. So that the following formula can be established:
dei;j ðxÞ
¼0 ð7Þ
dx
It is obviously impossible to solve Eq. (11) by simultaneous Eqs. (7), (8), (9) and
(10), because the constraint condition (9) and (10) themselves are interdependent and
not completely independent of each other. Therefore, we assume that a length pro-
portional correlation coefficient R 2 ½0; 1 the ratio between the joint length between the
starting point and the intersection of the rigid joint and the body wave curve and the
total length of the joint. Then, for an arbitrary value of R, the intersection of the joint j
and the motion curve ðxci;j ; yci;j Þ can be expressed by the following equation:
ðx xbi;j Þ2 þ ðy ybi;j Þ2 ¼ ðRlj Þ2
ð8Þ
y ¼ FB ðx; iÞ;
where: i ¼ 0. . .M 1; j ¼ 1. . .k
ðxci;j ; yci;j Þ can be obtained according to the above formula. Therefore, the linear
equation gi;j ðxÞ of the joint j can be easily obtained when knowing the coordinate of
starting point and the intersection coordinate. Rp ¼ pn ; p ¼ 0. . .:n is introduced to
calculate the minimum fitting error:
The fitting curve is obtained using the above equations, and then the rotation angle
equation of each joint of the biomimetic robotic fish can be obtained by fitting the
curve. The rotation angle equation is given as:
/i;j ¼ hi;j j¼1
ð10Þ
/i;j ¼ hi;j hi;j1 j ¼ 2. . .k
In order to control the motion of a robotic fish, the rotation angle /i;j at each
moment is integrated into a time varying equation /j ðtÞ. Since the joint rotation of the
robotic fish is a periodic motion that changes over time, the shape of the motion curve
of each joint is similar to a sine function. It can be expressed by Fourier series as:
X
1
/j ðtÞ ¼ an;j sinðnxt þ un;j Þ; j ¼ 1. . .k: ð11Þ
n¼1
Study on Swimming Curve Fitting of Biomimetic Carangiform Robotic Fish 267
The fitting curves of the traditional fish joint are plotted by using the angle control
equation, as shown in Fig. 4. While the minimum error method for the robotic fish joint
curves are shown in Fig. 5. One can see that there are no significant differences on the
shape of the joint fitting curve using traditional fitting method and the proposed
minimum error method. However, the initial phase angle and amplitude of each joint
movement are different. And the difference between these two parameters will directly
affect both the swimming posture and swimming speed of the robotic fish.
In order to verify the effect of the initial phase angle and amplitude of each joint on the
swimming posture and swimming speed of the biomimetic robotic fish, a three-joint
biomimetic carangiform robotic fish prototype is developed, as shown in Fig. 6, and
the following experiment was carried out in the experimental system shown in Fig. 7.
268 B. Lou et al.
The fitting curves shown in Figs. 4 and 5 are realized on the three-joint biomimetic
carangiform robotic fish prototype. Figure 8 is a set of continuous experimental pictures
using the traditional fitting curve method for experimentation. While Fig. 9 is a set of
Study on Swimming Curve Fitting of Biomimetic Carangiform Robotic Fish 269
continuous experimental pictures using the proposed minimum error fitting curve method
for experimentation. Compared with the two sets of pictures, one can see that using the
minimum error fitting curve to control the robotic fish swimming trajectory has a higher
similarity with the swimming trajectory of real fish, and the curve is more smoothly in the
process of movement. Comparing the two sets of pictures with the corresponding theo-
retical curve, one can find that the swimming curve of the robotic fish with the minimum
error fitting curve is closer to the theoretical curve and has better linearity. Moreover,
comparing the duration time taken by robotic fish after swimming the same distance, it
can obviously be found that the duration time of the robotic fish using the minimum error
fitting curve method is less than that of using the traditional method. The average velocity
is improved by 7%, and the swimming posture is getting more stable.
5 Conclusion
In this paper, the curve fitting method of biomimetic carangiform robotic fish has been
studied, and the experimental results have been validated. The following conclusions
can be drawn as: A minimum error curve fitting method has been proposed for the
three-joint biomimetic carangiform robotic fish. The method does not require the end
point of the joint falling on the swimming curve. Experimental results have proved that
the proposed method can make the fish swimming curve closer to that of the real fish.
The fitted motion curve has been applied to a robotic fish prototype, which has proved
that the proposed method can make the robotic fish swim more smoothly and improve
the swimming speed rapidly.
References
1. Yu, J., Ding, R., Yang, Q., Tan, M., Wang, W., Zhang, J.: On a bioinspired amphibious
robot capable of multimodal motion. IEEE/ASME Trans. Mechatron. 17(5), 847–856 (2012)
2. Lu, X.Y., Yin, X.Z.: Propulsive performance of a fish-like travelling wavy wall. Acta Mech.
175(1), 197–215 (2005)
3. Hong, C.: Kinematic mechanism research on the swimming and maneuvering of robot fish.
Ph.D. dissertation, University of Science and Technology of China, China (2006)
4. Sfakiotakis, M., Lane, D.M., Davies, J.B.C.: Review of fish swimming modes for aquatic
locomotion. IEEE J. Oceanic Eng. 24(2), 237–252 (1999)
5. Hu, T., Lin, L., Zhang, D., Wang, D., Shen, L.: Effective motion control of the biomimetic
undulating fin via iterative learning. In: 2009 IEEE International Conference on Robotics
and Biomimetics (ROBIO), Guilin, pp. 627–632 (2009)
6. Low, K.H., Willy, A.: Biomimetic motion planning of an undulating robotic fish fin. J. Vib.
Control 12(12), 1337–1359 (2006)
7. Liu, J., Dukes, I., Knight, R., et al.: Development of fish-like swimming behaviours for an
autonomous robotic fish. Univ. Bath 15(1), 6–9 (2004)
8. Lighthill, M.J.: Note on the swimming of slender fish. J. Fluid Mech. 9(2), 305–317 (1960)
9. Taylor, G.: Analysis of the Swimming of Microscopic Organisms. Proc. Roy. Soc. Math.
Phys. Eng. Sci. 209(209), 447–461 (1951)
10. Domenici, P., Blake, R.W.: The kinematics and performance of the escape response in the
anglefish (Pterophyllum eimekei). J. Exp. Biol. 156, 187–205 (1991)
11. Spierts, I.L., Leeuwen, J.L.: Kinematics and muscle dynamics of C- and S-starts of carp
(Cyprinus carpio L.). J. Exp. Biol. 202(Pt 4), 393 (1999)
12. Rosenberger, L.J.: Pectoral fin locomotion in batoid fishes: undulation versus oscillation.
J. Exp. Biol. 204(2), 379–394 (2001)
13. Ijspeert, A.J.: Central pattern generators for locomotion control in animals and robots: a
review. Neural Netw. 21(4), 642–653 (2008)
14. Chen, W.H., Ren, G.J., Zhang, J.B., Wang, J.H.: Smooth transition between difierent gaits of
a hexapod robot via a central pattern generators algorithm. J. Intell. Rob. Syst. 67(3–4), 255–
270 (2012)
15. Chuan, L.K., Kit, T.W.: Task learning utilizing curve fitting method for kinect based
humanoid robot. In: 2014 IEEE International Conference on Robotics and Biomimetics
(ROBIO 2014), Bali, pp. 1505–1511 (2014)
Study on Swimming Curve Fitting of Biomimetic Carangiform Robotic Fish 271
16. Liu, Y.-X.: The entity design and dynamic research on the two-joint robot fish. Master
dissertation, Harbin Institute of Technology, China (2007)
17. Liu, J., Hu, H.: Biological inspiration: from carangiform fish to multi-joint robotic fish.
J. Bionic Eng. 7(1), 35–48 (2010)
18. Liu, J., Hu, H.: Mimicry of sharp turning behaviours in a robotic fish. In: 2005 IEEE
International Conference on Robotics and Automation, pp. 3318–3323. IEEE (2005)
H1 Filter Designing for Wireless Networked
Control Systems with Energy Constraints
1 Introduction
Over the past twenty years, WNCS has been widely expended in various kinds of
applications including remote operations, aircraft, marine monitoring, intelligent
transportation, and manufacturing plants [1]. These processes are often linked to pass
the data collected through the sending of a real-time wireless network medium. WNCS
has numerous advantages to energy saving optimization and actual industrial control
system [2]. Nevertheless, a great many new problems and difficulties are created for the
wireless network when it is applied to industrial control system. As the wireless net-
work is an unstable shared channel, it will make stability analysis difficult and decrease
the system control performance [3–5]. And the limited energy of wireless sensor nodes
will reduce the lifetime of the WNCS. Consequently, it is of enormous scientific
significance value to do research on the problem of filtering technique and controller
design for WNCS with the energy constraints [3].
Two of the central issues are delay and packet-dropout, which has been extensively
investigated in the last decade. A lot of researchers have tried to solve the control
performance with delay and packet-dropout [4, 5]. Several good results have been
available in the existing literatures. For instance, Zhen et al. [6] researched the problem
of modeling and stabilization for WNCS with both time-varying delay and packet-
dropout, which are modeled as an asynchronous dynamic system with three subsystems.
And the sufficient condition for the system was obtained. Peng et al. [7] investigated the
problem that the energy efficiency could be much improved while preserving the desired
H1 performance. And the self-triggered sampling scheme was proposed. Dong et al. [8]
designed the robust filter for a class of discrete-time WNCS with communication
constraints and packet dropouts by using switched system theory. In conclusion, the
majority of the above researches center on the issues of the delay and packet losses. The
filtering technology is adopted in the external noise disturbance. On the foundation of
the research achievements above, we are aimed at investigating the problem of filter
designing and stabilization analysis for a class of WNCS with energy constraints.
The rest of paper is organized as follows. In Sect. 2, we propose a method to build a
new modeling of WNCS with energy constraints. In Sect. 3, the sufficient conditions of
the H1 filtering are given. In addition, and the gain used in the H1 filtering can be
derived by LMI technology. Then the effectiveness of the proposed method is proved
by the simulation results in Sect. 4. At last, the conclusions are given in Sect. 5.
2 Problem Formulations
The block diagram of a class of WNCS with filter is shown in Fig. 1. We can know that
wireless sensor nodes contains a fixed topology previously, and transmits data from
plant to filter nodes by reducing-frequency transmission.
Because the energy in WNCS is limited, the energy saving strategy is shown in
Fig. 1 to prolong the service life of wireless sensor. The time-varying sampling is used
to reduce the consumption of data acquisition of wireless sensor nodes, and then the
reducing-frequency transmission is used to save the energy consumption of data
transmission phase. The plant R1 can be described as follows:
8
< x_ ðtÞ ¼ AxðtÞ 1 wðtÞ
þB
yðtÞ ¼ CxðtÞ þ B 2 wðtÞ ð1Þ
:
zðtÞ ¼ C0 xðtÞ
274 L. Wei et al.
d2 d1
hk ¼ d1 þ eð Þ e ¼ 0; 1; 2; ; c: ð3Þ
c
So the range of the system sampling period hk is in the finite set X ¼ fd1 ; ;
d1 þ eðd2 d
c Þ; ; d2 g.
1
R hk
where A ¼ eAhk ; B1 ¼ 1.
eAS dsB
0
R Dh
Setting Dh ¼ hk d1 and Fk ¼ 1 ðeADh
eAS ds ¼ A IÞ, then we have
0
k þ I.
eADh ¼ AF
H∞ Filter Designing for Wireless Networked Control Systems 275
1
k þ I Þ ¼ eAd ~
A ¼ eAðd1 þ DhÞ ¼ eAd1 eADh ¼ eAd1 ðAF þ eAd1 AF k ¼ A þ M1 Fk N1 ð5Þ
~ ¼ eAd1 ; M1 ¼ eAd1 A;
N1 ¼ I.
where A
Similarly, we have
Z d1 Z hk
B1 ¼ 1 þ
eAS dsB 1
eAS dsB
0 d1
Z d1 Z hk d1
¼ AS 1 þ
e dsB 1
eAðS þ d1 Þ dsB
0 0 ð6Þ
Z d1 Z Dh
1
¼ 1 þ eAd
eAS dsB 1
eAS dsB
0 0
~ þ M2 F k N 2
¼B
R R
~ ¼ d1 eAS
where B dsB
1
1 ; M2 ¼ eAd 1 ; Fk ¼ Dh eAS
; N2 ¼ B
ds.
0 0
According to the above analysis, the discrete model of system (4) can be described
as:
8
< xðk þ 1Þ ¼ A ~ þ M1 Fk N1 xðkÞ þ B~ þ M2 Fk N2 wðkÞ
yðkÞ ¼ CxðkÞ þ B2 wðkÞ ð7Þ
:
zðkÞ ¼ C0 xðkÞ
R
where A
1
~ ¼ eAd
; M1 ¼ eAd1 A; ~ ¼ d1 eAS
N1 ¼ I; B dsB
1
1 ; M2 ¼ eAd 1 ; Fk ¼
; N2 ¼ B
R Dh AS
0
0 e ds
For the system (7), rmax ðAÞ is represent the largest singular value of system
T
parameters. Then Fk Fk k I, k [ 0 and satisfied the following Equation.
2
ermax ðAÞðd2 d1 Þ ermax ðAÞð0Þ
k¼ ð8Þ
rmax ðAÞ
N#ðkÞ ¼ diag½dðd 1Þ; dðd 2Þ dðd nÞ, d is the Kronecker delta function. The
data input of the filter for WNCS with energy constraints can be described as:
Considering the scheduling matrix N#ðkÞ is a random matrix with time varying, the
closed loop filtering error system can be described as a discrete system with n mode. In
order to design the H1 filter for the discrete control system (7), which is shown in
Eq. (10), we should get the following two conditions [7].
(1) When wðkÞ ¼ 0, the closed loop model of the filter error (11) is stochastic sta-
bility, if and only if
X
1
li m xðkÞT xðkÞ\1 ð12Þ
0
H∞ Filter Designing for Wireless Networked Control Systems 277
(2) At the condition of zero initial state, the filtering error system (11) is used to suppress
the disturbance level c [ 0, then the H1 performance c can be described as:
P
1 P
1
where keðkÞk22 ¼ eðkÞT eðkÞ , kwðkÞk22 ¼ wðkÞT wðkÞ.
k¼0 k¼0
3 Main Results
Our work is aiming at designing the filter for WNCS so as to solve the problems of
energy constraints. And prior to giving a detailed introduction of the main results, the
sufficient convergence condition for WNCS is given.
Theorem 1: Considering the system (11), given positive constant c [ 0. Then the
augmented system is asymptotical stable if there exist symmetric constant matrix
Pj [ 0, Pi [ 0 and
2 3
Pi
6 0 c2 I 7
6 7\0 ð14Þ
4 U B Pj 1
5
C 0 0 I
Hence the system (11) is asymptotically stable as well as having the H1 perfor-
mance index.
where
2 3
P1
6 7
~ 1 ¼ 6 0 c I
2
P 7
4 P2 P3 P4 5 ; P1 ¼ Pi
P5 0 0 I
278 L. Wei et al.
2 3
T ~ T T
J11 A þ V N#ðkÞ C þ J31 N#ðkÞ C U V I N#ðkÞ þ J31 ðI N#ðkÞ Þ
6 T~ T
T 7
P2 ¼ 4 J12 A þ V N#ðkÞ C þ J32 N#ðkÞ C U V I N#ðkÞ þ J32 ðI N#ðkÞ Þ 5
T ~ T T
J13 A þ J33 N#ðkÞ C 0 J33 ðI N#ðkÞ Þ
2 3
T ~ T
J11 B þ VN#ðkÞ B2 þ J31 N#ðkÞ B2
6 T~ T 7
P3 ¼ 4 J12 B þ VN#ðkÞ B2 þ J32 N#ðkÞ B2 5
T ~ T
J13 B þ J33 N#ðkÞ B2
P4 ¼ Pj J J T ; P5 ¼½ C0 W 0
T
2 3 2
T
3
J11 M1 0 0 J11 M2 0 0
P6 ¼ 4 0 0 0 5 ; P7 ¼ 4 0 0 05
0 0 0 0 0 0
2 3 2 3
0 0 0 0 ^1 0
N 0 0
6 0 0 0 07 6 0 N ^2 0 07
~ ¼6
M ~ ¼6
7; N 7
4 P6 P7 0 05 4 0 0 0 05
0 0 0 0 0 0 0 0
2 3 2 3 2 3
N1 0 0 N2 J11 J12 J13
^1 ¼ 4 0
N 0 ^ 2 ¼ 4 0 5; J ¼ 4 J22
0 5; N J22 0 5
0 0 0 0 J31 J32 J33
T T
And the parameters matrix of filter are Af ¼ J22 U; Bf ¼ J22 V and Cf ¼ W.
Remark: The optimal performance c can be obtained by solving the convex opti-
mization problem as follows:
J11 J33 ; Pi ; Pj ; U; V; W
pffiffiffiffiffiffiffiffiffi
So the optimal performance is c ¼ lmin .
4 Numerical Simulations
Assuming the minimum allowed sampling period of the system d1 ¼ 0:1s when
WNCS energy is sufficient, and the maximum allowed sampling period of the system
d2 ¼ 0:2s when WNCS energy is limited. The sampling period h1 ¼ 0:1s, h2 ¼ 0:15s
and h3 ¼ 0:2s. So the sampling period is randomly switched over in a finite set
X ¼ f0:1; 0:15; 0:2g. Setting #d ðkÞ ¼ f1; 2; 1; 2; g, N1 ¼ diag½ 1 0 and
N2 ¼ diag½ 0 1 , then by using Eq. (8), we have k ¼ 0:3014. When choosing the
noise wðkÞ ¼ 2e0:3k sinð0:5pkÞ and using theorem 2 and Matlab LMI Toolbox, we
obtain the optimal suppression level c ¼ 0:6427. At the same time, the feasible
solution of a set of H1 filter parameters under the optimal suppression level state is
obtained as follows.
0:1586 0:0983 1:6748 0:5342
Af ¼ ; Bf ¼
0:2366 0:1466 1:1469 0:3615
Cf ¼ ½0:6331 0:6664
Assuming that the initial values xf ð0Þ ¼ ½ 0 0 T and xð0Þ ¼ ½0:5 0:5T , the state
tracking curves of system and filter can be shown in Figs. 2 and 3 respectively.
1 0.5
x 1(k) x 2(k)
x 1f (k) x 2f (k)
0.5
x (k)and x (k)
x (k)and x (k)
2f
1f
0
2
1
-0.5
-0.5
0 10 20 30 40 0 10 20 30 40
k k
Fig. 2. State trajectory of x1 ðkÞ and x1f ðkÞ Fig. 3. State trajectory of x2 ðkÞ and x2f ðkÞ
In Figs. 2 and 3, x1 ðkÞ and x2 ðkÞ are the state of plant, and x1 f ðkÞ and x2 f ðkÞ are
the state of filter. It can be observed that xif ðkÞ can track xi ðkÞ effectively in 20 time
steps.
5 Conclusions
In this paper, the filtering problem for WNCS with energy constraints was considered.
By using time-varying sampling strategy, the energy consumption of wireless sensor
nodes in the data acquisition stage was reduced. And the mathematical model of
280 L. Wei et al.
Acknowledgments. This work was supported by natural science research programme of col-
leges and universities of an-hui province under grant KJ2016A062, natural science foundation of
an-hui province under grant 1608085MF146, foundation for talented young people of an-hui
polytechnic university under grant 2016BJRC008, visiting study Foundation for outstanding
young talent of an-hui educational committee under grant gxfxZD2016108, and the applied basic
research plan of nan-tong under grant GY12017015.
References
1. Ge, X., Yang, F., Han, Q.-L.: Distributed networked control systems: A brief overview. Inf.
Sci. 380, 117–131 (2017)
2. Zhang, X., Han, Q.: Event-triggered H∞ control for a class of nonlinear networked control
systems using novel integral. Int. J. Robust Nonlinear Control 27(4), 679–700 (2017)
3. Sadi, Y., SC, E.: Joint optimization of wireless network energy consumption and control
system performance in wireless networked control systems. IEEE Trans. Wirel. Commun. 16
(4), 2235–2248 (2017)
4. Bai, J., Renquan, L., Hongye, S., Xue, A.: Modeling and control of wireless networked
control system with both delay and packet loss. J. Franklin Inst. 352(10), 3915–3928 (2015)
5. Makled, E.A., Halawa, H.H., Daoud, R.M., Amer, H.H., Refaat, T.K.: On the perform ability
of hierarchical wireless networked control systems. Intell. Control Autom. 6(2), 126–133
(2015)
6. Hong, Z., Gao, J., Wang, N.: Output-feedback controller design of a wireless networked
control system with packet loss and time delay. Math. Probl. Eng. 1, 1–7 (2014)
7. Peng, C., Han, Q.: On designing a novel self-triggered sampling scheme for networked
control systems with data losses and communication delays. IEEE Trans. Ind. Electron. 63(2),
1239–1248 (2016)
8. Dong, H., Wang, Z., Gao, H.: Robust H1 filtering for a class of nonlinear networked systems
with multiple stochastic communication delays and packet dropouts. IEEE Trans. Signal
Process. 58(4), 1957–1966 (2010)
Distance Overestimation Error Correction
Method (DOEC) of Time of Flight Camera
Based on Pinhole Model
Le Wang, Minrui Fei, Hakuan Wang(&), Zexue Ji, and Aolei Yang
Abstract. Depth cameras with Time of Flight (ToF) technology are widely
used in machine vision and various measurement tasks. However, due to
hardware conditions and imaging characteristics, multiple errors limit the further
application of the ToF camera. This paper classified errors into errors caused by
non-imaging principle and error caused by imaging principle. In order to sim-
plify the experimental procedure and improve the efficiency of errors correction,
a simple and feasible method is used to correct errors caused by non-imaging
principle, and an evaluation function is proposed to determine the optimal ref-
erence distance, so as to select appropriate integration time and global offsets.
To tackle the radial distance error, Distance Overestimation Error Correction
method (DOEC) based on the principle of pinhole imaging is proposed to
further improve the accuracy of depth data. Finally, error correction methods
proposed in this paper are verified by experiments, and the segmentation of
different depth ranges is successfully achieved by using the modified data, prove
the effectiveness of the proposed methods.
1 Introduction
With the rapid development of machine vision system, different types of cameras play
significant roles in industrial intelligence more and more. Because of compact structure
and being less sensitive to illumination, time-of-flight (ToF) camera is widely applied
into many fields, such as mobile robot avoidance and navigation [1], object detection
and recognition [2, 3], 3D reconstruction and gesture recognition [4, 5].
Time-of-flight camera is a kind of active sensors, it emits near infrared (NIR) and
receive reflected light [6], and calculate the distance from the optical center of the
camera to an target object surface. There are some noise and errors exist in TOF camera
[7, 8], which cause negative impact on the accuracy of distance measurement [9]. Thus
it is essential to correct the data error of ToF camera at the initial procedure. There are
several source of errors in ToF camera, generally the error can be classified as sys-
tematic errors and non-systematic errors [10]. Systematic errors can be predicted and
corrected by different calibration methods, but the non-systematic errors are complexed
and unpredicted and generally removed by filtering.
Several phenomena from the idealistic model of ToF camera are analyzed, but the
temperature, multi-path reflections and ambient light effects are still hard to solve [11].
An error model is utilized to estimate ground and wall planes, acquisition geometry and
noise characteristic are analyzed in this error model [12]. A 2.5D pattern board with
holes is used to capture both color and ToF depth images for feature detection. The
transformation of the ToF camera is reset. In order to get accurate bias after ray
correction, k-means clustering and B-spline functions are utilized [1].
In ToF camera, the radial distance is acquired between the optical center to an target
object surface, which is longer than the actual distance. To remove the radial distance
error, this paper divides the various errors in the TOF camera into errors caused by the
non-imaging principle and error caused by the imaging principle. When compensating
errors caused by the non-imaging principle, data error caused by temperature changes
and integration time are analyzed and corrected, multipath errors and other noises are
compensated by taking offsets, and an evaluation function is proposed to determine the
optimal reference distance. To tackle the radial error caused by imaging principle,
Distance Overestimation Error Correction method (DOEC) based on the principle of
pinhole imaging is proposed. By analyzing the positional relationship between the
center pixel and other pixels in depth image, the angle relationship between the two is
obtained, and the radial distance is convert to vertical distance. Thus the right distance
of ToF camera is obtained.
The remainder of this paper is organized as follows. Section 2 is errors caused by
non-imaging principle in ToF camnera. Section 3 corrects the radial error caused by the
imaging principle. Section 4 states the experimental results.
TOF cameras are able to obtain distance information by utilizing the time of flight
measurement of photons. Photons are emitted by modulated infrared light. The mod-
ulated infrared light determines the phase shift u between reference signal and reflected
light. where c is the speed of light, f is the modulation frequency. The distance value d
is formulated as:
1 u
d¼ c ð1Þ
2 2pf
The temperature of the chip changes with the operating time, which will cause drift
of depth data, especially when the temperature of the chip changes significantly when
the ToF camera is just started. In order to reduce the deviation caused by temperature
changes, warm-up time of the chip is required before the measurement of ToF camera.
Observed from the experimental results of EPC660 chip, it takes about 15–20 min to
warm up before the depth image is acquired, and the obtained depth data is relatively
more stable and accurate.
DOEC Method of Time of Flight Camera Based on Pinhole Model 283
Integration Time (IT) is the time span when ToF camera computes multiple phase
of photons and obtain the distance. Thus the Integration Time is important to the
accuracy of ToF camera data. In this paper, when the integration time is 1500 ls, the
effect of integration time on the depth data is studied by taking pictures of flat walls at
different distances. As shown in Fig. 1(a), (b), (c), (d) is depth images acquired at
different distance. Different colors represent different distances and black represents
invalid points. It can be seen that when the integration time is fixed, the larger the
measurement distance is, the more invalid points are, and the more pixels that appear
jitter at the edge of the image, the accuracy of the measurement decreases. Therefore, it
is necessary to select an appropriate integration time.
Fig. 1. Depth image of different distance with integration time is 1500 ls (Color figure online)
Multipath errors and other noises can be eliminated through precise modeling, but
the implementation process is more complex. Some nonlinear errors will change with
distance, which are difficult to solve. In order to simplify the errors correction proce-
dure, this paper sets offsets to correct such errors.
Through the above analysis of the source of error, this paper corrects the error of
ToF camera under the range of 0–4 m. Since the nonlinear errors are related to the
distance, 1 m, 2 m, 3 m, and 4 m are utilized as the distance baseline for the experi-
ment. The experiment steps of errors correction caused by non-principle are shown as
follows, where 1 m is sets as the distance baseline firstly.
Step1: Warm up before starting. To reduce errors due to temperature changes,
image acquisition should start after 20 min so that the ToF camera has enough time
to warm up, meanwhile the data fluctuation range is small.
284 L. Wang et al.
Step2: Adjust the integration time. The integration time was adjusted enable the
amplitude at 1 m distance baseline reach about 1000LSB regarded as suitable
amplitude. Integration time did not change during the subsequent experiment.
Step3: Set the global offset. Since the data accuracy of the pixel at the optical center
is less affected by various errors than other pixels in the depth image, the trueness of
the data is higher. Therefore, the measured depth data at the optical center denoted
as basemeai is used to compute the global offset at the distance baseline of 1 m. i is
the representation of different distance baseline from 1 m to 4 m. The value of
global offseti can be formula as follow:
X
4
v ¼ ðacti meai Þ2 i ¼ 1; 2; 3; 4 ð3Þ
i¼1
If the difference between the actual distance and measured distance is smaller,
indicates that the accuracy of the measured data is higher, so the value of v is
smaller.
Step6: Repeat the above process by changing the distance baseline to 2 m, 3 m, and
4 m respectively.
Experiment results are shown in Table 1. When the distance baseline is 2 m, the
value of evaluation function v is the smallest than that of other distance baseline. It
indicates that the distance data obtained are more reliable, and errors caused by the
non- imaging principle have been improved to some extent. Thus the integration time
and global offset set at the distance baseline of 2 m are chosen as the better experiment
parameters in the 0–4 m operating range of ToF camera. Depth data acquired at
distance baseline of 2 m are also saved as data sample, which are used to correct the
errors caused by imaging principle in the next section.
DOEC Method of Time of Flight Camera Based on Pinhole Model 285
After correcting errors caused by the non-imaging principle, a more accurate depth
image is obtained. In this section, we analyze the error caused by the imaging principle
and propose a Distance Overestimation Error Correction method (DOEC) based on the
pinhole imaging principle.
Where hA is the angle between the straight line rA to camera plane, hC is the angle
between the straight line rC to camera plane. This kind of distance overestimation will
have a great impact on reconstruction of the 3D image.
A d
rA
A
B rB
C rC
Experiment has been done in front of smooth wall, the vertical distance between the
wall to the camera optical center Oc is 1 m. Figure 3 is the scatter diagram of the wall
after using the depth data collected from the experiment. It can be seen that the farther
the test point is away from the optic center Oc , the more obvious the overestimation of
the depth in the 3D image is, that means the depth value of points near the optical center
are closer to 1 m and the depth value near edge points are greater, even though they are
in same plane. So it is necessary to compensate such error in accurate measurement.
p
ob oc
r oa
θ
θ
p
f d
Fig. 4. Imaging model of ToF
d ¼ r cos h ð5Þ
The distance of each point can be corrected if h is knowable. h0 can be measured as:
O b p0
h0 ¼ arctanð Þ ð6Þ
f
DOEC Method of Time of Flight Camera Based on Pinhole Model 287
where Ob p0 is the image distance between p and image plane center Ob . The depth map
exists as a two-dimensional pixel matrix in the depth camera. Since there is a conversion
relationship between successive imaging plane and discrete image pixel coordinate
system, Ob p0 actually represents the true distance of pixel point in the discrete image
pixel coordinate system, and the position of p0 in the continuous imaging plane coor-
dinate system is known as (x, y). The coordinate transformation between the continuous
imaging plane and the discrete image pixel coordinate is shown in Fig. 5.
where u v represents the discrete pixel coordinate system, x y represents the
continuous image coordinate system, and from a continuous imaging plane to a discrete
image pixel coordinate system, p0 ðx; yÞ will be transformed into ðup0 ; vp0 Þ, which is
shown in (7):
x y
up 0 ¼ þ u0 vp 0 ¼ þ v0 ð7Þ
dx dy
where dx represents the true physical width of one pixel in the x-axis direction, and dy
represents the same in the y-axis direction. These two widths are intrinsic parameters of
the imaging chip and can be found by consulting the chip manual. If the chip is stable,
dx and dy will not change. ðu0 ; v0 Þ is the center point of the image plane, which is the
projection point of the camera aperture center on the image plane and can be obtained
by calibration. So Ob p0 can be calculated by (8).
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
ob p0 ¼ ðup0 u0 Þ2 þ ðvp0 v0 Þ2 ð8Þ
ouv u
ob (u0 , v0 ) x
dx
p dy
v y
Fig. 5. Coordinate transformation Fig. 6. Experiment result of DOEC
288 L. Wang et al.
Through the above steps, the distance overestimated has been corrected, final result
of the correction are shown in Fig. 6. It can be seen that depth data of points far from
the optical center point are corrected, and the measured data fluctuate around 1 m about
2 cm of 1 m. Thus the effectiveness of the DOEC has been proved.
4 Experiment Results
By correcting errors caused by the non-imaging principle and the error caused by
imaging principle, the accuracy of the depth data is greatly improved, and the depth
data in the same plane can be kept within a small fluctuation range. To prove the
effectiveness of the above method, objects on the desktop and the desktop are taken as
the experimental objects. By analyzing the scope of the depth data, the plane where the
desktop is located and the plane where the objects are located are divided.
Figure 7(b) is a grayscale image, as shown in Fig. 7(a) is a depth map after non-
imaging principle error correction, it can be seen that the overall image data after the
error correction caused by the non-imaging principle is relatively stable, but due to the
impact of errors caused by imaging principle, radiating radially from the center of the
depth image, the color gradually darkens from yellow to blue. This phenomenon
indicates that the depth data is different at different locations although they are in the
same plane. The accuracy of the depth data is not high, the fluctuation range is large,
and it is difficult to distinguish between the target object and the background. There-
fore, in order to further improve the accuracy of the depth data, it is necessary to correct
the depth overestimation error caused by the imaging principle.
Fig. 7. Experiment process and results of errors correction (Color figure online)
data after the overestimation error correction to the range of 0–255, so that the cor-
rected depth image is represented by a grayscale image, as shown in Fig. 7(c). Different
depth data corresponds to different gray values in the range of 0–255. The greater the
depth distance is, the greater the gray value is. It can be seen from the Fig. 7(c) that the
gray value on the same plane after the error correction is relatively uniform, indicating
that the fluctuation range of the depth data on the same plane is small, so the back-
ground and the target object can be separated according to the depth distance range of
different planes. Figure 7(d) is a divided background, and Fig. 7(e) is a divided target
object. Because the distance from the center of the camera to the desktop is greater than
the distance from the surface of the article, the gray value corresponding to the desktop
after correction is greater than the corresponding gray value of the surface of the target
object.
After change the number and shape of target objects, final results are shown in
Fig. 8. The background planes and the target objects are well separated, proved that the
two error correction methods are valid.
5 Conlusion
Due to the imaging characteristics and various factors in the external environment,
there are many kinds of errors exist in the depth data of ToF camera. Starting from the
imaging principle, this paper classifed errors into two categories: errors caused by the
non-imaging principle and errors caused by imaging principle. To improve errors
correction efficiency of errors caused by non-imaging principle, temperature changes,
amplitude intensities and integration time are analyzed and corrected in this paper,
multipath errors and other noises are compensated by taking offsets, an evaluation
function is proposed to determine the optimal reference distance, so as to select
appropriate integration time and global offsets. In the correction of errors caused by
290 L. Wang et al.
References
1. Pertile, M., Chiodini, S., Giubilato, R., Debei, S.: Calibration of extrinsic parameters of a
hybrid vision system for navigation comprising a very low resolution time-of-flight camera.
In: IEEE International Workshop on Metrology for Aerospace, pp. 391–396 (2017)
2. Lin, J., Liu, Y., Suo, J., Dai, Q.: Frequency-domain transient imaging. IEEE Trans. Pattern
Anal. Mach. Intell. 39, 937–950 (2016)
3. Adam, A., Dann, C., Yair, O., Mazor, S., Nowozin, S.: Bayesian time-of-flight for realtime
shape, illumination and albedo. IEEE Trans. Pattern Anal. Mach. Intell. 39, 851–864 (2017)
4. Anwer, A., Ali, S.S.A., Khan, A., Mériaudeau, F.: Underwater 3D scene reconstruction
using Kinect v2 based on physical models for refraction and time of flight correction. IEEE
Access 5, 1–11 (2017)
5. Shim, H., Lee, S.: Recovering translucent objects using a single time-of-flight depth camera.
IEEE Trans. Circuits Syst. Video Technol. 26, 841–854 (2016)
6. Francis, S.L.X., Anavatti, S.G., Garratt, M., Shim, H.: A ToF-camera as a 3D vision sensor
for autonomous mobile robotics. Int. J. Adv. Robot. Syst. 12, 1 (2015)
7. Illadequinteiro, J., Brea, V.M., López, P., Cabello, D., Doménechasensi, G.: Distance
measurement error in time-of-flight sensors due to shot noise. Sensors 15, 24–42 (2015)
8. Fürsattel, P., et al.: A comparative error analysis of current time-of-flight sensors. IEEE
Trans. Comput. Imaging 2, 27–41 (2016)
9. Belhedi, A., Bartoli, A., Bourgeois, S., Gay-Bellile, V.: Noise modelling in time-of-flight
sensors with application to depth noise removal and uncertainty estimation in three-
dimensional measurement. Comput. Vis. IET 9, 967–977 (2015)
10. Ghorpade, V.K., Checchin, P., Trassoudaine, L.: Line-of-sight-based ToF camera’s range
image filtering for precise 3D scene reconstruction. In: European Conference on Mobile
Robots, pp. 1–6 (2015)
11. Hertzberg, C., Frese, U.: Detailed modeling and calibration of a time-of-flight camera. In:
International Conference on Informatics in Control, Automation and Robotics, pp. 568–579
(2015)
12. Konno, Y., et al.: Accurate plane estimation based on the error model of time-of-flight
camera. In: Second IEEE International Conference on Robotic Computing, pp. 304–307
(2018)
Discrete Event-Triggered H1 State-Feedback
Control for Networked Control Systems
1 Introduction
compared with time-triggered method. With the reduction of transmission data, the
problems of limited bandwidth, packet loss and network delay in NCSs will be alle-
viated and the lifespan of the battery of the nodes will also increase.
However, compared with the theory of time sampling, the theoretical research on
ETM is relatively lacking, and the main reason caused that is the theoretical analysis on
the latter is much more complex. Even for a simple system, the complexity and
challenge for analysis and design is apparently. Time delay system method is one of the
main modeling and analysis methods of event triggered control system [1, 5–8]. The
linear system is modeled as a system model with internal delay, which facilitates the
consideration of network constraints such as induced delays and designing controllers.
In [1], theorems are proposed to preserve the desired control performance of the event-
triggered linear system and evaluate the upper bound of the event-triggered condition;
and to avoid the sensor nodes keeping radio on to wait for reception of the sampled
data, a novel self-triggered sampling scheme is developed based on event-triggered
theorems. In [5], a sufficient condition is derived to co-design both the desired dynamic
output feedback controls and the event-triggering parameters considering not all state
of the system is measurable. In [9], a more general forms of the event-triggered
paradigm than [5] refers is used for both state-feedback and static output-feedback
control; the maximum verifying period is obtained by solving a generalized eigenvalue
problem. For the linear system with constant delay, a state feedback controller based on
event triggered mechanism is designed, which can offer the optimal control perfor-
mance while reducing the control task executions and the data transmissions [6]. In [7],
an improved event-triggered mechanism is proposed which can obtain low transmis-
sion rate during the whole operation time by using both the state and state-independent
information. In [8], for the case that the statistical information on the external noises is
not precisely known, an event-triggered method to ensure the exponential stability as
well as prescribed H1 performance for the filtering error system are derived. However,
there are a few studies on event-triggered control considering time delay and packet
loss simultaneously, which is the motivator of the study in this paper.
Referring to the above literature, the time-delay system modeling method is applied
in this paper to the stability analysis and controller design of event-triggered systems.
Considering that most research is a continuous trigger mechanism depending on
hardware, a discrete event-triggered communication mechanism with variable weight,
which can be implemented by software and guarantee the minimum lower limit of the
release period in principle, is proposed in this paper. Take the effect of both time delay
and packet losses into account, the NCS is remodeled. The results show that the joint
design method of the event-triggered method and the state feedback controller pro-
posed in this paper can achieve longer average control period, i.e. reduce more
resources in NCSs.
Notation: Rn and N denote the n-dimensional Euclidean space and natural number,
respectively. The notation X [ 0ðX 0Þ stands for a real symmetric positive definite
(positive semidefinite). X T denotes the transpose of matrix X. Diag{}denotes the block-
diagonal matrix. “*” is used as the ellipsis for terms induced by symmetry. Sym { X }
represents X þ X T .
Discrete Event-Triggered H∞ State-Feedback Control 293
2 Problem Statement
2.1 System Description
Consider the following continuous-time linear system governed by
x_ ðtÞ ¼ AxðtÞ þ B^uðtÞ þ Bx xðtÞ
ð1Þ
zðtÞ ¼ CxðtÞ
where u^ðtÞ 2 Rm is the control input vector, xðtÞ 2 L2 ½0; 1Þ is the disturbance, and
zðtÞ 2 Rp is the control output; A, B, C, D and Bx are constant matrices with appro-
priate dimensions; the initial condition of the system (1) is given by xðt0 Þ ¼ x0 .
The structure of NCSs with network latency and ETM is constructed in Fig. 1.
uˆ(t )
x(t )
x(kh)
x(tk h)
The sensor collects the information of a plant; In this paper, it is used to monitor
changes in the state xðtÞ. The sampler detects the continuous signal acquired by the
sensor in a fixed period h, and obtains discrete signals xðkhÞ. The sampling time is set
as Ss ¼ f0; h; 2h; . . .gðk 2 NÞ. Whether the sampled data is transmitted to the con-
troller depends on the event-triggered condition predefined by event generator.
The control signal is transmitted in a single packet. First do not consider the packet
loss, the successful transmitted instant sequence of the event generator can be described
as Sg= {0, b1h, b2h, …} ðbk 2 NÞ. With the rapid development of electronic technology
and chips, and the wide use of high-speed digital processors, the computational delay is
negligible compared with transmission delays in the NCSs [2]. The network induced
delay can be expressed and sk ¼ ssc ðtk Þ þ sca ðtk Þ and sk 2 ½sm ; sM Þ, where ssc ðtk Þ,
sca ðtk Þ are the transmission delay between sampler to controller and controller to
actuator respectively; sm = min{sk }, sM = max{sk }. The input of the actuator is gen-
erated by the ZOH, and the holding time is t 2 ½tk ; tk þ 1 Þ(tk ¼ bk h þ sk ).
½xðbk h þ jhÞ xðbk hÞT W ½xðbk h þ jhÞ xðbk hÞ [ dNT WN ð2Þ
d
X ¼ [ Xik ð3Þ
ik ¼ 0
where d ¼ bk þ 1 bk 1,
½bk h þ ik h þ sk ; bk h þ ðik þ 1Þh þ sk Þ; ik ¼ 1; 2; . . .; d 1
Xik ¼
½bk h þ dh þ sk ; bk þ 1 h þ sk þ 1 Þ; ik ¼ d
where 0\b\1. data not be lost when bðtÞ ¼ 1. Then the final input signal translates
into: ^uðtÞ ¼ bðtÞKxðbk hÞ, and the spatial model of the system is as follows:
8
< x_ ðtÞ ¼ AxðtÞ þ B^uðtÞ þ Bx xðtÞ
zðtÞ ¼ CxðtÞ ð7Þ
:
^uðtÞ ¼ bðtÞKxðbk hÞ; t 2 ½tk ; tk þ 1 Þ
In combination with the above analysis, a closed loop system model based on event
triggered mechanism is obtained:
ðtÞ
x_ ðtÞ ¼ w1 ðtÞ þ ðbðtÞ bÞw 2 ð8Þ
zðtÞ ¼ CxðtÞ
In which, the initial function of the system is xðtÞ ¼ uðtÞ, t 2 ½g2 ; 0. Where
½xðt gðtÞÞ eðtÞ þ Bx xðtÞ, w ðtÞ ¼ BK ½xðt gðtÞÞ eðtÞ.
w1 ðtÞ ¼ AxðtÞ þ bBK 2
In this section, for a given disturbance attenuation level c, we will obtain a sufficient
condition on the design of the state feedback controller such that the system described
by (8) with the event generator (2) meets the following two requirements.
(1) The closed-loop system described by (8) with xðtÞ ¼ 0 is asymptotically stable;
(2) Under zero initial condition, the controlled output zðtÞ satisfies kzðtÞk2 c
kxðtÞk2 for any nonzero xðtÞ 2 L2 ½0; 1Þ.
where
T T T T T
e1 e1 e2 e2 e1 e1 e3 e3 e1 e1
U11 ¼ symfCT1 Pe1 g þ Q1 Q1 þ Q2 Q1 þ R1
e2 e2 e3 e3 e2 e2 e4 e4 e6 e6
T T
e6 T e6 e1 e1 e7 e7 T
R1 þ R2 R2 e1 eT2 Z1 ðe1 e2 Þ c2 eT10 e10 eT4 eT5 Z2 ðe4 e5 Þ
e7 e7 e6 e6 e8 e8
eT5 eT8 Z2 ðe5 e8 Þ þ eT5 eT8 Sðe4 e5 Þ þ eT4 eT5 Sðe5 e8 Þ eT9 We9
T T
þ d ðr 1Þe9 þ e5 W ðr 1Þe9 þ e5
g1 g
UT21 ¼ colf C1 ; g21 C1 ; C2 ; 1 C3 ; g21 C3 g; U22
n
3 3 o
¼ diag Z11 ; Z21 ; I; b ^1 Z 1 ; b^1 Z 1
1 2
Z Z t Z g1 Z t
g1 0 T
V3 ðtÞ ¼ x_ ðsÞZ1 x_ ðsÞdsdh þ g21 x_ T ðsÞZ2 x_ ðsÞdsdh; g21
3 g31 t þ h g2 tþh
¼ g2 g 1
g1 g
EfV_ 2 ðtÞg ¼ xT1 ðtÞQ1 x1 ðtÞ xT1 ðt ÞQ1 x1 ðt 1 Þ þ xT1 ðtÞQ2 x1 ðtÞ
3 3
T 2g1 2g1 T T g g
x1 ðt ÞQ1 x1 ðt Þ þ x2 ðtÞR1 x2 ðtÞ x2 ðt 2 ÞR1 x2 ðt 2 Þ ð13Þ
3 3 3 3
T T 2g2 2g2
þ x2 ðtÞR2 x2 ðtÞ x2 ðt ÞR2 x2 ðt Þ
3 3
Z t
g2 ^ T ðtÞZ1 w ðtÞ g1
EfV_ 3 ðtÞg ¼ 1 ½wT1 ðtÞZ1 w1 ðtÞ þ bw 2 2 x_ T ðsÞZ1 x_ ðsÞds
9 3 th31
Z tg1 ð14Þ
T ^ T
þ g21 ½w1 ðtÞZ2 w1 ðtÞ þ bw2 ðtÞZ2 w2 ðtÞ g21
2
x_ T ðsÞZ2 x_ ðsÞds
tg2
T T
^ ¼ bð1
where x1 ðtÞ ¼ xT ðtÞ xT ðt g31 Þ ; x2 ðtÞ ¼ xT ðtÞ xT ðt g32 Þ ; b bÞ.
The formula (14) contains two integral terms. In order to reduce conservatism, we
use Jensen’s inequality [7] and reciprocally convex method to scale the two integral
terms respectively:
Z h
g t
g iT h g i
1 _ T ðsÞZ1 x_ ðsÞds xðtÞ xðt 1 Þ Z1 xðtÞ xðt 1 Þ
x ð15Þ
3 g
t 31 3 3
R tg1 T
g21 tg2 x_ T ðsÞZ2 x_ ðsÞds ½xðt g1 Þ xðt gðtÞÞ Z2 ½xðt g1 Þ xðt gðtÞÞ
T
½xðt gðtÞÞ xðt g2 ÞT Z2 ½xðt gðtÞÞ xðt g2 Þ
þ sym ½xðt gðtÞÞ xðt g2 Þ S½xðt g1 Þ xðt gðtÞÞ
ð16Þ
_
For VðtÞ, plus and minus the corresponding items, obtains:
_
E VðtÞ _
¼ E VðtÞ eT ðtÞWeðtÞ þ eT ðtÞWeðtÞ zT ðtÞzðtÞ
ð17Þ
þ zT ðtÞzðtÞ c2 xT ðtÞxðtÞ þ c2 xT ðtÞxðtÞ
_
E VðtÞ vT ðtÞHvðtÞ zT ðtÞzðtÞ þ c2 xT ðtÞxðtÞ ð19Þ
where vðtÞ ¼ fxðtÞ; xðt g31 Þ; xðt g1 Þ; xðt gðtÞÞxðt g32 Þ; xðt 2g32 Þ; xðt g2 Þ;
g2 g2 ^ T 2 ^ T
eðtÞ; xðtÞg; H ¼ U11 þ 91 CT1 Z1 C1 þ g221 CT1 Z2 C1 þ CT2 C2 þ 91 bC 3 Z1 C3 þ g21 b C3 Z2
C3 ; C1 ¼ Ae1 þ bBKe
5 bBKe9 þ Bx e10 ; C2 ¼ Ce1 ; C3 ¼ BKe5 BKe9 ; ei ði ¼ 1; 2;
. . .; 10Þ is a block coordinate matrix, for example e2 ¼ ½0 I 0 0 0 0 0 0 0 0.
298 W. Shen et al.
~ 11
U
~
U21 ~ 22 \0
U
ð22Þ
Where
T T T T
e1 e1 e2 e2 e1 e1 e3 e3
~ 11 ¼symfeT C
U ~ ~1 ~1 ~2 ~1
1 1g þ Q Q þ Q Q
e2
e2 e3 e3 e2 e2 e4 e4
T T
T T
e1 e1 e6 e6 e1 e1 e7 e7 T
þ ~1
R ~1
R þ ~2
R ~2
R e1 eT2 Z~1 ðe1 e2 Þ
e6 e6 e7 e7 e6 e6 e8 e8
~ 9 c2 eT10 e10 eT4 eT5 Z~2 ðe4 e5 Þ eT5 eT8 Z~2 ðe5 e8 Þ þ eT5 eT8 ~Sðe4 e5 Þ
eT9 We
þ eT4 eT5 ~Sðe5 e8 Þ þ d ðr 1ÞeT9 þ eT5 W ~ ðr 1Þe9 þ e5
~ 21 ¼ colfg1 C
U ~ 1 ; g21 C ~ 2 ; g1 C
~ 1; C ~ 3 ; g21 C
~ 3g
3 3
n o
~ 22
U ¼ diag X Z~11 X; X Z~21 X; I; b ^1 X Z ~11 X; b ^1 X Z~ 1 X
2
C
~ 1 ¼ AXe1 þ bBYe ~
5 bBYe9 þ Bx e10 ; C2 ¼ CXe1 ; C3 ¼ BYe5 BYe9 :
Z~2 T Z2
~S ¼1 1[0 ð23Þ
Z~2 S Z2
~ 11
U U11
~ ~ 22 ¼ nT n\0 ð24Þ
U21 U U21 U22
4 Illustrative Examples
Assume that the systems described in Fig. 1 is the satellite control system [5]. The state
equation of the satellite system is described as follows:
8 2 3 2 3 2 3
>
> 0 1 0 0 0 1
>
> 6J J J J 7
k d k d 607 617
<
x_ ðtÞ ¼ 6
4 0
2 2 2 7
xðtÞ þ 6 7 6 7
4 0 5uðtÞ þ 4 1 5xðtÞ
0 15
2
0 ð25Þ
>
> k d k d
>
> J1 J1 J1 J1
1
J1 1
:
zðtÞ ¼ ½ 1 0 0 0 xðtÞ
In Fig. 2, the abscissa and the ordinate show the release instants for the signal and
the time interval between two transmission instants. From the diagram, we can see that
under the event-triggered communication mechanism, the control cycle of the system is
not fixed, and the maximum execution cycle is 1.6 s, which is larger than the sensor
sampling period h. For t 2 (0, 60 s), the number of data obtained from the sampler is
600. If the traditional periodic trigger control strategy is adopted, 600 packets have to
be transmitted, although some of which is unnecessary. However, the simulation results
show that the amount of transmitted data is 78 with proposed ETM, that is, the average
data transmission rate is 13.00%. Compared with the time-triggered mechanism, it
300 W. Shen et al.
saves 87.00% of the bandwidth resources. The frequency of task execution is reduced
in the NCS with ETM, but the control effect is not affected when some control data lost
(Fig. 3).
5 Conclusion
To save the NCSs with limited resources and increase its life-span, in this paper, we
have studied a kind of discrete event-triggered scheme for continuous-time system
when existing time-varying delay and packet loss constrains. The data is transmitted in
the form of a single packet in the network, and the loss of packets is described by a
stochastic variable that satisfies the Bernoulli distribution. For the resulting delay
system, a Lyapunov-Karsovskii functional with the delay information has been con-
structed. A new stabilization criterion in the sense of mean-square expressed in the
form of matrix inequalities with H1 norm bounds and H1 control design has been
developed. As shown by simulation results, the proposed method can significantly
reduce data transmission while normal operation of the system, despite delay and
packet loss.
Discrete Event-Triggered H∞ State-Feedback Control 301
Acknowledgments. This work was financially supported by the Science and Technology
Commission of Shanghai Municipality of China under Grant (No. 17511107002).
References
1. Peng, C., Han, Q.L.: On designing a novel self-triggered sampling scheme for networked
control systems with data losses and communication delays. IEEE Trans. Ind. Electron. 63
(2), 1239–1248 (2016)
2. Zhang, L., Gao, H., Kaynak, O.: Network-induced constraints in networked control systems
—a survey. IEEE Trans. Ind. Inform. 9(1), 403–416 (2013)
3. Liu, Q., Wang, Z., He, X., Zhou, D.H.: A survey of event-based strategies on control and
estimation. Syst. Sci. Control. Eng. 2(1), 90–97 (2014)
4. Trimpe, S., Campi, M.C.: On the choice of the event trigger in event-based estimation. In:
2015 International Conference on Event-based Control, Communication, and Signal
Processing (EBCCSP), pp. 1–8, 17–19 June 2015
5. Zhang, X.M., Han, Q.L.: Event-triggered dynamic output feedback control for networked
control systems. IET Control Theory Appl. 8(4), 226–234 (2014)
6. Wu, W., Reimann, S., Görges, D., Liu, S.: Suboptimal event-triggered control for time-
delayed linear systems. IEEE Trans. Autom. Control 60(5), 1386–1391 (2015). https://doi.
org/10.1109/TAC.2014.2347214
7. Li, F., Fu, J., Du, D.: An improved event-triggered communication mechanism andL∞con-
trol co-design for network control systems. Inf. Sci. 370–371, 743–762 (2016)
8. Hu, S., Yue, D.: Event-based H∞ filtering for networked system with communication delay.
Signal Process. 92(9), 2029–2039 (2012)
9. Chen, X., Hao, F.: Periodic event-triggered state-feedback and output-feedback control for
linear systems. Int. J. Control Autom. Syst. 13(4), 779–787 (2015)
10. Peng, C., Han, Q.L.: A novel event-triggered transmission scheme and L2 control co-design
for sampled-data control systems. IEEE Trans. Autom. Control 58(10), 2620–2626 (2013)
Brief Technical Analysis of Facial Expression
Recognition
Abstract. Facial expression recognition (FER) is the current hot research topic,
and it is widely used in the fields of pattern recognition, computer vision and
artificial intelligence. As it is an important part of intelligent human-computer
interaction technology, the FER has received widespread attention in recent
years, and researchers in different fields have proposed many approaches for it.
This paper reviews recent developments on FER approaches and the key
technologies involved in the FER system: face detection and preprocessing,
facial expression feature extraction and facial expression classification, which
are analyzed and summarized in detail. Finally, the state-of-the-art of the FER is
summarized, and its future development direction is pointed out.
1 Introduction
As an aspect of emotion recognition, the FER plays a very important role in inter-
personal communication. Facial expression is an important way to express emotion and
information exchange. According to research conducted by the social psychologist
Albert Mehrabian, in the daily human communication, emotional information trans-
mitted through facial expressions is as high as 55% of the total amount of information
[1], while information transmitted through voice and language accounts for 38% and
7% of the total amount of emotional information. The FER is the foundation of
emotional understanding, and it has injected new blood for artificial intelligence.
In 1971, Ekman and Friesen made groundbreaking work [2]. They defined six basic
types of facial expression for human expression recognition: happiness, sadness, sur-
prise, fear, anger, disgust. In 1978, Ekman and Friesen et al. developed the famous
Facial Action Coding System (FACS) to detect subtle changes in facial expressions [3].
In 1991, MASE and Kenji proposed a new theory that uses the optical flow method to
recognize facial expressions. Since then, researchers have used computer technology to
extract and classify facial expression features that have attracted widespread attention.
FER system includes three major technical steps: face detection and preprocessing,
facial expression feature extraction and facial expression classification. As shown in
Fig. 1, FER system is established to first detect and preprocess the face, which includes
image rotation correction, face location, and scale normalization of the expression
image, etc. The feature information that can represent the essence of the input
expression is extracted from the static images or the dynamic image sequences. In order
to avoid excessive dimensions, features such as dimension reduction and feature
decomposition are required in the process of extracting feature data. Finally, the
relationship between features is analyzed and the input facial expressions are classified
into corresponding categories. At present, the FER plays an important role in inter-
active game platforms, safe driving, intelligent recommendation, distance education,
criminal investigation and security, auxiliary medical and other fields, and it has
tremendous space for development and application.
Facial
Face Detection Facial
Expression
and Expression
Feature
Preprocessing Classification
Extraction
According to the relevant literatures on the FER studies in recent years, this paper
analyzes face detection and preprocessing, facial expression feature extraction,
expression classification technology. It reviews the research results of facial expression
recognition, and then looks forward to its development trend.
The most importance for FER is to be able to completely extract feature information of
the entire face on an original image. Before feature extraction, it is necessary to
preprocess the original image. Image preprocessing is mainly divided into two aspects:
removing the unwanted background of the facial expression image and image
normalization.
The development of face detection technology provides an effective way to remove
the complex background of the original image in the FER system. Viola and Jones used
cascaded classifiers on Haar-like features to detect face [4]. However, sometimes it is
failed to detect contours or partially occluded faces by this method. Due to the different
image acquisition methods, the size and format of the generated images are different.
The original images need to be normalized to make the full use of the images. The
image normalization includes three parts: face normalization, plane face rotation cor-
rection and depth face rotation correction. The problem of large difference in gradation
due to uneven illumination can also be solved by gray-scale normalization after the
geometric normalization is achieved. After a series of steps, the normalization of the
original picture is finally completed. For example, as shown in Fig. 2, facial images in
304 L. Xu et al.
the Japanese Female Facial Expression (JAFFE) database are detected and prepro-
cessed, and they are classified into seven basic expressions.
Fig. 2. Sample images of one subject with seven expressions on JAFFE database
Feature extraction plays an important role in the entire FER system. Under the premise
of ensuring the integrity of the original image information of the images, the approach
for extracting useful information of facial expressions can greatly improve the rate of
FER. The extraction of expression features is divided into static image feature
extraction and dynamic sequence image feature extraction according to the different
nature of the image. However, most feature extraction processes have excessive
dimensions or the high volume of data, and it is necessary to perform dimension
reduction, feature decomposition and etc. At present, various feature extraction algo-
rithms have been continuously proposed and improved.
which reflects the information of the bottom face of the face image, and particularly
shows the information of the local minute change. The appearance feature extraction
methods mainly include Local Binary Pattern (LBP) and Gabor Filter. LBP is used to
compare the local features of an image. It can extract and measure the texture infor-
mation of the local neighborhood in a grayscale image, and can effectively suppress the
influence of illumination. Chao et al. [7] proposed the expression-specific LBP with
symmetric extension features, and it can better extract important face points of the
human face. Gabor Filter is a classic method for facial expression feature extraction.
A set of filters are used to filter the image. The result can reflect the relationship
between local pixels. The Gabor filter can detect texture changes of the multi-scale and
multi-directional, and it is less affected by light changes. A novel approach for FER is
proposed, and the facial feature is extracted from image by applying Gabor wavelet
filter on an image and computing local transitional pattern (LTP) code from the
transformed image. Extensive experiments illustrate that it is effective and efficient for
the FER.
completed. Zafeiriou and Pitas [10] propose a new technique based on the elastic map
matching improves the performance of the selection of the most discriminant facial
landmarks for the FER.
separate the different face factors, so that the various types of recognition can be
performed in different subspaces to avoid interference from other factors. A dictionary-
based approach for the FER by decomposing expressions is proposed by Taheri et al.
[14]. They achieve expression decomposition and recognition by computing the facial
expression sparse code matrix.
The expression classification method and the expression feature extraction method are
closely related. It refers to defining a set of categories and designing a corresponding
classification mechanism to recognize expressions and classify them into correspond-
ing categories. At present, the main expression classification methods include Hidden
Markov Model (HMM), K-Nearest Neighbor (KNN), Support Vector Machines
(SVM), AdaBoost, Artificial Neural Network (ANN), and so on.
4.4 AdaBoost
AdaBoost is an iterative algorithm, and its basic principle of Adaboost algorithm is to
combine several weak classifiers (weak classifiers generally use a single-level decision
tree) to make it a strong classifier. It adopts an iterative idea that only one weak
classifier is trained per iteration and the weak classifier trained will participate in the
use of the next iteration. The AdaBoost algorithm which is proposed by Freund and
Schapire, is first practical boosting algorithm, and it is one of the most widely used and
studied [18]. Owusu et al. proposed a new method which is based on the AdaBoost,
and it can enhance accuracy and speed with the expression dataset reduced [19].
5 Conclusion
The FER is a cross-disciplinary research topic and has become a very popular research
topic in the fields of psychology, medicine, computer vision, pattern recognition, etc. It
has important application value in the field of intelligent human-computer interaction.
This paper describes the key technologies involved in facial expression recognition:
Brief Technical Analysis of Facial Expression Recognition 309
face detection and preprocessing, expression feature extraction and expression classi-
fication. The paper categorizes the research results on facial expression recognition
technology at home and abroad, and introduces the problems existing in some tech-
nologies and the latest technologies related to facial expression recognition.
Although many achievements have been made in the study, many problems still
need to be solved. For example, expression classification is too restrictive and human
expressions are not only six basic expressions, but there are also many complex and
superimposed expressions. Face expressions in dynamic images and facial expressions
in 3D real life are easily affected by illumination, occlusion, etc. The robustness of the
FER is poor, and the effectiveness of the recognition algorithm still needs improve-
ment. Therefore, many expression recognition algorithms have not yet been applied to
specific projects or products, and it is still a challenging topic. With the arrival of the
era of big data, the demand for real-time facial expressions has increased dramatically.
The application of the FER technology has a broader scope of development and uti-
lization value. In the future, facial expressions can be assisted by integrating sound,
heat, and environmental factors. Expression recognition system adds depth information
based on 2D to support 3D face analysis. Taking time changes into consideration helps
to improve expression recognition rates by analyzing the correlation between con-
secutive adjacent frames.
References
1. Mehrabian, A.: Communication without words. Psychol. Today 2, 53–55 (1968)
2. Ekman, P., Friesen, W.V.: Constants across cultures in the face and emotion. J. Pers. Soc.
Psychol. 17, 124–129 (1971)
3. Ekman, P.: Facial Action Coding System. A Technique for the Measurement of Facial
Action (1978)
4. Hu, P.: Application research on face detection technology based on OpenCV in mobile
augmented reality. Int. J. Sign. Process. Image Process. Pattern Recogn. 8, 249–256 (2015)
5. Shbib, R., Zhou, S.: Facial expression analysis using active shape model. Int. J. Sign.
Process. Image Process. Pattern Recogn. 8, 9–22 (2016)
6. Li, Y., Liu, W., Li, X., Huang, Q., Li, X.: GA-SIFT: a new scale invariant feature transform
for multispectral image using geometric algebra. Inf. Sci. 281, 559–572 (2014)
7. Chao, W.L., Ding, J.J., Liu, J.Z.: Facial expression recognition based on improved local
binary pattern and class-regularized locality preserving projection. Signal Process. 117, 1–10
(2015)
8. Ghimire, D.: Recognition of facial expressions based on tracking and selection of
discriminative geometric features. Int. J. Multimed. Ubiquitous Eng. 10, 35–44 (2015)
9. Dewan, M.A.A., Granger, E., Marcialis, G.L., Sabourin, R., Roli, F.: Adaptive appearance
model tracking for still-to-video face recognition. Pattern Recogn. 49, 129–151 (2016)
310 L. Xu et al.
10. Zafeiriou, S., Pitas, I.: Discriminant graph structures for facial expression recognition. IEEE
Trans. Multimed. 10, 1528–1540 (2008)
11. Siddiqi, M.H., Ali, R., Khan, A.M., Park, Y.T., Lee, S.: Human facial expression recognition
using stepwise linear discriminant analysis and hidden conditional random fields. IEEE
Trans. Image Process. 24, 1386–1398 (2015)
12. Zhao, X.: Facial expression recognition using local binary patterns and discriminant kernel
locally linear embedding. EURASIP J. Adv. Signal Process. 2012, 20 (2012)
13. Wang, S., Yang, H., Li, H.: Facial expression recognition based on incremental isomap with
expression weighted distance. J. Comput. 8, 2051–2058 (2013)
14. Taheri, S., Qiu, Q., Chellappa, R.: Structure-preserving sparse decomposition for facial
expression analysis. IEEE Trans. Image Process. 23, 3590–3603 (2014)
15. Wu, C., Wang, S., Ji, Q.: Multi-instance hidden Markov model for facial expression
recognition. In: IEEE International Conference and Workshops on Automatic Face and
Gesture Recognition, pp. 1–6 (2015)
16. Bahari, F., Janghorbani, A.: EEG-based emotion recognition using recurrence plot analysis
and K nearest neighbor classifier. In: Biomedical Engineering, pp. 228–233 (2013)
17. Wang, F., He, K., Liu, Y., Li, L., Hu, X.: Research on the selection of kernel function in
SVM based facial expression recognition. In: Industrial Electronics and Applications,
pp. 1404–1408 (2013)
18. Schapire, R.E.: Explaining AdaBoost. Springer, Heidelberg (2013)
19. Owusu, E., Zhan, Y., Mao, Q.R.: A neural-AdaBoost based facial expression recognition
system. Expert Syst. Appl. 41, 3383–3390 (2014)
20. Lopes, A.T., Aguiar, E.D., Souza, A.F.D., Oliveira-Santos, T.: Facial expression recognition
with convolutional neural networks: coping with few data and the training sample order.
Pattern Recogn. 61, 610–628 (2017)
Application of Intelligent Virtual Reference
Feedback Tuning to Temperature Control
in a Heat Exchanger
1 Introduction
Heat exchangers are extensively utilized in many industrial fields, including power
generation systems [1], chemical processes [2] and district heating systems [3]. Effi-
cient temperature control can improve the dynamic performance and enhance stability
of the system involving the heat exchanger significantly [4]. However, it is a major
challenge to control the fluid temperature being heated at a specific and stable reference
value because heat exchangers are extremely complex devices involving a large
number of thermal dynamic phenomena, for which the prediction of their operation
from first principles seems to be impossible [5]. Therefore, researchers proposed
various advanced approaches to design and improve temperature controllers for heat
exchangers over past few decades. However, most of those methods [4, 6, 7] are
designed based on mathematical models, which means that plenty of effort has to be
spent on system modelling. Furthermore, it is difficult to construct a perfect model and
some complex nonlinear systems are even unlikely to be modelled.
Virtual reference feedback tuning (VRFT) is a data-driven control (DDC) approach.
As a hotspot, VRFT approaches have been improved substantially and applied to a
wide range of control systems [8–10]. However, the optimal controller cannot be
designed successfully because the performance indices are not included in its objective
function in the standard VRFT method. Therefore, Wang et al. [11] developed an
intelligent VRFT (IVRFT), in which a potential performance metric, i.e., the reference
model parameter, was included in the controller design to achieve the optimal control
performance. Compared with VRFT approaches, the controller design problem in the
IVRFT is transformed from a parameter identification problem into an optimization
problem. Traditional derivative-based methods including the least square method,
which is adopted in VRFT approaches, cannot be used to solve such an optimization
problem.
Obviously, the optimization method has a significant impact on the controller
design based on IVRFT method [11]. Human learning optimization (HLO) is an simple
yet efficient optimization tool, which mimics the human learning process [12]. In
previous works, HLO has obtained the best-so-far results on two well-studied sets of
multi-dimensional knapsack problems, i.e. 5.100 and 10.100 [13], which shows its
advantages. In this paper, a multi-population cooperative human learning optimization
(MCHLO) is proposed as an optimization tool to design the optimal controller for a
water heat exchanger based on IVRFT.
The reminder of this paper is organized as follows. Section 2 introduces the IVRFT
briefly. MCHLO is described in Sect. 3. Section 4 presents the optimal controller
design based on IVRFT and MCHLO (IVRFT-MCHLO) in details. Then IVRFT-
MCHLO is utilized to control the temperature in a water heat exchanger and compared
with the other control approaches in Sect. 5. Finally, we draw the conclusions in
Sect. 6.
The VRFT approach [14] aims to solve a model-reference problem in discrete time.
The objective function adopted in the VRFT method does not include any performance
index, and thus it is almost impossible to design the optimal controller through the
original VRFT method. However, for practical engineering applications, the dynamic
performance is of great importance because it significantly influences the efficiency of
systems and the benefit of enterprises. Considering that the reference model determines
the closed-loop control performance of a controller tuned by the VRFT method, Wang
et al. [11] presented a novel intelligent VRFT approach in which the reference model
and the controller are both optimized. Consequently, the controller with the best per-
formance can be found in the framework of IVRFT.
Application of Intelligent Virtual Reference Feedback Tuning 313
ð1 AÞz1 d
Mðz1 Þ ¼ z ð1Þ
1 Az1
Apparently, the reference model with a relatively small A will have a better
dynamic performance. The controller design based on such a reference model is bound
to result in better control performance. Encouraged by that fact, a novel objective
function is proposed in IVRFT as Eq. (2)
where JVRFT denotes the objective function of the standard VRFT approach, and
k denotes the weight of A which should be small enough to make sure that JVRFT
dominates this objective function.
Social entities such as organizational units are often drawn by boundaries of knowledge
[15]. Individuals shape their own organizations and are also impacted by them. Levels of
cooperation have great impacts on knowledge sharing among organizational units [16].
It is natural that proper knowledge sharing can speed up the knowledge conversion
process. Inspired by this fact, a multi-population cooperative strategy is introduced into
HLO, and a novel multi-population cooperative human learning optimization is pro-
posed, where the whole population is divided into a master subpopulation and S slave
subpopulations as Fig. 1. The master population involves S individuals and the
remaining individuals are shared by the slave subpopulations equally.
Slave
subpop
Gib
SKD
Gib Gib
Slave SKD Master SKD Slave
subpop subpop subpop
In MCHLO, the cooperation between the master subpopulation and each slave
subpopulation works together. The master subpopulation and slave subpopulation are
evolved in parallel, which can take full advantage of the emerging solutions saved in
314 Y. Wen et al.
0; 0 r1 0:5
xij ¼ Rð0; 1Þ ¼ ð4Þ
1; else
2 3
ikd 1
6 ikd 2 7
6. 7
6. 7
6. 7
IKD ¼ 6 7; 1 i N ð5Þ
6 ikd i 7
6. 7
4. 5
.
ikd N
2 3 2 3
ikd i1 ik i11 ik i12 ik i1j ik i1M
6 ikd i2 7 6 ik i21 ik i22 ik i2j ik i2M 7
6 7 6 7
6 .. 7 6 .. .. .. .. 7
6. 7 6. . . . 7
6
ikdi ¼ 6 7¼6 7; 1 p L ð6Þ
ikd 7 6 ik ip1 ik ip2 ik ipj ik ipM 7
6 ip 7 6 7
6 .. 7 6. .. .. .. 7
4. 5 4 .. . . . 5
ikd iL ik iL1 ik iL2 ik iLj ik iLM
Application of Intelligent Virtual Reference Feedback Tuning 315
where ikdi is the IKD of individual i, L indicates the number of solutions stored in the
IKD, and ikdip is known as pth best solution for individual i. For the slave subpopu-
lation, when MCHLO performs the individual learning operator (ILO), it generates new
solutions according to the knowledge in the IKD as Eq. (7)
For the master subpopulation, the IKDmaster directly inherits Gib, i.e., the iteration-
best solution generated by the slave subpopulation.
where ikdmaster, i is the IKD of the ith individual in the master subpopulation. When an
individual in the master subpopulation yields the new candidate, it generates new
solutions from the knowledge stored in the IKDmaster as Eq. (9).
Social Learning Operator (SLO). As human beings are social animals, human
learning cannot be separated from interaction among each other. To accelerate the
learning processes, it is natural for an individual to learn from his/her own organization.
For the slave population in MCHLO, the social knowledge database (SKD) is utilized
to reserve the knowledge of the subpopulation as Eq. (10)
2 3 2 3
skd1 sk11 sk12 sk1j sk1M
6 skd2 7 6 sk21 sk22 sk2j sk2M 7
6 7 6 7
6 .. 7 6 .. .. .. .. 7
6 . 7 6 . . . . 7
SKD ¼ 6 7 6
6 skdq 7¼6 skq1
7; 1 q H
7 ð10Þ
6 7 6 skq2 skqj skqM 7
6 .. 7 6 . .. .. .. 7
4 . 5 4 .. . . . 5
skdH skH1 skH2 skHj skHM
where skdq is the qth solution, and H denotes the size of the SKD. For an individual in
the slave subpopulation, the social learning operator (SLO) is performed as Eq. (11)
xij ¼ sk qj ð11Þ
For the master subpopulation, only the best-so-far solution is stored in the
SKDmaster as Eq. (12). Each individual in the master subpopulation executes SLO as
Eq. (13)
xj ¼ sk j ð13Þ
316 Y. Wen et al.
In summary, an individual in MCHLO employs the RLO, ILO and SLO in the
search process to generate new candidates, which can be formulated as Eq. (14)
8
< RLO; if 0 r pr
xij ¼ ILO; if pr\r pi ð14Þ
:
SLO; if pi\r 1
where pr, (pi − pr) and (1 − pi) represent the probability of executing RLO, ILO and
SLO, respectively. Note that ILO and SLO for the master subpopulation are different
from those for each slave subpopulation, which has been described before.
where u(k) represents the manipulated variable at the kth sampling instant, e(k) refers to
the error between the process output and its set-point at the kth sampling instant, and
Kp, Ki and Kd are parameters of a PID controller. The corresponding controller transfer
function can be formulized as Eq. (16), and its parametric form is as Eq. (17)
Ki
C z1 ; h ¼ Kp þ 1
þ Kd 1 z1 ð16Þ
1z
1
C z ; h ¼ bT ðzÞh ð17Þ
Application of Intelligent Virtual Reference Feedback Tuning 317
where
T 1
b ðzÞ ¼ 1 1 z1 ð18Þ
1 z1
h ¼ ½ Kp Ki K d T ð19Þ
From Eqs. (1) and (16), uvir ðtÞ can be deduced as Eqs. (20)–(21)
vir
1
Ki 1
1 z1 d
u z ¼ Kp þ þ Kd 1 z z y ð20Þ
1 z1 ð1 AÞz1
2 3
Kp
uvir ðz1 Þ ¼ uT ðtÞh ¼ up ðtÞ ui ðtÞ ud ðtÞ 4 Ki 5 ð21Þ
Kd
where up ðtÞ, ui ðtÞ, ud ðtÞ denote the ratio of Kp, Ki and Kd, respectively.
In IVRFT-MCHLO, MCHLO is used to search for the best Kp, Ki, Kd and A by
following Eq. (2), in order to tune the controller with superior performance. The
implementation of IVRFT-MCHLO can be summarized as follows:
Step 1: initialize the algorithm parameters included in MCHLO;
Step 2: set the length of each decision variable; in this paper, each decision variable
is coded by 30 bits;
Step 3: initialize the master subpopulation and each slave subpopulation randomly;
Step 4: initialize the IKDs and SKD of each slave subpopulation, as well as the
IKDsmaster and SKDmaster;
Step 5: generate new solutions by performing learning operators as Eq. (14);
Step 6: evaluate the fitness of each individual according to Eq. (2);
Step 7: update the IKDs and SKD of each slave subpopulation, as well as the
IKDsmaster and SKDmaster according to the updating rules of MCHLO;
Step 8: output the optimal solution in SKDmaster if the termination criterion is met;
otherwise go back to step 5.
As Fig. 2 shows, the valve V that manipulates the input water is utilized to control the
temperature T of output water in a water heat exchanger. The input water is from a
storage tank, whose temperature can be kept at a certain point by an independent
controller. The behavior of this process can be modeled as Eq. (22) [19]. Note that
IVRFT-MCHLO and VRFT are model-free approaches, and the model is only utilized
to collect the I/O data and validate the controller designed by IVRFT-MCHLO.
FC
Cold water
Hot water
FH
TT
where IAE, tup, and OS refer to the integral of absolute error, rise time and overshoot,
respectively. w1, w2 and w3 are the corresponding weight coefficients which are the
recommended values in [21]. Note that the ideal model as well as the model with 0.5%
error is taken as a competitor in this paper. The control performance of the above-
mentioned approaches is depicted in Fig. 3, and the corresponding results are listed in
Table 1. Non-overshoot control is achieved by all approaches except the Z-N method.
It is undoubted that the results obtained by the IVRFT approaches are inferior to the
MPID method with the perfect model. However, it should be emphasized that the
results obtained by IVRFT-MCHLO are very close to that of MPID and significantly
better than that of MPID with 0.5% model error. IVRFT-MCHLO outperforms IVRFT
on each performance index except OS, which reflects that the optimization algorithm is
important to the control performance in the framework of IVRFT.
Table 1. Results of MCHLO-IVRFT and other control approaches on the temperature control
Methods Tup(s) OS(%) IAE Fitness
Z-N 12 21.88 17.5284 1123.5000
MPID with 0.5% model error 9 0 8.6685 17.6685
MPID 6 0 6.4707 12.4707
IVRFT-MCHLO 7 0 6.8822 13.8822
IVRFT 8 0 7.6210 15.6210
6 Conclusions
Heat exchangers widely exist in industrial processes. However, its temperature control
is still challenging in engineering applications. Therefore, we proposed an improved
intelligent virtual reference feedback tuning based on a novel multi-population coop-
erative human learning optimization algorithm (IVFRT-MCHLO), which can design
the optimal temperature controller of heat exchangers only based on the collected data
of the system efficiently. With the enhanced exploration and exploitation abilities,
MCHLO can be combined with the IVRFT to find out the optimal controller. Exper-
imental results display that IVRFT-MCHLO can be utilized to design the temperature
controller for a water controller successfully and achieve the desirable control
performance.
References
1. May, P., Ehrlich, H.-C., Steinke, T.: ZIB structure prediction pipeline: composing a complex
biological workflow through web services. In: Nagel, W.E., Walter, W.V., Lehner, W. (eds.)
Euro-Par 2006. LNCS, vol. 4128, pp. 1148–1158. Springer, Heidelberg (2006). https://doi.
org/10.1007/11823285_121
2. Anxionnaz, Z., Cabassud, M., Gourdon, C., Tochon, P.: Heat exchanger/reactors (HEX
reactors): concepts, technologies: state-of-the-art. Chem. Eng. Process.: Process. Intensif. 47,
2029–2050 (2008)
3. Tol, Hİ., Svendsen, S.: A comparative study on substation types and network layouts in
connection with low-energy district heating systems. Energy Convers. Manag. 64, 551–561
(2012)
4. Wang, Y., You, S., Zheng, W., Zhang, H., Zheng, X., Miao, Q.: State space model and
robust control of plate heat exchanger for dynamic performance improvement. Appl. Therm.
Eng. 128, 1588–1604 (2018)
́
5. Dıaz, G., Sen, M., Yang, K., McClain, R.L.: Dynamic prediction and control of heat
exchangers using artificial neural networks. Int. J. Heat Mass Transf. 44, 1671–1679 (2001)
320 Y. Wen et al.
6. Vasičkaninová, A., Bakošová, M., Čirka, Ľ., Kalúz, M., Oravec, J.: Robust controller design
for a laboratory heat exchanger. Appl. Therm. Eng. 128, 1297–1309 (2018)
7. Wang, S., Yu, X., Liang, C., Zhang, Y.: Enhanced condensation heat transfer in air-
conditioner heat exchanger using superhydrophobic foils. Appl. Therm. Eng. 137, 758–766
(2018)
8. Formentin, S., De Filippi, P., Corno, M., Tanelli, M., Savaresi, S.M.: Data-driven design of
braking control systems. IEEE Trans. Control Syst. Technol. 21, 186–193 (2013)
9. Formentin, S., Karimi, A.: Enhancing statistical performance of data-driven controller tuning
via L2-regularization. Automatica 50, 1514–1520 (2014)
10. Yan, P., Liu, D., Wang, D., Ma, H.: Data-driven controller design for general MIMO
nonlinear systems via virtual reference feedback tuning and neural networks. Neurocom-
puting 171, 815–825 (2016)
11. Wang, L., Ni, H., Yang, R., Pardalos, P.M., Jia, L., Fei, M.: Intelligent virtual reference
feedback tuning and its application to heat treatment electric furnace control. Eng. Appl.
Artif. Intell. 46, 1–9 (2015)
12. Sörensen, K., Sevaux, M., Glover, F.: A history of metaheuristics. In: Martí, R., Panos, P.,
Resende, M. (eds.) Handbook of Heuristics, pp. 1–18. Springer, Cham (2018). https://doi.
org/10.1007/978-3-319-07153-4_4-1
13. Wang, L., Yang, R., Ni, H., Ye, W., Fei, M., Pardalos, P.M.: A human learning optimization
algorithm and its application to multi-dimensional knapsack problems. Appl. Soft Comput.
34, 736–743 (2015)
14. Campi, M.C., Lecchini, A., Savaresi, S.M.: Virtual reference feedback tuning: a direct
method for the design of feedback controllers. Automatica 38, 1337–1346 (2002)
15. von Krogh, G.: Individualist and collectivist perspectives on knowledge in organizations:
implications for information systems research. J. Strat. Inf. Syst. 18, 119–129 (2009)
16. Hansen, M.T.: The search-transfer problem: the role of weak ties in sharing knowledge
across organization subunits. Adm. Sci. Q. 44, 82–111 (1999)
17. Wang, L., Ni, H., Yang, R., Fei, M., Ye, W.: A simple human learning optimization
algorithm. In: Fei, M., Peng, C., Su, Z., Song, Y., Han, Q. (eds.) LSMS/ICSEE 2014. CCIS,
vol. 462, pp. 56–65. Springer, Heidelberg (2014). https://doi.org/10.1007/978-3-662-45261-
5_7
18. Wang, L., Ni, H., Yang, R., Pardalos, P.M., Du, X., Fei, M.: An adaptive simplified human
learning optimization algorithm. Inf. Sci. 320, 126–139 (2015)
19. Normey-Rico, J.E., Camacho, E.F.: Unified approach for robust dead-time compensator
design. J. Process Control 19, 38–47 (2009)
20. Åström, K.J., Hägglund, T.: Revisiting the Ziegler-Nichols step response method for PID
control. J. Process Control 14, 635–650 (2004)
21. Wang, L., Yang, R., Pardalos, P.M., Qian, L., Fei, M.: An adaptive fuzzy controller based on
harmony search and its application to power plant control. Int. J. Electr. Power Energy Syst.
53, 272–278 (2013)
Design and Realization of 108MN
Multi-function Testing System
Abstract. With the rapid development of China economics, large buildings are
beginning to appear. In order to ensure safety of the large buildings, 108MN
Multi-Function Testing System (MFTS) built by HIT with 6-DOF and large
loading force was proposed by China State Construction Engineering Corpo-
ration (CSCEC), which can research and evaluate the seismic performance of
structures such as long columns building, cross shape joints, shear walls and 8
floors full scale building. This paper shows you the base function, system main
structure parameter and performance characteristics of MFTS, detailed design
process was given, inverse kinematics model based Hydraulic servosystem and
compound control model of force and displacement were built to analysis the
performance, simulation analysis results show that the MFTS can meet design
requirements and the proposed controller has good abilities. The related
experiments will be tested after MFTS is built.
1 Introduction
As everyone knows, earthquake has extensive damage for building structure etc. which
lead to loss of many people’s lives and impact on the local economy (e.g.
1994 Northridge earthquake which occurred on January 17 with 60 deaths and property
damage between $13 billion and $50 billion; 1995 Hyogo-ken Nambu earthquake
which occurred on January 17 with over 5000 deaths and property damage $150 billion
[1]; 2008 wenchuan earthquake which occurred on May 12 with 69225 deaths and
property damage RMB 8451 hundreds of millions [2]). In order to reduce the destroy
associated with large earthquakes, the seismic performance of building structures
should be evaluated. Because of the simulation models are often inconformity to
capture complex response, and the seismic characteristics is difficult to be precise
evaluated when multi-dimensional and continuously varying are acting on large and
complex building structure during earthquakes [3, 4], many large scale structural
seismic tables have been built and used for performance evaluation. The University of
Minnesota’s Multi-Axial Subassemblage Testing (MAST) [5] System provides a
powerful tool for investigating effects of earthquakes with 6-DOF control system and
high speed data acquisition system, the Multi-Axial Testing System (MATS) in the
NCREE [6], E-Defense [7] which is the largest 3-D shaking table in the world, the
Multi-Axial Full-Scale Sub-Structured Testing and Simulation Facility (MUST-SIM) is
a state-of-the-art physical and analytical simulation facility [8], SRMD [9] consists of
prestressed concrete reaction frame and a moving platen with four hydrostatic low
friction hydraulic has the ability to test dampers and isolator bearings. Above those
loading facilities have the capacity of multiaxis, multifunctional and full-scale tests
which greatly fulfil the experiment demand for buildings.
China is the country with frequent earthquakes, large high-rise buildings and
complex structural bridges are becoming more and more with the rapid economic
development, 108MN Multi-Function Testing System (MFTS) with high speed and
large-scale loading force was proposed and developed jointed by CSCEC (China State
Construction Engineering Corporation) and HIT (Harbin Institute of Technology),
which can provide real-time, full-scale, 6-DOF dynamic and static performance testing
for bearings and all kinds of structural components including shear walls, long col-
umns, and complex cross shape joints. MFTS is full scale and multi scope powerful
experiment tool in earthquake engineering research field when it is finished.
First the paper gives the main construct and characteristics of MFTS,then shows the
detailed design and analysis process,the inverse kinematics model based hydraulic
servosystem of MFTS is derived for understanding the characteristics of the system, a
mixed controller with loading force in out loop and displacement in inner loop is
designed to resolve the control problem of MFTS, simulation and experiment shows
the MFTS has good control performance.
2 MFTS Description
Figure 1 shows the main components of the MFTS, 6DOF loading system is the core of
MFTS. The MFTS is placed in a 25 m 25 m 9 m (length * width * depth) base
pit. In order to ensure the system performance requirement, Hydraulic Power Unit
(HPU) is provided with 1750 L/min at 21 MPa for system supply and 350 L/min at
14 MPa for pre-stage supply for servo valve and 700 L/min at 21 MPa for system
supply or at 31 MPa for accumulator charging. The accumulator is composed of
15,480 L accumulator volume and 19,440 L gas volume pressurized up to 31 MPa.
The oil tank is 68 m3 for oil storage. The 6DOF loading system is made up of 18
actuators, and a loading platform which can simultaneously impose different vertical
loading force and double direction horizontal displacement. Isolation system with 52
vibration isolators is placed between reinforced concrete base and concrete base to
prevent the impact on the surrounding facilities when the MFTS is working. Four lift
cylinders can drive the movable platform move to adjust the distance between 6-DOF
loading system and movable platform for adapting to different parts.
The performance parameter of the MFTS are summarized in Table 1. MFTS can
satisfy four kinds of specimens including long column, shear wall, complex cross shape
joints, and rubber bearings. Testing objectives range from slow-speed unidirectional
testing for basic performance characterization to high-speed, 3-D dynamic testing for
energy-based analysis. The clearance for specimen installation is up to 9.1 m 6.6 m
in plane and up to 10 m in height between movable platform and 6-DOF loading
Design and Realization of 108MN Multi-Function Testing System 323
system. In order to solve the problem of 108MN large loading force in Z direction,
Static pressure support actuators with very low friction are adopted which are fixed
together with the loading platform.
3 System Modeling
hold-down actuators, and a loading platform. Figure 2 shows the structure diagram of
6-DOF loading system. Two coordinate systems were built to analyses the performance
of 6DOF loading system with body coordinate system Op-xpypzp fixed on the loading
platform and the inertial coordinate system Ob-xbybzb fixed on the ground. In the initial
condition, two coordinate systems are overlapped, Ob and Op at the top surface center
of the loading platform.
x2 x1
7205mm
y4 y1
z4 z1
2250mm
xp(xb)
6000mm
y5 h3 h1 2250mm y2
z5 z2
yp(yb)Op (Ob)
y6 h4 h2 y3
z6 z3
2500mm
4000mm 4550mm 5550mm
zp(zb)
yp(yb) Op(Ob)
1090mm
850mm
8600mm
Figure 3 shows you the three kinds of articulated joints in the system: x direction
(a) is SPS (P-prismatic joint, S-spherical joint, E-epitaxy), y direction (b) is PSS, and
the z direction ((c) is EPS which including z1 * z6 and h1 * h4).
The inverse kinematics model of x direction is as following:
li ¼
Rðqo Þai bi þ qp
¼ kgi bi k ði ¼ 1; 2Þ ð1Þ
where ai, and gi are the (3 1) vectors of upper joint point Ai in two coordinate system
respectively, bi is the (3 1) vector of the lower joint point Bi in Ob-xbybzb. Ri is a
rotation matrix of transformation from Op-xpypzp to Ob-xbybzb. qp ¼ ½ x y z T is the
Design and Realization of 108MN Multi-Function Testing System 325
bi lni
pi bi
Ci
vai
lni gi Ob
Ai
bi
Bi Bi
Bi
pj ¼ pj cj ¼ gj bj cj ðj ¼ 1; 2; 3; 4; 5; 6Þ ð2Þ
According to the Eqs. (1), (2), and (3), 6DOF loading system’s inverse kinematic
can be obtained.
For y DOF, the velocity vcj of the pushrod and the velocity of y cylinder are as
following:
vci ¼ cTnj vaj ¼ cTnj q_ p þ cTnj w Raj ð6Þ
_l ¼ Jy q_ p x T ð7Þ
According to Eqs. (5), (7), (9), 6DOF loading system velocity Jacobian matrix J is
as following:
J ¼ ½ Jx Jy Jz ð10Þ
• •
check the control effect, Matlab/Simulink was used to develop the control system
model and XPC-target real time kernel are adopted to run control model, the sample
time is 1 ms.
4 Simulation Experiments
a) Y displacement
b) Z displacement
5 Conclusion
References
1. Elnashai, A.S., Di Sarno, L.: Fundamentals of Earthquake Engineering, pp. 18–19. Wiley
Press, Chicheste (2008)
2. http://www.baike.com/wiki/wenchuanearthquake
3. Gao, C., Zheng, S., Cong, D., Han, J., Yang, Z., Sun, J.: Modeling and control of the CSCEC
multi-function testing system. J. Earthq. Eng. 22, 1–14 (2016)
4. Nakata, N., Spencer, B.F.: Multi-dimensional hybrid simulation using a six-actuator self-
reaction loading system. In: The 14th World Conference on Earthquake Engineering, China
(2008)
5. http://nees.umn.edu
6. Chen, J.C., Lin, T.H., Chen, P.C., Lin, K.C., Tsai, K.C.: Advanced seismic testing using the
multi-axial testing system (MATS) in NCREE. In: Proceedings of the 3rd International
Conference on Advances in Experimental Structural Engineering, San Francisco (2009)
7. Xianchang, L.: E-defense and research tests. J. Earthq. Eng. Eng. Vib. 28(4), 111–116 (2008)
8. http://www.structures.ucsd.edu/
9. Shortreed, J.S., Benzoni, G.: Characterization and testing of the caltrans seismic response
modification device test system. Philos. Trans. Roy. Soc. Math. Phys. Eng. Sci. 359(1786),
1829–1850 (2001)
Research on Vibration Suppression for Boom
Luffing of Telescopic Boom Aerial Work
Platform
Abstract. With the increase in the length of the boom for aerial work vehicles,
the boom vibrates in rapid motion. Therefore, how to suppress the boom
vibration and achieve good stability and safety is significant. Based on the
research of telescopic boom aerial vehicles, the mathematical models of the link
hydraulic system of telescopic boom aerial vehicles are established. According
to the state equation of the boom system, the feedback gain vector is designed
based on the linear quadratic optimal control theory, the feedback value is fed
back to the signal input end of the solenoid valve to construct the closed loop
control system which restrains the boom vibration, and the simulation of the
closed loop control system is made to analyze the vibration suppression effect of
the boom. Through the control of luffing motion, a vibration control method of
telescopic boom aerial vehicle is put forward, which has a good reference
function for similar products.
1 Introduction
Aerial work platforms are widely used in building, manufacturing and maintenance.
The vibration will be generated on the platform when the boom luffing motion occurs.
Especially while the boom works with full boom extension, the vibration is obvious. So
it is necessary to study the vibration suppression for improving the quality of the aerial
work platform.
Now the researches mainly aimed at the study of the state equation, dynamic
analysis and vibration suppression. For example, the mathematical model of the folding
aerial platforms was established by Jing through the Lagrange equations [1]; Tan
proposed a rapid method of acquiring modal parameters for the high platform fire
engine [2]; Sawodny of the University of Stuttgart, Germany designed a kind of motion
planner for the boom elevation angle of the fire ladder car [3], and got the state space
equation of the boom system [4–6]. Based on study of vibration suppression, this paper
adds the electro-hydraulic proportional control model and a new control system is built.
© Springer Nature Singapore Pte Ltd. 2018
K. Li et al. (Eds.): ICSEE 2018/IMIOT 2018, CCIS 924, pp. 330–341, 2018.
https://doi.org/10.1007/978-981-13-2384-3_31
Research on Vibration Suppression for Boom Luffing 331
I0 ðsÞ
G1 ðsÞ ¼ ¼ Ka ð1Þ
Ii ðsÞ
QðsÞ Ksv
Gsv ¼ ¼ ð2Þ
IðsÞ ws22 þ 2f
wsv s þ 1
sv
sv
xp
K
A1 A2
mt FL
V1 V2
Bp
p1 p2
xv
qL qL
ps p0
Imax
Ks ¼ ð3Þ
Lmax
The luffing angle is achieved by controlling the displacement of the luffing hydraulic
cylinder, so the geometric relationship of the expansion and luffing angle of the
hydraulic cylinder can be derived. The structure dimension diagram of the telescopic
boom aerial platform is shown in Fig. 4.
d1 - the distance between A and B. The A indicates the hinges between the boom
and the turntable, B indicates the hinge point between the boom and the variable
luff hydraulic cylinder;
d2 - the distance between A and C. C indicates the hinge point between the
turntable and the variable luff hydraulic cylinder;
l - luffing hydraulic cylinder length;
h0 - When the boom is in horizontal position, the angle between AB and AC;
hðtÞ - luffing angle.
d12 þ d22 l2
cosðh þ h0 Þ ¼ ð4Þ
2d1 d2
334 R.-M. Teng et al.
According to the state equation of the boom system [12], the feedback gain vector is
designed by linear quadratic optimal control theory, and the feedback signal is input
into the electro-hydraulic proportional valve and the closed loop control system is
constructed to suppress the vibration of the boom. The control system is simulated in
the Simulink and the effect of vibration suppression is analyzed.
x_ ¼ Ax þ Bu
y ¼ Cx þ Du
among,
2 3 2 3
0 1 0 0 0
6 17:99 0:424 0 0 7 6 7
6
A ¼4 7; B ¼ 6 10355 7
0 0 0 1 5 4 0 5
0 0 538:9 0:232 65822
C ¼ ½ 0:00191 0 0:00095 0 ; D ¼ 0
Output deviation is
Because output y(t) and input u(t) are one-dimensional functions, the weight
function Q and R can be represented by the constant Q and R. The optimal objective
function (6) is
Z1
1
J¼ xT ðtÞðC T qCÞxðtÞ þ uT ðtÞðDT qD þ rÞuðtÞ þ 2xT ðtÞðCT qDÞuðtÞ dt ð6Þ
2
0
further,
Z1
1 T
J¼ x ðtÞQ1 xðtÞ þ uT ðtÞR1 uðtÞ þ 2xT ðtÞN1 uðtÞ dt ð7Þ
2
0
among, Q1 ¼ CT qC; R1 ¼ DT qD þ r; N ¼ C T qD
Research on Vibration Suppression for Boom Luffing 335
the residual vibration, and extracts the boom tip angular displacement curve in the four
stages. Figures 8 and 9 are respectively the boom angular displacement response curve
in rising stage and decline stage.
Research on Vibration Suppression for Boom Luffing 337
Fig. 8. The angel displacement response of boom tip at the rising stage
Fig. 9. The angel displacement response of boom tip at the decline stage
Through the response analysis of the angular displacement of the boom tip, it can
be found that the vibration of the boom is suppressed by the LQR optimal control. As
time increases, the suppression effect is more and more obvious. In addition, although
the angle displacement of the boom tip has some overshoot, the steady-state value is
basically same with the expected value, which eliminates the steady-state deviation.
And through data analysis, it can be found that the angular displacement response of
Research on Vibration Suppression for Boom Luffing 339
the boom tip is basically consistent when the LQR optimal control system stays at the
rising and decline stages.
Figure 12 is the response curve of the tip line speed at the decline stage. Under the
same variable luffing angle, the effect of LQR optimal control is almost consistent with
the rising stage, namely, the vibration luffing and vibration attenuation time are basi-
cally consistent.
Through the analysis of the line speed of the boom tip, the effect of the vibration
suppression of the boom line speed is not obvious in the first phase. As time increases,
the vibration suppression effect of the boom line speed based on the LQR optimal
control system is more obvious, and the line speed of the boom tip tends to be a stable
value of 0 after about four to five cycles. Compared with the uncontrolled system, the
system has a good effect on suppressing the vibration of the boom.
5 Conclusion
On the basis of the hydraulic system of the telescopic boom aerial vehicles, the
mathematical modeling about the electro-hydraulic proportional valve, the valve con-
trolled asymmetric hydraulic cylinder and the displacement sensor is carried out in this
paper for the luffing vibration; According to the structure dimension of the telescopic
boom aerial vehicle, the geometric relationship between the expansion of the hydraulic
cylinder and the elevation angle of the boom is established. And on the basis of the
state equation of the boom system, the feedback gain vector is designed by using the
linear quadratic optimal control theory. A complete system simulation model is built in
the Simulink and through the simulation analysis, the system based on the linear
quadratic optimal control system can eliminate the steady-state deviation between the
angular displacement of the boom tip and the expected angle displacement, and then
can achieve a good vibration suppression effect and provide ideas for faster and more
stable motion of the aerial platform vehicle.
Research on Vibration Suppression for Boom Luffing 341
References
1. Jing, Y.L., Xiong, J.Q., Wang, F.L.: Application of input shaper in anti-swing control of
folding aerial work vehicles. Noise Vibr. Syst. (4), 38–42 (2011). (in Chinese)
2. Tan, J.H.: Dynamic analysis and vibration control of fire truck boom based on platform.
Xiangtan University, Xiangtan (2014). (in Chinese)
3. Sawodny, O., Aschemann, H., Bulach, A., Hofer, E.P.: Online generation of trajectories
considering kinematic constraints. In: Proceedings of 7th IEEE International Conference on
Methods & Models in Automation & Robotics, Miedzydroje, Poland, pp. 7133–7188 (2001)
4. Sawodny, O., Lambeck, S., Hildebrandt, A.: Trajectory generation for the trajectory tracking
control of a fire rescue turntable ladder. In: Proceedings of the Third International Workshop
on Robot Motion and Control, RoMoCo 2002, pp. 411–416 (2002)
5. Zuyev, A., Sawodny, O.: Observer design for a flexible manipulator model with a payload.
In: Proceedings of 45th IEEE International Conference on Decision and Control, pp. 4490–
4495 (2006)
6. Kharitonov, A., Zimmert, N., Sawodny, O.: Active oscillation damping of the fire-rescue
turntable ladder. In: Proceedings of the IEEE Conference on Control Applications, pp. 391–
396 (2007)
7. Yang, Z.R., Hua, K.Q., Xu, Y.: Electro-Hydraulic Ratio and Servo Control. Metallurgical
Industry Press, Beijing (2009). (in Chinese)
8. Wu, Z.S.: Hydraulic Control System. Higher Education Press, Beijing (2008). (in Chinese)
9. Zhao, C.L.: Adaptive controller and its application in valve-controlled asymmetric cylinder
speed system. Harbin Institute of Technology, Heilongjiang (2006). (in Chinese)
10. Zhang, Y.: Discrete adaptive controller in valve-controlled asymmetric cylinder electro-
hydraulic servo system. Harbin Institute of Technology, Heilongjiang (2008). (in Chinese)
11. Zhang, L.C.: Adaptive controller and its application in valve-controlled asymmetric cylinder
force system. Harbin Institute of Technology, Heilongjiang (2006). (in Chinese)
12. Gao, L.C.: Research on vibration suppression of vertical arm system of aerial work vehicle.
Dalian University of Technology, Liaoning (2015). (in Chinese)
13. Wang, J.Z.: Study on amplitude suppression of arm boom for telescopic boom aerial work
truck. Dalian University of Technology, Dalian (2016). (in Chinese)
Intrusion Detection in SCADA System:
A Survey
Abstract. Nowadays, the industrial systems are more and more interconnected
with the outside world. However, the interconnection of Supervisory Control
and Data Acquisition (SCADA) systems with the outside world using Internet-
based standards introduce numerous vulnerabilities to these systems. Although
awareness is constantly rising, the SCADA systems are still exposed to serious
threats. In this paper, a review of Intrusion Detection and report results is
conducted in the surveyed works. In the end, we also discuss the potential
research directions on this topic.
1 Introduction
The importance of Supervisory Control and Data Acquisition (SCADA) systems has
increased in the past years [1]. And there are more and more cyber incidents in recent
years, which have caused serious damage to the critical infrastructures with very
important economic losses, especially after the Stuxnet incident in 2010 [2]. The
malware which was targeting the PLCs connected to a nuclear centrifuge, and its
ultimate goal was to sabotage the nuclear plant. Besides, a city water utility was hacked
(2011) [3], fully demonstrated the great threat of attacks to industrial control systems.
Intrusion Detection technology is becoming a major research direction for solving
the problem of attacks to SCADA systems which detect attacks by analyzing network
characteristics (such as network traffic, packet loss rate, etc.) after occurring of attacks.
During the last years, many papers about the Intrusion Detection in SCADA system
were reported [9–29]. Therefore, we will make a survey about them in this paper.
This paper is organized as follows: background on SCADA systems is reviewed in
Sect. 2. Taxonomy of detection approach is overviewed in Sect. 3. Intrusion Detection
methodologies are discussed in Sect. 4. This paper is ended with the final conclusion in
Sect. 6.
2 Background
have a better performance than those described in previous research used with several
other classifiers.
Lin et al. [10] propose a semantic analysis framework. The framework combines
Intrusion Detection System (IDS) with a power flow analysis that can estimate the
execution results of control commands.
Wan et al. [11] propose a method which use One Class Support Vector Machine
(OC-SVM) as the first layer to detect misbehaviors. For the process data behavior, they
use Reconstruction Error based on Kernel Principal Component Analysis (RE-KPCA)
method to detect data exceptions. This approach has a better performance in classifi-
cation accuracy than those methods mentioned in the paper.
Shitharth et al. [12] propose an Intrusion Weighted Particle-based Cuckoo Search
Optimization (IWP-CSO) method which combines Hierarchical Neuron Architecture
based Neural Network (HNA-NN) techniques is proposed in this paper. The IWP-CSO
method is used to select the best feature when the input data is clustered. And then, the
HNA-NN is used to classify the intrusions in SCADA system.
Sadhasivan et al. [13] propose an approach which uses the Linear Weighted
Cuckoo Search Optimization (LWCSO) algorithm to get the best features. A Proba-
bilistic Kernel Model (PKM) is used to update the weight function to form a cluster that
represents the best features.
Kleinmann et al. [14] use algorithms which first build a Discrete-Time Markov
Chain (DTMC). Then, the symbols are split into sets. And they use the Deterministic
Finite Automata (DFA) in the work to build a graph model.
Giuseppe et al. [15] introduce models that are developed with the Control Theory
perspective. This research highlights the role of determining the reaction that the
monitoring system may experience when it undergoes different types of cyberattacks.
Yang et al. [16] propose detection method based on expert knowledge. The
advantage of this method is that it does not take into account the actual operating
environment because it has clear advantages over existing more general proposals.
Cruz et al. [17] use an IT-OCSVM method which perform intrusion detection with
high accuracy and low overhead in a time step. The algorithm introduced at most a
96.3% detection accuracy and a 2.5% false alarm rate in the experiments.
Bermudez et al. [18] propose a system named FieldHunter. The system which
extracts fields and infers their types automatically. And it can detect realistic zero-day
attacks on SCADA networks, as result of the valuable information about network
protocol specification it provides.
Almalawi et al. [19] propose a method which uses a data-driven clustering method
and the DBSCAN is selected as the base clustering algorithms. And the method can
identify whether the given system is normal or critical.
Zhou et al. [20] propose a method which use the hidden Markov model. The
method is designed to distinguish between attacks and failures and overcome the
shortcomings of anomaly detection. The research limits the false alarm rate to less than
1.61%. Besides, the designed system has little influence on the performance of control
system.
Erez et al. [21] propose a new domain-aware intrusion detection system. The
system can detect unusual changes in SCADA control register values within
Modbus/TCP protocols, which use Incremental classification and Single-window
Intrusion Detection in SCADA System: A Survey 347
classification. The experimental result shows that the system have a 93% classification
accuracy. Besides, it has a 0.86% false alarm rate.
Nader et al. [22] propose a work which can detect anomaly. The anomaly will
happen after malicious have already passed the IDS. This paper uses machine learning
method for anomaly detection by using one-class classification method including
SVDD and KPCA.
Yang et al. [23] propose a system which provide a considerate method to solve
intrusion detection, by analyzing multiple attributes. And the system uses whitelists and
behavior-based concept, so as to make SCADA systems more safety.
Goldenberg et al. [24] propose a method which use the DFA algorithm by an HMI-
PLC channel. The DFA-based intrusion detection system delve into Modbus/TCP
packets and generate a very detailed model.
Carcano et al. [25] present a method which is based on the critical state analysis and
state proximity concepts. Because the detection is based on the system evolution
analysis, and not on the analysis of the attack evolution, the IDS can detect also “zero-
day attacks”.
Yoo et al. [26] propose a method which extracts the main fields from the
GOOSE/MMS packets using 3-phase preprocessing. And then they removed outliers
by an EM algorithm. Besides, it uses the OCSVM algorithm to learn and detection
normal behavior.
Hink et al. [27] propose a power system classification framwork by using machine
learning method. With the JRipper + AdaBoost approach, they could dependably
classify SCADA system with low false positive rates.
Pan et al. [28] propose a method which uses a Bayesian network graph and gets
data logs from the perspective of relay in the SCADA system. The resulting IDS was
used to classify data get from the perspective of each relay station.
Pan et al. [29] propose another approach which uses the common paths mining-
based method. The common paths is complex specifications which could describe
patterns of system behavior connected with SCADA systems. The IDS which is
mentioned in this paper correctly classifies 90.4% of tested instances. And the average
detection accuracy of this IDS for zero-day attack is 73.43%.
5 Discussion
frequently in the papers we survey. And then, the power systems are the main field that
investigators study in, which causes frequent using of Modbus protocol. Mississippi
State University SCADA Security Laboratory integrates control systems within
numerous critical infrastructure industries to create a testbed and datasets (MSU
SCADA datasets) [30]. These datasets are very widely used in various related studies.
6 Conclusion
We present a survey paper comprising three main contributions: (i) a review of current
proposals of Intrusion Detection in SCADA system, (ii) a taxonomy to classify existing
intrusion detection methodologies and (iii) a comparison of the surveyed works by
which we can see the research trends.
Intrusion Detection in SCADA system is still a developing field. With the devel-
opment of machine learning algorithms [31], researchers achieve a higher detection
accuracy. In the future research, we can combine the machine learning methodologies
with other approach (such as trust management systems [32, 33], adversarial learning
[34, 35] and game theory [36–38]) to get better practical results.
Acknowledgments. This work was partially supported by the National Natural Science Foun-
dation of China (Nos. 61502293, 61775058 and 61633016), the Shanghai Young Eastern Scholar
Program (No. QD2016030), the Young Teachers’ Training Program for Shanghai College &
University, the Science and Technology Commission of Shanghai Municipality (Nos. 18ZR141
5000 and 17511107002) and the Shanghai Key Laboratory of Power Station Automation
Technology.
References
1. Nader, P.: One-class classification for cyber intrusion detection in industrial systems. IEEE
Trans. Ind. Inf. 10(4), 2308–2317 (2015)
2. Chen, T.M., Abu-Nimeh, S.: Lessons from Stuxnet. Computer 44(4), 91–93 (2011)
3. Do, V.L., Fillatre, L., Nikiforov, I., et al.: Feature article: security of SCADA systems
against cyber-physical attacks. IEEE Aerosp. Electron. Syst. Mag. 32(5), 28–45 (2017)
4. Dong, Y., Zhou, P.: Jamming attacks against control systems: a survey. In: Yue, D., Peng,
C., Du, D., Zhang, T., Zheng, M., Han, Q. (eds.) LSMS/ICSEE -2017. CCIS, vol. 762,
pp. 566–574. Springer, Singapore (2017). https://doi.org/10.1007/978-981-10-6373-2_57
5. Cardenas, A., Amin, S., Sinopoli, B., et al.: Challenges for securing cyber physical systems.
In: First Workshop on Cyber-physical Systems Security, vol. 2010, pp. 363–369 (2010)
6. Kushner, D.: The Real Story of Stuxnet. IEEE Spectr. 50(3), 48–53 (2013)
7. Fovino, I.N., Coletta, A., Masera, M.: Taxonomy of security solutions for the SCADA
sector. Joint Research Centre of the European Commission (2010)
8. Bishop M.: Introduction to computer security. In: Euromicro International Conference on
Parallel, Distributed, and Network-Based Processing, pp. 170–174. IEEE (2004)
9. Lahza, H., Radke, K., Foo, E.: Applying domain-specific knowledge to construct features for
detecting distributed denial-of-service attacks on the GOOSE and MMS protocols. Int.
J. Crit. Infrastruct. Prot. 20, 48–67 (2017)
350 P. Zeng and P. Zhou
10. Lin, H., Slagell, A., Kalbarczyk, Z.T., et al.: Runtime semantic security analysis to detect
and mitigate control-related attacks in power grids. IEEE Trans. Smart Grid 9(1), 163–178
(2017)
11. Wan, M., Shang, W., Zeng, P.: Double behavior characteristics for one-class classification
anomaly detection in networked control systems. IEEE Trans. Inf. Forensics Secur. 12(12),
3011–3023 (2017)
12. Shitharth, S., Winston, D.P.: An enhanced optimization based algorithm for intrusion
detection in SCADA network. J. Comput. Secur. 70, 16–26 (2017)
13. Sadhasivan, D.K., Balasubramanian, K.: A novel LWCSO-PKM-based feature optimization
and classification of attack types in SCADA network. Arab. J. Sci. Eng. 42(8), 3435–3449
(2017)
14. Kleinmann, A., Wool, A.: Automatic construction of statechart-based anomaly detection
models for multi-threaded industrial control systems. ACM Trans. Intell. Syst. Technol. 8(4),
55 (2016)
15. Giuseppe, B., Miciolino, E.E., Pascucci, F.: Monitoring system reaction in cyber-physical
testbed under cyber-attacks. Comput. Electr. Eng. 59, 86–98 (2017)
16. Yang, Y., Xu, H.Q., Gao, L., et al.: Multidimensional intrusion detection system for IEC
61850-based SCADA networks. IEEE Trans. Power Deliv. 32(2), 1068–1078 (2017)
17. Cruz, T., Rosa, L., Proença, J., et al.: A cybersecurity detection framework for supervisory
control and data acquisition systems. IEEE Trans. Ind. Inf. 12(6), 2236–2246 (2017)
18. Bermudez, I., Iliofotou, M., et al.: Towards automatic protocol field inference. Comput.
Commun. 84(C), 40–51 (2016)
19. Almalawi, A., Fahad, A., Tari, Z., et al.: An efficient data-driven clustering technique to
detect attacks in SCADA systems. IEEE Trans. Inf. Forensics Secur. 11(5), 893–906 (2016)
20. Zhou, C., Huang, S., Xiong, N., et al.: Design and analysis of multimodel-based anomaly
intrusion detection systems in industrial process automation. IEEE Trans. Syst. Man Cybern.
Syst. 45(10), 1345–1360 (2015)
21. Erez, N., Wool, A.: Control variable classification, modeling and anomaly detection in
Modbus/TCP SCADA systems. Int. J. Crit. Infrastruct. Prot. 10, 59–70 (2015)
22. Nader, P., Honeine, P., Beauseroy, P.: Ip-norms in one-class classification for intrusion
detection in SCADA systems. IEEE Trans. Ind. Inf. 10(4), 2308–2317 (2014)
23. Yang, Y., Mclaughlin, K., Sezer, S., et al.: Multiattribute SCADA-specific intrusion
detection system for power networks. IEEE Trans. Power Deliv. 29(3), 1092–1102 (2014)
24. Goldenberg, N., Wool, A.: Accurate modeling of Modbus/TCP for intrusion detection in
SCADA systems. Int. J. Crit. Infrastruct. Prot. 6(2), 63–75 (2013)
25. Carcano, A., Coletta, A., Guglielmi, M., et al.: A multidimensional critical state analysis for
detecting intrusions in SCADA systems. IEEE Trans. Ind. Inf. 7(2), 179–186 (2011)
26. Yoo, H., Shon, T.: Novel approach for detecting network anomalies for substation
automation based on IEC 61850. Multimed. Tools Appl. 74(1), 303–318 (2015)
27. Hink, R.C.B., Beaver, J.M., Buckner, M.A., et al.: Machine learning for power system
disturbance and cyber-attack discrimination. In: International Symposium on Resilient
Control Systems, pp. 1–8. IEEE (2014)
28. Pan, S., Morris, T., Adhikari, U.: A specification-based intrusion detection framework for
cyber-physical environment in electric power system. Int. J. Netw. Secur. 17(2), 174–188
(2015)
29. Pan, S., Morris, T., Adhikari, U.: Developing a hybrid intrusion detection system using data
mining for power systems. IEEE Trans. Smart Grid 6(6), 3104–3113 (2015)
30. Morris, T., Srivastava, A., Reaves, B., et al.: A control system testbed to validate critical
infrastructure protection concepts. Int. J. Crit. Infrastruct. Prot. 4(2), 88–103 (2011)
Intrusion Detection in SCADA System: A Survey 351
31. Mitchell, T.M., Carbonell, J.G., Michalski, R.S.: Machine Learning. McGraw-Hill, New
York (2003)
32. Zhou, P., Chang, R., Gu, X., et al.: Magic train: design of measurement methods against
bandwidth inflation attacks. IEEE Trans. Dependable Secure Comput. PP(99), 1 (2018)
33. Zhou, P., Jiang, S., Irissappane, A., et al.: Toward energy-efficient trust system through
watchdog optimization for WSNs. IEEE Trans. Inf. Forensics Secur. 10(3), 613–625 (2015)
34. Lowd, D., Meek, C.: Adversarial learning. In: Eleventh ACM SIGKDD International
Conference on Knowledge Discovery in Data Mining, pp. 641–647. ACM (2005)
35. Dalvi, N., Domingos, P., Sanghai, S., et al.: Adversarial classification. In: Tenth ACM
SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 99–108.
ACM (2004)
36. Li, Y., Shi, L., Cheng, P., et al.: Jamming attacks on remote state estimation in cyber-
physical systems: a game-theoretic approach. IEEE Trans. Autom. Control 60, 2831–2836
(2015)
37. Yuan, Y., Yuan, H., Guo, L., et al.: Resilient control of networked control system under DoS
attacks: a unified game approach. IEEE Trans. Ind. Inf. 12(5), 1786–1794 (2016)
38. Huang, J.Y., Liao, I.E., Chung, Y.F., et al.: Shielding wireless sensor network using
markovian intrusion detection system with attack pattern mining. Inf. Sci. 231, 32–44 (2013)
39. Stouffer, K., Falco, J., Kent, K.: Guide to supervisory control and data acquisition (SCADA)
and industrial control systems security, Spin (2006)
Intelligent Servo Feedback Control
for Hydrostatic Journal Bearing
1 Introduction
Rotary equipment’s for precision applications have been long utilized hydrostatic
journal bearing due to their many useful features. These journal bearing offer greater
position accuracy, enhanced stability, and low friction. For the reasons mentioned
earlier, numerous researchers have conducted studies on hydrostatic journal bearings.
Matthew et al. [1] studied the dynamic performance using linear and non-linear
numerical methods for complaint cylindrical journal bearings. Kim et al. [2] derived
equations of perturbation and general Reynold equation. He accomplished this by
analyzing the behavior of journal bearings and studying effects of different recess
shapes on it. Thomsen and Klit [3] further focused their studies on analyzing the effects
of compliant layers over the behavior of hydrodynamic journal bearings. De Pellegrin
et al. [4] derived isothermal and iso-viscous models for hydrostatic journal bearing by
keeping in view the behavioral characteristics of tilt pad type hydrostatic journal
bearing. Phalle et al. [5] studied the effect of wear on performance of a hybrid journal
bearing system which employ membrane type of compensation. Above mentioned
research findings demonstrated to play a vibrant role in successful improvement of the
performance under normal operational environmental conditions in traditionally used
journal bearings.
Considering the operational conditions, these conditions severely restrict behavior
of journal bearings and also play an important part too. So numerous studies have been
conducted in this field by great number of researchers to enhance the dynamic char-
acteristics of these bearing, the rigidness of fluid film and rigid rotor’s stability. Bouaziz
et al. [6] studied the dynamic behavior of the misaligned rotor. Sharma et al. [7]
examined the wear-tear phenomena on a hybrid conical journal bearing with four
pockets. The research findings of this study proposed that with augmented radial
position accuracy and stationarity of traditional journal bearings, we can directly
influence the behavior of mechanical equipment under harsh operational conditions. It
is also tried to control oil film thickness in [8, 9] and some control strategies were
presented.
After this above conclusion, it is probable to increase the performance of a journal
bearing using an active control technology. Using several types of actuators like
piezoelectric, magnetic & hydraulic to control the vibrations in rotor (Nicoletti and
Santos [10]. The Morosi and Santos [11] have effectively validated that it is likely to
apply active lubrication to the gas journal. The Santos and Watanabe [12] further
explored the applications of the active lubrication in journal bearings, in order to get
enhanced damping coefficients and to have a reduced cross-coupling stiffness. With
that it considerably augmented the threshold for the stability of the journal bearing.
Santos suggested an active control for the control of vibration in a flexible rotor, using
an output feedback controller and computed gains of the controllers. Kim and Lee [13]
focused on improvement of dynamic pressure by examining the properties of a sealed
squeeze film type damper. And the results of their study showed that active control
technology is effective in dampers. Ho et al. [14] worked and analysed the effects of
thrust magnetic bearings on the stability and bifurcation of the flexible rotor with active
magnetic bearing system. Nicoletti and Santos [15] focused on the analysis of the
354 W. U. Rehman et al.
frequency response tilt type active hydrostatic bearing. Estupinan and Santos [16]
effectively studied multi-body systems which were inter-connected through thin films
of fluid and derived the mathematical modeling and showed that it can be easily
controlled by using several types of actuators. Haugaard and Santos [17] and Nicoletti
and Santos [15] studied the dynamic performance in theory and focused on active
lubrication of tilting-pad journal bearings. These research studies were concentrated on
customary analysis of stiffness in bearings and the damping in the frequency domain.
Most of the precision rotary equipment have been extensively and more commonly
employing these hydrostatic journal bearings, which led to the requirement of having a
bearing with improved performance and reliability. Santos and Watanabe [12] worked
on the extension of the stability-threshold of the hydrostatic journal bearing and studied
the viability of reduced vibrations in theory. With most of the researchers discovering
these areas, none of them have previously worked or studied on controlling the journal
position accuracy and precision of a hydrostatic journal bearing under external load
conditions and not much existing literature addresses this topic.
In our pursuit to realize our objectives, first we made mathematical model for
hydrostatic journal bearing using servo control and then two different control strategies
are presented. Self-tuning PID control and PID control. A numerical comparison has
been done in Result and discussion part which show effectives of proposed system.
Servo valve is key component of proposed system. It controls the flow towards bearing.
Flow through servo valve is describe by second order differential equation in [18].
The coefficients such as gain, frequency and damping are obtained from manufacturers
[19, 20].
Let suppose non linearities are ignored for derivation of a mathematical model of
Electro hydraulic servo valve (EHSV). The EHSV current is is related to spool dis-
placement xv by first order transfer function [21, 22].
Intelligent Servo Feedback Control for Hydrostatic Journal Bearing 355
Servo
valve
Buffer op amp
Comparator
Control Output
PRV
Sha
o
e
Oil recess 4 J
Senso
Servo
r
valve
PRV
Ps Computer
Filter
Oil Tank
sv x_ vi ¼ Ki ii xvi i 2 x; y ð1Þ
Flow for hydrostatic bearing along recess aligned in vertical and horizontal
direction is given by [23];
8 sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
>
> 2
>
>
< Q2 ¼ Cd wxvx q ðPs P0 Px Þ
>
sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ð2Þ
>
>
>
> 2
> Q4 ¼ Cd wxvx
: ðPs P0 þ Px Þ
q
Hydrostatic journal bearing which has thin land normally we obtain its transfer
function by analyzing effect of changes of pressure and oil film thickness into and out
of bearing recess. Whenever there is offset distance then squeezing effect come into
play which is given by [23],
dhi
Qsqueeze ¼ Ae ð4Þ
dt
Flow is due to effect of pressure at one end. Fluid between two circular surfaces
faces restriction due to shear stress. The motion of journal produces surface velocity
which is zero at boundary and is maximum at place where oil film thickness center is
present. Let suppose pressure across flow is constant then equilibrium force on element
of width b is
dp ds
pbdy p þ dx bdy sbdx þ s þ dy bdx ¼ 0
dx dy
This leads to
ds dp
¼ ð5Þ
dy dx
d 2 u dp
g ¼ ð6Þ
dy2 dx
Uy 1 dp
u¼ yh y2 ð7Þ
h 2g dx
1 bh3 dp
QPU ¼ Uhb ð8Þ
2 12g dx
Let suppose there is a recess or pocket of width b and length l, then pressure
gradient in an Eq. (8) will take the form of PL =l
1 bh3 PLi
QPU ¼ Uhb ð9Þ
2 12g l
Equation (9) is very useful; it is employed to flow that is due to surface velocity and
pressure. Pressure is obtained from external source of energy while surface velocity is
Intelligent Servo Feedback Control for Hydrostatic Journal Bearing 357
produced because of relative motion between two surfaces such as journal and bearing
surface.
Compressibility has good influence on bearing stiffness and damping. It depends
upon bulk modulus which is relationship between change in volume and change in
pressure, is described by relationship;
Ve
dV ¼ dPLi ð10Þ
be
dV Ve dPLi
¼ ð11Þ
dt be dt
Time rate of change of volume is called is flow for hydrostatic bearing termed as
compressible flow, so Eq. (11) takes form that is given by;
Ve dPLi
QCompressibility ¼ ð12Þ
be dt
One can get overall flow in hydrostatic bearing by employing equation of conti-
nuity [24, 25]. Where flow from servo valve is employed in term of squeezing flow,
pressure velocity flow and compressibility flow. Motion dynamics of hydrostatic
bearing can be described by considering factors such as; arrangement of masses, fluid
film stiffness and structure linkages. The equation of motion for hydrostatic bearing in
term of masses, stiffness and coefficients, is given by;
4 Design Strategy
In this paper two design strategies are proposed for HJBSC, one is PID control and
second is Intelligent Control.
optimal parameters. A fittness function is used to find precise and accurate values of
controller parameters. This fitness function finds sum of errors between refrence and
current value and it is given by;
8
>
> 0 x ar
<
ðx ar Þ=ðbr ar Þ ar \x\br
lr ¼ ð15Þ
>
> ðc xÞ=ðcr br Þ br \x\cr
: r
0 cr x
Where subscript z presents the type of the tuning parameter. It may be derivative,
proportional or integral.
Performance of PID & self Tuning PID with respect to Initial pressure
25.05
25
24.95
PID
Bearing Clearance (um)
24.85 P0=3MPa
P0=3MPa
24.80
24.75
24.70
24.65
24.60
24.55
1.05 1.1 1.15 1.2 1.25 1.3 1.35
Time (seconds)
tuning PID control. The external load of 800 N and 1600 N is applied at a time 1.1 s as
shown in Fig. 4. Result shows that under same external load, the stiffness and load
rejection performance of self-tuning PID control based active hydrostatic journal is
better than PID control based active hydrostatic journal bearing. This thing provides the
system good rotation and stationary characteristics as well provides rotor or shaft more
stability.
Performance of PID & self Tuning PID with respect to external loading
25.1
25
Bearing Clearance (um)
24.9
24.8
24.7
PID
24.6 Self-Tuning PID
1600N
800N
24.5
1600N
800N
24.4
1.09 1.1 1.11 1.12 1.13 1.14 1.15 1.16 1.17 1.18 1.19 1.20
Time (seconds)
25
24.95
Bearing Clearance (um)
24.90
Self-Tuning PID
24.85 PID
24.80 η=0.015
2.475
24.70
24.65
24.60
24.55
1.05 1.1 1.15 1.2 1.25 1.3
Time (seconds)
results under different speed as compared to PID controller as shown in Fig. 6. Servo
valve is key component to remove squeezing effect in active hydrostatic journal
bearing.
25
24.95
Bearing Clearance (um)
24.90
Self-Tuning PID
24.85 PID
24.80 RPM=1500
2.475
24.70
24.65
24.60
24.55
1.05 1.1 1.15 1.2 1.25 1.3
Time (seconds)
6 Conclusions
The current work presents an accurate and precise mathematical model that consists of
active hydrostatic journal bearing and servo feedback control. The proposed mathe-
matical model is checked under two different strategies. Experiments are carried out in
Matlab/Simulink. The problem of traditional hydrostatic journal bearing is eccentricity
under different loads which is removed by proposed active hydrostatic journal bearing.
Furthermore, Proposed active hydrostatic journal bearing is checked under two
Intelligent Servo Feedback Control for Hydrostatic Journal Bearing 363
different strategies such PID controller and intelligent Controller. Results show that
active hydrostatic journal bearing with intelligent Controller has better stiffness, load
rejection, rotor stability, less vibration and no wear under different dynamic conditions
of rotating velocity, temperature, initial oil pressure and varying load. Proposed active
hydrostatic journal bearing monitors bearing clearance in real time and adjust oil film
thickness in such a way so that eccentricity is zero under new equilibrium after being
applied load. Current work involves simulation work while future work will be to
validate proposed system on test bench.
References
1. Cha, M., Kuznetsov, E., Glavatskih, S.: A comparative linear and nonlinear dynamic
analysis of compliant cylindrical journal bearings. Mech. Mach. Theory 64, 80–92 (2013)
2. Kim, H., Jang, G., Ha, H.: A generalized Reynolds equation and its perturbation equations
for fluid dynamic bearings with curved surfaces. Tribol. Int. 50, 6–15 (2012)
3. Thomsen, K., Klit, P.: A study on compliant layers and its influence on dynamic response of
a hydrodynamic journal bearing. Tribol. Int. 44, 1872–1877 (2011)
4. De Pellegrin, D.V., Hargreaves, D.J.: An isoviscous, isothermal model investigating the
influence of hydrostatic recesses on a spring-supported tilting pad thrust bearing. Tribol. Int.
51, 25–35 (2012)
5. Phalle, V.M., Sharma, S.C., Jain, S.: Influence of wear on the performance of a 2-lobe
multirecess hybrid journal bearing system compensated with membrane restrictor. Tribol.
Int. 44, 380–395 (2011)
6. Bouaziz, S., Hili, M.A., Mataar, M., Fakhfakh, T., Haddar, M.: Dynamic behaviour of
hydrodynamic journal bearings in presence of rotor spatial angular misalignment. Mech.
Mach. Theory 44, 1548–1559 (2009)
7. Sharma, S.C., Phalle, V.M., Jain, S.: Influence of wear on the performance of a multirecess
conical hybrid journal bearing compensated with orifice restrictor. Tribol. Int. 44, 1754–
1764 (2011)
8. Rehman, W.U., Yuanxin, L., Guiyun, J., Yongqin, W., Rehman, S.U., Bibi, S., et al.:
Control of oil film thickness for hydrostatic journal bearing using PID disturbance rejection
controller. In: 2017 IEEE 3rd Information Technology and Mechatronics Engineering
Conference (ITOEC), pp. 543–547 (2017)
9. Rehman, W.U., Yuanxin, L., Guiyun, J., Yongqin, W., Yun, X., Iqbal, M.N., et al.: Control
of an oil film thickness in a hydrostatic journal bearing under different dynamic conditions.
In: 2017 29th Chinese Control And Decision Conference (CCDC), pp. 5072–5076 (2017)
10. Nicoletti, R., Santos, I.F.: Control system design for flexible rotors supported by actively
lubricated bearings. J. Vib. Control 14, 347–374 (2008)
11. Morosi, S., Santos, I.F.: Active lubrication applied to radial gas journal bearings. Part 1:
modeling. Tribol. Int. 44, 1949–1958 (2011)
12. Santos, I.F., Watanabe, F.Y.: Compensation of cross-coupling stiffness and increase of direct
damping in multirecess journal bearings using active hybrid lubrication: part I—theory.
J. Tribol. 126, 146–155 (2004)
364 W. U. Rehman et al.
13. Kim, K.-J., Lee, C.-W.: Dynamic characteristics of sealed squeeze film damper with a central
feeding groove. J. Tribol. 127, 103–111 (2005)
14. Ho, Y., Liu, H., Yu, L.: Effect of thrust magnetic bearing on stability and bifurcation of a
flexible rotor active magnetic bearing system. Trans. Am. Soc. Mech. Eng. J. Vib. Acoust.
125, 307–316 (2003)
15. Nicoletti, R., Santos, I.F.: Frequency response analysis of an actively lubricated rotor/tilting-
pad bearing system. In: ASME Turbo Expo 2004: Power for Land, Sea, and Air, pp. 735–
744 (2004)
16. Estupinan, E.A., Santos, I.F.: Linking rigid multibody systems via controllable thin fluid
films. Tribol. Int. 42, 1478–1486 (2009)
17. Haugaard, A.M., Santos, I.F.: Multi-orifice active tilting-pad journal bearings—harnessing
of synergetic coupling effects. Tribol. Int. 43, 1374–1391 (2010)
18. Merritt, H.E.: Hydraulic Control Systems. Wiley, Hoboken (1967)
19. Edelmann, H.: Schnelle Proportionalventile und ihre Anwendug, Sonderdruck aus Ölhy-
draulic und Pneumatik. Schnelle Proportionalventile und ihre Anwendug, Sonderdruck aus
Ölhydraulic und Pneumatik 30, 1 (1986)
20. Neal, T.: Performance estimation for electrohydraulic control systems. Moog Tech. Bull. 126
(1974)
21. Rehman, W.U., Wang, S., Wang, X., Shi, C., Zhang, C., Tomovic, M.: Adaptive control for
motion synchronization of HA/EHA system by using modified MIT rule. In: 2016 IEEE 11th
Conference on Industrial Electronics and Applications (ICIEA), pp. 2196–2201 (2016)
22. Rehman, W.U., Wang, S., Wang, X., Fan, L., Shah, K.A.: Motion synchronization in a dual
redundant HA/EHA system by using a hybrid integrated intelligent control design. Chin.
J. Aeronaut. 29, 789–798 (2016)
23. Rehman, W.U., Jiang, G., Luo, Y., Wang, Y., Khan, W., Rehman, S.U., et al.: Control of
active lubrication for hydrostatic journal bearing by monitoring bearing clearance. Adv.
Mech. Eng. 10, 1687814018768142 (2018)
24. Rehman, W.U., Wang, X., Wang, S., Azhar, I.: Motion synchronization of HA/EHA system
for a large civil aircraft by using adaptive control. In: 2016 IEEE Chinese Guidance,
Navigation and Control Conference (CGNCC), pp. 1486–1491 (2016)
25. Rehman, W.U., Nawaz, H., Wang, S., Wang, X., Luo, Y., Yun, X., et al.: Trajectory based
motion synchronization in a dissimilar redundant actuation system for a large civil aircraft.
In: 2017 29th Chinese Control And Decision Conference (CCDC), pp. 5010–5015 (2017)
Iterative Feedback Tuning
for Two-Degree-of-Freedom System
Abstract. This paper is concerned with Iterative Feedback Tuning (IFT) for
Two-degree-of- freedom (2-DOF) system. The IFT is a data-driven method for
tuning controller parameters, which uses the closed-loop system input-output
data directly, and without establishing a mathematical model for the controlled
system. The design of control system is a multi-objective problem, so a 2-DOF
control system naturally has advantages over a one-degree-of- freedom (1-DOF)
control system. When tuning 2-DOF system controllers, two-step method is
firstly concerned. While in this paper, the application of IFT method in a 2-DOF
control system is studied, which can make controllers’ parameters tuned at the
same time, and the IFT method is more accurate and efficient in tracking per-
formance and robustness. The feasibility and effectiveness of the method are
verified by numerical simulation and comparison.
1 Introduction
input-output data is available. Similar to other DDC methods [15–19], IFT does not
need to establish the mathematical model of the controlled system, and only uses the
input-output data of the closed-loop experiments of the system to tuning the controller
parameters directly. IFT theory was originally proposed for single-loop control sys-
tems, and with the development, it has been studied and extended to many advanced
control systems, such as multi-input-multi-output systems [20], internal model control
[21], cascade control [22] and so on. IFT method has been extended to various
industrial applications [23–26].
In this paper, the IFT method is studied to tuning the parameters of the PID
controllers in the 2-DOF control system. In IFT method, at each iteration, the exper-
imental input-output data of the system are utilized to tuning the parameters of the 2-
DOF PID controllers’ parameters directly. The feasibility and effectiveness of the
method are proved by the numerical simulation and comparison of the typical 2-DOF
control system.
The paper is organized as follows. Section 2 illustrates the basic theory of 2-DOF
system and IFT method. In Sect. 3, the IFT theory is extend to the 2-DOF system,
which contains the derivation of correlation variables. Section 4 gives two numerical
examples of the given method, and Sect. 5 is the summary of this paper and some
future work.
2 Preliminaries
Pd ðsÞ
Gd ðsÞ ¼ ð2Þ
1 þ PðsÞCðsÞHðsÞ
To simplify the problem, the following two assumptions that are appropriate for
many practical design problems with some exceptions, are introduced for the control
system shown in Fig. 1.
Iterative Feedback Tuning for Two-Degree-of-Freedom System 367
contoller plant
disturbance
Cf (s) d Pd (s)
controlled
set point manipulated
variable + variable + variable
r + + u + y
C(s) P(s)
-
H(s)
detector
Assumption 1: The detector has sufficient accuracy and speed for the given control
purpose, i.e.,
HðsÞ ¼ 1 ð3Þ
As a result, the control system shown in Fig. 1 can be simplified to the feed-
forward type of 2-DOF control system shown in Fig. 2 under assumptions 1 and 2 [3].
Cf (s) d
+ +
r + + u+ y
C(s) P(s)
-
The controller of the Fig. 2 system consists of C ðsÞ and Cf ðsÞ, where the C ðsÞ is a
conventional PID controller and Cf ðsÞ is a feed-forward compensation PID controller.
1
CðsÞ ¼ Kp 1 þ þ Td DðsÞ ð5Þ
Ti s
Where the Kp is the proportional gain, Ti is the integral time and Td is the differ-
ential time. a and b are 2-DOF parameters, and the above five parameters can be
adjusted. DðsÞ can be expressed by formula (7), where s ¼ Td =d, and d is a differential
gain which is a fixed value and does not affect the optimization of other parameters, but
the characteristics of the actual controlled object need to be considered.
s
DðsÞ ¼ ð7Þ
1 þ ss
368 H. Pan et al.
According to formula (8), the output of the closed-loop system can be divided into
two parts: one is the output caused by the set value r ðtÞ, which is recorded as y2r ðtÞ; and
the other is the output caused by the disturbance d ðtÞ, which is recorded as y2d ðtÞ.
CðsÞ þ Cf ðsÞ
y2r ðtÞ ¼ PðsÞrðtÞ ð9Þ
1 þ PðsÞCðsÞ
1
y2d ðtÞ ¼ PðsÞdðtÞ ð10Þ
1 þ PðsÞCðsÞ
d
+
r+ u+ y
C(s,θ ) P(s)
-
Where r ðtÞ is the system set value input; d ðtÞ is the disturbance; PðsÞ is the
unknown plant; uðtÞ is the control signal. The controller structure of C ðs; hÞ has been
selected beforehand, the controller parameter vector h is adjustable, and the parameter
vector h is differentiable. h ¼ ½h1 hn , n is the number of the parameters. In practice,
the controller C ðs; hÞ is usually chosen as the conventional PID controller, then n ¼ 3.
Iterative Feedback Tuning for Two-Degree-of-Freedom System 369
Where r ðtÞ is the system set value input; d ðtÞ is the disturbance; PðsÞ is the
unknown plant; uðtÞ is the control signal. The controller structure of C ðs; hÞ has been
selected beforehand, the controller parameter vector h is adjustable, and the parameter
vector h is differentiable. h ¼ ½h1 hn , n is the number of the parameters. In practice,
the controller C ðs; hÞ is usually chosen as the conventional PID controller, then n ¼ 3.
The expected closed-loop response of the system from set value input r ðtÞ to output
y1d ðtÞ, which is expressed by formula (13). Td1 ðsÞ is the expected closed-loop system
transfer function.
The tracking error between the actual output y1 ðtÞ and the expected output y1d ðtÞ is
~y ðtÞ.
1
The design goal of the controller C ðs; hÞ is to minimize the tracking error ~y1 ðtÞ
between the y1 ðtÞ and the y1d ðtÞ, so that the system performance criterion function J ðhÞ
can be chosen as quadratic criterion.
N h 2 i
1 X
JðhÞ ¼ y1 ðk; hÞ y1d ðkÞ ð15Þ
2N k¼1
When the sampling data length N is fixed, the value of J ðhÞ is only related to the
controller parameters vector h. The goal of the IFT method is to find the h when the
performance function J ðhÞ gets the minimum value, and define it as the optimal
controller parameter. To find the optimal h can be regarded as the problem of finding
the minimum value of the quadratic function J ðhÞ.
To obtain the minimum of J ðhÞ, a solution for h to the Eq. (17) is found:
N
@JðhÞ 1 X @y1 ðk; hÞ
¼ y ðk; hÞ yd ðkÞ
1 1
¼0 ð17Þ
@h N k¼1 @h
If the gradient @J ðhÞ=@h could be computed, then the solution of (17) would be
obtained iteratively by the following Algorithms:
370 H. Pan et al.
@Jðhi Þ
hi þ 1 ¼ hi ci R1
i ð18Þ
@h
Here, Ri is some appropriate positive definite matrix at iteration i, typically a Gauss-
Newton approximation of the Hessian of J ðhÞ, while ci is a positive real scalar that
determines the step size. The sequence ci , must obey some constraints for the algorithm
to converge to a local minimum of the performance criterion function J ðhÞ.
As stated, this problem is intractable since it involves expectations that are
unknown. However, such problem can be solved by using a stochastic approximation
algorithm of the form (18). The gradient @J ðhÞ=@h evaluated at the current controller
can be replaced by an unbiased estimate. In order to solve this problem, one thus needs
to generate the following quantities:
(i) the signals ~y1 ðtÞ and r ðtÞ;
(ii) the gradients @~y1 ðt; hÞ=@h and @y1 ðt; hÞ=@h.
Because the PðsÞ is unknown, the input-output data of the system are required
which are obtained by closed-loop system experiments.
The IFT for 1-DOF system and the typical structure of 2-DOF are given in previous
section. In this section, the IFT is extended for the 2-DOF case shown in Fig. 2.
T
Where h ¼ ½ hc hf . The error ~y2 ðtÞ between actual output and expected output
of the Fig. 3 system can be expressed by formula (20).
N h 2 i
1 X
JðhÞ ¼ y2 ðk; hÞ y2d ðkÞ ð21Þ
2N k¼1
The gradient @J ðhÞ=@h can be obtained from the derivation of h on both sides of
the Eq. (21).
N
@JðhÞ 1 X @~y2 ðk; hÞ
¼ ~y2 ðk; hÞ ð22Þ
@h N k¼1 @h
Obviously, by calculating the gradient @y2 ðt; hi Þ=@h to estimate @~y2 ðt; hÞ=@h. For
the following calculation, the controller parameters hc and hf are differentiated
respectively.
Notice that the model of the PðsÞ is unknown and there is no need to model it.
@y2 ðt; hi Þ=@hc and @y2 ðt; hi Þ=@hf are obtained by using the input-output data from three
experiments in IFT method.
372 H. Pan et al.
CðsÞ þ Cf ðsÞ 1
y21 ðt; hi Þ ¼ rðtÞPðsÞ þ dðtÞPðsÞ ð26Þ
1 þ PðsÞCðsÞ 1 þ PðsÞCðsÞ
Cf (s) r − y12
d
+ +
r=0 + + u+ y22 ( ŷ )′
2
C(s) C′(s)
2
P(s)
-
(ii) the second experiment is described in Fig. 4, set r2 ¼ 0 and r ðtÞ y21 ðt; hi Þ as
the input of the set value of the system. The output y22 ðt; hi Þ with data length N should
be collected. The gradient @~y2 ðt; hi Þ=@hf can be calculated by formula (28).
PðsÞ
y22 ðt; hi Þ ¼ rðtÞ y21 ðt; hi Þ þ dðtÞ ð27Þ
1 þ PðsÞCðhc Þ
Where
@Cðhc Þ PðsÞ
wðt; hi Þ ¼ dðtÞ þ y2d
1 ðt; hi Þ ð29Þ
@hc 1 þ PðsÞCðhc Þ
(iii) the third experiment is described in Fig. 5, to set r3 ¼ 0 and r ðtÞ as the input of
the set value of the system. Similarly, to get the output y23 ðt; hi Þ with data length N. The
gradient @~y2 ðt; hi Þ=@hf can be calculated by formula (30).
Iterative Feedback Tuning for Two-Degree-of-Freedom System 373
Cf (s) r
d
+ + y32 ( yˆ32 )′
r=0 + C(s) + u+
P(s) C′f (s)
-
PðsÞ
y23 ðt; hi Þ ¼ ðrðtÞ þ dðtÞÞ ð30Þ
1 þ PðsÞCðhc Þ
Where
@Cðhf Þ PðsÞ
vðt; hi Þ ¼ dðtÞ ð32Þ
@hf 1 þ PðsÞCðhc Þ
Through three experiment, the related input-output data are obtained, and can be
calculate for @~y2 ðt; hi Þ=@hc and @~y2 ðt; hi Þ=@hf .
When there is no noise or mean value of noise in the system is zero, or in another
word, the mean values of Eqs. (30) and (33) are zero, @y2 ðt; hi Þ=@hc and @y2 ðt; hi Þ=@hf
are unbiased estimates of @~y2 ðt; hi Þ=@hc and @~y2 ðt; hi Þ=@hf [14].
Based on the above analysis, the gradient @J ðhi Þ=@h can be rewritten as follows:
8 N
>
> @Jðhi Þ 1 X @~y2 ðk; hi Þ
>
> ¼ y ðk; hi Þ yd ðkÞ
2 2
< @hc N k¼1 @hc
ð33Þ
>
> @Jðhi Þ 1 X 2
N @~y ðk; hi Þ
2
>
> ¼ y ðk; hi Þ yd ðkÞ
2
:
@hf N k¼1 @hf
The positive definite matrix Ri can be obtained by using the Hessian estimation of
J ðhi Þ.
8 " T 2 #
>
> 1 X N
@~y 2
ðt; h Þ @~y ðt; h Þ
>
> R ¼
i i
>
< ci N @hc @hc
k¼1
" # ð34Þ
>
> 1X N
@~y2 ðt; hi Þ
T
@~y2 ðt; hi Þ
>
>
: Rfi ¼ N
>
@hf @hf
k¼1
374 H. Pan et al.
On the basis of formula (18), the formula is obtained according to the derivation
and transformation of the above relevant variables. Finally, the iteration of controller
parameter vector h is realized according to formula (35), and the optimal value is finally
achieved.
8
>
> 1 @Jðhi Þ
< hci þ 1 ¼ hci cci Rci @h
c
ð35Þ
>
> @Jðhi Þ
: hfi þ 1 ¼ hci cfi R1
fi
@hf
4 Numerical Example
Based on the above theoretical derivation, this section extends the application of IFT in
2-DOF system by numerical simulation, to compare the feasibility and effectiveness of
this method. In the following discussion, two numerical examples are chosen for the
simulation, which come from reference [3].
1
P1 ðsÞ ¼ eLs ð36Þ
1 þ Ts
2 3
hc1
hf 1
hc ¼ 1 1 s 4 hc2 5; hf ¼ 1 s
ð37Þ
s 1 þ ss 1 þ ss hf 2
hc3
From the data in Fig. 6, it can be seen that the value of the performance criterion
function J ðhÞ obtained from the previous iterations decreases at a relatively large rate,
and the range of parameter changes is relatively large. But with the iteration i increases,
the J ðhÞ decreases slowly and tends to be stable.
The above initial parameters and optimization parameters are brought into the
simulation experiment respectively, and the optimized controller parameters in the
literature are taken as reference. To set up r ðtÞ ¼ 1, with the starting time is 0 s and the
running time is 3.5 s. The simulation figure is shown in a diagram in Fig. 7. In the
following figures, Ini represents the initial parameters of the controller, IFT represents
the optimal parameters, and Ref represents the parameters in the reference [3].
Iterative Feedback Tuning for Two-Degree-of-Freedom System 375
0.25
J(θ)
0.2
0 2 4 6 8 10 12
5.5
c1
θ
5
0 2 4 6 8 10 12
12
c2
11
θ
10
0 2 4 6 8 10 12
0.4
c3
0.3
θ
0.2
0 2 4 6 8 10 12
3.6
3.4
f1
θ
3.2
0 2 4 6 8 10 12
0.45
0.4
f2
θ
0.35
0 2 4 6 8 10 12
Iterative times (i)
1.2 0.3
Ini
IFT
1 0.25
Ref
0.8 0.2
0.6 0.15
y
y
Output
Output
0.4 0.1
0.2 0.05
0 Ini 0
IFT
Ref
-0.2 -0.05
0 0.5 1 1.5 2 2.5 3 3.5 0 0.5 1 1.5 2 2.5 3 3.5
Time(s) (t/s) Time(s) (t/s)
Fig. 7. Comparison of set point response and disturbance response in numerical examples 1.
Setting up r ðtÞ ¼ 0, and adding the disturbance d ðtÞ ¼ 1 to the system. The start
time is 0 s, the simulation figure is shown in b diagram in Fig. 7. It can be seen from
Fig. 7 that the comprehensive control effect of the IFT parameters is better than that of
the initial parameters and the reference parameters.
Td2 ðsÞ ¼ 1, and the structure of PID controllers are shown in formula (38) in numerical
example 1.
1
P2 ðsÞ ¼ 2
eLs ð38Þ
ð1 þ TsÞ
According to the reference data in the literature, the initial controller parameters are
as follows: hc1 ¼ 10, hc2 ¼ 10:5, hc3 ¼ 2:2, hf 1 ¼ 8, hf 1 ¼ 1:7. The running condition
is the same as numerical example 1. Iterative data is shown in Fig. 8.
0.4
J(θ)
0.3
0.2
0 5 10 15 20 25 30
12
11
c1
θ
10
0 5 10 15 20 25 30
20
15
c2
θ
10
0 5 10 15 20 25 30
4
3
c3
θ
2
0 5 10 15 20 25 30
10
9
f1
θ
8
0 5 10 15 20 25 30
2.5
2
f2
θ
1.5
0 5 10 15 20 25 30
Iterative times (i)
From the data in Fig. 8, it can be seen that the value of the performance criterion
function J ðhÞ obtained from the previous iterations decreases relatively small, and the
range of controller parameters is relatively stable.
The initial parameters and optimization parameters are brought into the simulation
experiment respectively, and the optimized controller parameters in the literature are
taken as reference. Setting up r ðtÞ ¼ 1 with the starting time is 0 s, and the running
time is 7 s. The simulation result is shown in a graph in Fig. 9.
Setting up r ðtÞ ¼ 0, and adding the disturbance d ðtÞ ¼ 1 to the system. The sim-
ulation result is shown in b graph in Fig. 9.
According to Fig. 9, to conclude that the comprehensive control effect of the IFT
parameters is better than that of the initial parameters and the reference parameters,
though the anti-disturbance is not as good as reference parameters.
Iterative Feedback Tuning for Two-Degree-of-Freedom System 377
1.4 0.14
Ini Ini
IFT 0.12 IFT
1.2 Ref
Ref
0.1
1
0.08
0.8 0.06
Output (y)
Output (y)
0.04
0.6
0.02
0.4
0
0.2
-0.02
0 -0.04
0 1 2 3 4 5 6 7 0 1 2 3 4 5 6 7
Time(s) (t/s) Time(s) (t/s)
Fig. 9. Comparison of set point response and disturbance response in numerical examples 2.
5 Summary
In this paper, we extend the IFT method to tuning the parameters of the PID controllers
in the 2-DOF structure control system, and the typical feed-forward 2-DOF control is
considered as the research object. Compared with the conventional TSTM, the 2-DOF
PID controllers can be tuned by using the input-output data of the closed-loop control
system without establishing a mathematical model, also two PID controllers’ param-
eters can be tuned at the same time, which makes the controller parameter tuning more
direct and accurate. The simulation results show that the tracking performance of the
system is better than that of the TSTM, and the anti-interference performance of the
system is slightly worse than that of the TSTM. However, the IFT method has less
oscillation and faster stability. It is feasible and effective to tune the parameters of 2-
DOF PID controllers by integrating the control effect of the whole system.
The following two points should be noted when applying the IFT method in tuning
the 2-DOF parameters:
(i) It is necessary to select an appropriate reference model, otherwise, the local
optimum is easy to appear in the IFT iteration, and the control effect of the
obtained parameters is not satisfactory.
(ii) It is necessary to select the appropriate initial value of the controller parameters,
so that the whole control system can be in a stable state, and all the controller
parameters can reach the optimum or near the optimum synchronously, which
requires some experience in the controlled system.
Currently, the IFT is still under study.
References
1. Taguchi, H., Araki, M.: Two-degree-of-freedom PID controllers — their functions and
optimal tuning. J. IFAC Proc. Volumes 33(4), 91–96 (2000)
2. Horowitz, I.M.: Synthesis of Feedback Systems. Academic Press, Cambridge (1963)
3. Araki, M., Taguchi, H.: Two-degree-of-freedom PID controller. Int. J. Control Autom. Syst.
1(4), 18–25 (2003)
4. Yukitomo, M., Shigemasa, T., Baba, Y.: A two degrees of freedom PID control system, its
features and applications. In: Control Conference, IEEE, 2004, pp. 456–459 (2004)
5. Araki, M.: Two-degree-of-freedom control system – I. J. Syst. Control 29, 649–656 (1985)
6. Ye, S.U., Han, P., Wang, D.F.: Chaos optimization of two degree-of-freedom PID control.
J. Comput. Simul. 21(11), 164–166 (2004)
7. Taguchi, H.: Two-degree-of-freedom PID controllers - their functions and optimal tuning.
In: IFAC Digital Control: Past, Present and Future of PID Control, 5–7 April 2000, Terrassa,
Spain (2000)
8. Taguchi, H., Kokawa, M., Araki, M.: Optimal tuning of two-degree-of-freedom PD
controllers. In: Proceedings of the 4th Asian Control Conference, 25–27 September 2002,
Singapore (2002)
9. Taguchi, H., Doi, M., Araki, M.: Optimal parameters of two-degrees-of-freedom PID control
systems. J. Trans. Soc. Instrum. Control Eng. 23(9), 889–895 (2009)
10. Lecchini, A., Campi, M.C., Savaresi, S.M.: Virtual reference feedback tuning for two degree
of freedom controllers. Int. J. Adapt. Control Signal Process. 16(5), 355–371 (2002)
11. Hjalmarsson, H., Gunnarsson, S., Gevers, M.: A convergent iterative restricted complexity
control design scheme. In: 1994 Proceedings of the IEEE Conference on Decision and
Control, pp. 1735–1740 (1994)
12. Xu, J.X., Hou, Z.S.: Notes on data-driven system approaches. J. Acta Autom. Sinica 35(6),
668–675 (2009)
13. Hjalmarsson, H., Gevers, M., Gunnarsson, S.: Iterative feedback tuning: theory and
applications. J. IEEE Control Syst. 18(4), 26–41 (1998)
14. Hjalmarsson, H.: Iterative feedback tuning—an overview. J Int. J. Adapt. Control Signal
Process. 16(5), 373–395 (2002)
15. Arimoto, S., Kawamura, S., Miyazaki, F.: Bettering operation of dynamic systems by
learning: a new control theory for servomechanism or mechatronics systems. In: 2007 IEEE
Conference on Decision and Control, pp. 1064–1069 (1984)
16. Han, Z.G., Hou, Z.S.: Robust model-free learning adaptive control for nonlinear systems.
Control Decis. (2), 137–142 (1995)
17. Spall, J.C.: Multivariate stochastic approximation using a simultaneous perturbation gradient
approximation. IEEE Trans. Autom. Control 37(3), 332–341 (1992)
18. Campi, M.C., Lecchini, A., Savaresi, S.M.: Brief virtual reference feedback tuning: a direct
method for the design of feedback controllers. Automatica 38(8), 1337–1346 (2002)
19. Safonov, M.G., Tsao, T.C.: The unfalsified control concept: a direct path from experiment to
controller. In: Francis, B.A., Tannenbaum, A.R. (eds.) Feedback Control, Nonlinear
Systems, and Complexity. LNCIS, vol. 202, pp. 196–214. Springer, Heidelberg (1995).
https://doi.org/10.1007/BFb0027678
20. Nakamoto, M.: Parameter tuning of multiple PID controllers by using iterative feedback
tuning. In: 2004 SICE 2003 Conference, IEEE, pp. 183–186. IEEE (2003)
21. Bruyne, F.D.: Iterative feedback tuning for internal model controllers. Control Eng. Pract. 11
(9), 1043–1048 (2003)
Iterative Feedback Tuning for Two-Degree-of-Freedom System 379
22. Tesch, D., Eckhard, D., Bazanella, A. S.: Iterative feedback tuning for cascade systems. In:
2017 Control Conference, pp. 495–500. IEEE (2017)
23. Ren, Q., Xu, J., Li, X.: A motion control approach for a robotic fish with iterative feedback
tuning. In: 2015 IEEE International Conference on Industrial Technology, pp. 40–45. IEEE
(2015)
24. Navalkar, S.T., Wingerden, J.W.V.: Iterative feedback tuning of an LPV Feedforward
controller for wind turbine load alleviation. IFAC PapersOnLine 48(26), 207–212 (2015)
25. Rădac, M.B., Precup, R.E., Petriu, E.M.: Stable iterative feedback tuning method for servo
systems. In: 2011 IEEE International Symposium on Industrial Electronics, pp. 1943–1948.
IEEE (2011)
26. Precup, R., Mosincat, I., Radac, M.: Experiments in iterative feedback tuning for level
control of three-tank system. In: 2010 IEEE Mediterranean Electrotechnical Conference,
Melecon 2010, pp. 564–569. IEEE (2010)
IoT Systems
Data Monitoring for Interconnecting
Microgrids Based on IOT
Abstract. The real time monitoring is very important for the interconnecting
microgrids. The power production status is not only grasped by watching the
real time data, but also some optimization approach can be employed to improve
operation mode thus dispatching power reasonably. But now a full range of
monitoring of interconnecting microgrids is not implemented yet. In the present
age of big data, the computing capability and storage space have been devel-
oping fast. These advanced techniques enable the large-scale data monitoring
system. The big data from monitoring can be analyzed and applied into the
optimization of production process. Motivated by this, a real-time data moni-
toring system simulation platform is developed. Specially, ThingSpeak platform
is validated in our work.
1 Introduction
The Microgrid (MG) is considered to be the most promising intelligent power man-
agement system that can be operated in parallel or on an island. The power generated
by a single microgrid is limited and can only provide a maximum load capacity of
approximately 10 MVA. However, several MGs can be interconnected to provide more
power to meet greater load requirements. It also has more redundancy and ensures
better power supply reliability [1]. Multiple interconnected microgrids are often
referred to as Multi Microgrid (MMG) [2]. The MMG system connects multiple
individual microgrids and other distributed generation to a medium voltage distribution
network. Currently, each MG is usually based on the local control strategy and does not
coordinate with other microgrids according to local requirements and goals [3, 4].
However, it is expected that future subtransmission and distribution systems will
consist of several interconnected microgrids and form a complex grid [5].
In an interconnected microgrid, each microgrid will be connected to the next
microgrid by High Voltage DC (HVDC) transmission based on Voltage-Sourced
Converter (VSC). With the rapid development of fully-controlled power electronic
devices and power electronics, high-voltage direct current (HVDC) systems based on
voltage source converters (VSC-HVDC) and pulse width modulation technologies have
been extensively studied and applied [6]. The VSC-HVDC system provides an eco-
nomical and reliable solution for large-capacity long-distance power transmission.
Recently, the development process of the VSC-HVDC system shows that they are a
better alternative to the traditional thyristor-based HVDC system [7].
A large range of modeling and control of the VSC-based HVDC system have been
published in the last few years [8]. From these studies, it becomes clear that the control
and the system impedance can have the impact on the stability of the system. Usually,
we need to pre-assess their impact on the system stability before connecting to the main
grid [9, 10]. Interconnecting together multiple microgrids can lead to undesirable
dynamic behaviors, that is why we must monitor some of the important parameters.
It is, therefore, necessary to data monitoring based on IOT. In particular, this paper
investigates interconnecting together multiple microgrids, thus provide a method for
monitoring data. The remote monitoring center can read the data stored in ThingSpeak
from any microgrid at any time. The main objective in this paper is to use thingspeak
on the microgrid data monitoring, and the remote monitoring center can access
microgrid data at any time for data analysis.
2 System Model
A single microgrid system operated in islanded mode and data monitoring center can
monitor data by the communications network. In other words, each microgrid can
transmit data in real time, and for the data monitoring center can always get any time
required data from any microgrid. Figure 1 shows the comprehensive architecture of
IOT framework with ThingSpeak cloud service for microgrid data monitoring. Where
P1, P2, P3 respectively are active power of each microgrid, Q1, Q2, Q3 respectively are
reactive power of each microgrid, and Udc is DC voltage of VSC-HVDC link.
Thingspeak
P1 Udc1 P2 Udc2 P3
Q1 Q2 Q3
VSC- VSC-
HVDC Link HVDC Link
VSC-
HVDC Link
Udc3
Overview of the control system of a VSC converter and interface to the main circuit
shows an overview diagram of the VSC control system and its interface with the main
circuit.
The converter 1 and converter 2 controller designs are identical. The two controllers
are independent with no communication between them. Each converter has two degrees
of freedom. In our case, these are used to control: P and Q in station 1 (rectifier); Udc
and Q in station 2 (inverter).
3 Results
Fig. 5. the data from microgrid 1 is displayed in ThingSpeak platform with simulation.
(a) microgrid 1 Udc, (b) microgrid 1 P, (c) microgrid 1 Q
There are matlab code to read data stored in ThingSpeak channel as follows:
endDate=datetime('5/14/2017
08:00:00','InputFormat','MM/dd/yyyy
hh:mm:ss');
startDate=datetime('5/12/2017
06:00:00','InputFormat','MM/dd/yyyy hh:mm:ss');
dateVector=[startDate, endDate];
channelID =270840;
readKey ='8NQ2MRCG53WICFTS';
[data,time]=
ThingSpeakRead(channelID,'field',[1,2,3],'DateRang
e',dateVector,'ReadKey',readKey);
plot(time, data)
(a) (b)
(c)
Fig. 6. read the data from ThingSpeak platform (a) microgrid 1 Udc, (b) microgrid 1 P,
(c) microgrid 1 Q
Data Monitoring for Interconnecting Microgrids Based on IOT 389
4 Conclusions
References
1. Chowdhury, S., Chowdhury, S.P., Crossley, P.: Microgrids and Active Distribution
Networks, Renewable Energy Series, vol. 6. Institution of Engineering and Technology,
Stevenage (2009)
2. Gil, N.J., Lopes, J.A.P.: Hierarchical frequency control scheme for islanded multi-microgrids
operation. In: IEEE Lausanne Power Tech Lausanne, pp. 473–478 (2007)
3. Lopes, J.A.P., Hatziargyriou, N., Mutale, J., Djapic, P., Jenkins, N.: Integrating distributed
generation into electric power systems: A review of drivers, challenges and opportunities.
Lecct. Rower Syst. Rest. 77(9), 1189–1203 (2007)
4. Buayai, K., Ongsakul, W., Mithulananthan, N.: Multi-objective micro-grid planning by
NSGA-II in primary distribution system. Eur. Trans. Elect. Power 22(2), 170–187 (2012)
5. Koyanagi, K., et al.: Electricity cluster-oriented network: a gridindependent and autonomous
aggregation of micro-grids. In: Proceedings of the International Symposium: Modern
Electric Power System, Wroclaw, Poland, pp. 1–6, September 2010
6. Aditya, S.K., Das, D.: Load-frequency control of an interconnected hydro-thermal power
system with new area control error considering battery energy storage facility. Int. J. Energy
Res. 24, 525–538 (2000)
7. Usama, M.U., Kelle, D., Baldwin, T.: Utilizing spinning reserves as energy storage for
renewable energy integration. In: Power Systems Conference, pp. 1–5 (2014)
8. Ktiraei, F., Iravani, R., Hatziargyriou, N., Dimeas, A.: Microgrids Management: Controls
and Operation Aspects of Microgrids. IEEE Power and Energy Magazine (2008)
9. Beerten, J., Cole, S., Belmans, R.: Modeling of multi-terminal VSC HVDC systems with
distributed DC voltage control. IEEE Trans. Power Syst. 29(1), 34–42 (2014)
10. Lu, W., Ooi, B.-T.: Optimal acquisition and aggregation of offshore wind power by
multiterminal voltage-source HVDC. IEEE Trans Power Deliv. 18, 201–206 (2003)
11. Maureira, G.A.M., Oldenhof, D., Teernstra, L.: ThingSpeak—an API and Web Service for
the Internet of Things. World Wide Web, 7 November 2015
A Hybrid Routing Control Mechanism
for Dual-Mode Communication of Streetlight
Information Acquisition System
Abstract. In order to improve network stability and data acquisition success rate
of streetlight information acquisition network, a hybrid routing control mecha-
nism is proposed. Combining the distribution and access mode of streetlight node,
and signal attenuation differences between wireless and power line carrier, a dual-
tree network is constructed. Firstly, wireless and carrier next hop nodes are
respectively calculated in dual-tree network by post-order traversal algorithm.
Secondly, next hop node of forwarding packet is selected from wireless and carrier
next hop node by the index of transformation rate of signal. The test results show
that hybrid routing control mechanism can improve network stability and data
acquisition success rate of streetlight information acquisition network.
1 Introduction
Min Xiang, born in 1974, Ph.D., Professor. His research interests include Smart Grid, wireless
sensor network and industrial internet of things.
Xudong Zhao, born in 1993, M.S., His research interests include Smart Grid and industrial
internet of things.
Yongmin Sun, born in 1988, M.S., His research interests include Smart Grid and industrial
internet of things.
In recent years, there has been relevant research on the application of dual-mode
communication technology in streetlight information acquisition network. In [4], a
comparison between PLC and wireless used in streetlight control system is proposed,
and the advantages and disadvantages about them are introduced. In [5], a design of
remote control system based on PLC in the internet of things is proposed. The node can
get the RSSI decision channel by obtaining the power carrier and the radio receiving
signal intensity. In [6], a method of channel selection and routing strategy is proposed
based on the ladder algorithm. The node makes the decision channel through the field
strength and the pre divided channel priority.
In order to improve network stability and data acquisition success rate of streetlight
information acquisition network, a hybrid routing control mechanism is proposed.
Firstly, a dual-tree network is constructed. Secondly, routing of streetlight node for-
warding packets is calculated by post-order traversal algorithm and index of trans-
formation rate of signal (TROS).
Master station
Remote
communications Ethernet GPRS CDMA Others
Concentrator
PLC PLC
Local dual-mode
communications
Wireless Wireless
Streetlight
Start
N
Received packet?
Calculate TROSw and TROSp
Y
N Y
Wait for a moment Dual-tree network has
been built? Next hop node is Next hop node is
wireless next hop carrier next hop
Y node node
End
0 7 55 63 71 79 80 87
32H Source Address C HOP N 16H
MHR MAC payload MFR
(2) Network initialization stage. When streetlight node and concentrator node are
powered on, firstly the network will initialized. The dual-mode communication
modules of concentrator and streetlight node are initialized on software and
hardware. For example, IPv6 protocol stack, software state machine and hardware
I/O. The IPv6 address of concentrator node is set by information from the master
station. After its IPv6 address is configured, the wireless routing forwarding tree
and carrier routing forwarding tree is built by concentrate node, and the wireless
and carrier network beacon frame is broadcasted.
(3) Network discovery stage. Streetlight node needs to wait for a while and judge
whether network beacon frame is received. If streetlight node hasn’t received
network beacon frame since node began to wait, it will broadcasts a route request
frame to its neighboring neighbor nodes and requests a network beacon frame. If
the network beacon frame is received, streetlight node will analyze network
beacon frame to obtain hop count, the content of control domain C, and source
address. Streetlight node calculates RSSI in data link layer. At the same time, the
streetlight receive IPv6 network prefix by neighbor discovery protocol (NDP).
(4) Access control stage. The streetlight node firstly evaluates communication per-
formance of the network to be accessed. The streetlight node sets wireless and
carrier routing metric by hop count and RSSI. When streetlight node evaluates the
communication performance of wireless network, node focuses on RSSI rather
394 M. Xiang et al.
When streetlight node firstly joins the network, its parent node assigns a two-byte
short address Ac to current streetlight node by Eq. 2 and Eq. 3 [8, 9]. The Ap is two-
byte short address of parent node, and the n is the number of child node of parent node.
In Eq. 3. The d is the hop count of current node. After having set two-byte short
address, the streetlight node can obtain the PANID and obtain 64-bit IPv6 network
prefix of parent streetlight node.
A dual-tree network is built by one concentrator and seven streetlight nodes, which
is shown in Fig. 4. The concentrator node is the root node of this dual-tree network,
and each streetlight node has one wireless and carrier parent node. When streetlight
node C joins dual-tree network, streetlight node C doesn’t choose concentrator node as
wireless parent node because there is a barrier between concentrator node and street-
light node C, but streetlight node C still choose concentrator node as carrier parent node
because barrier can’t affect carrier signal. In dual-tree network, some streetlight nodes
have the same wireless and carrier parent nodes while some streetlight nodes have
different wireless and carrier parent nodes. When concentrator node and seven street-
light nodes all use wireless channel to communicate, dual-tree network can be regarded
as wireless routing forwarding tree, and the wireless next hop count of the streetlight
node C is two.
A B C D
Concentrate
Rode
PLC
Street Wireless
Barrier
Rode
I G FH E
calculated between current node and wireless and carrier next hop node. Finally, next
hop node is calculated by TROS and wireless and carrier next hop.
(1) Calculate wireless and carrier next hop node. Wireless next hop node of for-
warding packet is calculated by post-traversing wireless routing forwarding tree,
and carrier next hop node of forwarding packet is calculated by post-traversing
carrier routing forwarding tree. The process of streetlight node post-traversing
routing forwarding tree is shown in Fig. 5.
Start
Post-traverse a node
Receive and Y
This node is the current
process
node?
Post-traverse a node
Y
N
Y Y
End
In Fig. 5, when current node post-traverses a node in the routing forwarding tree, it
will judge whether the node is destination node of packet. If this node isn’t the
396 M. Xiang et al.
destination node, current node will continue post-traversing a node and judging. If this
node is the destination node, current node will post-traverse a node and judge whether
the node is the child node of current node. If the node is child node of current node, the
node is next hop node. If not, current node won’t stop post-traversing a node and judge
whether the node is child node of current node until the node is current node or child
node of current node.
(2) Calculate TROS. In the streetlight information acquisition network, there are
different transmit powers of wireless and carrier communication modules, so the
wireless and carrier RSSI can’t be compared to judge the communication quality
of wireless and carrier channels, and the index of TROS is defined. The current
streetlight node calculates TROS by Eq. 4. The rssi is the received signal strength
of the channel by which current streetlight node communicates with node N, and
the w is the transmit power value in milliwatts. The current streetlight node should
calculate wireless and carrier transformation rate of signal.
10 lg w rssi
TROSðNÞ ¼ 100%: ð4Þ
10 lg w
(3) Hybrid routing control. When current streetlight node receives packet and the
destination address of received packet is the current node, the node receive and
process this packet. When forwarding received packet, the current streetlight node
calculates the next hop node Nnext by Eq. 5. The node N1 is the wireless next hop
node, and the node N2 is the carrier next hop node.
8
< N1 ; TROSðN1 Þ [ TROSðN2 Þ
Nnext ¼ N2 ; TROSðN1 Þ\TROSðN2 Þ : ð5Þ
:
random; TROSðN1 Þ ¼ TROSðN2 Þ
In order to test the performance of hybrid routing control mechanism, a test platform is
set up. This test platform consists of ten streetlight nodes, one concentrator node and
master station. Each streetlight node and concentrator node uses dual-mode commu-
nication module which has wireless and carrier communication mode. In the test, there
are walls between two streetlight nodes, and test platform is shown in Fig. 6.
Related network parameters is shown in Table 1. The Nw is the total barriers in test
area, the Rw is the communication radius of wireless modules, and the Lm is the biggest
permissible number of child nodes.
(1) Parent node update times. In order to verify the ability of improving network
stability by hybrid routing control mechanism (HRCM), HRCM is compared with
carrier communication scheme proposed in [10] and the wireless communication
A Hybrid Routing Control Mechanism for Dual-Mode 397
Master station
Streetlight node
Street
15m
20m
scheme proposed in [11]. The devices shown in test platform are adopted to build
network of streetlight information acquisition by three communication schemes,
and concentrator node calculates the parent node update times. The parent node
update times under different time is shown in Fig. 7. The parent node update times
under different numbers of wall is shown in Fig. 8.
In Fig. 7, the parent node update times in the daytime is more than it at night, but
the parent node update times of HRCM is less than the parent node update times of
Fig. 7. The parent node update times under Fig. 8. The parent node update times under
different time different numbers of wall
398 M. Xiang et al.
PLC in the same time. In Fig. 8, with the increase of the numbers of wall, the parent
node update times gradually increase, but the parent node update times of HRCM is
less than the parent node update times of wireless in the same number of streetlight
nodes. Combining the distribution of streetlight equipment, access method of streetlight
equipment and signal attenuation differences between micro power wireless and power
line communication, a dual-tree network is proposed. The dual-tree has the less parent
node update times.
(2) Packet delivery rate. In the network of streetlight information acquisition, the
packet delivery rate can directly reflect the communication performance of net-
work. In order to verify the ability of improving acquisition success rate streetlight
information, HRCM is compared with PLC and wireless by the index of packet
delivery rate. The devices shown in test platform are adopted to build network of
streetlight information acquisition by three communication schemes, and con-
centrator node calculates packet delivery rate. The packet delivery rate under
different time and wall number is shown in Table 2.
In Table 2, with the increasing number of streetlight nodes, the packet delivery rate
of wireless is gradually reduced, and the packet delivery rate of PLC in the day time is
less than at night. The packet delivery rate of HRCM is the highest in three commu-
nication schemes. In the network built by HRCM, streetlight node evaluates wireless
and carrier channel communication performance by TROS index, and chooses the
better channel to transfer packet.
5 Conclusions
In order to improve network stability and data acquisition success rate of streetlight
information acquisition network, a hybrid routing control mechanism is proposed.
Firstly, a dual-tree network is constructed by combining the distribution of streetlight
equipment, access mode of streetlight equipment, and signal attenuation differences
between wireless and power line carrier. Secondly, when node forwards packet, it
firstly post-traverses wireless and carrier routing forward tree to calculate wireless and
carrier next hop node; node secondly calculates next hop node by index of TROS and
A Hybrid Routing Control Mechanism for Dual-Mode 399
wireless and carrier next hop node. The test results show that hybrid routing control
mechanism can improve network stability and data acquisition success rate of street-
light information acquisition network.
Acknowledgments. This work is supported by Key R&D program of common key technology
innovation for key industries in Chongqing (cstc2017zdcy-zdyfX0032).
References
1. Bai, C.-L., Ma, J.: Intelligent street lamp monitoring system based on internet of things.
J. Appl. Electron. Tech. 40(3), 82–85 (2014)
2. Yan, Z.-L., Gao, S.-J., Huang, X.-Y.: Application of wireless sensor network based on IPv6-
based low-power wireless personal area network in smart street lights. J. China Electric
Power 23, 78–79 (2016)
3. Gao, Y.-H., Liang, X.-Y.: Design of intelligent streetlight control system based on Zigbee.
J. Modern Electron. Technol. 19, 29–32 (2013)
4. Zhu, M.: Comparison of Zigbee and power line carrier in street lamp single lamp control
system. J. China New Commun. 17(19), 39–40 (2015)
5. Li, Z.-Z., Du, Y., Fan, X.-Z.: Research and design of dual mode heterogeneous
communication based on ladder algorithm. J. Electro Tech. Appl. 17, 118–121 (2015)
6. Kong, Y.-H., Li, J.-C., Chen, Z.-X.: Design and Realization of an Adaptive Dual-mode
Communication Module Based on RSSI. J. Sci. Technol. Eng. 16(23), 203–207 (2016)
7. Zhu, Y.-N., Liu, J., Xu, Q.: Application of dual-mode communication in bidirectional
interaction of smart meters. J. Appl. Electron. Tech. (z1), 210–213 (2015)
8. Qi, Z.-Y., Li, Z.-F.: Microgrid monitoring network based on Zigbee communication.
J. Comput. Eng. 43(4), 79–83 (2017)
9. Zhou, S.-Y., Gao, J.-Z., Zou, S.-H.: Energy-efficient data storage with non-uniform nodes
distributed wireless sensor networks. J. Chongqing Univ.: Nat. Sci. Edn. 40(9), 57–66
(2017)
10. Xiang, M., He, J.-X., Du, Y.-H.: An IPv6 based-tree routing control mechanism for low-
voltage power line communication. J. Power Syst. Technol. 40(06), 1874–1880 (2016)
11. Huang, Z.-C., Yuan, F., Li, Y.: Implementation of IPv6 over low power wireless personal
area network based on wireless sensor network in smart lighting. J. Comput. Appl. 34(10),
3029–3033 (2014)
An RF Energy Harvesting Approach for
Secure Wireless Communication in IoT
Abstract. Mobile terminals and base stations are being deployed glob-
ally resulting in an increase in carbon emissions. Energy harvesting is
attracting considerable attention and an environment- friendly commu-
nication system is desirable. In this paper, we demonstrate such a sys-
tem. A power beacon provides energy for a transmitting source and for a
relay using time switching radio frequency energy harvesting technique.
We demonstrate that when an eavesdropper tries to wiretap the signals
transmitted from the source and the relay the system remains secure.
The closed-form expression for secrecy outage probability is derived. Our
results show how the performance varies with the SNR of the power bea-
con and the distance from the relay and the source to the power beacon.
In addition, we show that the duration of energy harvesting process in
relation to the time used for transmitting information has a significant
impact on the system security.
1 Introduction
In early Internet of Things (IoT) applications batteries are being used in plenty
of sensor devices and embedded equipment but their lifespan is limited and they
need replacement and maintenance. Most batteries end up in landfills and this
has a bad impact on the environment. To reduce the use of batteries, energy
harvesting (EH) techniques are gaining increasing interest. Devices and equip-
ment can derive energy by themselves using energy harvesting technology from
environmental resources and prolong their lifetime. Ideally sensors could derive
enough energy to become battery-less! When performing a harvesting process,
energy is gathered from the external environment, i.e., solar, wind, vibration,
electromagnetic, thermoelectric phenomena. In this way, sufficient energy could
be harvested to drive a wireless sensor network (WSN) for an IoT application.
To compare the amount of harvested energy from these resources, we show some
figures from [4,5] in Table 1. Although solar, vibration, thermal have more power
density, they are not always available, moreover, the cost of energy such as solar
is still high.
Compared with other sources in Table 1, due to the increase in mobile ter-
minals and base stations, electromagnetic energy is available anywhere (both
c Springer Nature Singapore Pte Ltd. 2018
K. Li et al. (Eds.): ICSEE 2018/IMIOT 2018, CCIS 924, pp. 400–407, 2018.
https://doi.org/10.1007/978-981-13-2384-3_37
An RF Energy Harvesting Approach for Secure Wireless Communication 401
However, DF relays require more power to operate due to its complexity. Relay
nodes can only be equipped with finite energy to work for a functioning time
[1]. Replacing and charging batteries is not desirable. This motivate us to con-
sider exploiting RF energy harvesting for relays too. We could envisage a sensor
network where sensors and relays are now battery-less.
In future sensor applications, security is also a crucial problem that we can-
not ignore. Because of the broadcast nature of wireless channel, information is
vulnerable to eavesdropping [6]. Traditional way to secure wireless information
is to deploy cryptographic techniques at higher layers despite consuming huge
power for encrypting, decrypting data and burdening the protocol stack which
is to be avoided in energy constrained sensor networks. At the lower layer, phys-
ical layer security (PLS) has gained popularity in recent years to secure wireless
communication.
In this research, we propose a mathematical approach and system model
to obtain the system performance, i.e. secrecy outage probability (SOP). In
the system, an information source (e.g. a sensor) transmits desired signals to a
destination (e.g. a base station or sink) with the help of a relay node. In the
future we envisage IoT scenarios where the sensors and relay nodes are battery-
less. To enable such a system, a power beacon is assumed to provide energy for
the information source and the relay node. Then the information source and
the relay forward the signals with the harvested energy. To assess security, an
eavesdropper is also considered in the system, and it can wiretap signals from
the source and the relay. In our model, an outage occurs when either the system
is not reliable or not secure, hence we assess the secrecy outage probability as a
performance parameter. Our contributions are summarized as:
– We propose a system model including an information source, a relay, a des-
tination, an eavesdropper and a power beacon. The information source and
relay node harvest energy from the power beacon using TS protocol.
– The impacts of the location and the signal-to-noise ratio (SNR) of power
beacon on the system performance are investigated.
– The fraction of time dedicated for energy harvesting in a TS frame α versus
the time dedicated for information transmission to the relay and destination
has a direct impact on performance. This is investigated to find the optimum
value of alpha.
– Closed-form expression for SOP is derived. In addition, Monte Carlo method
is used to evaluate simulation results.
2 System Model
We consider a network with a power beacon B, an information source S, a DF
relay R, a destination D and an eavesdropper E, as shown in Fig. 1. All the
nodes are equipped with one antenna and all the channels are independent and
identically distributed. The additive white Gaussian noise (AWGN) at Rk and D
has zero mean and variance N0 . Assuming all the channels are Rayleigh fading
and the channel power gains follow exponential distribution with parameter λAB ,
where A ∈ {B, S, R} and B ∈ {S, R, D, E}.
An RF Energy Harvesting Approach for Secure Wireless Communication 403
Power Beacon
Cloud
LegiƟmate Link
IllegiƟmate Link
Eavesdropper
Fig. 2. TS Protocol
In the considered system, E can wiretap the information from S and R. Assume
that E is equipped with energy and B does not provide energy for E. S and R
use different code books to enhance the security performance. Hence, the secrecy
capacity can be obtained as
where C1s and C2s are the achievable secrecy rate of S → R and R → D, expressed
as follows:
+ +
1−α 1 + γ1M 1 + γ1M
C1s = log2 = log2 , (6)
2 1 + γ1E 1 + γ1E
+ +
1−α 1 + γ2M 1 + γ2M
C2s = log2 = log2 , (7)
2 1 + γ2E 1 + γ2E
where Cs is the secrecy capacity of the system, and Rth is the secrecy target
rate (Rth > 0). An outage occurs when either the system is not reliable or not
secure.
From (12), we have
where
1 + γM ξ|hBS |2 |hSR |2
γPRS = min ,
1 + γE ξ|hBS |2 |hSE |2
1 + γM ξ|hBR |2 |hRD |2
. (15)
1 + γE ξ|hBR |2 |hRE |2
Rth
FγPRS (β) is the CDF of γPRS , and β = 2 .
From (14) we have the following lemma.
4 Numerical Results
In this section, the simulation results using Monte Carlo approach are evaluated
to prove the accuracy of above performance analysis. The ’Sim’ curves are the
simulation results and ‘Ana’ curves are analytical results. In the figures, we can
observe that both the simulation curves and analytical curves match very well.
Moreover, the parameters are fixed as: γE = 20dB, Rth = 0.2 bits/s/Hz and
η = 0.5. It is assumed that the position of the nodes are located in Cartesian
coordinate system as S = (0, 0), R = (2, 0), D = (3, 0) and E = (1, −4) respec-
tively. The location of B is not fixed because the impact of the location of B is
investigated
later. Hence, the distance between two nodes can be described as
dAB = (xA − xB )2 + (yA − yB )2 where A, B = {S, B, D, R, E}. A and B have
the co-ordinates (xA , yA ) and (xB , yB ). It is assumed that average SNR of each
link is dependent on the path loss as 1/λX = 1/dpl X , where pl is the path loss
exponent. In this section, pl = 4 is assumed.
In Fig. 3, the impacts of γM and the position of B are investigated. We can
observe that with the increase of γM , the system has a better performance due
to the decrease of SOP. The position of B varies from (1, 0), (1, 1) to (1, 2)
406 C. Yin et al.
which means it moves vertically, further from S and R. We can observe from
the figure that the distance between B to R and S increases, the system has a
worse performance because the SOP increases. This demonstrates the fact that
as the distance between B to R and S increases, R and S harvest less energy for
transmission, thus decreasing the system performance.
In Fig. 4, SOP is plotted as function of α. We assume that γE = 20 dB
and B = (1, 0). It can be observed that SOP is extremely high when α is too
small or too large. This is because when α is small, S and R can not harvest
enough energy. In addition, when α is large, the transmit period will be short, so
the secrecy capacity is low. Therefore, we should design α carefully in order to
achieve a secure wireless system. In our examples, α = 0.6 is nearly an optimum
point.
5 Conclusion
In this paper, an EH system with a DF relay and a power beacon is proposed.
The information source and the relay can harvest energy from the power beacon
and then forward the signal to the destination. The eavesdropper can wiretap the
signals transmitted from the source and the relay. Time switching protocol based
EH is used. The closed-form expression for SOP is derived. The results have
shown that γM and the distance between the power beacon and the source and
the relay can affect the performance of the considered EH system significantly.
When the power beacon locates closer to the source and the relay as well as γM
is high, the system performs significantly better. Moreover, the time dedicated
to harvesting energy in a TS frame α can affect the secrecy performance of the
considered system significantly. Hence, α should be designed carefully. We show
the optimum value in our examples. In summary, we have demonstrated that it
is possible to harvest enough energy from RF sources to transmit information, to
extend the range via a relay and more significantly for the system to be resilient
to an eavesdropper attack.
References
1. Akhtar, F., Rehmani, M.H.: Energy replenishment using renewable and traditional
energy resources for sustainable wireless sensor networks: a review. Renew. Sustain.
Energy Rev. 45, 769–784 (2015)
2. Benedict, F.P., Maji, P., Roy, S.D., Kundu, S.: Secrecy analysis of a cognitive radio
network with an energy harvesting AF relay. In: 2017 International Conference
on Wireless Communications, Signal Processing and Networking (WiSPNET), pp.
1358–1363. IEEE (2017)
3. Bloch, M., Barros, J., Rodrigues, M.R., McLaughlin, S.W.: Wireless information-
theoretic security. IEEE Trans. Inf. Theory 54(6), 2515–2534 (2008)
4. Tran, L.G., Cha, H.K., Park, W.T.: RF power harvesting: a review on designing
methodologies and applications. Micro Nano Syst. Lett. 5(1), 14 (2017)
5. Vullers, R., van Schaijk, R., Doms, I., Van Hoof, C., Mertens, R.: Micropower energy
harvesting. Solid-State Electron. 53(7), 684–693 (2009)
6. Yin, C., Nguyen, H.T., Kundu, C., Kaleem, Z., Garcia-Palacios, E., Duong, T.Q.:
Secure energy harvesting relay networks with unreliable backhaul connections. IEEE
Access 6, 12074–12084 (2018)
7. Yin, C., Nguyen, N.P., Garcia-Palacios, E., Tran, X.N., Le-Tien, T.: Secure energy
harvesting communications with relay selection over nakagami-m fading channels.
Mob. Netw. Appl. 1–8 (2017)
8. Zhang, J., Nguyen, N.P., Zhang, J., Garcia-Palacios, E., Le, N.P.: Impact of primary
networks on the performance of energy harvesting cognitive radio networks. IET
Commun. 10(18), 2559–2566 (2016)
Low-Cost, Extensible and Open Source
Home Automation Framework
1 Introduction
The field of IoT has been the topic of much discussion and development over
the past few years, and has arguably suffered from its own hype. The stunning
array of low-cost sensors, actuators and controllers presents a panoply of possible
applications - and in so doing, runs the risk of an over-abundance of choice.
This is reflected in the lack of harmonisation or standardisation throughout the
field [1].
However, some standard tools and techniques have started to emerge which
are helping to increase interoperability between ecosystems and devices [10] -
an important aspect of the interconnected world which is striven for. In this
fluid and constantly developing landscape, a strong strategy for platform devel-
opment focuses upon wide compatibility, technological agnosticism and quick
reconfigurability [6].
Useful technology can be defined by the impact it has upon our lives, and the
home environment presents a prime opportunity for this impact. Some schools of
c Springer Nature Singapore Pte Ltd. 2018
K. Li et al. (Eds.): ICSEE 2018/IMIOT 2018, CCIS 924, pp. 408–418, 2018.
https://doi.org/10.1007/978-981-13-2384-3_38
Low-Cost, Extensible and Open Source Home Automation Framework 409
architecture consider the home as a machine for living and in this vein, the most
important functions of that machine could be considered as suitable targets for
IoT technologies.
The temperature of a dwelling is one of the most important aspects of com-
fort (and even safety, in extreme conditions) and usually consumes the largest
proportion of the total energy expended in the home - both in empirical terms
and in cost. Central heating control is a natural starting point for many home
IoT systems, which often build out their functionality from this foundation [4].
Proactive energy management of a home plays a large role in setting household
running costs and influences environmental impact - and is greatly assisted and
optimised through the use of smart interconnected systems.
A selection of available IoT platforms will be overviewed (with a focus on
Open-Source) in Sect. 2, then this paper will introduce the framework in Sect. 3,
including the details of the software packages, then go on to describe the demon-
stration system in Sect. 4, with information about the implementation in hard-
ware and software. Section 5 offers some observations and thoughts on the results
of using this framework in a practical setting, and the work is briefly concluded
in Sect. 6.
2 IoT Platforms
The framework that will be presented here is more akin to an architecture than
any individual products and so it is most fruitful to consider the major Home
IoT ecosystems in use today rather than attempt to review the wide range of
individual products available.
Amazon Alexa, Google Home and Apple Homekit, nest products and Sam-
sung Smartthings make up a large segment of the market and give a fair repre-
sentation of the shape of closed-source Home IoT technologies. They are charac-
terised by expensive components, centralised servers for management and beau-
tiful, highly developed interfaces that require minimal effort from the end-user
to set up.
Initially, each company tried their best to create a closed ecosystem with a
monopoly on compatible products, however consumer pressure is forcing them to
gradually open their platforms to more devices [5]. The support that each offers
is still far from universal, ultimately restrictive in what can be accomplished and
largely pre-defines how the user interacts with their system.
Open source solutions are available for IoT and there are many, and ever-
changing choices (as is custom in this arena). These solutions vary from pre-
senting a complete management suite down to aggregation of control panels for
disparate devices/systems. Some closed-source projects also aim to integrate dis-
parate IoT ecosystems and generally monetise this effort by offering enhanced
features for a price.
The general advantages of an open-source approach are transparency of oper-
ation (helping to safeguard security and privacy), the ability to customise the
elements as desired and a reduction in cost. In a less tangible fashion, the com-
munity surrounding an open-source project is often passionate about what they
410 C. Cameron and K. Li
are building and using, which is commonly reflected in the level of help that is
available from developers and peers.
Many of these platforms provide built-in support for a wide range of end
devices and hubs, along with a community that continues to develop plug-ins.
Without a specific application in mind, it is counter-productive to attempt to
definitively rank the different projects - Table 1 shows a comparison of the impor-
tant characteristics and features of a variety of systems.
3.1 Overview
Interface Hub
Web GUI
Database
Mosquitto Broker
RF24SN nRF24L01+
WiFi LAN
Nodes
WiFi
nRF
2.4GHz
3.2 Software
its interrelations will be explained, along with some of the notable qualities of
each component.
Communication - The nodes and hub communicate using the MQTT protocol
which implements a topic based publish/subscribe model. The hub acts as a
broker (which routes and buffers all network messages) and each client wishing
to communicate connects to this hub. A single message is composed of a topic
and a payload, each of which is a plaintext string of variable length.
The hub makes use of the free and open-source broker mosquitto, and the
only requirement placed upon nodes is that they must be able to communi-
cate via MQTT - however this does not necessarily need to be accomplished
on the node itself as the hub can implement a translation layer (note RF24SN
in Fig. 1). Complex data structures may be transferred using JSON encoding,
which enables rich communication or lightweight messaging without the need for
multiple communication protocols.
This choice of communication protocols and data structures reflects the flex-
ible nature of the framework - using the most established lightweight stan-
dards maximises interoperability and minimises the burden upon hardware which
enables the widest range of devices to be integrated into a synthesised system [7].
Logic - Node management, business logic, data processing and display are all
accomplished via Node-RED - a programming tool originally developed at IBM
that was open-sourced in 2013 and continues to grow in popularity and func-
tionality. This flow-based programming language runs on top of Node.js and
enables unparalleled development speed and reconfigurability. Instead of tradi-
tional code based programming, the user drags and drops nodes on to a canvas
and wires them together create functionality. In the most basic conception, nodes
are classified into Input, Function and Output - with a rich set of features already
included in the base installation.
Program operation is defined by messages which propagate through a given
set of nodes - transformations of that message or actions triggered by it are
defined by the logic that the user has wired into the program. Each set of nodes
may include branches, conditional blocks or message routing to enhance the
complexity of the program. By defining behaviour in this way, sections of a
program naturally decouple and the relationship between different elements can
be clearly visually seen. The created ‘codebase’ is organised into one or more
labelled tabs which are scoped individually.
There is a thriving open-source community constantly contributing to Node-
RED, both with custom nodes and also full fledged programs (known as Flows)
- the integration of which is often single click effort, again underscoring the
ease-of-use for novice users.
This flow based paradigm is becoming an increasing focus in IoT research
due to its quick reconfigurability and ease of use [2] - applications are not static
and IoT implementations are made considerably more responsive and useful by
enabling a larger proportion of the user base to modify their operation without
needing to learn a complex programming language.
Low-Cost, Extensible and Open Source Home Automation Framework 413
More advanced users can write custom functions in Javascript, create custom
widgets using html/Angular code, group functionality into subflow nodes for easy
re-use, and make completely new nodes with customised functionality. Scoped
variables and objects are also available for buffering and communication of data
within the program.
Node-RED is an excellent match for the requirements of a reconfigurable logic
and data handling framework, and is being increasingly recognised as a powerful
tool [8]. The fact that it is accessible to new users with excellent community
support further enhances its strong position as the heart of a home automation
framework (Fig. 2).
NODES DEPLOY
PALLET (Commit
and Run)
Custom
Subflows
FLOW
FLOW SWITCHING
CANVAS TABS
(Program
Area)
Data Storage - Small sets of data can be stored as JSON encoded text files
by Node-RED, however this strategy quickly becomes unsustainable for a highly
connected home - especially when data analytics are desirable to enhance the
intelligence and utility of an IoT system. Therefore the free and open-source
MySQL relational database is used for data storage.
4 Demonstration System
4.1 Hub
A Raspberry Pi 3 Model B serves as the hub for the system, integrating WiFi
with an additional NRFL4201+ radio for wireless communication. The hub runs
the RF24SN service that implements an MQTT translation layer between the
MQTT Broker and NRFL2401+ radios. A preconfigured Raspberry Pi disk
image minimises set up time.
414 C. Cameron and K. Li
4.2 Nodes
The node types can be divided into three groups - sensors (regularly report
some feature of the environment), actuators (listen for commands from the
hub and then take some action) and actors (perform some action in response
to an event but are not available for command unless externally triggered).
A mix of Arduino Nano and Lolin32 Lite microcontrollers are used to imple-
ment nodes. Battery powered nodes use their onboard ADC to read the battery
voltage and report back to the hub each time they communicate. If the battery
voltage falls below a programmed threshold they will automatically power down
and issue an alert (Fig. 3).
(a) Arduino Nano, NRFL2401+ (b) Lolin32 Lite (ESP32) with battery and
radio and DS18B20 tempera- door latch driver circuitry
ture sensor
Fig. 3. (a) Temperature Sensor node and (b) Finger Print Scanner controller
Definition of Operations:
Sensors
Order of Operations: MEASURE - REPORT - LISTEN - SLEEP
Sensors are configured to take readings of some quantity of interest, and are
expected to be available to the hub at regular intervals.
Low-Cost, Extensible and Open Source Home Automation Framework 415
Actuators
Order of Operations: LISTEN - ACT - REPORT - SLEEP
Actuators are expected to be always online, with minimised latency - and there-
fore are generally mains powered. The hub will issue commands to these nodes
and receive a report of the outcome of the node action. In the case of important
systems, the nodes should have suitable fallback/offline behaviour to cater for
network failures.
Actors
Order of Operations: ACT - REPORT - LISTEN - SLEEP
Actors are distinct from actuators because the hub does not command them
directly - there is no requirement for a minimum contact interval. They are semi-
autonomous devices whose functionality is enhanced by a server connection but
not dependent upon it.
Data Acquisition and Storage - Data from the nodes is reported at regu-
lar intervals via MQTT messages. Upon receipt in Node-RED, the message is
timestamped and its topic set to an appropriate label. If the data is a sensor
reading then a MySQL query is generated and the information is inserted into
the appropriate database.
A NODESTATES object tracks the status of all nodes in the system (includ-
ing battery voltage, last transmission and online/offline status), and is updated
upon receipt of information from each component. Instantaneous or otherwise
non-persistent data can be delivered directly to output nodes (charts, gauges,
text) and will not bloat the database with undesired information.
Data Processing - The system calculates a windowed derivative to estimate
the rate of heating or cooling for each reported location, which can be displayed
as a graph or as an instantaneous estimate.
Hysteresis reporting is also used to remove noise and smooth the graph of
historical temperature data (which is queried from the MySQL database at a
regular interval).
These are simple examples and it should be noted that NodeRED and
MySQL can integrate with external tools for more advanced analytics.
System Operations
Node Error Identification:
The NODESTATES object is interrogated on a regular basis to set a flag for
radio errors, low voltage and offline state. A human readable string is produced
to alert the user that some corrective action should be taken.
Heating Controller:
A multi-room thermostat regulates the heat (using customisable width hystere-
sis) based upon the lowest reported temperature in a user defined set of rooms
and thus ensures a comfortable temperature in all the living areas.
416 C. Cameron and K. Li
Error
Indicator
Gauges
Showing
Rate of
Change
Temperature
(ºC per Hour)
History
for Each
(2 Days)
Node
Current
Temperature, Gauges
Setpoint and Showing
Estimated Current
Time to Temperature
Reach Target of Each
Node
ON/OFF
Central
Table
Heating
showing
System
time since
last
Setpoint transmission
Adjustment (in minutes)
from each
System node
Mode
Hysteresis:
Heating or
Cooldown
Control
Reference
Room Reset
NRF Radio
manually
Demand
and State of
Boiler
Relays
5 Discussion
The low cost NRFL2401+ 2.4 GHz radio system from Nordic Semiconductor does
not support WiFi or encryption and its use here should not be considered an
endorsement of this outdated technology which has largely been supplanted by
low cost devices integrating 802.11 WiFi. Espressif has made a large impact with
their low-cost WiFi microcontrollers, firstly with the ESP8266, followed by the
upgraded ESP32. The SoCs can be procured for less than £2 and development
boards for less than £5 - this price point makes the aforementioned NRF24L01+
practically obsolete for IoT applications and due to the fact that it must be paired
with a microcontroller, the total cost is usually higher than a single ESP SoC.
Examining the collected time series of temperatures (along with knowledge
of the house activities) has made it clear that rich knowledge is available from
this data - it is possible to easily see when the shower has been used, to deter-
mine if doors between rooms are open or closed, the thermal characteristics and
performance of individual rooms are readily apparent and faults in the heat-
ing system leave unique features. The potential for intelligent behaviour when
more advanced analytics are applied is quite significant - additional sensors and
measurands will only increase this. Minimising the fuel used for heating through
smarter scheduling and recommendations for targeted insulation efforts (iden-
tified by modelling the performance of the various house areas) could make a
significant impact on energy bills.
6 Conclusion
A framework based upon flexibility, wide compatibility and quick reconfiguration
was proposed for Home IoT applications. Using open-source software through-
out reduces cost and increases transparency of each component, safeguarding
personal data.
418 C. Cameron and K. Li
Due to the changeable and unpredictable nature of IoT, the framework does
not seek to impose limitations or anticipate end uses - instead it offers a gener-
alised set of tools to easily accomplish tasks and respond to changing application
requirements.
A demonstration system which has improved the performance of the home in
which it was installed has been described, and some basic real-time analysis of the
data has been incorporated without the need for any third party tools. Informal
observation of the data suggests that there is significant scope for knowledge
extraction from low-grade sensor information and with more advanced analytics
it would be possible to make suggestions for physical improvements that target
heating efficiency, as well as automatically minimising the fuel required.
References
1. Bandyopadhyay, D., Sen, J.: Internet of things: applications and challenges in tech-
nology and standardization. Wirel. Pers. Commun. 58(1), 49–69 (2011)
2. Belsa, A., Sarabia-Jácome, D., Esteve, M.: Flow-based programming interoperabil-
ity solution for IoT Platform Applications (2018)
3. Connolly, A.: Freedom of Encryption. Real-World Crypto, pp. 102–103 (2018)
4. Dharur, S., Hota, C., Swaminathan, K.: Energy efficient IoT framework for Smart
Buildings, pp. 793–800 (2017)
5. Leminen, S., Westerlund, M., Rajahonka, M., Siuruainen, R.: Towards IOT ecosys-
tems and business models. In: Andreev, S., Balandin, S., Koucheryavy, Y. (eds.)
NEW2AN/ruSMART -2012. LNCS, vol. 7469, pp. 15–26. Springer, Heidelberg
(2012). https://doi.org/10.1007/978-3-642-32686-8 2
6. Lu, C.H.: Improving system extensibility via an IoT-interoperable platform for
dynamic smart homes. In: Proceedings of the 2017 IEEE International Conference
on Applied System Innovation: Applied System Innovation for Modern Technology,
ICASI 2017, pp. 1300–1303 (2017)
7. Naik, N.: Choice of effective messaging protocols for IoT systems: MQTT, CoAP,
AMQP and HTTP. In: 2017 IEEE International Symposium on Systems Engineer-
ing, ISSE 2017 - Proceedings (2017)
8. Rajalakshmi, A., Shahnasser, H.: Internet of Things using Node-Red and alexa.
In: 2017 17th International Symposium on Communications and Information Tech-
nologies (ISCIT), pp. 1–4 (2017). http://ieeexplore.ieee.org/document/8261194/
9. Ruiz, M.C., Mcarmenruizuclmes, E., Olivares, T., Teresaolivaresuclmes, E., Lopez,
J.: Evaluation of Cloud Platforms for Managing IoT Devices (2017)
10. Zitnik, S., Jankovic, M., Petrovcic, K., Bajec, M.: Architecture of standard-based,
interoperable and extensible IoT platform. In: 24th Telecommunications Forum,
TELFOR 2016, pp. 1–4 (2017)
Spectrum Utilization of Cognitive Radio
in Industrial Wireless Sensor
Networks - A Review
1 Introduction
In the current information-intensive society, the development of distributed mon-
itoring and control system and industrial and factory automation is featured by
the characteristics of flexibility, integration, robotization and intelligentization.
Industrial applications have no longer been confined to closed plant environ-
ment, the networks tend to coexist with other industrial wireless systems and
existing commercial wireless systems [1]. Given the growing number of inter-
connected industrial systems and devices for agile industrial manufacturing,
industrial wireless sensor networks (IWSNs) are playing an increasingly more
important role. In an industrial environment, wireless sensor nodes are installed
on on-site equipment and used to monitor diverse parameters such as tempera-
ture, pressure, humidity, location and vibration [2]. The advantage of IWSNs are
appealing over traditional wired communication systems such as cost-efficiency,
c Springer Nature Singapore Pte Ltd. 2018
K. Li et al. (Eds.): ICSEE 2018/IMIOT 2018, CCIS 924, pp. 419–428, 2018.
https://doi.org/10.1007/978-981-13-2384-3_39
420 M. Yin et al.
self-organization, scalability and mobility [3]. It has been seen as a vital com-
ponent in the Industry 4.0 framework, and can be used for smart factories,
networked manufacturing and industrial internet of things.
Current spectrum allocation mechanism assigns wireless spectrum to licensed
operators based on the static spectrum allocation policy. The cost of the license
is very high, however, only parts of the spectrum is utilized effectively. Most
of the distributive spectrum is utilized intermittently. According to report of
Federal Communications Commission (FCC), the current utilization of a licensed
spectrum varies from 15% to 85% [4]. The inefficient use and spectrum scarcity
definitely hinder the development of wireless communication systems. The rest
of industrial and home applications have to compete for the even more crowded
free Industrial Scientific Medical (ISM) bands not only with each other but also
with other wireless communication devices in the same area.
Cognitive Radio (CR) technology is recognized as an emerging technology
to solve the dilemma spectrum utilization is facing: the contradiction between
rapid-growing number of wireless devices and the scarcity of available spectrum
resources. The definition of CR by FCC is “Cognitive radio: A radio or system
that senses its operational electromagnetic environment and can dynamically and
autonomously adjust its radio operating parameters to modify system operation,
such as maximize throughput, mitigate interference, facilitate interoperability,
access secondary markets [5].” In this paradigm, sensor nodes can sense the idle
frequency bands (namely, the spectrum holes) and transmit data by re-using
the frequency bands occupied by the authorized users. Thus, the utilization of
existing spectrum resources is enhanced. The authorized users are recognized as
primary users (PUs) while sensors nodes are secondary users (SUs). SUs carry
radio devices that have the ability of scan and sense of the surrounding spec-
trum utilization. Once spectrum holes found, SUs can access the idle spectrum
opportunistically by altering their spectrum bands subsequently [6].
In this paper, we start with a discussion on the challenges of industrial envi-
ronments, focusing on different interference sources, coexistence, and heteroge-
neous environment. Next, an overview of spectrum sensing techniques is given.
This is then followed by a presentation on a range of methods relative to spec-
trum utilization. Finally, we discuss some open problems and future trends of
IWSNs.
Industrial environment is often more complicated than public and private envi-
ronments. It has higher quality of service (QoS) requirements than that applica-
tions at homes and offices. Reliability, latency, and availability are some major
aspects and they can be quite specific for different applications. For example,
monitoring system is usually time-sensitive. Data with long latency may lead to
wrong decisions.
Spectrum Utilization of Cognitive Radio in IWSN - A Review 421
Traditionally, the applications of the IWSN systems can be classified into three
categories: safety systems, control systems, and monitoring systems [10]. In these
systems, a wide range of data are collected or processed over a given region for
a long duration and the data are exploited thoroughly to make certain conclu-
sions. In recent years, the data sources are more abundant which may include
on-demand or live video streaming, audio, and still images. These multimedia
applications set high demands for transmission rate and bandwidth. Moreover,
in a dense network environment such as indoor plants, the media control access
mechanism should be carefully designed because different sensor nodes in the
area are likely to access a channel simultaneously. In addition, the topology of
the network and connection of the nodes change over time due to the link fail-
ures and battery depletion. The idea of heterogeneous sensor networks thus come
naturally with two of more nodes working with different energy and hardware
complexity.
(1) Matched Filter Detection: The matched filter (also known as coherent detec-
tor), is considered as the optimum method because of its accuracy. As the
Spectrum Utilization of Cognitive Radio in IWSN - A Review 423
parameters of the transmitted signals are known, the matched filter accord-
ingly maximizes the received signal-to-noise ratio (SNR). The presence or
absence of PU is determined by the comparison between the result of corre-
lation and a predetermined threshold. However, matched-filtering also has
limitations. It requests the features of PU signal in advance, which include
bandwidth, operating frequency, modulation type and order, pulse shap-
ing, and frame format [11]. Also, the computational complexity and power
consumption are very high.
(2) Energy Detection: Energy detection is performed in both time and frequency
domain. It compares the energy received by PU signal with a threshold based
on the noise estimation. The method is simple and has low computational
cost because it does not need a priori information of PUs signal. But, the
drawback of the method is obvious as well. The noise needs to be accurate
which is difficult to estimate [12]. It is also not suitable to use the method
in low SNR environment and for detecting spread spectrum signals.
(3) Cyclostationary Feature Detection: The cyclostationary feature detector dif-
ferentiates the PU signal from the additive noise since the method deals with
the inherent cyclostationary properties or features of a signal. A cyclosta-
tionary signal is said to be periodic when considering the mean and auto-
correlation [13]. The technique is simple and effective. Only one or two
periodic features of the PU signal is needed. It can be used to detect at very
low SNR condition. The robustness to noise of the method performs better
than energy detector, nevertheless, the computational complexity is higher
than that of energy detector.
4 Spectrum Utilization
4.1 Clustering
4.4 Optimization
The challenges of power and hardware limitations of sensor nodes should be over-
looked as its inherent characteristics. Combining CR functionalities with sensor
nodes is a self-evident choice when energy and network lifetime is co-considered.
The existing networks and protocols are seldom aware of cognitive capability,
many open optimization issues are left for research. For example, authors in [24]
investigated the optimal packet size for a cognitive wireless sensor network. It
defines the system and gives the simulation result of the optimal packet size for
the proposed system with maximum energy efficiency while maintaining reason-
able interference level for the licensed PU. In [25], a multi-objective optimization
framework is proposed for reducing CO2 emissions. The cooperative cognitive
system jointly optimizes the relay assignment and power allocation with two
conflicting objectives which are to maximize the sum-capacity and to minimize
the CO2 emissions. Although a hybrid evolutionary scheme of EDA is proposed,
the real-time performance is poor due to the priori knowledge of the distribution
of relay nodes and the constraints of greedy algorithm.
mathematical models of random phenomena, and their values will change with
the affecting factors. [26] proposes a hybrid sensing method using a continuous
time Markov chain model to adjust the parameters in avoid of contradictory
factors. According to the characteristics of PU and SUs, the best sensing period
is found. Considering that sensing results are affected by uncertain noise and
contradictory factors, [27] introduces the partially observable Markov decision
processes to build the channel model. It enables the cognitive nodes with the
ability of channel sensing, channel switching, and data transmission.
5 Conclusion
In this paper, the potential benefits and current studies of using CR technology
to improve the spectrum utilization in IWSN have been discussed. We give a brief
introduction of the CR technology and its different sensing techniques, namely
signal processing and cooperative sensing techniques. Further, the potential tech-
niques from the aspects of network architecture, spectrum access, interference
management, spectrum sensing and spectrum sharing are presented. However,
there are still some open problems in this area. One thing should be noted is
that there is no universal test platform and test environment to implement the
approaches, evaluate the cooperation mechanism and verify the accuracy. Hence,
standardized protocols and practical methods need to be proposed.
References
1. Chiwewe, T.M., Mbuya, C.F., Hancke, G.P.: Using cognitive radio for interference-
resistant industrial wireless sensor networks: an overview. IEEE Trans. Ind. Inform.
11(6), 1466–1481 (2015)
2. Gungor, V.C., Hancke, G.P.: Industrial wireless sensor networks: challenges, design
principles, and technical approaches. IEEE Trans. Ind. Electron. 56(10), 4258–4265
(2009)
3. Cena, G., Seno, L., Valenzano, A., Zunino, C.: On the performance of ieee 802.11E
wireless infrastructures for soft-real-time industrial applications. IEEE Trans. Ind.
Inform. 6(3), 425–437 (2010)
4. Akyildiz, I.F., Lee, W.Y., Vuran, M.C., Mohanty, S.: A survey on spectrum man-
agement in cognitive radio networks. IEEE Commun. Mag. 46(4), 40–48 (2008)
5. Federal Communications Commission: Notice of proposed rule making and order:
facilitating opportunities for exible, efcient, and reliable spectrum use employing
cognitive radio technologies. ET Docket, February 2005
6. Joshi, G.P., Nam, S.Y., Kim, S.W.: Cognitive radio wireless sensor networks: appli-
cations, challenges and research trends. Sensors 13(9), 11196–11228 (2013). http://
www.mdpi.com/1424-8220/13/9/11196
7. Low, K.S., Win, W.N.N., Er, M.J.: Wireless sensor networks for industrial environ-
ments. In: International Conference on Computational Intelligence for Modelling,
Control and Automation and International Conference on Intelligent Agents, Web
Technologies and Internet Commerce (CIMCA-IAWTIC 2006), vol. 2, pp. 271–276,
November 2005
8. Werb, J., Newman, M., Berry, V., Lamb, S., Sexton, D., Lapinski, M.: Improved
quality of service in IEEE 802.15.4 mesh networks. In: Proceedings of International
Workshop on Wireless and Industrial Automation, pp. 1–4 (2005)
9. Qu, F., Zhang, J., Shao, Z., Qi, S.: Research on resource allocation strategy of
industrial wireless heterogeneous network based on IEEE 802.11 and IEEE 802.15.4
protocol. In: 2017 3rd IEEE International Conference on Computer and Commu-
nications (ICCC), pp. 589–594, December 2017
10. Somappa, A.A.K., Ovsthus, K., Kristensen, L.M.: An industrial perspective on
wireless sensor networks x2014; a survey of requirements, protocols, and challenges.
IEEE Commun. Surv. Tutor. 16(3), 1391–1412 (2014)
11. Zhang, X., Chai, R., Gao, F.: Matched filter based spectrum sensing and power
level detection for cognitive radio network. In: 2014 IEEE Global Conference on
Signal and Information Processing (GlobalSIP), pp. 1267–1270, December 2014
12. Xuping, Z., Jianguo, P.: Energy-detection based spectrum sensing for cogni-
tive radio. In: 2007 IET Conference on Wireless, Mobile and Sensor Networks
(CCWMSN07), pp. 944–947, December 2007
13. Aparna, P.S., Jayasheela, M.: Cyclostationary feature detection in cognitive radio
using different modulation schemes (2012)
14. Bera, D., Maheshwari, S., Chakrabarti, I., Pathak, S.S.: Decentralized coopera-
tive spectrum sensing in cognitive radio without fusion centre. In: 2014 Twentieth
National Conference on Communications (NCC), pp. 1–5, February 2014
15. Pritom, M.M.A., Sarker, S., Razzaque, M.A., Hassan, M.M., Hossain, M.A., Ale-
laiwi, A.: A multiconstrained QoS aware MAC protocol for cluster-based cognitive
radio sensor networks. Int. J. Distrib. Sens. Netw. 11(5), 262871 (2015). https://
doi.org/10.1155/2015/262871
428 M. Yin et al.
16. Rauniyar, A., Shin, S.Y.: A novel energy-efficient clustering based cooperative spec-
trum sensing for cognitive radio sensor networks. Int. J. Distrib. Sen. Netw. 11(6),
198456 (2015). https://doi.org/10.1155/2015/198456
17. Singhal, D., Barjatiya, S., Ramamurthy, G.: A novel network architecture for cogni-
tive wireless sensor network. In: 2011 International Conference on Signal Process-
ing, Communication, Computing and Networking Technologies, pp. 76–80, July
2011
18. Li, X., Wang, D., McNair, J., Chen, J.: Residual energy aware channel assignment
in cognitive radio sensor networks. In: 2011 IEEE Wireless Communications and
Networking Conference, pp. 398–403, March 2011
19. Chen, T., Zhang, H., Maggio, G.M., Chlamtac, I.: Topology management in
CogMesh: a cluster-based cognitive radio mesh network. In: 2007 IEEE Interna-
tional Conference on Communications, pp. 6516–6521, June 2007
20. Zhang, H., Zhang, Z., Dai, H., Yin, R., Chen, X.: Distributed spectrum-aware
clustering in cognitive radio sensor networks. In: 2011 IEEE Global Telecommuni-
cations Conference - GLOBECOM 2011, pp. 1–6, December 2011
21. Jamal, A., Tham, C.K., Wong, W.C.: Event detection and channel allocation in
cognitive radio sensor networks. In: 2012 IEEE International Conference on Com-
munication Systems (ICCS), pp. 157–161, November 2012
22. Mougy, A.E., Ibnkahla, M.: Achieving end-to-end goals of WSN using weighted
cognitive maps. In: 37th Annual IEEE Conference on Local Computer Networks,
pp. 328–331, October 2012
23. El Mougy, A., Ibnkahla, M.: A cognitive WSN framework for highway safety based
on weighted cognitive maps and q-learning. In: Proceedings of the Second ACM
International Symposium on Design and Analysis of Intelligent Vehicular Networks
and Applications, DIVANet 2012, pp. 55–62. ACM, New York (2012). https://doi.
org/10.1145/2386958.2386967
24. Oto, M.C., Akan, O.B.: Energy-efficient packet size optimization for cognitive radio
sensor networks. IEEE Trans. Wirel. Commun. 11(4), 1544–1553 (2012)
25. Naeem, M., Ashrafinia, S., Lee, D.: Estimation of distribution algorithm for green
resource allocation in cognitive radio systems. In: 2012 6th International Confer-
ence on Signal Processing and Communication Systems, pp. 1–7, December 2012
26. Khattab, A., Perkins, D., Bayoumi, M.A.: Opportunistic spectrum access: from
theory to practice. IEEE Veh. Technol. Mag. 7(2), 62–68 (2012)
27. Han, J.A., Jeon, W.S., Jeong, D.G.: Energy-efficient channel management scheme
for cognitive radio sensor networks. IEEE Trans. Veh. Technol. 60(4), 1905–1910
(2011)
28. Wang, B., Wu, Y., Liu, K.R.: Game theory for cognitive radio networks: an
overview. Comput. Netw. 54(14), 2537–2561 (2010). http://www.sciencedirect.
com/science/article/pii/S1389128610001064
29. Yuan, W., Leung, H., Chen, S., Cheng, W.: A distributed sensor selection mech-
anism for cooperative spectrum sensing. IEEE Trans. Sig. Process. 59(12), 6033–
6044 (2011)
30. Huang, J., Berry, R.A., Honig, M.L.: Auction-based spectrum sharing. Mob. Netw.
Appl. 11(3), 405–418 (2006). https://doi.org/10.1007/s11036-006-5192-y
31. Gandhi, S., Buragohain, C., Cao, L., Zheng, H., Suri, S.: A general framework for
wireless spectrum auctions. In: 2007 2nd IEEE International Symposium on New
Frontiers in Dynamic Spectrum Access Networks, pp. 22–33, April 2007
A Network Coding Against Wiretapping
Attacks of the Physical Layer Security Based
on LDPC Code
1 Introduction
With the rapid development of wireless communication, more and more attention was
paid to how to realize the safe transmission of information. As the lowest layer of
communication network, the physical layer has attracted attention for its security
problems due to the eavesdropping hazards. In 2010, Wyner [1] proposed the eaves-
dropping channel model.
The traditional network security problem relies on the method of cryptography, but
there are some disadvantages such as large amount of computation. Therefore, the
security of information theory has attracted people’s attention. Security in the infor-
mation theory sense is security in the absolute sense. In the past, it was generally
believed that security in the sense of information theory could not be realized, so,
security in the sense of information theory was not considered in the actual commu-
nication system. Shannon has pointed out that when the quality of the eavesdropping
channel is worse than that of the legitimate user channel, the physical layer security can
© Springer Nature Singapore Pte Ltd. 2018
K. Li et al. (Eds.): ICSEE 2018/IMIOT 2018, CCIS 924, pp. 429–440, 2018.
https://doi.org/10.1007/978-981-13-2384-3_40
430 Y. Zheng and J. Fu
be realized by certain coding methods [2, 3]. The LDPC code can approach Shannon’s
limit and realize the security of information theory. Network Coding (NC) came into
being in 2000, which was proposed by Dr. Ahlswede [5] from Hong Kong based on
Shannon’s theory of “maximum flow and minimum cut”, it was proved theoretically
that nodes in the network can complete the storage and forwarding function in the
traditional relay network, it can also carry out some coding processing on the trans-
mitted signals to achieve the theoretical maximum of network multicast, such as
exclusive OR (XOR) in the finite domain. In 2002, Cai and Yeung [6] first studied
security network coding, it was assumed that in a single-source network, an eaves-
dropper can eavesdrop on a subset of a channel that the sender and receiver do not
know about. This network code is considered secure when the eavesdropper does not
receive any information transmitted over the network through the eavesdropper. In the
literature, the wire-tap model (CSWN) is proposed, this model is aimed at eaves-
droppers with certain eavesdropper capability, when the number of eavesdropper
channels is less than a certain value, no useful information of source information can be
obtained, this model has been proved to be safe in information theory. In 2006, the
application of network coding ideas was extended to the physical layer, Shengli Zhang
from Chinese university of Hong Kong proposed the concept of Physical layer network
coding (PNC) in literature [4]. The essence is to make full use of the superposition
characteristic of radio waves to improve the throughput of the network. The specific
method is to make use of the superposition characteristic of radio waves, and then carry
out XOR coding processing on the signal that needs to be relayed, finally, send it to the
destination node.
The Progressive Edge Growth (PEG) algorithm is a greedy algorithm. Currently,
the most commonly used LDPC construction method is PEG algorithm proposed by
Hu et al. [11] it’s a classic random construction method, it is considered to be the
optimal algorithm in the design of short and medium code. The method presented in
this paper is based on the improved PEG algorithm and a structured construction
approach is used, it’s made up of a circulant replacement unit matrix, the matrix is
sparser, the constructed check matrix is not limited by column weight and can effec-
tively reduce the demand for memory space, the matrix has no short loop and the check
part matrix has the structure of quasi dual-diagonal lines, it has the characteristics of
fast coding.
The core idea of physical layer network coding is under the premise of strict carrier
synchronization and code element synchronization, introducing the appropriate map-
ping mechanism, after the physical layer of relay node superimposes and receives the
electromagnetic waves arriving at the same time, by modulation mapping, this
superimposed electromagnetic wave is mapped to bit information flow on the network
layer, it is equivalent to directly in the network layer carried out XOR coding.
A Network Coding Against Wiretapping Attacks 431
The physical layer network coding is shown in Fig. 3. In this scheme, only two
times gaps are needed to realize the mutual transmission of information. The first time
slot is S1 and S3 sending messages to R at the same time. X2 ðX2 ¼ X1 X3Þ is
encoded by the physical layer network. The second time slot is R broadcasting X2 to
432 Y. Zheng and J. Fu
S1 and S3, and S1 and S3 solve each other’s information according to their own
information. Therefore, compared with the traditional scheme, the throughput is
improved by 100%.
In the formula, s1 ðtÞ and s3 ðtÞ are the band communication Numbers sent by N1
and N3 within a symbol period, ai and bi (i = 1or3) are the information bits modulated
by QPSK. R received the communication number with the direction of R is a pair of
signals that are orthogonal to each other, namely ða1 þ a3 ÞcosðxtÞ and
ðb1 þ b3 ÞsinðxtÞ. QPSK signal is regarded as two orthogonal BPSK signals, the same
analysis is done for the homologous component and the orthogonal component.
Let the sending bit data be mi , The range of bit data of source node S1 and S2 in the
network layer corresponding to the modulated electromagnetic wave signal is
ei 2 f1; 1g, its modulation mapping mode is:ei ¼ 2mi 1, electromagnetic waves
are sent by the physical layers of the two nodes. That is, in the case of complete
synchronization, the physical layer of relay node receives the superposition of the
electromagnetic wave amplitude of the signal sent by two signal source nodes, whose
range is ðe1 þ e2 Þ 2 f2; 0; 2g, equivalent to the specific or value directly carried out at
the network layer, the network coding function at the physical layer is realized.
The relay node modulates the mapped signal and sends it, the two source nodes are
demodulated after receiving, then the encoded bit data m ¼ m1 m2 is geted,
according to the signals sent respectively m1 and m2 , S1 and S2 get the messages sent
by the other party through XOR operation.
A Network Coding Against Wiretapping Attacks 433
Node S1
QPSK N Node R
LDPC encode
ModulaƟon
SoŌ decision
decoding
PNC map decode
QPSK
LDPC encode
ModulaƟon
Node S2
Fig. 4. Block diagram of physical layer network and LDPC co-coding system
In Fig. 4, the transmission of the entire system consists of two stages: Multiple
access phases (MAC) and broadcast phases (BC). At the multi-access stage, the source
nodes S1 and S2 simultaneously send coded modulated signals to the relay R, but the
relay R receives signals from S1 and S2 and mixed signals with noise; During the
broadcast phase, relay R simultaneously sends network-coded messages to nodes S1
and S2.
Legimate
Sender X main channel Y user
Encoder Decoder
(Gaussian noise)
Hacking channel
(Gaussian noise)
Eavesdropper Z
yn ¼ xn þ vb ; z n ¼ xn þ ve ð2Þ
yn ¼ G u þ vb ; z n ¼ G u þ ve ð3Þ
Where W is the confidential message being sent; H(W) is the information contained
in the source; Signal through the channel transmission with noise to eavesdroppers,
source of information loss is HðWjZ n Þ.
Wyner used conditional entropy ðHðWjZ n ÞÞ to represent the eavesdropper’s
uncertainty about confidential information. In practical communications, it is difficult to
calculate the uncertainty of the eavesdropper. Therefore, the error rate of eavesdropper
is approximately defined as equivalent instead of uncertainty. It’s not hard to prove
from the definition of information entropy that the code error rate of the eavesdropper is
0.5 when the uncertainty is maximum. That is, the decoding error rate of the eaves-
dropper is equal to 0.5. Wyner showed that as long as the quality of the main channel is
better than that of the eavesdropping channel, then a coding scheme that meets this
requirement can always be found. but the encoding method that meets the requirement
has a transmission upper limit, namely secret capacity. I. Csiszar and S.K.L eung’s
study shows that the secret channel capacity of gaussian channel is the difference
between the capacity of legal channel and the capacity of eavesdropping channel. that
is, the upper bound of the channel secret transmission rate. Therefore, the security
capacity is the maximum transmission efficiency when the eavesdropper has the
greatest uncertainty. The reduction of the minimum bit error rate (the increase of
reliability) is achieved by adding redundancy, this must reduce the transmission rate,
confidentiality conditions try to limit this redundancy. However, reliability and con-
fidentiality can be well compromised, and the measurement index is channel confi-
dentiality capacity.
A Network Coding Against Wiretapping Attacks 435
l-k k n+k-l
d s p
Fig. 6. The code words
code book is designed, each message corresponds to the sub code book one by one, and
the random number generator is used to generate the random code word that grows into
l − k.
The LDPC security coding scheme based on AWGN channel is as follows:
In the verification matrix H, element “0” accounts for the majority, while element
“1” accounts for the rest, which is a sparse construction matrix. The check matrix is
designed as system code, which is divided into the information part matrix H1 and the
parity part matrix H2 . The information part constructs the cyclic replacement matrix
based on PEG algorithm. the calibration part adopts the structure of quasi dual-
diagonal, thus, the constructed LDPC code has the characteristics of low complexity
and fast coding. The structure of H is as follows:
H ¼ ½H1 H2 ð5Þ
Ix represents the unit matrix of P P that moves x bits to the right, where
x 2 ð0; 1; . . .; p 1Þ. X equals 0 is the identity matrix I.
H2 is constructed as follows:
2 3
I I 0 0 0
6 .. 7
6 0 I . 0 07
6 .. .. 7
6 Ix 0 I . . 7
6 7
H2 ¼ 6 .. 7 ð7Þ
6 . I I 07
6 0 7
6 .. 0 I I 7
4 . 0 5
0 0 I
I 0
Where, I and 0 respectively represent the identity matrix and zero matrix of P P.
Because H2 is the structure of bidiagonal, it satisfies the nonsingularity. And it can be
used directly for simple and fast coding.
When the information bits are multiplied by the generated matrix, the code word C
encoded by LDPC can be obtained, which is denoted as C = (d, s, p). By the definition
of check matrix H C T ¼ 0:
A Network Coding Against Wiretapping Attacks 437
2 3
dT
H C T ¼ ½ H1 H2 4 sT 5 ¼ 0 ð8Þ
dpT
The actual transmission efficiency of sub codes is (l-k)/n. For legitimate users, the
actual transmission efficiency of LDPC code is l/(n + k), it is less than the channel
capacity, therefore, the legitimate user can simultaneously translate the real message s
and the randomly generated message d with the decoding error probability approaching
0. For the eavesdropper, it is hoped that all the decoding ability can be used for the
correct sub code translation (d, p). For the sub code (d, p), where the message is d, the
transmission efficiency of the sub code ðl kÞ=n ¼ CðSNR2 Þ is required. So
Therefore, for the eavesdropper, the actual transmission efficiency of the code word
is greater than the capacity of the eavesdropper channel. According to Shannon’s
theorem, the probability of eavesdropper decoding error cannot approach 0.
Finally, the classical Belief Propagation (BP) decoding algorithm is used to com-
plete the LDPC coding algorithm.
To verify the effect of this method, MATLAB was used to simulate LDPC code in the
gaussian channel. The security network coding scheme designed by the eavesdropping
channel model should meet the requirements of any small bit error rate of the legitimate
receiver (approximately 0) and any large bit error rate of the eavesdropper (approxi-
mately 0.5). Setting the number of iterations to 5, constructed the matrix H, so that
there are 8 cyclic permutations in each row, and 4 cyclic permutations in each column,
the order p of the cyclic replacement matrix is 127, so the matrix has n þ k l ¼ 508
rows and n þ k = 1016 columns. During the simulation, a 127-bit real message and a
random message with l k ¼ 381 bits were generated at a time. The code word of 508
438 Y. Zheng and J. Fu
bits message bits and 508 bits check bits was then encoded with LDPC, and the code
length is 1016 bits.
After generating a 127-bit binary number from the binary random sequence gen-
erator, this binary number was received and encoded by the LDPC coding function
module, after coding, the data obtained was modulated with QPSK modulation module,
and then the modulated data entered the AWGN channel module to generate noise
signal, the simulation of the signal generation and transmission process was completed.
Then, QPSK demodulation module was used to demodulate the noise signal at the
receiving end, the receiver transcoded the demodulated signal data through the BP
decoding module, the signal from the transmission to the reception of the entire process
was completed. Figure 7 is the simulation curve of the signal to noise ratio (SNR) and
bit error rate of the main channel (The LDPC code constructed in this paper is com-
pared with PEG algorithm under the same code length and the same parameters).
Fig. 7. The relationship between the main channel error rate and SNR
As can be seen from Fig. 7, with the increase of the main channel SNR, the main
channel error rate decreases. When the SNR is less than 1.5 dB, the encoding scheme
presented in this paper is like the error code performance of PEG algorithm; with the
increase of SNR, the code word performance constructed in this paper is better than
that of PEG code, at the BER of 105 , the codes get an improvement of 0.2 dB
compared with the PEG algorithm.
To study the relationship between the SNR and BER of the eavesdropping channel,
if the SNR of the main channel is fixed, and the SNR of the eavesdropping channel is
variable, the SNR1 of the main channel is taken as 14, SNR2 takes 15 different values
(0.5, 0.4, 0.3, 0.2, 0.1, 0.075, 0.05, 0.02, 0.01, 0.0085, 0.005, 0.0035, 0.002, 0.001).
The simulation curves of SNR and BER of gaussian channel are obtained:
Bit rate l=ðn þ kÞ ¼ 1=2 \CðSNR1 Þ ¼ 0:5log2 ð1 þ SNR1 Þ 2, at this point, the
LDPC code’s actual transmission efficiency is far less than the channel capacity of
main channel. As shown in Fig. 8, the lower the SNR of the eavesdropping channel,
the higher the bit error rate of the eavesdropper is. That is, the closer the error prob-
ability of the eavesdropper is to 0.5, the more uncertain the eavesdropper is, and the
more secure the security encoding scheme is. Moreover, when the SNR of the
A Network Coding Against Wiretapping Attacks 439
eavesdropping channel increases, the code rate should be reduced, because the main
channel SNR is fixed, the security capacity decreases with the increase of the SNR2 .
Therefore, when the eavesdropper error rate reaches the maximum, the code rate can be
increased to make the transmission rate close to the secret capacity of the channel.
Based on channel security, the maximum speed transmission can be realized.
Fig. 8. The relationship between the BER and SNR of the eavesdropping channel
5 Conclusion
This paper mainly introduces the construction of LDPC code verification matrix and
the realization process of security coding, quasi dual-diagonal matrix reduces the
coding complexity, the structured unit circulant replacement matrix effectively reduces
the memory space, unrestricted column weight improves LDPC code decoding per-
formance. Finally, a simple verification simulation experiment was conducted for
LDPC code through MATLAB, under the AWGN and BP decoding algorithm, the
simulation results show that the proposed LDPC codes get an improvement of 0.2 dB
compared with the PEG algorithm at the BER of 106 , under the same BER condition,
this method is closer to zero. In the eavesdropping channel, as the SNR of the
eavesdropping channel decreases, the bit error rate approaches 0.5. This method can
effectively prevent eavesdropping.
Acknowledgments. This work was financially supported by the Science and Technology
Commission of Shanghai Municipality of China under Grant (No. 17511107002).
References
1. Shannon, C.E.: Communication theory of secrecy systems.In: International Conference on
Autonomous Agents and Multiagent Systems. International Foundation for Autonomous
Agents and Multiagent Systems, pp.1951–1952 (2015)
2. Wyner, A.D.: The wire-tap channel. Bell Labs Tech. J. 54(8), 1355–1387 (2014)
3. Csiszar, I., Korner, J.: Broadcast channels with confidential messages. IEEE Trans. Inf.
Theory 24(3), 339–348 (2003)
440 Y. Zheng and J. Fu
4. Zhang, S., Liew, S.C., Lam, P.P.: Hot topic: physical-layer network coding. In: International
Conference on Mobile Computing and Networking, MOBICOM 2006, Los Angeles, Ca,
USA, pp. 358–365, September 2006
5. Ahlswede, R., Cai, N., Li, S.Y.R., et al.: Network information flow. IEEE Trans. Inf. Theory
46(4), 1204–1216 (2002)
6. Cai, N., Yeung, R.W.: Secure network coding. In: 2002 Proceedings of the IEEE
International Symposium on Information Theory, p. 323. IEEE (2008)
7. Wang, J., Wang, W.: Probability Theory and Mathematical Statistics. Advanced Education
Press, Beijing (2015)
8. Mahdavifar, H., Vardy, A.: Achieving the secrecy capacity of wiretap channels using polar
codes. IEEE Trans. Inf. Theory 57(10), 6428–6443 (2011)
9. Dorf, R., Simon, M., Milstein, L., et al.: Digital communication. IEEE Trans. Acoust.
Speech Signal Process. 32(1), 190 (2017)
10. Liu, L., Yan, Y., Ling, C.: Achieving secrecy capacity of the gaussian wiretap channel with
polar lattices. IEEE Trans. Inf. Theory 64(3), 1647–1665 (2018)
11. Hu, X.Y., Eleftheriou, E., Arnold, D.M.: Regular and irregular progressive edge-growth
tanner graphs. IEEE Trans. Inf. Theory 51(1), 386–398 (2005)
Neural Networks and Deep Learning
A Weighted KNN Algorithm Based
on Entropy Method
1 Introduction
K-Nearest Neighbor (KNN) algorithm, as the supervised learning method, has been a
relatively mature classification method in theory and is commonly used in the classi-
fication of machine learning and data mining classification. The classification of the
KNN algorithm is realized by the training sample k, an arbitrary number greater than 3,
that is closest to the classification object. The spatial distance between the samples
generally adopts the Euclidean distance measurement method. All the samples are
classified into the same category when least three samples belong to the category [1].
KNN algorithm relies on training samples, one of the commonly used classification
algorithms, becoming widely used in classification problems owning to its simplifi-
cation, high efficiency, avoiding training and establishing the predictive model [2–6].
There is an inductive bias in KNN [7], and the classification results may be inaccurate
due to the uneven distribution of samples. The optimizing of KNN algorithm have
developed three angles, such as feature attribute pruning, rough set, improved distance
formula. The clustering, principal component analysis, etc., as the common ways of the
feature attribute pruning, have the disadvantage of possibly cutting too much in the
process of pruning. The disadvantage of rough set classification is that it is sensitive to
noise data. The improved distance formula is a common direction of improvement. The
traditional KNN algorithm adopts Euclidean distance. Some scholars have proposed
using chi-square distance and Markov distance instead of Euclidean distance calcula-
tion. These improved ideas are based on the traditional KNN algorithm to improve the
classification accuracy.
The high dimension of the data is one of the reasons for the low classification
accuracy, some attributes in the attribute dimension are of little reference value to the
classification results, this part of the noise data reduces the accuracy of the classifi-
cation. When the traditional KNN algorithm calculates the similarity between samples
by various distance formulas, it will appear that the feature index with large data
difference has a great influence on the classification result, and the weight of the
classification feature index has not been considered. Therefore, considering the weight
of feature index in the classification of the algorithm is an important direction of the
improvement of the KNN algorithm. Some scholars have improved the algorithm in
this direction. For example, a KNN algorithm based on particle swarm optimization
(PSO) weighting is proposed, by using PSO-KNN algorithm and traditional KNN
algorithm, GA weighted KNN algorithm are used to classify the sample data, which
verifies that PSO-KNN algorithm is superior to traditional KNN algorithm and GA
weighted KNN algorithm [8]. Using attribute correlation to calculate weight improves
the nearest neighbor algorithm of K [9]. In reference [10], the traditional KNN algo-
rithm only takes into account the distance between the same attributes, neglecting the
correlation between the attributes, which leads to the problem of low classification
accuracy, thus, the KNN algorithm of feature entropy correlation difference is pro-
posed. A KNN algorithm for information entropy of attribute values is introduced in
reference [11], which based on the sample with the same maximum attribute value
between the samples to be classified and training samples to classify the samples to be
classified, suitable for the classification of non-numerical attribute values. Literature
[12] introduces the sparse learning theory and proposes a K-valued adaptive SA-KNN
algorithm. In reference [13], an improved algorithm for KNN algorithm is proposed,
which uses the combined feature entropy to calculate the information entropy, form a
new feature attribute, and then classify the samples. In order to solve the problem of
low classification accuracy brought about by multidimensional information, using
information entropy to reduce the dimension of the feature attribute and using the
attribute with large contribution degree to classify the KNN algorithm in reference [14].
A KNN algorithm based on attribute weighted entropy is proposed in reference [15],
the method of normalized entropy value is used to calculate the weight of characteristic
attributes, and the factor of attribute weight is added to calculate the distance between
samples.
To sum up, most of the KNN optimization algorithms which join weights are based
on entropy method to give weights to attributes. Entropy method is the most primitive
entropy nondimensionalization standard, which is the normalization method, and the
interval stability and translation independence of this method are weak. In the nondi-
mensionalization method of the entropy method, the extreme value processing method
is the best [16]. Therefore, based on the existing research, the KNN algorithm (EEM-
KNN algorithm) based on the extreme value entropy method is proposed, and the
EEM-KNN classification is used to classify 4 data sets in the Jupyter Notebook
A Weighted KNN Algorithm Based on Entropy Method 445
interactive WEB end based on Python language, to test the rationality of the algorithm.
The improved algorithm is superior to the traditional KNN algorithm in improving the
classification accuracy.
2 KNN Algorithm
The KNN algorithm is a case-based learning algorithm. Given a set of local query
vector sets and a set of feature vector sets, the algorithm searches K nearest local
feature vectors. The working mechanism of the KNN algorithm is to first calculate all
the distances between the query vector and all database vectors, then sort the calculated
distances, and select the smallest K reference amounts from the ranking to determine
the classification of the query vectors [17].
3 EEM-KNN Algorithm
(1) tij indicates the proportion of the i-th sample data under the j-th index to the total
sample under the j-index
aij mj
tij ¼ Pn ð2Þ
j¼1 ðaij mj Þ
X
n
hj ¼ tij ln tij ð3Þ
j¼1
ej ¼ khj ð4Þ
1
k¼
ln m
gj ¼ 1 e j ð5Þ
The entropy method can effectively avoid the influence of subjective randomness
on the importance of the index. It depends on the sample data and has a strong
theoretical basis. The performance of extreme value method in the nondimensional-
ization method of entropy value method is the best. Therefore, this paper uses the
extreme entropy method to give weight to the sample characteristic index [17].
(4) Sort the calculated distance and select k training samples nearest to the sample to
be classified.
(5) According to the category of k selected training samples, the largest sample of k
samples is used as the category of samples to be classified.
(6) According to the classification results of the samples to be classified and com-
pared with the actual category, the accuracy of the classifier is calculated. The
formula for calculating the accuracy of the classifier is as follows:
B
p¼ 100% ð8Þ
C
In the formula: B is the correct number of test samples and C is the total number of
test samples.
4 Example Verification
In this paper, the improved algorithm is verified by data, and the KNN classifier is
written with python on the Jupyter Notebook interactive WEB. The data set is imported
into the classifier, and different K values are selected to verify the data. The whole
verification process is divided into two parts. In the first part, the original KNN
algorithm is used to verify the data. In the second part, the improved KNN algorithm is
used to compare the accuracy of different K values with the results of the first
part. Finally, the experimental results are analyzed.
Table 2. Test accuracy of different data sets for different K values (%)
Data sets Algorithm K=3 K=5 K=8
Tobacco leaves KNN 92.1594 93.3741 93.2211
EEW-KNN 95.0244 96.6005 94.246
Iris KNN 95.6733 95.456 95.2647
EEW-KNN 96.6321 96.01 95.923
Wine KNN 73.1842 73.7775 72.8113
EEW-KNN 90.9656 93.6511 90.425
Glass-identification KNN 70.9336 68.4688 69.3213
EEW-KNN 82.3282 88.6889 80.0948
It can be seen from Table 2 that in the data set test of the original KNN algorithm,
the classification accuracy of the data sets numbered 3 and 4 is low, mostly around
70%. while for data set 1 and data set 2, the classification accuracy is above 90%. From
the classification results of the two classification algorithms, it can be seen that the
classification accuracy of the four data sets is improved by using the improved KNN
classification algorithm, and the changes in the wine data set are most obviously. The
largest increase of accuracy is the test in the Glass-Identification data set when K = 5.
There is a low accuracy of classification owning to more sample features contained
in the data sets 3 and 4, leading to strong interference feature data. In order to improve
the accuracy of classification, low weight is given to the feature data with poor overall
performance to reduce their impact on the classification results when improved KNN
algorithm is adopted.
In order to evaluate the classification effect of the improved KNN algorithm more
intuitively, a broken line graph of the classification accuracy with different K values
was plotted. When K = 3, the accuracy rate result is shown in Fig. 1. When K = 5, the
result is shown in Fig. 2, and when K = 8, the result is shown in Fig. 3.
It can be seen from Figs. 1, 2 and 3 that the classification accuracy of the improved
algorithm is greatly improved no matter how much K value is taken by adding the
eigenvector weight of entropy method. It can be seen from the Fig. 1 that the KNN
algorithm with Euclidean distance is more effective in the classification of data set Iris.
When K = 3, the classification accuracy is the highest. For data set 1, the classification
accuracy is the highest when K = 5. Figures 1, 2 and 3 show that the stability of the
improved algorithm in the data training process is not very good, and the accuracy of
the algorithm varies greatly.
A Weighted KNN Algorithm Based on Entropy Method 449
65 Glass-IdenƟficaƟon KNN
60 Glass-IdenƟficaƟon EEW-KNN
1 2 3 4 5 6 7 8 9 10
55 Glass-IdenƟficaƟon EEW-KNN
1 2 3 4 5 6 7 8 9 10
65 Glass-IdenƟficaƟon KNN
60 Glass-IdenƟficaƟon EEW-KNN
1 2 3 4 5 6 7 8 9 10
5 Conclusion
References
1. Li, L., Weinberg, C.R., Darden, T.A., et al.: Gene selection for sample classification based
on gene expression data: study of sensitivity to choice of parameters of the GA/KNN
method. Bioinformatics 17(12), 1131 (2001)
2. Beaudoin, A., Bernierp, Y., Guindon, L., et al.: Mapping attributes of Canada’s forests at
moderate resolution through. Can. J. Forest Res. 44(5), 521–532 (2014)
3. Pandya, D.H., Upadhyay, S.H., Harsha, S.P.: Fault diagnosis of rolling element bearing with
intrinsic mode function of acoustic emission data using APF-KNN. Expert Syst. Appl. 40
(10), 4137–4145 (2013)
4. Aslam, M.W., Zhu, Z., Nandi, A.K.: Automatic modulation classification using combination
of genetic programming and KNN. IEEE Trans. Wirel. Commun. 11(8), 2742–2750 (2012)
A Weighted KNN Algorithm Based on Entropy Method 451
5. Steenwijk, M.D., Pouwels, P.J., Daams, M., et al.: Accurate white matter lesion
segmentation by k nearest neighbor classification with tissue type priors (kNN-TTPs).
Neuroimage Clin. 3(3), 462–469 (2013)
6. Homaeinezhad, M.R., Atyabi, S.A., Tavakkoli, E., et al.: ECG arrhythmia recognition via a
neuro-SVM-KNN hybrid classifier with virtual QRS image-based geometrical features.
Expert Syst. Appl. 39(2), 2047–2058 (2012)
7. Tan, S.: An effective refinement strategy for KNN text classifier. Expert Syst. Appl. 30(2),
290–298 (2006)
8. Ren, J.T., Zhuo, X.L., Xu, S.C.: PSO based feature weighting algorithm for KNN. Comput.
Sci. 34(5), 187–189 (2007)
9. Wang, Z.; Wang, K.: Improve KNN algorithm based on entropy method. Comput. Eng.
Appl., 45, 129–131,160 (2009)
10. Zhou, J., Liu, J.: KNN algorithm based on feature entropy correlation difference. Comput.
Eng. 37(17), 146–148 (2011)
11. Tong, X., Zhou, Z.: Enhancement of K-nearest neighbor algorithm based on information
entropy of attribute value. Comput. Eng. Appl. 46(3), 115–117 (2010)
12. Sun, K., Gong, Y., Deng, Z.: An efficient K-value adaptive SA-KNN algorithm. Comput.
Eng. Sci. 37(10), 1965–1970 (2015)
13. Zhou, J., Liu, J.: Improved K-nearest neighbor algorithm for feature union entropy. Comput.
Appl., 31(7), 1785–1788,1792 (2011)
14. Wei, L., Zhao, X., Zhou, X.: An enhanced entropy-K-nearest neighbor algorithm based on
attribute reduction. In: 2014 4th International Conference on Computer Engineering and
Networks, pp 381–388 (2014)
15. Xiao, X., Ding, H.: Enhancement of K-nearest neighbor algorithm based on weighted
entropy of attribute value. In: 2012 5th International Conference on BioMedical Engineering
and Informatics, Chongqing, pp. 1261–1264 (2012)
16. Zhu, X., Wei, G.: Discussion on the non-dimensional method’s good standard in entropy
method. Stat. Decis. 02, 12–15 (2015)
17. Mejdoub, M., Amar, C.B.: Classification improvement of local feature vectors over the KNN
algorithm. Multimed. Tools Appl. 64, 197–218 (2013)
18. Zou, Z.H., Yun, Y., Sun, J.N.: Entropy method for determination of weight of evaluating
indicators in fuzzy synthetic evaluation for water quality assessment. J. Environ. Sci. 18(5),
1020–1023 (2006)
Control Strategy and Simulation for a Class
of Nonlinear Discrete Systems with Neural
Network
Peng Liu(&)
1 Introduction
Due to its robustness and simplicity, the PID controller has been widely used in
industry. For ordinary control processes, they show satisfactory control results. How-
ever, due to the unpredictability of the dynamic process, the controller parameters are
not adjusted in time, which will lead to poor control performance. Many of the changes
in parameters are caused by nonlinearity. Since 2010s, using artificial intelligence to
study nonlinear adaptive PID control has become a research hotspot. The use of neural
networks has been shown to be successful in identifying and controlling complex
nonlinear dynamics systems [4, 5]. In particular, a new PID adaptive control scheme is
proposed for a nonlinear system [6]. The model discussed is a class of nonlinear and
time-varying systems. Therefore, the PID control is applied to the dominant model to
achieve satisfactory control effect. The neural network uses radial basis function net-
work. Documentation [7–10] Several different BP learning algorithms are proposed,
such as an online neural network learning algorithm for time varying processing, fast
learning algorithm based on neuronal spatial gradient decline and LM algorithm.
The overall arrangement of this article is as follows: In Sect. 2, the control system
is expounded. In Sect. 3, we proposed the learning algorithm. The design and the rules
for the PID tuning are explained. Simulation result and conclusions are given in
Sects. 4 and 5, respectively.
2 Problem Statement
which u, y are the control and the output, respectively, n is non-negative integer
(n m), m is non-negative integer, g is an unknown nonlinear function. Assume that
the output rðjÞ is single boundary.
The purpose of the control problem is to determine a bounded control u so that the
system output y can tracks the required value r.
Denotes the tracking error between the required value r and the system output y as
where kp , ki and kd are the parameters which are tuned by the BPNN algorithm with
optimal learning rate.
In the proposed methodology, the adjustment criterion for the weights w of the neural
network is the negative gradient of the variance performance index function. The
variance performance index of systems (1) be defined as:
1
EðjÞ ¼ ½rðj þ 1Þ yðj þ 1Þ2
2 ð4Þ
1 2
¼ e ðj þ 1Þ
2
The numbers of the wki and vlk are arbitrary. The vlk has the input regression vector,
composed of the desired input, the past states of the plant output and the control signal.
Three neurons in the output layer are jp ; ji ; and jd . The structure of NN is shown in
Fig. 1.
The NN weights are computed by the following equation
@E
Dwlk ðjÞ ¼ r ð5Þ
@wlk
454 P. Liu
@E
Dwlk ðjÞ ¼ r þ aDwlk ðj 1Þ ð6Þ
@wlk
where 0 a\1 is the momentum constant to increase the learning speed. It controls
the feedback loop acting around Dwlk ðj 1Þ.
NN
PID Pl
where netl is the total input of the neuron k; wlk ðjÞ is the neuron weights between the l
and k.
The actual Dwki ðjÞ changes from the hidden node k to the input node i are defined as
where outk is the output of the k; wki ðjÞ is the neuron weights between the hidden
neuron k and the input neuron i. f is the sigmoid function can be defined by
1
f ðnetk Þ ¼ ð12Þ
1 þ expðnetk Þ
P
Step4: Give some positive real constant e. If eðjÞ2 \e, then stop and output uðjÞ
j
in (3), else go to step 3.
Step5: Letting j ¼ j þ 1, go to step 2.
4 Simulation
Next, we use the MATLAB to simulate the above algorithm. In this example, the
nonlinear system is described by
5yðjÞ
yðj þ 1Þ ¼ þ u3 ðjÞ ð13Þ
2:5 þ yðjÞ
The desired values are sinð0:25pjÞ for 0 j 40 and sinð0:5pjÞ for 40\j 80.
Each simulation, using 0 to 80 steps as a simulation cycle. NN consists of 1 middle
layer with 10 nodes, and 3 output nodes. Which r ¼ 0:1, a ¼ 0:4.
The desired value figure, the error between the system outputs and expected values
are shown in Fig. 3, respectively.
From Fig. 3 it can be seen that the plant outputs is perfectly coincident with the
desired values.
456 P. Liu
Fig. 3. The system outputs (solid lines) and the expected values (dashed lines). (a) The desired
value, (b) The plant outputs, (c) The error between the system outputs and expected values
Control Strategy and Simulation for a Class 457
5 Conclusion
Neural Network PID control of nonlinear systems using a fast learning algorithm is
present. By comparing the error of the system output and the expected value, the
weight of NN is reversed. In the simulation, the output of the system quickly finds the
expected value, so the effectiveness of the learning algorithm is proved.
References
1. Al-Assadi, S.A.K., Al-Chalabi, L.A.M.: Optimal gain for proportional integral derivation
feedback. IEEE Control Syst. Mag. 7(2), 16–19 (2007)
2. Hwang, C., Hsiao, C.-Y.: Solution of a non-convex optimization arising in PI/PID control
design. Automatica 38(6), 1895–1904 (2012)
3. Daley, S., Li, G.P.: Optimal PID tuning using direct search algorithms. Comput. Control
Eng. J. 10(3), 51–56 (2009)
4. Juditsky, A., Hjalmarsson, H., Benveniste, A., Delyon, B., Ljung, L.: Nonlinear black-box
models in system identification. Mathematical foundations. Automatica 31(3), 1725–1750
(2015)
5. Narendra, K.S.: Neural networks for control: theory and practice: mathematical foundations.
Proc. IEEE 84(1), 1385–1406 (2016)
6. Huang, S.N., Tan, K.K., Lee, T.H.: A combined PID/adaptive controller for a class of
nonlinear systems. Automatica 37(6), 611–618 (2011)
7. Zhao, Y.: On-line neural network learning algorithm with exponential convergence rate.
Electron. Lett. 32(1), 1381–1382 (2016)
8. Zhou, G., Si, J.: Advanced neural network training algorithm with reduced complexity based
on Jacobian deficiency. IEEE Trans. Neural Netw. 9(3), 448–453 (2017)
9. Parisi, R., Di Claudio, E.D., Orlandi, G., Rao, B.D.: A generalized learning paradigm
exploiting the structure of feedforward neural networks. IEEE Trans. Neural Netw. 7(2),
1450–1459 (2016)
10. Hagan, M.T., Menhaj, M.B.: Training feedforward neural networks with the Marquardt
algorithm. IEEE Trans. Neural Netw. 5(1), 95–99 (2014)
11. Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning representations of back-propagation
errors. Nature 323(2), 533–536 (2016)
Research on Joint Nondestructive Testing
Based on Neural Network
1 Introduction
Material hardness is one of the key factors influencing the use of materials. The
traditional method for measuring the surface hardness of materials is to use the
indentation test method to calculate the surface hardness value by measuring the size of
the indentation. However, the indentation test method is a destructive test method that
will cause irreversible damage to parts and test blocks and affect subsequent use. With
the continuous emergence of related advanced technologies and instruments, non-
destructive testing technology has gradually been applied to the characterization and
evaluation of the mechanical properties of materials [1], avoiding the irreversible
damage to materials. Conventional nondestructive testing methods including: ultra-
sonic testing, magnetic particle testing, eddy current testing, penetration testing, and ray
testing. In addition, other nondestructive testing methods such as incremental perme-
ability method, metal magnetic memory method, Barkhausen noise method, etc. are
also used.
Ultrasonic non-destructive testing is the most widely used, most frequently used,
and most studied non-destructive testing technology at the present stage [2]. It is widely
used in product quality evaluation, defect evaluation, and performance monitoring [3].
Some scholars also use this method to qualitatively or quantitatively evaluate the
surface hardness of materials. Magnetic Barkhausen Noise (MBN) is also a widely
used non-destructive testing method that can be applied to the detection of mechanical
properties of ferromagnetic materials and used as an internal organizational structure
criterion [4].
There are many influencing factors in the non-destructive testing process, and the
detection results and the required parameters may have a nonlinear mapping rela-
tionship [5]. The use of traditional methods is more complicated and not effective [6].
Due to its excellent feature extraction capability, self-learning ability, and high fault-
tolerance capability, deep neural networks can achieve tasks that cannot be quickly
completed by many conventional methods [5–10]. Through deep neural networks, data
fusion, performance prediction, and intelligent identification of parameters can be
performed [11–17].
In this paper, ultrasonic non-destructive testing method and Barkhausen noise
method were used to detect 45 steel specimens under different heat treatment condi-
tions, and a deep neural network was used to establish a complex mapping relationship
between the two signals and surface hardness, and the accuracy of joint detection was
verified.
The Barkhausen detection process was carried out using a portable Barkhausen
detector developed by Nanjing University of Aeronautics and Astronautics and a self-
made probe.
0:204F
HBW ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ; ð1Þ
pDðD D2 d 2 Þ
Among them, D is the alloy ball diameter, d is the indentation diameter, and F is the
experimental force (take 1000 kgf).
2h
v¼ ; ð2Þ
s
Where v is the speed of ultrasonic sound, h is the thickness of the sample, and t is
the sound time.
Fig. 3. The topology of the network. The nodes are fully connected with other notes in the
adjacent layer.
Before the network training, the two input and output data are Z-score normalized.
Randomly increase the Gaussian noise perturbation in the six groups of measured data,
increase the number of samples, and make the network have better robustness and
generalization ability.
462 J. Tan et al.
Due to the unevenness of the heat-treated test blocks, the Brinell hardness, ultrasonic
velocity, and Barkhausen noise signals measured at different test points on the surface
will also be different. For the entire multidimensional mapping calibration process, the
network training is more robust because it provides more noise-containing data in
different heat treatment states, making the trained characterization network have better
generalization capabilities. In the experiment, it is assumed that the thickness of each
test piece (line cut) measured by the vernier caliper is the same. The following table
shows the thickness measurement results (Table 1):
Table 1. The hardness of test blocks under six different heat treatment conditions.
Groups A N 600 °C 400 °C 200 °C WQ
Thickness/mm 18.95 18.97 18.98 19.00 18.96 18.97
For the measurement of ultrasonic speed, Barkhausen noise signal, and Brinell
hardness, 200 points were selected for each test block, a total of 1200 sets of data.
600
500
400
Hardness/HBW
300
200
100
600
500
400
Hardness/HBW
300
200
100
0 20 40 60 80 100
barkhausen noise/RMS
The root mean square of the Barkhausen noise signal measured by the verification
block is 19.8, and the hardness calculated by the function is 364.6 HBW with an error
of 15.45%.
464 J. Tan et al.
500
400
Hardness/HBW
300
200
100 100
5780 80 S
Ult 58005820 60 M
ras /R
oni 58405860 40 ise
cv
elo 20 no
/(m city 58805900 n
/s) se
5920
0
hau
r k
Ba
The above figure shows the three-dimensional scatter distribution of three param-
eters. The left and right planar projections have the same distribution as a single non-
destructive testing method.
Through the deep neural network, the input is the ultrasonic detection signal and
the Barkhausen detection signal, the output is the surface hardness, and the amount of
data is expanded by ten times to 12,000 training data using Gaussian noise. After 5000
iterations of training, the network converged. The multi-dimensional mapping curve
established should be on the surface shown in the figure below (Fig. 7).
In fact, the size of the ultrasonic wave and the magnetic domain distribution that
causes the magnetic Barkhausen signal are related to the microstructure of the material.
There is a correlation between the two detection signals. When considering the cor-
relation according to the actual detection data, the multidimensional mapping The
relationship curve is as shown in the Fig. 8.
The root mean square values of ultrasonic velocity and Barkhausen noise signal
measured by the verification block were normalized and brought into the calibration
Research on Joint Nondestructive Testing Based on Neural Network 465
depth network. The surface hardness was 333.7 HBW, the relative error was 6.96%,
and the accuracy was improved.
Multiple methods combined with non-destructive testing extract material properties
from multiple signals and increase the tolerance to test noise. This can not only
improve the detection accuracy, but also can avoid to some extent that some test pieces
are not sensitive to the change of the parameters of a certain method, which leads to the
466 J. Tan et al.
detection failure. In the future, more parameters will be jointly modeled through more
methods, and feature detection and deep-learning capabilities will be used to implement
joint detection in a broader sense.
4 Conclusion
The hardness of 45 steel specimens under six different heat treatment conditions was
tested using the lossy method. The ultrasonic and Barkhausen noise methods were used
for the quantitative non-destructive evaluation of hardness. Through a large number of
experiments to obtain data, the ultrasonic-hardness, Barkhausen noise-hardness map-
ping relationship was calibrated, and the test block was used to evaluate the calibration
curve error. Afterwards, an ultrasonic-Barkhausen joint detection method based on
deep learning network was proposed and a joint detection network model was estab-
lished. After training, the network can accurately represent the hardness of 45 steel. The
experimental results show that the joint detection method to characterize the hardness
of 45 steel is feasible and has higher accuracy, and this joint detection method provides
a new direction for the comprehensive and intelligent evaluation of the mechanical
properties of materials in the future.
Acknowledgments. This work was supported by the National Key Research and Development
Program of China, no. 2016YFB1100205.
References
1. Xia, X.L., Zhang, Y.Y., Yang, H.X.: Development of mechanical fault diagnosis technology
under the condition of modern science and technology. J. Xuzhou Inst. Technol. 21(9), 67–
69 (2006). (in Chinese)
2. Cheng, S.B., Yao, J., Cai, P.: Application of ultrasonic measuring technology in the rigidity
measurement. Chin. Meas. Technol. 30(1), 12–13 (2004). (in Chinese)
3. Vasudevan, M., Palanichamy, P., Venkadesan, S.: A novel technique for characterizing
annealing behavior. Scr. Metall. Mater. 30(11), 1479–1483 (1994)
4. Puppin, Ezio: Statistical properties of Barkhausen noise in thin Fe films. Phys. Rev. Lett. 84
(23), 5415 (2000)
5. Chen, Y.H., Li, M.X.: Application of artificial neural network to ultrasonic nondestructive
testing. J. Appl. Acoust. 3, 40–44 (1996). (in Chinese)
6. Lorenz, M., Wielinga, T.S.: Ultrasonic characterization of defects in steel using multi-SAFT
imaging and neural networks. NDT and E Int. 26(3), 127–133 (1993)
7. Liu, W.J., Wang, X.M.: Study on the defect determination method of ultrasonic
nondestructive testing based on neural network. J. Dalian Univ. Technol. 5, 548–552
(1998). (in Chinese)
8. Yang, L.J., Ma, F.M., Gao, S.W.: Quantitative recognition of pipeline defects based on
neural network and data fusion. Nondestr. Test. 28(6), 281–284 (2006). (in Chinese)
9. Dai, G., Qiu, F., Chen, R.G., et al.: Artificial neural network intelligent evaluation method of
tank bottom corrosion status. Nondestr. Test. 34(6), 5–7 (2012). (in Chinese)
10. Wang, P., Zhu, L., Zhu, Q., et al.: An application of back propagation neural network for the
steel stress detection based on Barkhausen noise theory. NDT and E Int. 55(3), 9–14 (2013)
Research on Joint Nondestructive Testing Based on Neural Network 467
11. Haykin, S.: Neural Network and Learning Machines. Pearson, London (2008)
12. Arel, L., Rose, D.C., Karnowski, T.P.: Deep machine learning a new frontier in artificial
intelligence research. Comput. Intell. Mag. 5(4), 13–18 (2010)
13. Deng, L., Yu, D.: Deep Learning: Method and Application. China Machine Press, Beijing
(2015). (in Chinese)
14. Yu, D., Deng, L., Seide, F.T.B., et al.: Discriminative pretraining of deep neural networks.
US, US9235799 (2016)
15. Zheng, W.H., Wang, Z.H.: Ant Colony algorithm and application in inspection of concrete
structure defects. Nondestr. Test. 35(5), 4–7 (2013). (in Chinese)
16. Yan, W.X.: Research on deep learning and its application on the casting defects automatic
detection. South China University of Technology (2016). (in Chinese)
17. Lu, H.: Study on fault diagnosis of flexible manufacturing system based on neural network.
Chang’an University, Xi’an, China (2016). (in Chinese)
A New Real-Time FPGA-Based
Implementation of K-Means Clustering
for Images
1 Introduction
However, when real-time processing is required, for videos or for images with very
high resolution, achieving real time performance of the standard Lloyd’s algorithm
becomes a problem, because of the iterative processing of the complete image [5].
Thus, an alternative way of implementing K-Means clustering is needed for real-time
processing. The standard Lloyd’s algorithm is naturally parallel [6]. We can improve
the performance either by using different hardware (e.g. multiple CPUs, or a GPU, or
an FPGA), or by using a faster algorithm. In this paper, we propose a new imple-
mentation for implementing K-means clustering which achieves significant speed-up
both by algorithm innovation and by hardware implementation on an FPGA.
The FPGA architecture can achieve real time performance for video processing or for
large image resolutions. The algorithm is considerably faster than the standard Lloyd’s
algorithm. Indeed, the algorithm gives significant speed-up even on a CPU. The main
contributions of the paper are as follows:
(1) A new histogram-based implementation of K-means clustering. Our method gives
the same result as the standard Lloyd’s algorithm for image data, but is much
more efficient.
(2) An FPGA implementation on the Xilinx ZedBoard. Our method is 7.6 times faster
but uses 50% of the resources compared to the Lloyd’s FPGA implementation.
(3) A comparison of two different FPGA implementations: our new Histogram-based
K-means method, and the standard Lloyd’s K-means method, plus a performance
comparison with three FPGA implementations from the literature.
2 Related Work
A. Standard Lloyd’s Algorithm for K-means Clustering
The K-means problem is to find K clusters such that the means lc of each cluster
Clc satisfy Eq. (1).
K X
X
arg min kpx lc k2 ð0 c \ K Þ ð1Þ
Clc c¼1 px 2Clc
1 X
ltcþ 1 ¼
Clt pClc ð3Þ
c pi 2Clt
c
Since the algorithm needs to access the data set to calculate the means and distances
at each iteration, it is necessary to keep a copy of the original data set in memory
throughout the FPGA implementation.
B. FPGA Implementation of K-means Clustering
A number of researchers have already studied the FPGA implementation of K-
means clustering. In [7], Lavenier feeds the image into the distance calculation units,
and indexes every pixel in another array. Due to limited resources on the FPGA, the
image is stored in off-chip memory, and must then be streamed multiple times. This
entails extra communication overheads between the FPGA and off-chip memory.
Estlick et al. implemented K-Means in hardware using software/hardware co-design.
The distance calculation is computed in hardware and the rest of the algorithm is
completed on a CPU to avoid using a huge amount of hardware resources [8]. And in
[4], Yuk-Ming and Choi etc. use three FPGAs which work together to deal with the
large amount of data because of the limited resources on each FPGA. In [9], Benkrid
implemented a highly parameterized parallel system, and his implementation is not
dedicated to image clustering but shows the potential of using FPGAs to accelerate the
clustering process. Takashi, Tsutomu et al. implemented K-means on a Xilinx FPGA,
using KD-Tree, but they are limited by the resources on the FPGA, and they have to
store the image in an off-chip SRAM. They report a performance of 30FPS on an RGB
24-bit image [10, 11]. In [11], the approach replaces the Euclidean distance into the
multiply-less Manhattan distance and reduces the word length to save resources. The
approaches above focus on achieving greater efficiency by adjusting the word-length,
simplifying the distance calculation or using a heterogeneous platform.
Others try to reduce the amount of data. In [1], a filter algorithm is applied first
before K-means to reduce the amount of data. Tsutomu reduces the scanning data size
by doing partial K-means, whereby the algorithm is applied on 1/8, 1/4, and 1/2
sampled data. This simplifies the K-means algorithm to get better performance but also
introduces inaccuracy into the result [12].
A New Real-Time FPGA-Based Implementation of K-Means Clustering 471
We first note that the K-means problem for image processing is more constrained than
K-means for other kinds of data. The range of the input data is discrete, e.g. for an 8-bit
grayscale image, the grey levels are from 0 to 255.
Secondly, the clustering of pixels does not need to take the location of pixels into
account. Therefore, we can represent the input image by the image histogram (for the
purposes of clustering). The image histogram (which does not change throughout the
iterative clustering process) is effectively a compressed version of the image. For any
grey level value g, all pixels with that value can be clustered at the same time, with a
single calculation.
Thirdly, we can represent each cluster c merely as a sub-range of the histogram,
from column Bc to column B(c+1)−1 inclusive, where Bi is the lower bound of cluster i.
Fourthly, given the means of each cluster lc, the new bounds, which define the new
clusters, will be simply half way between the respective means. Figure 2 shows a
typical histogram and the representation of four clusters.
For an Nx Ny image, and doing L iterations, the complexity of the standard
Lloyd’s algorithm is Nx Ny L. The complexity of our new histogram-based
algorithm will be Nx Ny + 256 L (assuming 256 grey levels), where the Nx Ny
component is for calculating the histogram. For example, for a 512 512 image with
L = 20, the standard method will require around 5.2M steps, whereas our histogram-
based method will require around 0.267M steps – an improvement, in theory, of the
472 T. Deng et al.
order of 20x. In practice, there will be the additional overhead of image input and
perhaps reconstruction of the output image and the number of iterations will be image-
dependent. These factors will typically have the effect of reducing the apparent speed-
up.
For an 8-bit image, the histogram is a one-dimensional vector with 256 elements,
where each element in the histogram represents the number of pixels having that grey
level. After generating the histogram, bounds are needed to divide the histogram into
different clusters. Each cluster is defined by a pair of bounds, which define the part of
the histogram corresponding to the cluster. If we have K clusters, there will be K + 1
bounds, where B0 = 0, and BK = 256. Cluster c (0 c < K) is the part of the his-
togram from column Bc to B(c+1)-1 inclusive. The inner bounds (0 < c < K) can be
initialized in different ways (see later). Then, the means for the current clusters are
calculated as defined by Eq. (4).
BCP
þ 1 1
ð Hi iÞ
i¼BC
lc ¼ BCP
ð0\C\KÞ ð4Þ
þ 1 1
ð Hi Þ
i¼BC
Once the updated means are calculated, the next iteration first assigns the clusters
by merely updating the bounds. To assign a pixel to the correct new cluster, we have to
see which mean it is closest to. Consider a pixel with value g (say, 150) in an image
with a histogram such as in Fig. 2, and the means are [0.0, 27.5, 90.5, 165.0, 220.0,
256.0]. Column g will be part of the cluster whose mean is closest to g. We can achieve
this simply by moving the bounds so that each bound is half way between the two
respective means. The updating of the clusters is defined by Eq. (5).
A New Real-Time FPGA-Based Implementation of K-Means Clustering 473
ðlC1 þ lC Þ
BC ¼ ð0 C\K Þ ð5Þ
2
The iteration can then be completed by updating the means using Eq. (4).
We use integers to hold the bounds, so that they correspond to actual pixel grey
levels. In our FPGA implementation, we use simple fixed point to represent the means.
Termination of the iterative algorithm could be brought about under several con-
ditions. The total distance can be readily calculated as in the standard Lloyd’s algo-
rithm. However, in our preferred approach, the termination condition is that none of the
bounds are changed by the updating (Eq. (5)). This has the advantage that the total
(squared) distance does not need to be calculated. A third option is to set a fixed
number of iterations (say, 50), which can be sufficient for many applications in practice.
This number may be application–specific, and can be set at runtime.
There are several ways to initialize the K-means algorithm such as Random, or
MacQueen and Kaufman, as mentioned in [14]. Actually, in our histogram method, we
initialize the cluster bounds rather than the means. We provide two particular ways to
do this. The first is simply to divide the grey level range into K equal parts. The second
way is to divide all the pixels into K equally sized clusters. The latter can be achieved
by using the cumulative histogram and setting the bounds to the columns which have
the appropriate height.
Dividing the histogram range into K equal parts takes the least resources in a
hardware implementation. However, in an image with an unevenly distributed his-
togram, some clusters may be empty (for instance, if the image is overexposed or too
dark). The danger condition of calculating the mean of a zero-element cluster can be
overcome with a suitable condition; however, if we have two adjacent zero-element
clusters, the bounds would not move and those clusters would not grow. Thus, the
safest way to initialize the clusters is to set each cluster initially to have the same
number of pixels (approximately), using the cumulative histogram. We can build the
cumulative histogram from the histogram during initialization. A refinement which is
easy to implement is for the user to supply an initial, application-specific estimate of the
relative sizes of each cluster as a parameter.
4 System Implementation
transferred using the AXI-Stream interface. The means and bounds of each cluster are
held in a BRAM.
Our architecture is parameterized to provide some flexibility for the user. Our
K-means hardware can support any number of clusters up to a defined maximum (in
our case, 10 clusters). Within this limit, the actual number of clusters can be selected,
and changed during the run time by setting the control registers in the control bus.
Extending the maximum number requires re-synthesis the hardware and pay more
hardware resources.
The method for setting the initial value of the bounds, as mentioned above, can be
selected at runtime via a control register as well.
For comparison purposes, we run both the Lloyd’s and our K-means at both
100 MHz and 150 MHz. The ARM output 100/150 MHz clock is divided into
50 MHz and 25 MHz. 50 MHz is used in the camera and display, and 25 MHz is used
in the data transfer through the FIFOs and frame buffers. The FIFOs are used to help
synchronise data across the clock domain between the frame buffer and our K-means
hardware.
Both Lloyd’s and our histogram-based algorithm can run at 150 MHz. However,
the hardware resources for both these higher speed implementations are noticeably
higher than for the 100 MHz version. In the next section, we discuss the performance
and resources in detail.
We compare the performance of two designs: the standard Lloyd’s algorithm and our
new histogram-based algorithm. For the hardware versions, we consider several fac-
tors: hardware usage, the clock cycles required to process an image (4 clusters), and the
A New Real-Time FPGA-Based Implementation of K-Means Clustering 475
actual number of frames per second (FPS). The utilization of the different designs is
shown in Table 1. It can be seen that the histogram-based algorithm takes about half
the LUT (Look up table) and FF (Flip-Flop) resources.
Since even at 100 MHz our histogram-based architecture is able to achieve real
time performance, the clustering itself does not demand using the fastest clock speed;
and since the 100 MHz version uses rather less hardware resources, its use would
therefore normally be preferred (Table 2).
This paper has presented a new K-means clustering algorithm, based on iteratively
processing the image histogram rather than the image. The main contributions are as
follows:
1. A new histogram-based K-means clustering algorithm, which transforms the 2-D
image into a 1-D histogram. The histogram retains all the details which are needed
to perform K-means clustering. The method is equivalent to Lloyd’s, and gives the
same results.
2. An FPGA implementation of the new histogram-based K-means clustering algo-
rithm. On an FPGA, the histogram-based takes approximately half the resources
compared to Lloyd’s, and a speed-up of approximately 7.6. Even at 100 MHz and
with 50 iterations, we achieve 82 FPS for 640 480 images.
Future work will look at providing greater flexibility in the FPGA architecture,
including a wider range of cluster initialization strategies, and some possible further
optimizations. Using the histogram-based algorithm to cluster RGB color image will be
challengeable, we are going to find a way to optimize the algorithm and make it
suitable for RGB color image.
References
1. Kanungo, T., Mount, D.M., Netanyahu, N.S., et al.: An efficient K-means clustering
algorithm: Analysis and implementation. IEEE Trans. Pattern Anal. Mach. Intell. 24(7),
881–892 (2002)
2. Ng, H.P., Ong, S.H., Foong, K.W.C., et al.: Medical image segmentation using K-means
clustering and improved watershed algorithm. In: Proceedings of 2006 IEEE Southwest
Symposium on Image Analysis and Interpretation, pp. 61–65. IEEE (2006)
3. Duda, R.O., Hart, P.E., Stork, D.G.: Pattern Classification and Scene Analysis, 2nd edn.
Wiley Interscience, Hoboken (1995)
4. Choi, Y.M., So, H.K.H.: Map-reduce processing of K-means algorithm with FPGA-
accelerated computer cluster. In: Proceedings of the 2014 IEEE 25th International
Conference on Application-Specific Systems, Architectures and Processors, ASAP, pp.
9–16. IEEE (2014)
5. Neshatpour, K., Koohi, A., Farahmand, F., et al.: Big biomedical image processing hardware
acceleration: a case study for K-means and image filtering. In: 2016 IEEE International
Symposium on Circuits and Systems, ISCAS, pp. 1134–1137. IEEE (2016)
6. Farivar, R., Rebolledo, D., Chan, E., et al.: A parallel implementation of K-means clustering
on GPUs. In: PDPTA 2008, vol. 13, no. 2, pp. 212–312 (2008)
7. Lavenier, D.: FPGA implementation of the K-means clustering algorithm for hyperspectral
images. In: Proceedings of Los Alamos National Laboratory LAUR (2000)
8. Estlick, M., Leeser, M., Theiler, J., et al.: Algorithmic transformations in the implementation
of K-means clustering on reconfigurable hardware. In: Proceedings of the 2001 ACM/SIGDA
Ninth International Symposium on Field Programmable Gate Arrays, pp. 103–110. ACM
(2001)
A New Real-Time FPGA-Based Implementation of K-Means Clustering 477
9. Hussain, H.M., Benkrid, K., Erdogan, A.T., et al.: Highly parameterized K-means clustering
on FPGAs: comparative results with GPPs and GPUs. In: Proceedings of 2011 International
Conference on Reconfigurable Computing and FPGAs, ReConFig, pp. 475–480. IEEE
(2011)
10. Saegusa, T., Maruyama, T., Yamaguchi, Y.: How fast is an FPGA in image processing? In:
Proceedings of International Conference on Field Programmable Logic and Applications,
FPL 2008, pp. 77–82. IEEE (2008)
11. Saegusa, T., Maruyama, T.: An FPGA implementation of K-means clustering for color
images based on Kd-tree. In: Proceedings of International Conference on Field Pro-
grammable Logic and Applications, FPL 2006, pp. 1–6. IEEE (2006)
12. Maruyama, T.: Real-time K-means clustering for color images on reconfigurable hardware.
In: Proceedings of 18th International Conference on Pattern Recognition, ICPR 2006, vol. 2,
pp. 816–819. IEEE (2006)
13. Khan, S.S., Ahmad, A.: Cluster center initialization algorithm for K-means clustering.
Pattern Recogn. Lett. 25(11), 1293–1302 (2004)
14. Pena, J.M., Lozano, J.A., Larranaga, P.: An empirical comparison of four initialization
methods for the K-means algorithm. Pattern Recogn. Lett. 20(10), 1027–1040 (1999)
15. Khawaja, S.G., Akram, M.U., Khan, S.A., et al.: A novel multiprocessor architecture for
K-means clustering algorithm based on network-on-chip. In: Proceedings of 2016 19th
International Multi-topic Conference, INMIC, pp. 1–5. IEEE (2016)
Neural Network Identification of an Axial
Zero-Bias Magnetic Bearing
Abstract. System identification for magnetic bearings can be used for various
purposes, such as aggressive control design, on-line monitoring and early
diagnosis of upcoming faults in rotating machinery, etc. Especially for the zero-
bias active magnetic bearing (AMB) system, which omits the bias current/flux to
reduce power loss and rotor heating, whereas brings out the problem of highly
nonlinearity. Here, we applied the NARX neural network to the identification of
an axial zero-bias magnetic bearing system. The I/O data are acquired through a
stabilized magnetic bearing test rig, which are more accurate than data that
generated by simulation. Two kinds of training methods are compared in dif-
ferent working conditions. The results show a good ability of generalization,
which prove the feasibility of the NN identification.
1 Introduction
As shown in Fig. 2, the zero-bias AMBs are comprised of the rotor, displacement
sensors, the controller, power amplifiers and electromagnets. The displacement sensors
detect the rotor position and send the position signals to the controller. After AD
conversion and data filtering, the controller calculates the control value and outputs it
as PWM wave. The power amplifiers connect the controller with the electromagnets,
which act as an electronic switch and meanwhile drive the EM.
Compared with conventional AMBs, there is no constant bias current and only one
electromagnet being electrified at any moment. In an attractive type bearing configu-
ration of electromagnets demonstrated in Fig. 1, when the rotor is deviating from the
balanced position due to a disturbance, the coil current of the EM that the rotor is
approaching reduces to zero, while the other EM on the opposite is switched on to
produce attractive magnetic force to drag the rotor back to the equilibrium position and
stabilized. The motion equation for the one-DOF zero-bias AMB system is derived as
m€x ¼ f1 f2 : ð1Þ
where m is the rotor mass and f1 , f2 are the nonlinear control force of each EM, which
can be denoted as
ki21 ki22
f1 ¼ 2
f2 ¼ : ð2Þ
ðx0 þ xÞ ðx0 xÞ2
where k is a constant ð¼ l0 N 2 S=4Þ, i1 ; i2 are the EM’s currents, x0 is the nominal air
gap and x is the rotor displacement. The total force to the rotor f is
f1 ðx [ 0; i2 ¼ 0Þ
f ¼ : ð3Þ
f2 ðx\0; i1 ¼ 0Þ
Neural Network Identification of an Axial Zero-Bias 481
As demonstrated, the relationship between the magnetic force and the control
current is nonlinear. Meanwhile, only one control input is effective at any moment
which is depended on the rotor position [6, 7].
where uðtÞ and yðtÞ denote the I/O of the network at time t, nu and ny represent the input
and output order, and function f is a nonlinear function. When function f can be
approximated by a multilayer perception, the resulting system is called a NARX net-
work. As shown in Fig. 3, the NARX network mainly consists of the input layer, the
hidden layer, the output layer, and the TDL. The output of the hidden layer i at time
k can be described as
X
nu ny
X
pi ðkÞ ¼ fsig ð wir ðkÞuðk rÞ þ wil ðkÞyðk lÞ þ bi Þ: ð5Þ
r¼1 l¼1
where fsig is the Sigmoid function of the hidden layer, nu is the length of time delay of
the external input, and ny is the length of time delay of the output feedback. wir ðkÞ is
the weight value between the input u(t − r) and the hidden unit i at time k; wil ðkÞ is the
weight value between the output feedback y(t − l) and the hidden unit i at time k; bi is
482 Q. Liu et al.
the threshold value of the unit i. Hence, the output yj ðkÞ of the output layer j can be
expressed as
X
N
yi ðkÞ ¼ wji ðkÞdi þ Tj : ð6Þ
i¼1
where wji ðkÞ is the weight value between the hidden unit i and the output layer j; Tj is
the threshold value of the output layer j; N is the number of the hidden units. Therefore,
the expression of NARX network can be expressed by combining Eqs. (5) and (6).
The NN configurations are demonstrated in Fig. 4. There are two kinds of con-
figurations. One is called parallel architecture and the other is series-parallel archi-
tecture. For the former structure, the output of the NARX network is fed back to the
input of the feedforward neural network as part of the standard NARX architecture. For
the latter one, the true output is used instead of feeding back the estimated output,
which possesses two advantages. Firstly, the input to the feedforward network is more
accurate. Secondly, the resulting network has a purely feedforward architecture, and
static backpropagation can be used for training.
As shown in Fig. 4, the aim of the NN training algorithm is to minimize the error
between the sample output and the corresponding NN model output estimation. The
training algorithm applies some parameter-searching strategy, such as the Levenberg-
Marquardt (LM) and back-propagation (BP) training function with momentum
weight/bias learning function, to adjust the parameters of all neurons in the NN. These
parameters mainly contain two groups, i.e. bias and weight.
The performance criteria for the network model can be classified into two types:
mean square error (MSE) and error range, which can be described as follows
1X N
^ 2
MSE ¼ ðyðiÞ yðiÞÞ : ð7Þ
N i¼1
^ ^
Dd = [min(yðiÞ yðiÞÞ; maxðyðiÞ yðiÞÞ: ð8Þ
^
where yðiÞ is the output of the zero-bias AMB system, and yðiÞ is the prediction of the
model output. N is the number of the samples. The smaller the MSE and Dd are, the
more accurate the network model is.
As for the training algorithms for the NN, there are many choices. Here we compare
two kinds of algorithms, i.e. the LM and Scaled Conjugate Gradient (SCG). The LM
interpolates between the Gauss–Newton algorithm (GN) and the method of gradient
descent. For well-behaved functions and reasonable starting parameters, the LM tends
to be a bit slower than the GN. Nevertheless, the LM is more robust than the GN, which
means that in many cases it finds a solution even if it starts very far off the final
minimum. In addition, a supervised learning algorithm (SCG) with super linear con-
vergence rate is applied, which is based upon Conjugate Gradient Methods. SCG uses
second order information from the neural network yet demands only O(N) memory
usage (N is the number of weights in the network) [12].
4 Simulation Results
Two group of data were acquired, the first was under the steady state with 2000
timesteps and a sampling time of 1 ms (Fig. 6). The second group covered an external
disturbance and the relaxation process, which contained 6000 timesteps with a sam-
pling time of 1 ms (Fig. 10). The number of hidden neurons is ten, and both the input
order and output order of the NARX model are two.
Fig. 6. Snapshot of: (a) input data sequences; (b) output data sequences (steady state)
The simulation results of the steady state were displayed in Figs. 7 and 8. The total
time of training, validation, and test for LM is 1 s, whereas that of the SCG is more
than 2 s. Specifically, Fig. 7. is about the MSE, which represents the average squared
difference between actual output and the targets [14]. Lower values are better, and zero
means no error. Figure 8 is about the error histogram. Obviously, the majority of the
error concentrated on zero, however, the error distribution was more dispersed using
LM. Moreover, the overall regression (R) values for LM is 0.97065, while that of SCG
is 0.9552. The regression measures the correlation between the output and target. An R
value of 1 means a close relationship, 0 a random relationship. Overall, LM has a far
Neural Network Identification of an Axial Zero-Bias 485
faster convergence speed and smaller regression value, whereas the epoch iteration is
more. Besides, both algorithms are sensitive to initial parameters.
Fig. 7. Mean squared error and epochs number of the steady state: (a) LM; (b) SCG
Fig. 8. Error histogram with 20 bins (steady state): (a) LM; (b) SCG
As for the second group, the simulation results of the steady state are demonstrated
in Figs. 9, 10, 11 and 12. The total time of training, validation, and test for LM is
around 10 s, whereas that of the SCG is more than 20 s. The LM algorithm has smaller
MSE and regression value compared with SCG. Overall, both algorithm can approach
the nonlinear system very well. Although the LM requires large calculation, it prevails
in convergence speed and regression.
486 Q. Liu et al.
Fig. 9. A snapshot of: (a) input data sequences; (b) output data sequences (under disturbance)
Fig. 10. Mean squared error and epochs number of: (a) LM; (b) SCG (under disturbance)
Fig. 11. Error histogram with 20 bins (under disturbance): (a) LM; (b) SCG
Neural Network Identification of an Axial Zero-Bias 487
Fig. 12. The regression values of the training, validation, test process (under disturbance):
(a) LM; (b) SCG
5 Conclusion
Zero-bias magnetic bearing is a highly-nonlinear system and the design of the con-
troller needs to include the nonlinear characteristics. Based on the fact that the NN can
approach any nonlinear mapping, here we present the system identification of the zero-
bias AMB with a NARXNET. Two kinds of training algorithms are compared in
different operation conditions. The results show the feasibility and effectiveness of NN
identification.
References
1. Kato, Y., Yoshida, T., Ohniwa, K.: Self-sensing active magnetic bearings with zero-bias
current control. Electr. Eeg. Jpn. 165, 69–76 (2008)
2. Rong, H., Zhou, K.: Nonlinear zero-bias current control for active magnetic bearing in power
magnetically levitated spindle based on adaptive backstepping sliding mode approach. Proc.
Inst. Mech. Eng. Part C: J. Mech. Eng. Sci. 231, 3753–3765 (2017)
3. Gahler, C., Mohler, M., Herzog, R.: Multivariable identification of active magnetic bearing
systems. JSME Int. J. Ser. C Mech. Syst. Mach. Elem. Manuf. 40, 584–592 (1997)
4. Cho, Y.M., Srinavasan, S., Oh, J.-H., Kim, H.S.: Modelling and system identification of
active magnetic bearing systems. Math. Comput. Model Dyn. 13, 125–142 (2007)
5. Gähler, C., Herzog, R.: Identification of magnetic bearing systems. Math. Model. Syst. 1,
29–45 (1995)
6. Sivrioglu, S., Saigo, M., Nonami, K.: Adaptive backstepping control design for a flywheel
zero-bias AMB system. In: Proceedings of 2003 IEEE Conference on Control Applications,
CCA 2003, pp. 1106–1111. IEEE (2003). B 7
488 Q. Liu et al.
7. Kato, Y., Yoshida, T., Ohniwa, K., Miyashita, O.: A self-sensing active magnetic bearing
with zero-bias-current control. In: European Conference on Power Electronics and
Applications, p. 9 (2005)
8. Mohd-Mokhtar, R., Liuping, W.: System identification of MIMO magnetic bearing via
continuous time and frequency response data. In: IEEE International Conference on
Mechatronics, ICM 2005, 10–12 July 2005, p. 191–196 (2005)
9. Srinivasan, S., Cho, Y.M.: Modeling and system identification of active magnetic bearing
systems. In: Proceedings of International Conference on Control Applications, pp. 252–260
(1995)
10. Jayaswal, P., Verma, S.N., et al.: Development of EBP-artificial neural network expert
system for rolling element bearing fault diagnosis. J. Vib. Cont. 17, 1131–1148 (2010)
11. Siegelmann, H.T., Horne, B.G., Giles, C.L.: Computational capabilities of recurrent NARX
neural networks. IEEE Trans. Syst. Man Cybern. Part B Cybern. 27, 208–215 (1997).
A Publication of the IEEE Systems Man & Cybernetics Society
12. Møller, M.F.: A scaled conjugate gradient algorithm for fast supervised learning. Neural
Netw. 6, 525–533 (1993)
13. Gibson, N.S., Heeju, C., Buckner, G.D.: H∞ control of active magnetic bearings using
artificial neural network identification of uncertainty. In: 2003 IEEE International
Conference on Systems, Man and Cybernetics, vol.1442, pp. 1449–1456 (2003)
14. Hamamreh, J., Dama, Y.: Self-organizing schedulers in LTE system for optimized pixel
throughput using neural network (2013)
Precision Measurement and
Instrumentation
A Novel Bench of Quarter Vehicle Semi-active
Suspension
1 Introduction
bench is promising but difficult to achieve. Some researchers investigated the devel-
opment of vehicle suspension test rig, e.g., Liu et al. [3] constructed a performance
testing system that can measure the absorption efficiency, vibration frequency, phase
difference and corresponding vibration wave form of suspension. It provides a basis for
comprehensive evaluation and fault diagnosis of suspension performance. Based on the
existing automobile suspension test platform, Chen et al. [4] designed a high precision
force sensor and the adjustment module of the output signal of the speed sensor, which
improved the accuracy and real time of the test system. By simplifying the actual
vehicle, Geng et al. [5] established the physical model of the two degree of freedom
Mcpherson suspension system. On this basis, the research and construction of the
performance test bench for the Mcpherson suspension system were designed and
constructed. Omar et al. [6] designed a universal suspension test rig for electrohy-
draulic active and passive automotive suspension system, and demonstrated the
effectiveness of the platform through simulation and experimentation.
However, these suspension test rig has less adaptive ability. There is a difficulty in
tuning damping coefficient. They are not suitable to study the semi-active suspension
and control algorithms. Consequently, the contribution of this study is to develop a
bench for study vehicle suspension. According to the vehicle suspension performance
evaluation criteria, a two-degree-of-freedom vehicle suspension physical model was
developed. This bench can not only be used for passive suspension training to match
the test of passive vehicle suspension stiffness and damping, but also for semi-active
suspension testing. By adopting rapid control prototype (RCP) technology, the com-
parison of different control algorithms on the impact of suspension performance can be
realized, the cost and time of the control algorithm test can be saved. The effectiveness
of the quarter car suspension bench system was verified through experiments.
The organization of this study is as following. Test standard for vehicle suspension
performance is addressed in the Sect. 2. In Sect. 3, the theoretic analysis for the bench
test system of a quarter car suspension is analyzed. The design of the bench test system
of a quarter car suspension is proposed in the Sect. 4. The prototype test and discussion
are investigated in the Sect. 5. Section 6 concludes this study.
Vehicle suspension evaluation methods can be roughly divided into subjective sensory
evaluation and objective physical quantity evaluation methods. Since subjective
evaluation methods vary from person to person, the discrete type is very large, and it is
unfavorable to explain the performance of the automotive suspension. With the
development of measurement technology and signal processing technology, the
objective physical quantity evaluation method has been widely used.
A Novel Bench of Quarter Vehicle Semi-active Suspension 493
where aw is the equivalent vibration value for the entire exposure time (m=s2 ); T is the
vibration analysis time (s).
When the crest factor is greater than 12, the method of average quadripartite value
can better estimate the severity of the high crest factor vibration, and vibration dose
value (VDV) can be obtained.
Z T 12
VDV ¼ a4w ðtÞdt m=s1:75 ð2Þ
0
hX i14
VDV ¼ VDVi4 ð4Þ
For the same vibration with a crest factor less than 12, EVDV is basically the same
as VDV. If the difference between the two values is less than 25%, the RMS method
can be used [7].
of the suspension. The RMS value of the dynamic load of the tire directly affects the
grounding of the vehicle. The excessive dynamic load of the tire will seriously affect
the safety of the grounding of the vehicle, which will degrade the steering stability.
Based on the above evaluation indicators, it can be seen that the ride stability and
comfort of the vehicle are affected by the sprung mass acceleration and suspension
dynamic deflection. Stability of vehicle operation is affected by tire dynamic load.
According to different vehicle suspension requirements, different vehicle suspension
parameters are selected to obtain the required suspension performance.
The working principle of the vehicle suspension system. The unevenness of the road
surface is transmitted to four dampers of suspension systems through the wheels, and
then transmitted to the vehicle body through the suspension system. The vehicle system
in motion is a multi-degree-of-freedom vibration system. Research on the car sus-
pension is based on the dynamic model of the suspension. The vehicle model can
accurately predict the impact of the suspension parameters on its performance, thus it
can reflect the performance of real suspension system. However, the establishment of
an accurate model is a complex process. The establishment of a complete vehicle
model requires more relevant parameters, and the accuracy of the parameters will
inevitably affect the accuracy of the model.
The quarter-vehicle suspension model is the most basic model for the research of
the suspension system. Although it cannot be used for the attitude control of the
vehicle, it is sufficient to reflect the relevant indicators that reflect the performance of
the suspension: the sprung mass acceleration and the sprung mass displacement, sus-
pension dynamic deflection and tire dynamic load. In addition, the quarter vehicle
suspension model involves fewer parameters, which facilitates in-depth study on the
suspension system. Therefore, the test object of the vehicle suspension characteristic
test system in this paper is the quarter vehicle suspension model [9, 10].
ms ms xs
xs
ks
cs fd
mu z
mu z
kt kt
d d
Fig. 1. 1⁄4 car passive suspension model Fig. 2. 1⁄4 car semi-active suspension model
A Novel Bench of Quarter Vehicle Semi-active Suspension 495
In the figure: ms , xs are body masses and displacements, respectively. mu , z are tire
mass and displacement, respectively. ks is the stiffness of the suspension. kt is the tire
stiffness. cs is the viscous damping coefficient of the suspension. d is random road
input. fd is the damping force of a MR damper.
Where €xs , x_ s are the sprung mass accelerations and speeds, respectively. €z, z_ are un-
sprung mass accelerations and speeds, respectively.
For a MR semi-active suspension, the corresponding passive damper can be
replaced with a magneto-rheological damper, and its kinetic equation is.
in which,
The transfer function of the sprung mass displacement with respect to the road
surface input:
Xs ðsÞ kt ðCs s þ ks Þ
HDI ¼ ¼ ð10Þ
DðsÞ Ds
The transfer function of the suspension deflection relative to the road surface input,
Xs ðsÞ ZðsÞ kt ms s2
HDE ¼ ¼ ð11Þ
DðsÞ Ds
496 X. Dong et al.
The transfer function of the tire dynamic load with respect to the road surface input,
The quarter car semi-active suspension bench system is mainly designed to simulate the
vibration characteristics of the quarter vehicle suspension system. Its overall structure is
shown in the Fig. 3.
The quarter vehicle semi-active suspension bench system consists of two parts, the
hardware system part and the sensing and software system part. The hardware system
part completes all physical movements of the vehicle suspension system. The sensing
scheme and software system part complete signal acquisition, feedback and control of
hardware.
The quarter vehicle semi-active suspension bench system working principle is as
follows, the electric servo vibration excitation system sends excitation signals for
simulating the road surface excitation to the vibration table, and the vibration table
produces corresponding vertical vibration, the vibration is transmitted to the inner
frame by the tire assembly and the damper assembly. The inner frame is vibrated by the
roller to move relative to the guide rail of the outer frame, thus the vertical vibration of
dSPACE
d Wheel assembly
Control algorithm
interactive platform
Vibration table control force
Driving source
Signal Input/output port
Fig. 3. The quarter vehicle semi-active suspension bench system block diagram
A Novel Bench of Quarter Vehicle Semi-active Suspension 497
the inner frame is achieved. Through the sensors at the specified position of the inner
frame, the status information of different parts of the suspension is transmitted to
dSPACE. The signal is converted by dSPACE and transmitted to the control algorithm
interaction platform. The control algorithm interactive platform outputs the control
signal to the damper after the signal is calculated and processed. The control signal is
applied through the current driver to the coil of the semi-active MR damper, thus the
suspension damping is changed and the suspension control is achieved.
4.1.1 The Vibration Control Cabinet and the Vibration Table Part
The vibration control cabinet includes servo amplification system and vibration table
system, etc. The servo amplifier system is designed to drive a vibration table, including
a power amplifier, a power amplifier cabinet, a system control box, etc. By modulating
the signal, conducting electrical control, measuring and controlling the output voltage
and output current of the power amplifier, the excitation function of the vibration table
can be completed.
The vibration table system is mainly composed of an excitation part, a moving coil
and its supporting and guiding, a vibration isolation mechanism and a bearing. In the
moving coil and its support and guide, an air spring is used for support.
The pulley guide is used between the internal and external frames. The pulley is
evenly arranged along the vertical direction of the guide rail, a certain pre-pressure is
exerted between the pulley and the guide rail to eliminate the shake due to the
movement of the inner frame.
With the aim of guaranteeing that the sprung mass is close to the real vehicle, the
counterweights are evenly and symmetrically arranged on the top platform of the inner
frame according to requirements. The counterweight and the top platform of the inner
frame are fixed by set screws to prevent vibration.
the damper piston rod and the frame bottom plate of the vehicle suspension, the output
force value of the damper can be monitored in real time, the real-time correction of the
control signal can be realized.
According to the function and structure of the semi-active suspension characteristic test
platform system of the vehicle as description in the third section, the prototype of the
test platform system is designed and constructed as shown in Fig. 4.
1 6
7
9
3
10
4
Fig. 4. Auto semi-active suspension characteristic test platform system prototype schematic
6 Conclusion
In this paper, a universal and efficient standard bench of quarter car semi-active sus-
pension is proposed, designed and configured, and the effectiveness of the bench is
verified through experiments. Four main conclusions can be drawn as followings.
(1) During the design process of the quarter car semi-active suspension bench, it is
ensured that the bench is close to the real vehicle model, and that the testing result
of the test platform is relatively high relative to the spoilage of the real vehicle
model.
(2) The bench can be equipped with a passive suspension training platform to match
the stiffness coefficient and the damping coefficient of passive suspension.
(3) The bench based on dSPACE system can realize RCP technology and multiple
real-time tests of control algorithms, thus avoiding the time and cost loss caused
by designing and manufacturing the control algorithm hardware.
(4) By arranging the force sensor on the MR damper, real-time tracking and cor-
rection of the control force value can be realized, and the real-time performance of
the control effect can be detected.
Acknowledgment. The authors disclosed receipt of the following financial support for the
research, authorship, and/or publication of this article: This work was financially supported by
the National Natural Science Foundation of People’s Republic of China (Project NO. 51675063),
this research is also supported by graduate research and innovation foundation of Chongqing,
China (Grant No. CYB17023). These supports are gratefully acknowledged.
A Novel Bench of Quarter Vehicle Semi-active Suspension 501
References
1. Dong, X., Zhang, Z., Yu, J., Tuo, F.: Partial fault sliding mode fault-tolerant control for
magneto-rheological semi-active suspension actuators considering temperature effect. In:
13th National Rheology Conference Proceedings, pp. 36–38, Xi’an (2016)
2. Dong, X., Yu, M., Liao, C., Chen, W., Huang, S.: Research on adaptive fuzzy logic control
for automobile magneto-rheological semi-active suspension. J. China J. Highw. Transp. 19
(2), 111–115 (2006)
3. Liu, Y., et al.: Study on testing method and testing system of vehicle suspension
performance. J. Highw. Transp. Res. Dev. 23(9), 126–130 (2006)
4. Chen, G., et al.: Research and design of the suspension performance test system for
automobiles. J. Chang. Univ. Sci. Technol. (Nat. Sci. Ed.) 32(3), 388–391 (2009)
5. Geng, X., Xu, W., Li, B., Wang, Z., Chen, Q.: The design of McPherson suspension system
performance test platform. J. Mach. Des. Manuf. (11), 163–165 (2012)
6. Omar, M., et al.: A universal suspension test rig for electrohydraulic active and passive
automotive suspension system. J. Alex. Eng. J. 56(4), 1–11 (2012)
7. ISO 2631-1: Mechanical vibration and shock-Evaluation of human exposure to whole-body
vibration-part 1: General requirements. S, pp. 3–10 (2003)
8. Miao, Y.: Study on control system of automobile magneto-rheological semi-active
suspension, pp. 46–51. Chongqing University, Chongqing (2003)
9. Tuo, F.: Research on fault-tolerant control of MR semi-active suspension with partial fault in
actuator considering temperature effect, pp. 32–34. Chongqing University, Chongqing
(2015)
10. Zhao, M., Liu, X., Pan, G.: Study on magneto-rheological damper model and semi-active
suspension control. J. Vib. Eng. 17, 1051–1053 (2004)
11. Dong, X.: Human simulated intelligent control of automobile magneto-rheological semi-
active suspension, pp. 44–49. Chongqing University, Chongqing (2006)
Research on Nondestructive Testing
for the Depth of Hardening Zone
Ping Chen, Tao Xiang(&), Song Guo, Jie Fang, and Xin Xu
1 Introduction
The basic principle of eddy current nondestructive testing: the exciting coil will gen-
erate an alternating magnetic field around the coil after energizing the exciting coil. The
eddy current shown in Fig. 1(a) will be generated, when the work-piece is placed in the
alternating magnetic field. However, the eddy currents will in turn generate a magnetic
field in the opposite direction compared with the excitation magnetic field so that the
original magnetic field in the coil is reduced, thus causing the change of the coil
impedance. Since the distribution of the eddy current depends on excitation conditions.
Therefore, when some exciting parameters are kept constant, the change of exciting
coil impedance or the voltage in another dedicated detecting coil can be measured so
that the researcher calculate the parameter that the test can need.
The eddy current in the measured work-piece is unevenly distributed because the
frequency of the exciting current, the conductivity and the magnetic permeability of the
tested work-piece can influence the eddy current generated in the work-piece. The eddy
(a) (b)
current density is the largest and the densest on the work-piece surface, and the eddy
current density gradually decreases with the increase of depth. The attenuation
amplitude shows an exponential decay function with the increase of depth and also
shows the skin effect of eddy current, as shown in Fig. 1(b) [4].
The distance that the eddy current penetrates into the work-piece is called the depth
of penetration. Usually the depth that the eddy current density is attenuated to 1/e
(37%) of work-piece surface’s eddy current density is defined as the standard pene-
tration depth, also called the skin depth. Its calculation formula is as follows [4]:
1
d ¼ pffiffiffiffiffiffiffiffiffiffi : ð1Þ
pf lr
In the formula, d is the standard penetration depth (skin depth), r is the conduc-
tivity, l is the magnetic permeability, and f is the frequency of the exciting current.
According to the calculation formula, it can be known that the skin depth is
inversely proportional to the 1/2 order power of the current frequency for a given
quenched work-piece. Therefore, the reasonable detecting frequency can be selected by
504 P. Chen et al.
the calculation formula so that the skin depth includes the entire quenched layer and the
detection made can be more accurate and faster.
The measurement of hardening depth by using eddy current is essentially based on
establishing mathematical model between the hardening depth and the induced voltage
of detecting coil, so as to achieve the goal of measuring the hardening depth. For
example, the calculation formula of induction voltage for cylindrical quenching parts is
as follows [5]:
pd 2
Vm ¼ ð2pf ÞNs l l H0 108 : ð2Þ
4 r eff
Where Vm is the induced voltage of detecting coil, Ns is the number of coil turn, f is
the frequency of the exciting current, lr is the relative magnetic permeability, leff is the
effective magnetic permeability, and H0 is the magnetic field strength.
It can be known from the above formula that the induced voltage increases with the
increase of excitation frequency, but the excitation frequency can also affects the
effective permeability. So the induced voltage will show a non-linear increasing trend
when the excitation frequency increases. Therefore, the mathematical model between
the hardening depth and the detecting coil’s induced voltage is established under the
condition that the frequency of excitation current remains unchanged by this formula,
and the depth of hardening zone is finally measured by solving the mathematical
model. For example, in the literature [6], the hardening depth of 40Cr steel was
measured by using this method. Through the analysis of the measured eddy current, a
mathematical model between the hardening depth and the induced voltage was
established. The maximum relative deviation of the measured results of the hardening
depth was less than 6% compared with the metallographic method on average.
form backscattered echo. Therefore, the depth of hardening zone can be calculated by
measuring the time when the ultrasonic wave is emitted to the receiving echo.
At present, the most widely used method in the field of ultrasonic nondestructive
testing is the piezoelectric ultrasonic nondestructive testing technology. Figure 2 shows
the detection principle. After the ultrasonic wave enters into work-piece by the cou-
plant, the backscattering phenomenon will occur when ultrasonic wave passes through
the interface. Therefore, the time that the ultrasonic wave transmits from the work-
piece surface to the interface can be measured by the detection of backscatter signals,
and then the depth of hardening zone is calculated by the path formula.
According to the theory of acoustic scattering, the size of microstructure and the
frequency of ultrasonic wave both have a great influence on the detection. Therefore, it
is necessary to determine the frequency range that can generate backscattered echo
according to the Eq. (3) before testing [1]. It can not only improve the efficiency of
experimental detection, but also make the detection more accurate.
v
dk¼ : ð3Þ
f
B
Bm
1
Br
2
-Hm -Hc Hc H
3 O 3 Hm
2
1 2 The reversible
-Br Barkhausen jump
1 3 The irreversible
-Bm Barkhausen jump
In order to obtain the value of the magnetic coercive force to characterize the depth
of hardening zone, the magnetic hysteresis loop is measured by the RC indirect
measurement method shown in Fig. 4. According to the following formula [18]:
CR2
B¼ UB : ð4Þ
SN2
The magnetic induction intensity in the tested piece can be calculated by measuring
the output voltage UB at the two ends of the integration capacitor C. Where B is the
magnetic induction intensity, C is the integration capacitor, R2 is the Integral resistance,
S is the effective magnetic flux area of detecting coil and N2 is the turn number of
detecting coil.
I1 Integration capacitor C UB
Power source
I2
Sampling resistance R1 Workpiece IIntegration resistance R2
After measuring the magnetic induction intensity B under different the external
magnetic field intensity, the magnetic hysteresis loop can be obtained and the value of
magnetic coercive force can be similarly obtained. Since the magnetic coercive force is
closely related to the carbon content and the grain size (the grain size increases
gradually from the work-piece surface to the interface) of material, the depth of
hardening zone can be obtained by establishing the mathematical model between the
magnetic coercive force and the hardening depth and then solving the model. For
example, the carburizing and quenching hardening layer depth is measured by mag-
netic coercive force in the literature [18], and the relationship between the magnetic
coercive force and the quenching depth was established.
Workpiece
The method shown in Fig. 5 is used to measure the Barkhausen Noise. After
establishing the mathematical model between the strength of noise signal and the depth
of hardening zone, and solving the mathematical model can finally achieve the goal of
measuring the depth of hardening zone. For example, the Barkhausen effect was used
to detect the depth of hardening zone in the literature [25]. The fitting straight line
equation of the strength of noise signal and the depth of hardening zone was
MP ¼ 14:688H 0:6828 , and the relative error of the fitting straight line equation was
measured to be within 4% at last.
The detecting methods of hardening depth that are referred to above are all non-
destructive testing method. These methods can directly detect the work-piece and the
operation is simple and quick. What’s more, it can achieve the goal of large-scale
detection for the quenched work-piece and eliminate the disadvantage that the tradi-
tional method needs to damage the specimen and needs to do sampling detection.
So these methods have a wide application prospect.
Although nondestructive testing of the hardening depth has many advantages, there
are still many limitations. For example, the eddy current nondestructive testing tech-
nology can only detect work-piece with the shallow depth of hardening zone because of
the skin effect. If the tested work-piece has the large depth of hardening zone, the
measurement result will be inaccurate due to the skin effect. What’s more, the eddy
current nondestructive testing technology can only be used for conductive materials.
With regard to ultrasonic nondestructive testing technology, the excitation frequency of
ultrasonic wave has great influence on the detection. If the inappropriate excitation
frequency is selected, the scattered echo signal may be extremely weak and difficult to
recognize. So the reasonable excitation frequency must be selected in the experiment.
In addition, the couplant must be also used between the piezoelectric ultrasonic probe
and the work-piece to measure. Then, the type of couplant, its thickness, and other
factors must also be taken into account. This makes the detection more complicated.
Since the nondestructive testing technology based on magnetic coercive force is limited
to the detection principle, it can only detect the work-piece made of ferromagnetic
material and be limited by the shape of the detected work-piece. According to the
literature [25], the nondestructive testing technology based on Barkhausen effect also
limits the detecting depth because of the detection principle. When the hardening depth
is more than 4 mm, the signal strength of the Barkhausen Noise no longer changes.
This makes the Barkhausen effect be limited when detecting the large depth of hard-
ening zone. In addition, these nondestructive testing methods all require that the
detection probe is separately designed for the work-piece of complex shape, which also
brings about many difficulties in detection.
Research on Nondestructive Testing for the Depth of Hardening Zone 509
In view of the above deficiency, the development of nondestructive test methods for the
depth of hardening zone can be considered according to the following points. For the
quenched work-piece with shallow depth of hardening zone, the eddy current nonde-
structive testing and the Barkhausen Noise nondestructive testing can be used, and the
optimization research is carried out for two kinds of nondestructive testing technology
on this basis. For the optimization research, it can be considered from the following
two aspects:
(1) For traditional eddy current nondestructive testing technology, the detection of
depth is limited due to the skin effect. So pulsed eddy current testing technology
can be considered to take place of traditional eddy current testing technology. The
pulse eddy current nondestructive testing technology is a new detection method
developed on the basis of traditional eddy current nondestructive testing tech-
nology. It uses pulsed current signal as excitation signal. Compared with single-
frequency excitation signal, the frequency spectrum of pulsed excitation signal
has a wider range. The secondary induced electric field is easier to identify and
detect. The final induced voltage signal carries more abundant and diverse
information, which can increase its detection depth.
(2) For the nondestructive testing of hardening depth based on Barkhausen effect, a
mathematical model between the hardening depth and grain size can be consid-
ered to establish. Because the Barkhausen Noise signal carries grain size infor-
mation, this can make the depth of hardening zone that is measured further
increase by the grain size information carried in the noise signal.
For the quenched work-piece with deep hardened layer, ultrasonic nondestructive
testing technology and nondestructive testing technology based on magnetic coercivity
can be considered to use, and the optimization research was carried out for two kinds of
nondestructive testing technology on this basis. The optimization research can be
considered from the following two aspects:
(1) For the piezoelectric ultrasonic testing technology, it is difficult to control the
thickness of couplant. Therefore, the electromagnetic ultrasonic transducer can be
used to replace the piezoelectric ultrasonic transducer. Electromagnetic ultrasonic
transducer is a new type ultrasonic receiving and transmitting device. It relies on
the electromagnetic effect to transmit and receive ultrasonic. What’s more, it
doesn’t need the couplant in the detection process and contact with the work-piece.
So it can avoid the problems and disadvantages because of using the couplant.
(2) For the nondestructive testing technology based on magnetic coercive force, the
detection process is limited to the detection probe. When the shape of work-piece
is complicated, different structure size’s probes need to be designed for different
work-pieces each time the detection is performed. Therefore, it can consider the
miniaturized and flexible design of the probe, which can make the probe adapt to
most of the different shapes of work-piece. If the problem of probe can be solved,
the nondestructive testing technology based on magnetic coercive force can be
applied more widely.
510 P. Chen et al.
The above four detection methods all measure the depth of hardening zone by
establishing a mathematical model of the special signal parameters and the depth of
hardening zone. But they all ignore the relationship between grain size and the depth of
hardening zone. Therefore, the next research can focus on the relationship between the
grain size and the depth of the hardened layer. That is, the mathematical model between
the hardening depth and grain size is established by the grain size information carried in
the detected signal. Finally, solving the mathematical model achieves the goal of
measuring the depth of hardening zone.
When the nondestructive testing technology is used to measure the depth of
hardening zone of quenched work-piece, it can greatly improve the production effi-
ciency of actual quenched work-piece and greatly reduce the cost of metallographic
detection because of its nondestructive, convenient, rapid, efficient and low-cost
characteristics. Therefore, there is extremely important economic value and industrial
value to carry out further research, exploration and optimization on the nondestructive
testing method for the depth of hardening zone.
References
1. Shi, J.C.: Nondestructive testing technology of hardened layer of 45 steel based on ultrasonic
wave. Anhui University of Technology (2016)
2. Guan, Y.Q.: The ultrasonic measurement of the depth of hardening zone. J. Constr. Mach. 6,
106–107 (2015)
3. Xu, G.Q., Zhong, X.L., Zhan, S.F.: Ultrasonic testing technology for quenching depth of
cummins crankshaft. 28(4), 196–197 (2016)
4. Cheng, X.M.: Research on carburized layer characteristics and eddy current testing system of
carburized materials. Wuhan University of Technology (2003)
5. Li, J.W.: Nondestructive Testing Manual. China Machine Press, Beijing (2012)
6. Wang, C.Y., Chen, T.Q., Song, L., et al.: Eddy current nondestructive testing of the
hardening depth of 40Cr steel. J. Nondestruct Test. 31(6), 8–13 (2007)
7. He, X.P., Tian, Y.P., Zhang, H.P.: Ultrasonic nondestructive evaluation on the grain size of
metal materials. J. Tech. Acoust. 6, 445–451 (2013)
8. Lobkis, O.I., Rokhlin, S.I.: Characterization of polycrystals with elongated duplex
microstructure by inversion of ultrasonic backscattering data. J. Appl. Phys. Lett. 96(16),
161905 (2010)
9. Weaver, R.L.: Diffusivity of ultrasound in polycrystals. J. Mech. Phys. Solids 38, 55–86
(1990)
10. Papadakis, E.P.: Ultrasonic attenuation caused by scattering in polycrystalline media. In:
Mason, W. (ed.) Physical Acoustics, vol. IV, no. Part B, pp. 269–328. Academic Press, New
York (1968)
11. Merkulov, L.G.: Investigation of ultrasonic scattering in metals. Sov. J. Tech. Phys. 26, 59–
69 (1956)
12. Hirsekorn, S.: The scattering of ultrasonic waves by polycrystals. J. Acoust. Soc. Am. 72,
1021–1031 (1982)
13. Beecham, D.: Ultrasonic scatter in metals its properties and its application to grain size
determination. J. Ultrason. 4(2), 67–76 (1966)
14. Fay, B.: Theoretical considerations of ultrasound backscatter. J. Acta Acust. United Acust.
28(19), 354–357 (1973)
Research on Nondestructive Testing for the Depth of Hardening Zone 511
15. Fay, B., Brendel, K., Ludwig, G.: Studies of inhomogeneous substances by ultrasonic back-
scattering. J. Ultrasound Med. Biol. 2(3), 195–198 (1976)
16. Du, H., Lonsdale, C., Oliver, J., et al.: Evaluation of railroad wheel steel with lamellar
duplex microstructures using diffuse ultrasonic backscatter. J. Nondestruct. Eval. 32(4), 331–
340 (2013)
17. Zhao, Z.K., Kong, C.H., Bao, W.B.: The principle and application of ultrasonic backscatter
hardening layer depth detector. In: 8th National Electromechanical Technology Annual
Meeting and Machinery Industry Energy Saving and Emission Reduction Process
Technology Seminar, pp. 95–101, Hangzhou (2014)
18. Li, D.P.: Research on detection technology of carburized and hardened depth of the magnetic
coercivity. Nanchang Hangkong University (2015)
19. Luo, X., Wu, W., Li, D.P., et al.: Detection of 20CrMnTi steel carburizing hardening depth
by magnetic coercive force. J. Nondestruct. Test. 38(7), 47–50 (2016)
20. Sun, J.M.: Study on measurement technology of stress measurement system on coercive
force measuring for ferromagnetic materials. Shenyang University of Technology (2016)
21. Lv, Z.: Ferromagnetic material stress testing technology research based on the coercivity.
Shenyang University of Technology (2016)
22. Zhou, Y.: Based on coercivity of ferromagnetic materials stress detecting technology.
Shenyang University of Technology (2015)
23. Zhang, L.: Coercive force of ferromagnetic materials testing technology research. Shenyang
University of Technology (2014)
24. Wang, W.T.: Research on the key technology of nondestructive testing based on Barkhausen
effect. Zhengzhou University (2017)
25. Liu, Y.B., Shi, H.J., Meng, F.Q., et al.: The research for Barkhausen noise to test the
hardened depth. J. Coal Mine Mach. 10, 24–25 (2001)
26. Liu, Y.B., Yang, Z.Y., Shi, X.J., et al.: Nondestructive determination of hardening depth in
an induction hardened bearings by Barkhausen noise. In: 1th International Mechanical
Engineering Conference, p. 547. Machinery Industry Press, Shanghai (2000)
27. Du, F.M., Xu, X.C., Xu, Y.H., et. al.: Nondestructive testing of hardening depth with
Barkhausen effect. In: 8th National Acoustic Emissions Academic Symposium, pp. 62–65,
Shanghai (1999)
28. Yuan, H.J., Fu, J., Li, H.M., et al.: Application status of nondestructive testing technology
based on Barkhausen effect. J. Acta Metrol. Sin. 37(01), 75–78 (2016)
29. Pan, L., Zhang, S.J., Wan, G., et al.: The detection of ferromagnetic material hardness based
on Barkhausen effect. J. Nondestruct. Test. 38(12), 28–31 (2016)
30. Zhang, Y.: Study on nondestructive testing of high strength steel plate based on Barkhausen
effect. Huazhong University of Science and Technology (2015)
Distributed Fusion Estimation Based
on Robust Kalman Filtering for 3D
Spatial Positioning System
Li Liu1(B) , Wenju Zhou1(B) , Aolei Yang2 , Jun Yue1 , and Xiaofeng Zhang1
1
School of Information Science and Electrical Engineering, Ludong University,
Yantai 264025, China
[email protected], [email protected]
2
Shanghai Key Laboratory of Power Station Automation Technology,
School of Mechatronic Engineering and Automation,
Shanghai University, Shanghai 200072, China
1 Introduction
In recent decades, to solve the problem of high-precision measurement and posi-
tioning, the space measurement and positioning systems with relevant measure-
ment methods have been developed rapidly. With the increasing of the network
scale, the scope of the measurement and positioning systems has been growing
much larger; meanwhile, the coverage mode has been growing much more flexible.
Therefore, the application of distributed network structure can able to increase
L. Liu—This work was supported by Natural Science Foundation of China
(61772253, 61633016, 61473182, 61472172), Key Research and Development Project
of Yantai (2017ZH061), Key Project of Science and Technology Commission of
Shanghai Municipality (15411953502), Key Technical of Key Industries of Shan-
dong (2016CYJS03A02-1), and Key Research and Development Project of Shandong
(2016ZDJS06A05).
c Springer Nature Singapore Pte Ltd. 2018
K. Li et al. (Eds.): ICSEE 2018/IMIOT 2018, CCIS 924, pp. 512–521, 2018.
https://doi.org/10.1007/978-981-13-2384-3_48
Distributed Fusion Estimation Based on Robust Kalman Filtering 513
i
in which, Pk|k−1 denotes the one-step prediction error covariance at the time
instant of k − 1.
Since the local estimate x̂ik+1|k satisfies as following:
Δ
and the initial value is set to X0 = μ0 μT0 + P0 [7].
Then, Wki and Vki are obtained by the following means, respectively.
i T
Wki = E x̃ik|k−1 wkT = Bk−1 Qk−1,k − Kk−1
i
Sk (7)
T
Vki = E x̃ik|k−1 vki i
= Bk−1 Sk−1,k+1 i
− Kk−1 i
Rk−1,k (8)
i T T i T
where Ξki = Ak Pk|k−1i
Hk + Bk Wki Hk + Ak Vki + Bk Sk,k+1
i
and Ξ̃ki =
T T
Hki Vki + Vki Hki .
i
Substituting Eqs. (9) into (4), error covariance matrix Pk+1|k is rewritten as
i i
T
Pk+1|k = Ak Pk|k−1 ATk + Γki − Kki Ξki (10)
T
Note that Γki = Bk Qk BkT + Ck Xk CkT + Ak Wki BkT + Bk Wki ATk . Based on
the above analysis, the matrix Σk can be calculated as following:
T T
Σki,j = Hki Pk|k−1
i,j
Hkj + Hki Vki,j + Hkj Vkj,i + Rki,j (11)
The weighted fusion strategy based on the Kalman one-step prediction x̂ik+1|k
and filter gain Kki , the predicted error covariance between the i-th and j-th
subsystems, at the time instant k can be obtained as following:
i T −1
Kki = Ξki Hki Pk|k−1
i
Hk + Rki + Ξ̃ki (13)
T
i,j i,j
Pk+1|k = Ak Pk|k−1 ATk + Bk Qk BkT + Ck Xk CkT + Kki Pk Kkj
T T
+Ak Wki BkT + Bk Wkj ATk − Ak Mki Kkj (14)
T T
− Bk LTk Kkj − Kki Mkj ATk − Kki Lk BkT
where,
Pk = E Ỹk ỸkT
i T
Wki = Bk−1 Qk−1,k − Kk−1 i
Sk
Xk+1 = Ak Xk ATk + Ck Xk CkT + Bk Qk BkT
+ Ak Bk−1 Qk−1,k BkT + Bk Qk,k−1 Bk−1 T
ATk (15)
n
i,q T
Mki = Pk|k−1 (Hkq ) + Vki,q θkq
q=1 T
n
T
Lk = (θkq ) Hkq Wkq + Sk−1,k
q
q=1
Next, the weighted fusion strategy is used for fusing the cross-correlated
covariance, which is defined as
T −1
J = Ixk − x̂k|k−1 Σk|k−1 Ixk − x̂k|k−1 (16)
Since the optimal Kalman filter uses the weighted fusion method of linear
minimum variance [12], the corresponding optimal information fusion estimates
i
covariance Pk|k−1 satisfies Pk|k−1 ≤ Pk|k−1 , i = 1, 2, · · · , n.
(a) State estimation from RKFDFE, OSF(b) State estimation of Sensor 1 from
and WRKF. RKFDFE.
5 Conclusion
Based on the principle of spatial positioning of linear CCD, the 3D photoelectric
sensor has been designed. The spatial target information has been measured,
520 L. Liu et al.
and the trajectory has been tracked and estimated. This paper has studied the
state estimation problem of distributed uncertain systems with cross correla-
tion noise and transmission delays. To deal with the stochastic transmissions
and reduce the network communication burden, the reorganized weighted fusion
estimation and innovation sequence method has been proposed. It has realized
the information interaction between any two neighboring subsystems due to the
communication constraints. Finally, numerical experiments have shown that the
proposed distributed fusion estimation algorithm has a better dynamic tracking
effect for state estimation, to improve the spatial positioning accuracy.
References
1. Ge, X., Han, Q.-L.: Distributed event-triggered H∞ filtering over sensor networks
with communication delays. Inf. Sci. 291, 128–142 (2015)
2. Li, D., Kar, S., Moura, J.M., Poor, H.V., Cui, S.: Distributed Kalman filtering over
massive data sets: analysis through large deviations of random Riccati equations.
IEEE Trans. Inf. Theory 61(3), 1351–1372 (2015)
3. Song, E., Xu, J., Zhu, Y.: Optimal distributed Kalman filtering fusion with sin-
gular covariances of filtering errors and measurement noises. IEEE Trans. Autom.
Control 59(5), 1271–1282 (2014)
4. Hu, J., Wang, Z., Chen, D., Alsaadi, F.E.: Estimation, filtering and fusion for
networked systems with network-induced phenomena: New progress and prospects.
Inf. Fus. 31, 65–75 (2016)
5. Zhong, X., Mohammadi, A., Premkumar, A., Asif, A.: A distributed particle filter-
ing approach for multiple acoustic source tracking using an acoustic vector sensor
network. Sig. Process. 108, 589–603 (2015)
6. Song, B., Wu, Z.-G., Park, J.H., Shi, G., Zhang, Y.: H∞ filtering for stochastic
systems driven by Poisson processes. Int. J. Control 88(1), 2–10 (2015)
7. Feng, J., Wang, Z., Zeng, M.: Distributed weighted robust Kalman filter fusion for
uncertain systems with autocorrelated and cross-correlated noises. Inf. Fus. 14(1),
78–86 (2013)
8. Song, H., Yu, L., Zhang, W.-A.: Distributed consensus-based Kalman filtering in
sensor networks with quantised communications and random sensor failures. IET
Sig. Process. 8(2), 107–118 (2014)
9. Zhang, W., Zhao, Y., Sheng, L.: Some remarks on stability of stochastic singular
systems with state-dependent noise. Automatica 51, 273–277 (2015)
10. Tian, T., Sun, S., Li, N.: Multi-sensor information fusion estimators for stochastic
uncertain systems with correlated noises. Inf. Fus. 27, 126–137 (2016)
11. Costa, O., Benites, G.: Robust mode-independent filtering for discrete-time Markov
jump linear systems with multiplicative noises. Int. J. Control 86(5), 779–793
(2013)
12. Yan, L., Li, X.R., Xia, Y., Fu, M.: Optimal sequential and distributed fusion for
state estimation in cross-correlated noise. Automatica 49(12), 3607–3612 (2013)
13. Qi, W., Zhang, P., Deng, Z.: Robust weighted fusion Kalman filters for multisensor
time-varying systems with uncertain noise variances. Sig. Process. 99, 185–200
(2014)
14. Zhang, P., Qi, W., Deng, Z.: Hierarchical fusion robust Kalman filter for clustering
sensor network time-varying systems with uncertain noise variances. Int. J. Adapt.
Control Sig. Process. 29(1), 99–122 (2015)
Distributed Fusion Estimation Based on Robust Kalman Filtering 521
15. Qi, W.-J., Zhang, P., Deng, Z.-L.: Robust sequential covariance intersection fusion
Kalman filtering over multi-agent sensor networks with measurement delays and
uncertain noise variances. Acta Automatica Sinica 40(11), 2632–2642 (2014)
16. Rezaei, H., Esfanjani, R.M., Sedaaghi, M.H.: Improved robust finite-horizon
Kalman filtering for uncertain networked time-varying systems. Inf. Sci. 293, 263–
274 (2015)
Image Processing
A Novel 3D Head Multi-feature Constraint
Method for Human Localization Based
on Multiple Depth Cameras
1 Introduction
At present, various methods [1–3] have been proposed in addressing this research
area, the majority of which are based on 2D images. Ma et al. [4] first obtained the
motion vector by the optical flow method, and then applied it into a shape model for
human detection and tracking. However, such methods are easily affected by the
external environment. The depth image has many advantages over the 2D gray image
and the color image. First, the depth image is robust to the changes in color and
illumination. In addition, the depth image directly represents the target’s 3D infor-
mation. Therefore, researchers began to use depth information to study the detection
and positioning of moving targets. The RDSF algorithm (Relational Depth Similarity
Features) proposed by Ikemura et al. [5] makes full use of the difference in the fre-
quency distribution of the depth values of different objects, and divides the depth map
into blocks that do not overlap each other. Yu et al. [6] proposed SLTP (Simplified
Local Ternary Patterns) to perform histogram statistics on the edge information of
depth images. Wang et al. [7] proposed a depth feature extraction algorithm PDSS
(Pyramid depth self-similarities) due to strong local self-similarity of depth information
of the human body. Spinello et al. [8] proposed HOD depth feature extraction algo-
rithm that mimics the HOG algorithm, and performs histogram statistics on the mag-
nitude and direction of depth difference to extracts depth features. Shen et al. [9]
improved the HOG features and proposed an HDD feature extraction algorithm. The
algorithm extends the gradient direction from the original 180-degree statistical space
to 360-degree, and makes statistics on the gradients in different directions more precise.
When shooting vertically, the head area is closest to the camera, and the human body
target can be detected and positioned by recognizing the head area. Xia et al. [10]
proposed a model-based human detection system. The algorithm uses 2D head contour
model and 3D head surface geometry model for human detection. The literature [11]
relies on the relationship between the pixels at the edge of the head region to create a
head template to realize the recognition of the head. Meanwhile, the author also pro-
posed a human template matching algorithm [12] to further improve the accuracy of the
head recognition. The algorithm has a strong recognition effect on multi-person
environment and strong ability to eliminate human-like targets.
Aiming at the practical needs of large space, this paper proposes a 3D head multi-
feature constraint method based on multiple depth cameras to detect and locate the
human body. According to the multi-camera spatial structure, NCP is used to determine
the positioning data of multiple cameras to obtain the final body positioning
information.
The rest of the paper is organized as follows: Sect. 2 presents the human posi-
tioning method based on multiple depth cameras; Sect. 3 proposes the nearest center
point method. The Sect. 4 analyses the experimental results, followed by that Sect. 5
concludes the paper.
Figure 1 shows the flow chart of human positioning method based on multiple depth
cameras. Using multiple cameras avoids the disadvantages of a single camera with a
small coverage and improves the detection accuracy.
A Novel 3D Head Multi-feature Constraint Method 527
where Bðx; yÞ is the background image, Ii ðx; yÞ is the i-th frame image, Ti ðx; yÞ is the
binary image, and it is the effective target point.
Then the depth image of the corresponding area is extracted, as shown in
Formula (2).
528 F. Zhou et al.
Fig. 2. Spatial structure and background images of 5 cameras. (a) spatial structure; (b–f)
background images.
Ii ðx; yÞ; Di ðx; yÞ T
Si ðx; yÞ ¼ ð2Þ
0; Di ðx; yÞ\T
where Si ðx; yÞ is the effective human region, and Si ðx; yÞ Dnear ; Dfar .
in Fig. 3(a). The depth of the edge is greater than the depth of the middle. The convex
surface can be considered to be composed of an infinite number of circular arc curves,
and for a single circular arc curve, the slope of positive and negative changes only once
during the process from one side to the other. According to this feature, a CSHA-based
head feature discriminate method is proposed. The n points on the edge of the con-
vexity d1 ; d2 ; d3 . . .dn are selected as shown in Fig. 3(b). Then the n points are not
repeatedly linked to obtain n=2 circular arc curves, as shown in the colored curves in
Fig. 3(b), and the number of points that conform to the law of slope change is counted.
Finally the head convexity similarity Scon is calculated.
N
Scon ¼ ð3Þ
n=2
(a) (b)
Fig. 3. Circular arc convexity of head area: (a) circular arc convexity; (b) circular arc curves.
In the 2D image, an approximate ellipse target head region will also be obtained, as
shown in the red circle in Fig. 6. To simplify the calculation, the rectangular region of
M N is taken as the head Harea, as shown in the blue rectangle in Fig. 4. The pixels
in the rectangular area Hi ðx; yÞ can be represented as
1; Di ðx; yÞ 2 Harea
Hi ðx; yÞ ¼ ð4Þ
0; Di ðx; yÞ 62 Harea
P M1
N1 P
16 Hi ðx; yÞ
x¼0 y¼0
Ssqu ¼ ð5Þ
ðChead Þ2
where Chead is the perimeter of a rectangular region. When the head is more like a
standard circle, the rectangle resembles a square, and the value of Ssqu is closer to 1.
Finally, CSSH is defined by combining these two features, as shown in Formula (6).
where k1; k2 are the weight coefficients, and Tk 2 ½0; 1 is the threshold. When Scs Tk ,
the current area satisfies the CSSH feature. Otherwise, it does not satisfy the feature.
The spatial density of head (SDH) proposed in this paper is based on a fixed ratio
relationship between the head region of the human body and the space area of the
shoulder and head, as shown in Fig. 5. This feature does not change with height and
weight. When a person walks in the scene, there will be no huge deformation on the
shoulders and above. The spatial density of the head Dhead can be calculated with
Formula (7).
VHead
Dhead ¼ ð7Þ
VRec
where VRec is the volume of space where the human head is located; VHead is the
volume of the head, which can be approximated as the sphere volume, and it can be
calculated with Formula (8).
A Novel 3D Head Multi-feature Constraint Method 531
0v ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi13
u P
u Ti ðx; yÞ
4 B tðx;yÞ2hz C
VHead ¼ pB
@
C
A ð8Þ
3 p
When VHead Vk and Vk 2 ½0; 1, the current region satisfies the SDH feature, and
otherwise, it do not satisfy the feature.
The effects of over-fitting and under-fitting caused by training can be prevented by
combining the above two features, and the detection accuracy of the human body can
be improved.
After
determining the head area of the human body, the central point of the head region
lx ; ly is selected as the positioning point of the human body, as shown in formula (9).
P M1
N1 P
x Hi ðx; yÞ
x¼0 y¼0
lx ¼
P M1
N1 P
Hi ðx; yÞ
x¼0 y¼0
ð9Þ
P M1
N1 P
y Hi ðx; yÞ
x¼0 y¼0
ly ¼
P M1
N1 P
Hi ðx; yÞ
x¼0 y¼0
Since multiple cameras may detect the head region at the same time, it is necessary
to determine which camera data is valid. In this paper, the nearest center point method
is proposed to select an effective location point. The pixel distance between the center
point of the head and the center point of the image ðu0 ; v0 Þ is calculated, and the closest
point is selected as the final human body target location point, as shown in formula
(10). This can prevent the problem of incomplete depth data when the target deviates
from the center of the image, thus ensuring the accuracy of positioning information.
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
2ffi
D¼ ðlx u0 Þ2 þ ly v0 ð10Þ
In order to verify the stability of the system and the effectiveness of proposed algo-
rithm, the positioning test of the human target was performed in the actual space. In this
experiment, a total of 10 individuals were tested in sequence. Firstly, 5000 frames of
depth images were collected for training, including head-up, head-down, head-turn,
squat, and other movement states. Some of the positive and negative samples used in
this paper are shown in Figs. 6 and 7 respectively, and all training samples are nor-
malized to the same size ð32 32Þ. The algorithm is written in C++ and compiled and
run in Visual Studio 2013. The test computer uses an Intel core i5-4570 processor
clocked at 3.20 GHz. The five TOF depth cameras used in this paper have a resolution
of 320 240, and they are fixed at a height of 3 m above the ground. The lens of the
middle camera is parallel to the ground, and the remaining four cameras are oriented in
the four directions towards the east, west, south and north respectively, and have a
certain angle with the horizontal plane.
SVM classifier combining the SDH and CSSH features is used to determine the
human head region. The results are shown in Fig. 8, and the point on the head is the
positioning point of this target.
In addition, the detection effect of human detection algorithm combining HOD with
SVM is compared with the algorithm combining SDH and CSSH feature constrains, as
shown in Table 1.
It can be seen from Table 1 that the 3D head feature constraint method proposed in
this paper can effectively improve the detection precision of the human target, and the
total running time of the algorithm is controlled within 100 ms, e.g. 10 frames of depth
images can be processed per second.
In order to further verify the accuracy of location, the relationship between the
outputs of the 3D coordinates of the center of the head and the location in the actual
space is compared in this paper. Figure 9(a) shows the change in the height of a person
with 1.70 m height, and he is positioned in real time when walking in the space. The
results show that the height error is 3 cm, while Fig. 9(b) shows the 2D track of
human target is compared with the track of walking in real space. It can be found that
the positioning is more accurate near the central point (0, 0), while in the place close to
A Novel 3D Head Multi-feature Constraint Method 533
(a ) (b )
(c ) (d )
Fig. 8. Human location: (a) normal walking; (b) head-up; (c) head-down; (d) squat.
Table 1. Comparison of accuracy and real time of two algorithms for human detection
Algorithm Training Test False False Detection Running
samples samples positive negative rate time
rate rate
HOD + SVM 3000 270 4.67% 4.83% 95.33% 82 ms
HOD + SVM + SDH + CSSH 3000 350 2.87% 3.96% 97.13% 94 ms
534 F. Zhou et al.
(a) (b)
Fig. 9. Comparison between experimental measurements and actual values: (a) height distri-
bution of human body; (b) track of human head.
the corner, there is an obvious deviation. The main reason is that when people approach
the wall, the information of the target is not complete.
5 Conclusion
In order to address the drawbacks of algorithms for large scene space location, a 3D
head multi-feature constraint method based on multiple depth cameras is proposed in
this paper to detect and locate the human target. Multiple depth cameras are jointly
applied to obtain depth images of the human body and a large number of depth images
of human head are divided to extract the HOD features of the area. Then the head and
shoulder classifier is trained by SVM, and SDH and CSSH features are used to further
improve the accuracy of human detection and positioning. Finally, the nearest center
point method is used to determine the final human body positioning information.
Experimental results show that the system has high stability and competitive real-time
performance, and the proposed algorithm also has good robustness and accuracy,
which is of great potential to future specific applications.
References
1. Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: IEEE
Computer Society Conference on Computer Vision & Pattern Recognition, pp. 886–893
(2005)
2. Dalal, N., Triggs, B., Schmid, C.: Human detection using oriented histograms of flow and
appearance. In: Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006. LNCS, vol. 3952,
pp. 428–441. Springer, Heidelberg (2006). https://doi.org/10.1007/11744047_33
3. Schwartz, W.R., Kembhavi, A., Harwood, D., Davis, L.S.: Human detection using partial
least squares analysis. In: IEEE International Conference on Computer Vision, pp. 24–31
(2010)
A Novel 3D Head Multi-feature Constraint Method 535
4. Ma, J., Ren, F.: Detect and track the dynamic deformation human body with the active shape
model modified by motion mectors. In: IEEE International Conference on Cloud Computing
and Intelligence Systems, pp. 587–591 (2011)
5. Ikemura, S., Fujiyoshi, H.: Real-time human detection using relational depth similarity
features. In: Kimmel, R., Klette, R., Sugimoto, A. (eds.) ACCV 2010. LNCS, vol. 6495,
pp. 25–38. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-19282-1_3
6. Yu, S., Wu, S., Wang, L.: SLTP: a fast descriptor for people detection in depth images. In:
IEEE Ninth International Conference on Advanced Video and Signal-Based Surveillance,
pp. 43–47 (2012)
7. Wang, N., Gong, X., Liu, J.: A new depth descriptor for pedestrian detection in RGB-D
images, pp. 3688–3691 (2012)
8. Spinello, L., Arras, K.O.: People detection in RGB-D data. In: International Conference on
Intelligent Robots & Systems, vol. 38, no. 2, pp. 3838–3843 (2011)
9. Shen, Y., Hao, Z., Wang, P., Ma, S., Liu, W.: A novel human detection approach based on
depth map via kinect. In: IEEE Conference on Computer Vision and Pattern Recognition
Workshops, pp. 535–541 (2013)
10. Xia, L., Chen, C. C., Aggarwal, J.K.: Human detection using depth information by kinect.
In: Computer Vision and Pattern Recognition Workshops, pp. 15–22 (2011)
11. van Oosterhout, T., Bakkes, S., Kröse, B.: Head detection in stereo data for people counting
and segmentation (2011)
12. van Oosterhout, T., Englebienne, G., Kröse, B.: RARE: people detection in crowded
passages by range image reconstruction. Mach. Vis. Appl. 26(5), 561–573 (2015)
Head Ternary Pattern-Head Shoulder Density
Features Pedestrian Detection Algorithm
Based on Depth Image
1 Introduction
pedestrians, Mu et al. [4, 5] proposed two variants of LBP: Semantic-LBP (S-LBP) and
Fourier LBP (F-LBP). Considering the basic features detected by the color images are
susceptible to light, Cosmo et al. [6] used the relative stability of human body structure to
propose a color self-similarity (CSS) feature, which can be combined with the HOG
feature to greatly improve the detection performance.
In recent years, with the development of hardware technology, depth cameras can
capture three-dimensional information of the shooting scene with a greater range and
higher accuracy. In terms of the fact, depth cameras can meet the requirements of
system stability on the industrial site, and therefore have more advantages in obstacle
positioning and detection. Compared with the color image, the depth information only
depends on the three-dimensional features of the object, which has nothing to do with
the light intensity and contrast. All the above advantages let depth cameras bypass the
problems faced by existing pedestrian detection algorithms. Xia et al. [7] proposed a
template-matching method based on depth image for human detection, using Canny
operator for edge extraction, and then calculating template matching of depth image to
locate the human body. Spinello and Arras [8] attempted to use depth information
instead of gray information for calculating traditional HOG feature, and proposed a
Histogram of Oriented Depth (HOD). Zhang et al. [9] proposed the Water-fill algorithm
which is based on the flow characteristics of water in reality. This algorithm obtains the
local extremum of the head by imitating the flow direction of the water droplets and the
stopping point of the water droplets and has a high recognition rate. The literature [10]
relies on the relationship between the pixel points at the edge of the head region to
create a head template and to achieve head recognition.
Based on the above research, this paper proposes an HTP-HSD features pedestrian
detection algorithm based on depth vision collected by ToF cameras. After the pre-
process of original images, the HTP-HSD combined features is used for pedestrian
detection and localization.
The structure of this paper is organized as follows: this section is an introduction;
specific scheme of pedestrian detection is introduced in Sect. 2; experiments for the
algorithm and analyzes the results are conducted in Sect. 3; Sect. 4 is a summary.
SVM Object
Recognition
Based on these problems, spatial domain and morphological filtering in image pre-
processing can weaken the fluctuation of depth information, fill in the missing depth
information, and thus improve the quality of the image.
3D Dilation and Erosion Based on Probability Density
Traditional morphological processing mostly uses fixed templates to process images in
binary images. Therefore, when dealing with depth images with complex scenes, the
traditional morphological filtering has more limitations [11]. In this paper, a 3D dilation
and erosion method based on probability density is proposed, which is used to fill
missing depth information in depth images and remove isolated points and noises.
First, the depth histogram of the original image is counted, and then the depth
interval between the obstacle and the pedestrian is located. Second, the original image
is projected onto the corresponding depth interval DR . The standard deviation of the
normal distribution is ra and expected Dði; jÞ defined as follows:
P
i;j ½Pði; jÞ Dði; jÞ
Dðx; yÞ ¼ P ð1Þ
i;j Pði; jÞ
1
k ¼ pffiffiffiffiffiffi ð3Þ
2pra 1
Let the dilation factor a and the erosion factor b are defined as:
P
i;j Pði; jÞ Hði; j; ra Þ
a¼ ð4Þ
mn1
b¼1a ð5Þ
Head Ternary Pattern-Head Shoulder Density Features 539
Let TP and TF be thresholds of dilation and erosion respectively. When the dilation
factor of the invalid point a [ TP , the depth of the point will be filled according to the
Eq. (1). And when the erosion factor of valid point b [ TF , this point will be removed.
An adaptive threshold is obtained by learning the probability distribution of all pixels
with the dilation factor and the erosion factor in the neighborhood of the center point.
The application of this method can effectively suppress noise of the depth image after
depth projection and can fill in the black hole to improve the effect of preprocessing.
Increasing the number of iterations of 3D dilation and erosion can better fill in the black
holes, remove noise, and enhance the effect of image preprocessing in the depth image.
The experiments show that, in Fig. 2, the human profile is more obvious traditional
method. It is more conducive to the final identification and classification.
Fig. 2. Comparison of effects of different dilation and erosion methods. (a) the original image.
(b) the image after 3D Dilation and Erosion Based on Probability Density processing. (c) the
image after traditional template-based processing.
Scene Segmentation
The objects themselves have a certain spatial thickness, therefore the use of depth
information for segmentation can effectively separate the complex foreground and
background in the depth image. Because the interference around the object within a
single scene after segmentation is relatively small, it is beneficial to improve the
accuracy of obstacle location. The positioning of obstacles uses the following
equations:
P
X P
Y
i Dði; jÞ
i¼1 j¼1
Cx ¼ ð7Þ
P
X P
Y
Dði; jÞ
i¼1 j¼1
540 H. Wang et al.
P
X P
Y
j Dði; jÞ
i¼1 j¼1
Cy ¼ ð8Þ
P
X P
Y
Dði; jÞ
i¼1 j¼1
where Dði; jÞ is the depth value of the point ði; jÞ, X and Y respectively denotes the
total number of pixels in the horizontal and vertical directions. The region of the
obstacle can be completely located by using the region growing method.
The result of image segmentation using the depth information is shown in Fig. 3.
Fig. 3. The result of image segmentation. (a) the original depth image. (b) three images after
segmentation.
1; Dði; jÞ 2 DR
Pði; jÞ ¼ ð10Þ
0; Dði; jÞ 62 DR
DR denotes the depth interval where the object is located. The extraction result of
HSD feature is shown in Fig. 4.
The Area
of Head
The Area of
Pedestrian
In the positioned region of head, HTP feature are extracted from its outline.
Because the detected region and the number of pixels is small, the stability of the
gradient features can be improved by using the Sobel operator to calculate the gradient
of the depth around the center point. The extraction HTP feature is as shown in Fig. 5.
where
8
< 1; Dx [ Td
tx ¼ 0; jDxj Td ð12Þ
:
1; Dx Td
542 H. Wang et al.
8
< 1; Dy [ Td
ty ¼ 0; jDyj Td ð13Þ
:
1; Dy Td
in which Dxc ¼ ð2Dx þ Dðx 1Þ þ Dðx þ 1ÞÞ=4 is denoted as the horizontal gradient
and Dyc ¼ ð2Dy þ Dðy 1Þ þ Dðy þ 1ÞÞ=4 is denoted as the vertical gradient of the
center pixel calculated by Sobel operator and Td denotes the threshold for adjusting the
HTP feature.
Using the HTP-HSD feature can detect pedestrians accurately and effectively
because this feature is less affected by the environment. It is suitable for implementing
pedestrian detection in application scenarios with complex backgrounds. The ToF
camera directly obtains the depth information of each pixel in the scene. The adoption
of scale information avoids the disadvantage of the long time-consuming disadvantage
of the traditional image detection that requires the construction of an image pyramid to
obtain scale-invariant features.
wT x þ b ¼ 0 ð14Þ
where x denotes the point on the classification hyperplane, and w denotes the vector
perpendicular to the classification hyperplane, and b denotes the displacement which is
used to improve the flexibility of classified hyperplane. In order to make all the training
samples outside two parallel hyperplane spacing regions, it is necessary to ensure that
all samples x1 xn satisfies:
wT xi þ b 1; or
ð15Þ
wT xi þ b 1
According to the content of Sects. 2.1 and 2.2, the HTP-HSD features are extracted
from massive depth images collected by ToF cameras, which are preprocessed by 3D
Head Ternary Pattern-Head Shoulder Density Features 543
dilation and erosion based on probability density. HTP-HSD features are trained by the
SVM classifier to achieve pedestrian detection and classification.
3 Experiments
In order to verify the effectiveness of the algorithm, 2400 depth images collected in the
actual application scene are divided into three parts in this experiment, in which 1440
training samples are for SVM training, and 480 cross verified samples are for validating
model validity, and the rest 480 test samples are for testing the generalization capability
of the model. The results of some different positions about obstacles and humans are
shown below (Fig. 6):
Fig. 6. The original image, segmented image, HTP-HSD features image and result of pedestrian
detection in different situation. (a) pedestrian in front of an obstacle. (b) pedestrian between two
obstacles. (c) pedestrian image is incomplete. (d) pedestrian standing sideway.
After training 1440 training samples, the parameters of the SVM were obtained.
Then cross verified samples are used to verify that whether the model parameters are
reasonable. The precision is 98.81% and there is no false alarm which indicates that the
trained SVM model is effective. Test samples are used to test the generalization ability
of the pedestrian detection method proposed in this paper. The false detection rate is
0.74% which shows that the HTP-HSD pedestrian detection algorithm has better
generality and universal applicability. The result and analysis of experiments are shown
in Tables. 1 and 2.
544 H. Wang et al.
Precision, Recall, and F Score are usually used in machine learning to evaluate
classification models and algorithms. The definitions are as follows:
TruePositive
Precision ¼ ð16Þ
TruePositive þ FalsePositive
TruePositive
Recall ¼ ð17Þ
TruePositive þ FalseNegative
PR
FScore ¼ 2 ð18Þ
PþR
Experimental results show that the HTP-HSD features can be used for pedestrian
detection based on depth images. By analyzing the false detection, images that are false
alarmed or missed were taken during outdoor operations. It is inspected that the HTP
feature cannot be matched due to that the intense sunlight caused abnormal fluctua-
tions. In normal situations, the HTP-HSD feature can accurately classify pedestrian.
This method meets the requirements of high detection rate and low false alarm rate in
application scenarios.
The test platform is a personal computer with an i5-4570 CPU and the development
environment is Windows. The software development platform is Microsoft Visual
Studio 2013. The average time of each image is 35 ms, and 28 frames of depth images
can be processed every second.
4 Conclusion
References
1. Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: 2005 IEEE
Computer Society Conference on Computer Vision and Pattern Recognition, pp. 886–893.
IEEE, San Diego (2005)
2. Malysiak, D., Markard, M.: Increasing the efficiency of GPU-based HOG algorithms
through tile-images. In: Nguyen, N.T., Trawiński, B., Fujita, H., Hong, T.P. (eds.) ACIIDS
2016. LNCS (LNAI), vol. 9621, pp. 708–720. Springer, Heidelberg (2016). https://doi.org/
10.1007/978-3-662-49381-6_68
3. Liu, L., Fieguth, P., Wang, X., Pietikäinen, M., Hu, D.: Evaluation of LBP and deep texture
descriptors with a new robustness benchmark. In: Leibe, B., Matas, J., Sebe, N., Welling, M.
(eds.) ECCV 2016. LNCS, vol. 9907, pp. 69–86. Springer, Cham (2016). https://doi.org/10.
1007/978-3-319-46487-9_5
4. Xu, F., Zhang, J., Wang, J.Z.: Microexpression identification and categorization using a
facial dynamics map. IEEE Trans. Affect. Comput. 8, 254–267 (2017)
5. Mu, Y., Yan, S., Liu, Y., Huang, T., Zhou, B.: Discriminative local binary patterns for
human detection in personal album. In: 2008 IEEE Conference on Computer Vision and
Pattern Recognition, pp. 1–8. IEEE, Anchorage (2008)
6. Cosmo, D.L., Salles, E.O.T., Ciarelli, P.M.: Pedestrian detection system based on HOG and
a modified version of CSS. In: Seventh International Conference on Machine Vision (ICMV
2014), pp. 5–15. SPIE, Milan (2015)
7. Xia, L., Chen, C.C., Aggarwal, J.K.: Human detection using depth information by kinect. In:
2011 Conference on Computer Vision and Pattern Recognition, pp. 15–22. IEEE, Colorado
Springs (2011)
8. Spinello, L., Arras, K.O.: People detection in RGB-D data. In: 2011 International
Conference on Intelligent Robots and Systems, pp. 3838–3843. IEEE, San Francisco (2011)
9. Zhang, X., Yan, J., Feng, S., Lei, Z., Yi, D., Li, S.Z.: Water filling: unsupervised people
counting via vertical kinect sensor. In: 2012 IEEE Ninth International Conference on
Advanced Video and Signal-Based Surveillance, pp. 215–220. IEEE, Beijing (2012)
10. Hou, Y.-L., Pang, G.K.H.: Multi-cue-based crowd segmentation in stereo vision. In: Real,
P., Diaz-Pernil, D., Molina-Abril, H., Berciano, A., Kropatsch, W. (eds.) CAIP 2011. LNCS,
vol. 6854, pp. 93–101. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-
23672-3_12
11. Hasan, S.M.A., Ko, K.: Depth edge detection by image-based smoothing and morphological
operations. J. Comput. Des. Eng. 3, 191–197 (2016)
12. Qi, X., Fei, M., Hu, H., Wang, H.: A novel 3D expansion and corrosion method for human
detection based on depth information. In: Fei, M., Ma, S., Li, X., Sun, X., Jia, L., Su, Z.
(eds.) LSMS/ICSEE -2017. CCIS, vol. 761, pp. 556–565. Springer, Singapore (2017).
https://doi.org/10.1007/978-981-10-6370-1_55
13. Wang, X., Han, T.X., Yan, S.: An HOG-LBP human detector with partial occlusion
handling. In: 2009 IEEE 12th International Conference on Computer Vision, pp. 32–39.
IEEE, Kyoto (2009)
Maximally Stable Extremal Regions Improved
Tracking Algorithm Based on Depth Image
Abstract. In order to solve the problem that traditional Camshift algorithm can
easily fail to track overlapping targets and multiple similar depth targets, a new
improved maximally stable extremal regions (MSER) algorithm is presented in
this paper. Firstly, the suspected target contour is extracted and similarity
analysis is performed. Secondly, the improved MSER algorithm is used to
confirm the target contour and update the similarity library. Finally, combined
with the physical properties unique to the depth image and based on the Kalman
filter, it is possible to predict the tracking target’s moving position. The
experimental results show that the real-time performance and recognition rate
are improved, and robustness to the situation of target overlap and occlusion is
better with the improved MSER algorithm.
1 Introduction
Traditional digital image processing technology [1] has gradually matured, but it is
greatly affected by light, and meanwhile stability is difficult to control at times, which
greatly restricts its further development. The depth information of the target can be
reflected directly and identified with the structural information due to that the depth
image [2] is not affected by the light intensity nor the surface characteristics of the
object, and binocular imaging can be replaced by depth images, because of which more
and more applications are developed.
Nowadays, there are many classic methods, e.g. eigen pictures and eigen shapes
[3], using the depth information for target tracking, among which the continuous
adaptive mean shift (Camshift) [4] algorithm attracts much attention. The algorithm
effectively solves human target deformation and achieves better results in the simple
background environment. However, when there are multiple targets of similar depth in
the image, simple layering alone cannot be multi-targeted.
The MSER algorithm [5] marks a series of discernible detection areas as suspicious
target areas for the current frame image by processing the images, and determines the
exact boundary of the detected area. However, when the pixel change is not obvious,
the algorithm will not be able to extract valid information, resulting in tracking failure.
Kalman filter [6], which is low computational complex and better real time, is an
algorithm calculating the linear minimum variance estimate of the state sequence of the
dynamic system.
Based on the content above, a new MSER improved tracking algorithm based on
depth image is proposed in this paper. Firstly, the depth image is extracted and the
contour curve and local surface features [7] are extracted, and the similarity is analyzed.
The contour will be considered invalid and ignored when the similarity is low and
considered valid when the similarity is high. If the similarity stays in the middle, the
improved MSER algorithm will be used to determine the target contour, and to extract
the contour features of the confirmed target, and to update the similarity database, so as
to improve the accuracy of the similarity analysis. The Kalman filter at last is used to
predict the position of the target in the three-dimensional scene which keeps the target’s
motion consistent and coherent.
The depth image is usually converted into a grayscale image [8] for intuitive display,
but the meaning of the depth image and the ordinary grayscale image is obviously
different. Every point of the depth image represents the distance between the point and
the depth camera, which is a solid physical meaning that will be used to determine the
position of the target in a three-dimensional scene.
There are many popular methods to acquire depth image such as stereo matching
[9] method of a binocular camera, ToF method [10, 11] based on the principle of time-
of-flight, depth measurement method based on the structure light [12] and the moiré
fringe characteristics. These methods have their own strengths in different fields. In this
paper, depth information is collected with time-of-flight method, which has high real-
time performance more advantages compared to the other measurement methods, and is
less influenced by the environment and less complex.
Target depth
image
Extract target
outline
L M
Mser
Similarity
improved
analysis
algorithm
H
Throw away
Update
similarity
library
Kalman filter
Forecast
target
location
Target
tracking
X
K
½hi ðkÞ hj ðkÞ2
Ci;j ¼ Cðpi ; qj Þ ¼ 0:5 ð2Þ
k¼1
hi ðkÞ þ hj ðkÞ
where hj ðkÞ denotes the point qi shape histogram of the target Q what satisfies
Maximally Stable Extremal Regions Improved Tracking Algorithm 549
X
HðpÞ ¼ Cðpi ; qp ðiÞÞ ð3Þ
The larger the value of H is, the lower the similarity will be, and vice versa. Target
Depth map and outline map are simply shown in Fig. 2.
0 1
2 5 8 4 6
2 1
5 8
4 6
When using the recursive method to find the root node, the complexity of each
algorithm is oðnÞ, which greatly increases the overall algorithm’s consumption time.
Therefore, a path compression method is used to reduce the complexity (see Fig. 4).
When all the algorithms are completed, only the final connected component remains.
The neighborhood quadtree method is used to extract the MSER in order to save the
originally connected regions at different times in the image and improve the efficiency
of the algorithm.
3 3
1 1 2 3 1 2 3
1 1 1 2 1 2 1 2 1 2
3 3
Before nodes are merged, function Make Set is used to complete the creation of the
initial node, and the presence of neighboring nodes within the four-neighborhood of the
pixel should be checked. The invalid region is deleted via the stability condition, and
the maximum stable extremum region that meets the conditions is obtained.
When the areas are merged, all the extremum regions in the image are obtained, but
not all of the regions are the maximum stable extremum region. For invalid areas, the
specific removal rules are as follows:
(1) Removal of oversized areas: the threshold number of pixels occupied by the
largest area is set to 3000 through experimental analysis of the tested images pixel
numbers in this paper.
(2) Removal of small areas: Similar to the method of large areas, the analysis of
experimental results shows that the threshold number of pixels occupied by the
minimum area is set to 400.
(3) Removal of larger change rate areas and smaller area: The area change rate is set
according to (3.4), and the threshold change rate is set to (0.5, 1). Extremum areas
that exceed this range will be cleared. Denote Ra and Rb as the areas of adjacent
areas respectively that change with depth values and that satisfy
jjRa j jRb jj
0:5\ \1 ð4Þ
jRa j
When the maximum stable extremum area is found, the remaining area in the depth
image is considered as the moving object detection area. By performing the AND
Maximally Stable Extremal Regions Improved Tracking Algorithm 551
operation with the suspected target contour previously analyzed by similarity, the target
contour to be tracked can be confirmed.
Zk ¼ ½xk yk zk T ð6Þ
So the Kalman filter can be used to predict the motion of the target in the Z axis.
The highest point as a reference point is shown in Fig. 5.
552 H. Wang et al.
Camera
The highest
point G
Tracking goals
Overlapping and occlusion experiments were carried out using the ToF camera in order
to validate the accuracy of the improved MSER tracking algorithm. The ToF camera
was placed on the top of the testing area to collect the images of the moving targets.
Tracking effect images with two and three targets can be seen in Fig. 6.
The overlapped targets will be considered as one with traditional Camshift for that
the targets to be tracked cannot be distinguished from the disturbance of similar
characteristics and failed to be tracked when the target is sheltered (see Fig. 7). The
traditional Camshift algorithm works well when the targets do not overlap. However,
when the targets overlap, the target cannot be distinguished correctly, and the two
targets are mistaken as one. These problems are solved with an improved MSER
tracking algorithm in this paper where the targets can be precisely tracked when there
are many targets or the targets are sheltered (Tables 1 and 2).
5 Conclusion
In this paper, using the unique physical properties of depth images, an improved
tracking algorithm for MSER is proposed, and Kalman filter is used to achieve accurate
prediction and tracking of target locations. The target’s contour is extracted and con-
firmed iteratively. The accuracy of the similarity analysis is improved with the simi-
larity database update. However, this algorithm based on the contour curve and local
area feature matching in the depth image has a large amount of computation and
increases the complexity. At the same time, the effect is good in the stable area while
many problems remain in the complex ones which will be improved in the future
research.
References
1. Baqersad, M., Hamedi, A., Mohammadafzali, M., Ali, H.: Asphalt mixture segregation
detection: digital image processing approach. Adv. Mater. Sci. Eng. (2017)
2. Shotton, J., et al.: Efficient human pose estimation from single depth images. IEEE Trans.
Pattern Anal. Mach. Intell. 35, 2821–2840 (2013)
3. Biskup, M., Procaccia, E.B.: Eigenvalue vs perimeter in a shape theorem for self-interacting
random walks (2018)
4. Hou, H.N., Wu, T.: Pedestrian tracking algorithm based on camshift multi-feature adaptive
fusion. Electron. Qual. 3, 007 (2017)
5. Donoser, M., Bischof, H.: Efficient maximally stable extremal region (MSER) tracking. In:
2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition,
pp. 553–560 (2006)
6. Nguyen, T., Mann, G.K.I., Vardy, A., Gosine, R.G.: Developing a cubature multi-state
constraint Kalman filter for visual-inertial navigation system. In: Conference on Computer
and Robot Vision, pp. 321–328 (2017)
7. Lo, T.W.R., Siebert, J.P.: Local feature extraction and matching on range images: 2.5D
SIFT. Comput. Vis. Image Underst. 113, 1235–1250 (2009)
8. Qiao, B., Jin, L., Yang, Y.: An adaptive algorithm for grey image edge detection based on
grey correlation analysis. In: International Conference on Computational Intelligence and
Security, pp. 470–474 (2017)
9. Sun, C.: A Fast Stereo Matching Method. Digital Image Computing Techniques and
Applications, pp. 95–100 (1997)
10. Eichhardt, I., Jankó, Z., Chetverikov, D.: Novel methods for image-guided ToF depth
upsampling. In: IEEE International Conference on Systems, Man, and Cybernetics (2017)
11. Bevilacqua, A., Stefano, L.D., Azzari, P.: People tracking using a time-of-flight depth
sensor. In: IEEE International Conference on Video and Signal Based Surveillance, p. 89
(2006)
12. Sitnik, R.: New method of structure light measurement system calibration based on adaptive
and effective evaluation of 3D-phase distribution. In: Proceedings of SPIE - The
International Society for Optical Engineering, vol. 5856, pp. 109–117 (2005)
13. Salve, S.G., Jondhale, K.C.: Shape matching and object recognition using shape contexts. In:
IEEE International Conference on Computer Science and Information Technology, pp. 483–
507 (2010)
14. Astrachan, O.: Bubble sort: an archaeological algorithmic analysis. ACM (2003)
Dynamic Hand Gesture Recognition Based
on the Three-Dimensional Spatial Trajectory
Feature and Hidden Markov Model
1 Introduction
With the progress of science and technology and the development of computer
application technology, the traditional mouse, keyboard and other human-computer
interaction means have been unable to meet the requirements of current users.
Therefore, how to develop a natural, friendly and harmonious human-computer
interaction which conforms to the human’s behaviour becomes a hot issue in the
present research. In recent years, hand gesture recognition has attracted much attention
as one of the most natural and friendly human-computer interaction [1, 2]. And it has
been widely applied to various fields such as virtual reality, unmanned vehicles and
smart home, etc.
Hand gesture segmentation is the first and crucial step of the dynamic hand gesture
recognition. It means detecting and segmenting the motion hand region of interest
accurately from the complex background. The result of the hand gesture segmentation
will directly affect the recognition rate of dynamic hand gestures.
In this paper, a TOF camera with a resolution of 320 240 is used to obtain depth
images. Considering the actual situation, the hand gestures are closer to the camera than
the background and the human body. Therefore, the method based on the depth
threshold is usually applied for the hand gesture segmentation. In this paper, an
adaptive hand gesture segmentation method with the combination of the frame dif-
ference method and the depth threshold is put forward. This method not only over-
comes the influence of complex fore-background and different illumination, but also
realizes the adaptive hand gesture segmentation. The specific steps are as follows:
Step 1: Detect the depth image of the motion region of the current frame.
It ðx; yÞ jIt ðx; yÞ It1 ðx; yÞj [ T1
Dðx; yÞ ¼ ð1Þ
0 jIt ðx; yÞ It1 ðx; yÞj T1
Where It ðx; yÞ is the depth image of the current frame, It1 ðx; yÞ is the depth image
of the previous frame, Dðx; yÞ is the depth image of the motion region.
Step 2: Calculate the average depth d of the depth image of the motion region, and
take the average d as the depth threshold T for the hand gesture segmen-
tation, as shown in the formula (2):
1 X X
T¼d¼ Dðxi ; yj Þ ð2Þ
240 320 i¼240 j¼320
Where Tðx; yÞ is the binary image of the segmented hand gesture region, d is the
given error of the depth threshold.
The result of the hand gesture segmentation is shown in Fig. 1. Figure 1(a) shows
the depth image from the TOF camera. Figure 1(b) shows the segmented hand region
by the above method. It can be seen that the hand gesture region is completely
segmented.
558 K. Liu et al.
(a) (b)
Fig. 1. (a) The original depth image (b) the binary image of the segmented hand gesture region
Where m10 and m01 are the first-order geometric moments of the contour image and
m00 is the zero-order geometric moment of the contour image.
Then the trajectory points are extracted every three frames denoted as Pðx; y; zÞ,
where z represents the depth value of the point.
100 50
50 100
y
0
150
0
100 200
200
200
y 0 x 0 100 200 300
x
100 100
50 50
z
0 0
0 100 200 300 0 50 100 150 200
x y
Assuming that the hand gesture trajectory points obtained at t 1 and t are
Pt1 ðxt1 ; yt1
; zt1 Þ and Pt ðxt ; yt ; zt Þ respectively. The direction angle of Pt1 Pt is
defined as hxyt ; hyzt ; hxzt , where hxyt , hyzt , and hxzt respectively are the orientation
angles of the trajectory points Pt1 and Pt in the XOY, YOZ, and XOZ plane. The
calculation method is shown in formula (5), (6) and (7):
8
>
> arctanðyxtt y xt1 Þ þ p
t1
x2 x1 \0
<
yt yt1
hxyt ¼ arctanðxt xt1 Þ þ 2p y2 y1 \0 : ð5Þ
>
>
: arctanð t t1 Þ
y y
otherwise
xt xt1
9
8 10
7 11
5 1
2
4
3
Maximum probability
Recognition result
An HMM is trained for each given dynamic hand gestures, then the motion tra-
jectory direction angle feature vector of a sample is taken as the input observation
sequence of the HMM. Finally calculate PðOjki Þ (0 i V, where V is the number of
the defined dynamic hand gestures). The hand gesture of the corresponding HMM with
the highest probability is considered as the recognition result. The dynamic hand
gesture recognition flow chart is as shown in Fig. 4.
Since all the samples are treated uniformly by using the Baum-Welch classic
training algorithm, PðOjki Þ may eventually converge to the local optimal value
resulting in the poor recognition of some samples. On the basis of this circumstance,
some improvements have been made in this paper. The training process of the hand
gesture samples is linked with the test results of the model, and the HMM is corrected
by using the test results of the model. If some hand gesture samples are misrecognized
as other gestures, then those samples will be input into the corresponding HMM again
for a second training to correct the model parameters. Finally, after performing this
operation several times, the optimal parameters of the HMM is determined.
Based on the principles of the practicality, simplicity and identifiability of the hand
gestures and conforming to people’s habits, this paper defines four basic dynamic hand
gestures, as shown in Fig. 5, which are the trajectories of the four specific dynamic
hand gestures respectively. The small solid black circle represents the starting position
of the hand gesture, and the direction of the black solid arrow is the direction of the
motion trajectory of the dynamic hand gesture, wherein the gesture 4 represents
moving hand from far to close.
Fig. 5. The trajectories of the four specific dynamic hand gestures: (a) gesture1, (b) gesture2,
(c) gesture3, (d) gesture4
562 K. Liu et al.
In order to verify the accuracy of the algorithms in this paper, a real-time dynamic
hand gesture recognition experiment was performed. In the experiment, a total of ten
volunteers were arranged to stand in front of the 3D-TOF camera and repeated four
kinds of dynamic gestures shown above 20 times each. Therefore, each volunteer need
to complete 80 gestures and the total test number for each gesture is 200.
At the same time, the recognition accuracy of dynamic hand gesture based on the
3D trajectory features is evaluated by comparing with the 2D trajectory features used in
the past. The comparison result is shown in Fig. 6.
Fig. 6. Comparison of hand gesture recognition accuracy based on 3D trajectory features and
2D trajectory features.
Dynamic Hand Gesture Recognition 563
The experimental results show that when recognizing gesture 2, gesture 3 and
gesture 4, the recognition accuracy is certainly improved by extracting the 3D trajec-
tory features proposed in this paper instead of the previous 2D trajectory features. This
shows that the depth information does increase the recognition accuracy.
From the table above, it can be seen that there is no dramatic change of the
recognition rate of the dynamic hand gestures when the background and illumination
change. Therefore, the proposed algorithms for the dynamic hand gesture recognition
have a good robustness to the different backgrounds and illumination.
5 Conclusion
In this paper, the dynamic hand gesture recognition method based on 3D-TOF camera
is principally investigated. In the preprocessing stage, an adaptive hand gesture seg-
mentation method combining the frame difference method and depth threshold is
adopted to successfully overcome the influence of complex background and different
illumination. In the process of the dynamic hand gesture recognition, the 3D spatial
trajectory feature vectors of the dynamic gestures are input into the HMM for training
and recognition. In addition, the misidentified samples are trained for the second time
to improve the training method of the HMM. The experimental results show that the
dynamic gesture recognition algorithms based on 3D trajectory features and HMM are
very effective for improving the accuracy of the hand gesture recognition and
increasing the range of the hand gesture recognition, and have a good robustness to
different backgrounds and illumination.
In the actual human-computer interaction, more abundant hand gesture changes are
needed. At the same time, more 3D space trajectory features should be explored to
improve the accuracy and generalizability of the algorithm presented in this paper.
564 K. Liu et al.
References
1. Rautaray, S.S., Agrawal, A.: Vision based hand gesture recognition for human computer
interaction: a survey. Artif. Intell. Rev. 43, 1–54 (2015)
2. Song, Y., Demirdjian, D., Davis, R.: Continuous body and hand gesture recognition for
natural human-computer interaction. ACM Trans. Interact. Intell. Syst. (TiiS)-Spec. Issue
Affect. Interact. Nat. Environ. 2, 1–28 (2012)
3. Liu, K., Chen, C., Jafari, R., Kehtarnavaz, N.: Fusion of inertial and depth sensor data for
robust hand gesture recognition. IEEE Sens. J. 14, 1898–1903 (2014)
4. Munaro, M., Basso, A., Fossati, A., Gool, L.V., Menegatti, E.: 3D reconstruction of freely
moving persons for re-identification with a depth sensor. In: 2014 IEEE International
Conference on Robotics and Automation (ICRA), pp. 4512–4519. IEEE, Hong Kong (2014)
5. Hsieh, C.-C., Liou, D.-H.: Novel Haar features for real-time hand gesture recognition using
SVM. J. Real-Time Image Process. 10, 357–370 (2015)
6. Rahmani, H., Mahmood, A., Huynh, D.Q., Mian, A.: Real time action recognition using
histograms of depth gradients and random decision forests. In: IEEE Winter Conference on
Applications of Computer Vision, pp. 626–633. IEEE, Steamboat Springs (2014)
7. Yang, X., Zhang, C., Tian, Y.: Recognizing actions using depth motion maps-based
histograms of oriented gradients. In: Proceedings of the 20th ACM international conference
on Multimedia, pp. 1057–1060. ACM, Nara (2012)
8. Premaratne, P., Yang, S., Vial, P., Ifthikar, Z.: Dynamic hand gesture recognition using
centroid tracking. In: Huang, D.-S., Bevilacqua, V., Prashan, P. (eds.) ICIC 2015. LNCS,
vol. 9225, pp. 623–629. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-22180-
9_62
9. Oreifej, O., Liu, Z.: HON4D: histogram of oriented 4D normals for activity recognition from
depth sequences. In: 2013 IEEE Conference on Computer Vision and Pattern Recognition,
pp. 716–723. IEEE, Portland (2013)
10. Lubo, G., Xin, M., Haibo, W., Jason, G., Li, Y.: Chinese sign language recognition with 3D
hand motion trajectories and depth images. In: Proceeding of the 11th World Congress on
Intelligent Control and Automation, pp. 1457–1461. IEEE Shenyang (2014)
11. Yang, J., Yuan, J., Li, Y.F.: Flexible trajectory indexing for 3D motion recognition. In: 2015
IEEE Winter Conference on Applications of Computer Vision, pp. 326–332. IEEE,
Waikoloa (2015)
12. Corti, A., Giancola, S., Mainetti, G., Sala, R.: A metrological characterization of the kinect
V2 time-of-flight camera. Robot. Auton. Syst. 75, 584–594 (2016)
13. Li, L., Xiang, S., Yang, Y., Yu, L.: Multi-camera interference cancellation of time-of-flight
(TOF) cameras. In: 2015 IEEE International Conference on Image Processing (ICIP),
pp. 556–560. IEEE, Quebec City (2015)
14. Kumar, P., Gauba, H., Roy, P.P., Dogra, D.P.: Coupled HMM-based multi-sensor data
fusion for sign language recognition. Pattern Recogn. Lett. 86, 1–8 (2017)
15. Raheja, J.L., Minhas, M., Prashanth, D., Shah, T., Chaudhary, A.: Robust gesture recognition
using Kinect: a comparison between DTW and HMM. Opt.-Int. J. Light. Electron Opt. 126,
1098–1104 (2015)
16. Liu, K., Chen, C., Jafari, R., Kehtarnavaz, N.: Multi-HMM classification for hand gesture
recognition using two differing modality sensors. In: 2014 IEEE Dallas Circuits and Systems
Conference (DCAS), pp. 1–4. IEEE, Richardson (2014)
17. Zhang, X.H., Wang, J.J., Wang, X., Ma, X.L.: Improvement of dynamic hand gesture
recognition based on HMM algorithm. In: 2016 International Conference on Information
System and Artificial Intelligence (ISAI), pp. 401–406. IEEE, Hong Kong (2016)
Author Index