Handbook of Machine Learning for Computational Optimization: Applications and Case Studies (Demystifying Technologies for Computational Excellence) 1st Edition Vishal Jain (Editor) download
Handbook of Machine Learning for Computational Optimization: Applications and Case Studies (Demystifying Technologies for Computational Excellence) 1st Edition Vishal Jain (Editor) download
https://ebookmeta.com/product/international-human-rights-law-and-
practice-3rd-edition-ilias-bantekas/
Power and Popular Protest Susan Eckstein (Editor)
https://ebookmeta.com/product/power-and-popular-protest-susan-
eckstein-editor/
https://ebookmeta.com/product/android-programming-1st-edition-
bill-phillips-brian-hardy/
https://ebookmeta.com/product/the-gathering-quantum-
prophecy-2-michael-carroll/
https://ebookmeta.com/product/medical-internet-of-things-
techniques-practices-and-applications-1st-edition-anirban-mitra/
https://ebookmeta.com/product/building-a-career-in-cybersecurity-
the-strategy-and-skills-you-need-to-succeed-1st-edition-yuri-
diogenes/
Lives of the eminent philosophers Diogenes Laertius
https://ebookmeta.com/product/lives-of-the-eminent-philosophers-
diogenes-laertius/
Handbook of Machine
Learning for Computational
Optimization
Demystifying Technologies for
Computational Excellence: Moving
Towards Society 5.0
Series Editors
Vikram Bali and Vishal Bhatnagar
This series encompasses research work in the feld of Data Science, Edge Computing,
Deep Learning, Distributed Ledger Technology, Extended Reality, Quantum Computing,
Artifcial Intelligence, and various other related areas, such as natural language pro-
cessing and technologies, high-level computer vision, cognitive robotics, automated
reasoning, multivalent systems, symbolic learning theories and practice, knowledge rep-
resentation and the semantic web, intelligent tutoring systems, AI, and education.
The prime reason for developing and growing out this new book series is to focus
on the latest technological advancements – their impact on the society, the challenges
faced in implementation, and the drawbacks or reverse impact on the society due to
technological innovations. With the technological advancements, every individual has
personalized access to all the services, all devices connected with each other commu-
nicating among themselves, thanks to the technology for making our life simpler and
easier. These aspects will help us to overcome the drawbacks of the existing systems
and help in building new systems with latest technologies that will help the society in
various ways, proving Society 5.0 as one of the biggest revolutions in this era.
Edited by
Vishal Jain, Sapna Juneja, Abhinav Juneja, and
Ramani Kannan
First edition published 2022
by CRC Press
6000 Broken Sound Parkway NW, Suite 300, Boca Raton, FL 33487-2742
© 2022 selection and editorial matter, Vishal Jain, Sapna Juneja, Abhinav Juneja, and Ramani
Kannan; individual chapters, the contributors
Reasonable eforts have been made to publish reliable data and information, but the author and
publisher cannot assume responsibility for the validity of all materials or the consequences of
their use. Te authors and publishers have attempted to trace the copyright holders of all material
reproduced in this publication and apologize to copyright holders if permission to publish in this
form has not been obtained. If any copyright material has not been acknowledged please write and
let us know so we may rectify in any future reprint.
Except as permitted under U.S. Copyright Law, no part of this book may be reprinted, reproduced,
transmitted, or utilized in any form by any electronic, mechanical, or other means, now known or
hereafter invented, including photocopying, microflming, and recording, or in any information
storage or retrieval system, without written permission from the publishers.
For permission to photocopy or use material electronically from this work, access www.copyright.
com or contact the Copyright Clearance Center, Inc. (CCC), 222 Rosewood Drive, Danvers, MA
01923, 978-750-8400. For works that are not available on CCC please contact mpkbookspermissions@
tandf.co.uk
Trademark notice: Product or corporate names may be trademarks or registered trademarks and are
used only for identifcation and explanation without intent to infringe.
DOI: 10.1201/9781003138020
Typeset in Times
by codeMantra
Contents
Preface...................................................................................................................... vii
Editors....................................................................................................................... xi
Contributors ............................................................................................................xiii
v
vi Contents
Index...................................................................................................................... 279
Preface
Machine learning is a trusted technology over decades and has fourished on a
global scale touching the lives of each one of us. The modern-day decision making
and processes are all dependent on machine learning technology to make matured
short-term and long-term decisions. Machine learning is blessed to have phenomenal
support from the research community, and have landmark contributions, which is
enabling machine learning to fnd new applications every day. The dependency of
human processes on machine learning-driven systems is encompassing all spheres of
current state-of-the-art systems with the level of reliability it offers. There is a huge
potential in this domain to make the best use of machines in order to ensure the opti-
mal prediction, execution, and decision making. Although machine learning is not a
new feld, it has evolved with ages and the research community round the globe have
made remarkable contribution for the growth and trust of applications to incorporate
it. The predictive and futuristic approach, which is associated with machine learning,
makes it a promising tool for business processes as a sustainable solution. There is
an ample scope in the technology to propose and devise newer algorithms, which are
more effcient and reliable to give machine learning an entirely new dimension in dis-
covering certain latent domains of applications, it may support. This book will look
forward to addressing the issues, which can resolve the modern-day computational
bottom lines which need smarter and optimal machine learning-based intervention
to make processes even more effcient. This book presents innovative and improvised
machine learning techniques which can complement, enrich, and optimize the exist-
ing glossary of machine learning methods. This book also has contributions focusing
on the application-based innovative optimized machine learning solutions, which
will give the readers a vision of how innovation using machine learning may aid in
the optimization of human and business processes.
We have tried to knit this book as a read for all books wherein the learners and
researchers shall get insights about the possible dimensions to explore in their spe-
cifc areas of interest. The chapter-wise description is as follows:
Chapter 1 explores the basic concepts of random variables (single and multiple),
their role and applications in the specifed areas of machine learning.
Chapter 2 demonstrates Wigner-Ville transformation technique to extract the
time-frequency domain features from typical and atypical EMG signals – myopathy
(muscle disorder) and amyotrophic lateral sclerosis (neuro disorder). Nature inspired
feature selection algorithms, whale optimization algorithm (WOA), genetic algo-
rithm (GA), bat algorithm (BA), fre-fy optimization (FA), and particle swarm opti-
mization (PSO) are utilized to determine the relevant features from the constructed
features.
Chapter 3 presents various algorithms of machine learning (ML), which can be
used for breast cancer detection. Since these techniques are commonly used in many
areas, they are also used for making decisions regarding diagnostic diagnosis and
clinical studies.
vii
viii Preface
Chapter 4 measures the effciency and thoroughly explores the scope for opti-
mal utilization of the input resources owned by depots of the RSRTC. The new
slack model (NSM) of DEA is used as it enumerates the slacks for input and output
variables. The model satisfes the radial properties, unit invariance, and translation
invariance. This study enables policy-makers to evaluate inputs for consistent output
up to the optimum level and improve the performance of the ineffcient depots.
Chapter 5 presents a binary error control coding scheme using weight-based
codes. This method is quite used for classifcation and employs the K nearest neigh-
bor algorithms. The paper also discussed the role of distance matrix with hamming
code evaluation.
Chapter 6 exhibits MRI images of the framed brain to create deep neural system
models that can be isolated between different types of heart tumors. To perform this
task, deep learning is used. It is a type of instrument-based learning where the lower
levels responsible for many types of higher-level defnitions appear above the differ-
ent levels of the screen.
Chapter 7 focuses on creating an affordable and effective warning system for driv-
ers that is able to detect the warning sign boards and speed limits in front of the mov-
ing vehicle, and prompt the driver to lower to safer speeds if required. The software
internally works on a deep learning-based modern neural network YOLO (You Only
Look Once) with certain modifcations, which allows it to detect the road signs really
quickly and accurately on low-powered ARM CPUs.
Chapter 8 presents an approach for the classifcation of lung cancer based on the
associated risks (high risk, low risk, high risk). The study was conducted using a lung
cancer classifcation scheme by studying micrographs and classifying them into a
deep neural network using machine learning (ML) framework.
Chapter 9 presents a statistical feedback evaluation system that allows to design an
effective questionnaire using statistical knowledge of the text. In this questionnaire,
questions and their weight are not pre-decided. It is established that questionnaire-
based feedback systems are traditional and quite straightforward, but these systems
are very static and restrictive. The proposed statistical feedback evaluation system
is helpful to the users and manufacturers in fnding the appropriate item as per their
choices.
Chapter 10 presents an experimental work based on the data collected on various
parameters on the scientifc measuring analytical software tools Air Veda instru-
ment and IoT-based sensors capturing the humidity and temperature data from atmo-
spheric air in certain interval of time to know the patterns of pollution increment or
decrement in atmosphere of nearby area.
Chapter 11 concerns with neural network representations and defning suitable
problems for neural network learning. It covers numerous substitute designs for the
primitive units making up an artifcial neural network, such as perceptron units,
sigmoid unit, and linear units. This chapter also covers the learning algorithms for
training single units. Backpropagation algorithm for multilayer perceptron training is
described in detail. Also, the general issues such as the representational capabilities
of ANNs, overftting problems, and substitutes to the backpropagation algorithm are
also explained.
Preface ix
Chapter 12 proposes a system which will make use of the machine learning
approach to predict a student’s performance. Based on student’s current performance
and some measurable past attributes, the end result can be predicted to classify them
among good or bad performers. The predictive models will make students aware who
are likely to struggle during the fnal examinations.
Chapter 13 presents a study that assists in assessing the awareness status of people
on the TB towards its mitigation and serves as contribution to the feld of health infor-
matics. Indeed, the majority of participants claimed that they had low awareness on
the TB and its associated issues in their communities. Though, the participants were
from Kano state, a strategic location in the northern part of Nigeria, which means
that the result of the experiment can represent major opinions of northern residents.
Chapter 14 deals with psychological data related to depression, anxiety, and stress
data to study how the classifcation and analysis is carried out on imbalanced data.
The proposed work not only contributes on providing practical information about
the balancing techniques like SMOTE, but also reveals the strategy for dealing
with working of many existing classifcation algorithms like SVM, Random Forest,
XGBoost etc. on imbalanced dataset.
Chapter 15 proposes the construction of segmented mask of MRI (magnetic reso-
nance image) using CNN approach with the implementation of ResNet framework.
The understanding of ResNet framework using layered approach will provide the
extensive anatomical information of higher-dimensional image for precise clinical
analysis for curative treatment of patients.
Editors
Vishal Jain is an Associate Professor in the Department of CSE at Sharda University,
Greater Noida, India. He has earlier worked with Bharati Vidyapeeth’s Institute of
Computer Applications and Management (BVICAM), New Delhi, India (affliated
with Guru Gobind Singh Indraprastha University, and accredited by the All India
Council for Technical Education). He frst joined BVICAM as an Assistant Professor.
Before that, he has worked for several years at the Guru Premsukh Memorial College
of Engineering, Delhi, India. He has more than 350 research citation indices with
Google scholar (h-index score 9 and i-10 index 9). He has authored more than 70
research papers in reputed conferences and journals, including Web of Science and
Scopus. He has authored and edited more than 10 books with various reputed pub-
lishers, including Springer, Apple Academic Press, Scrivener, Emerald, and IGI-
Global. His research areas include information retrieval, semantic web, ontology
engineering, data mining, ad hoc networks, and sensor networks. He was recipient of
a Young Active Member Award for the year 2012–2013 from the Computer Society
of India, Best Faculty Award for the year 2017, and Best Researcher Award for the
year 2019 from BVICAM, New Delhi.
Sapna Juneja is a Professor in IMS, Ghaziabad, India. Earlier, she has worked as
a Professor in the Department of CSE at IITM Group of Institutions and BMIET,
Sonepat. She has more than 16 years of teaching experience. She completed her
doctorate and masters in Computer Science and Engineering from M.D. University,
Rohtak, in 2018 and 2010, respectively. Her broad area of research is Software
Reliability of Embedded System. Her areas of interest include Software Engineering,
Computer Networks, Operating System, Database Management Systems, and
Artifcial Intelligence etc. She has guided several research theses of UG and PG
students in Computer Science and Engineering. She is editing book on recent tech-
nological developments.
xi
xii Editors
Ramani Kannan is currently working as a Senior Lecturer, Center for Smart Grid
Energy Research, Institute of Autonomous system, University Teknologi PETRONAS
(UTP), Malaysia. Dr. Kannan completed Ph.D. (Power Electronics and Drives) from
Anna University, India, in 2012; M.E. (Power Electronics and Drives) from Anna
University, India, in 2006; B.E. (Electronics and Communication) from Bharathiyar
University, India, in 2004. He has more than 15 years of experience in prestigious
educational institutes. Dr. Kannan has published more than 130 papers in various
reputed national and international journals and conferences. He is the editor, co-edi-
tor, guest editor, and reviewer of various books, including Springer Nature, Elsevier
etc. He has received award for best presenter in CENCON 2019, IEEE Conference on
Energy Conversion (CENCON 2019), Indonesia.
Contributors
Shivi Agarwal Renu Jain
Department of Mathematics University Institute of Engineering and
BITS Pilani Technology
Pilani, Rajasthan, India CSJM University
Kanpur, Uttar Pradesh, India
Amala Ann K. A.
Data Science Department Abhinav Juneja
CHRIST (Deemed to be University) KIET Group of Institutions
Bangalore, India Ghaziabad, Uttar Pradesh, India
xiii
xiv Contributors
CONTENTS
1.1 Introduction ......................................................................................................2
1.2 Random Variable ..............................................................................................3
1.2.1 Defnition and Classifcation.................................................................3
1.2.1.1 Applications in Machine Learning ........................................4
1.2.2 Describing a Random Variable in Terms of Probabilities....................4
1.2.2.1 Ambiguity with Reference to Continuous
Random Variable ...................................................................5
1.2.3 Probability Density Function................................................................6
1.2.3.1 Properties of pdf ....................................................................6
1.2.3.2 Applications in Machine Learning ........................................7
1.3 Various Random Variables Used in Machine Learning...................................7
1.3.1 Continuous Random Variables .............................................................7
1.3.1.1 Uniform Random Variable.....................................................7
1.3.1.2 Gaussian (Normal) Random Variable....................................8
1.3.2 Discrete Random Variables ................................................................ 10
1.3.2.1 Bernoulli Random Variable ................................................. 10
1.3.2.2 Binomial Random Variable ................................................. 11
1.3.2.3 Poisson Random Variable .................................................... 12
1.4 Moments of Random Variable........................................................................ 13
1.4.1 Moments about Origin........................................................................ 13
1.4.1.1 Applications in Machine Learning ...................................... 13
1.4.2 Moments about Mean ......................................................................... 14
1.4.2.1 Applications in Machine Learning ...................................... 14
1.5 Standardized Random Variable...................................................................... 15
1.5.1 Applications in Machine Learning..................................................... 15
1.6 Multiple Random Variables ............................................................................ 16
1.6.1 Joint Random Variables...................................................................... 17
1.6.1.1 Joint Cumulative Distribution Function (Joint CDF)........... 17
1.6.1.2 Joint Probability Density Function (Joint pdf) .................... 17
1.6.1.3 Statistically Independent Random Variables ....................... 18
1.6.1.4 Density of Sum of Independent Random Variables............. 18
1.6.1.5 Central Limit Theorem ........................................................ 19
DOI: 10.1201/9781003138020-1 1
2 Handbook of Machine Learning
1.1 INTRODUCTION
Predicting the future using the knowledge about the past is the fundamental objective
of machine learning.
In a digital communication system, a binary data generation scheme referred to as
differential pulse code modulation (DPCM) works on the similar principle, where,
based on the past behaviour of the signal, its future value will be predicted, using a
predictor. A tapped delay line flter serves the purpose. More is the order of the pre-
dictor, better is the prediction, i.e. less is the prediction error.[1]
Thus, machine learning, even though not being referred to by this name earlier,
was/is an integral part of technical world.
The same prediction error with reference to a DPCM system is now being addressed
as confdence interval in connection with machine learning. Less prediction error
implies a better prediction, and as far as machine learning is concerned, the prob-
ability of the predicted value to be within the tolerable limits of error (which is the
confdence interval) should be large, which is a metric for the accuracy of prediction.
The machine learning methodology involves the process of building a statistical
model for a particular task, based on the knowledge of the past data. This collected
past data with reference to a task is referred to as data set.
This way of developing the models to predict about ‘what is going to happen’,
based on the ‘happened’, is predictive modelling.
In detective analysis also, ‘Happened’, i.e. past data, is used, but there is no neces-
sity of predicting about ‘Going to happen’.
For example, 30–35 years back, Reynolds-045 Pen ruled the market for a long
time, specifcally in South India. Presently, the sales are not that much signifcant.
If it is required to study the journey of the pen from past to present, detective anal-
ysis is to be performed, since there is no necessity of predicting its future sales.
Similarly, a study of ‘Why the sales of a particular model of an automobile vehicle
came down?’ also belongs to the same category.
The data set referred above is used by the machine to learn, and hence, it is also
referred to as training data set. After learning, the machine faces the test data. Using
the knowledge the machine gained through learning, it should act on the test data to
resolve the task assigned.
In predictive modelling, if the learning mechanism of the machine is supervised
by somebody, then the mode of learning is referred to as supervised learning. That
supervising ‘somebody’ is the training data set, also referred to as labelled training
data, where each labelled data element such as Di is mapped to a data element D0 .
Such many pairs of elements are the learning resources for the machine, and are used
Random Variables in Machine Learning 3
to build the model, using which the machine predicts, i.e. this knowledge about the
mapped will help the machine to map the test data pairs (input-output pair).
It can be inferred about the supervised learning that there is a target variable
which is to be predicted.
Example: Based on the symptoms of a patient, it is to be predicted whether
he/she is suffering from a particular disease. To enable this prediction, the past
history or statistics such as patients with what symptoms (similar) were categorized
under what disease. This past data (both symptoms and categorization) is the train-
ing data set that supervises the machine in its process of prediction. Here, the target
variable is the disease of the patient, which is to be predicted.
In unsupervised learning mechanism, the training data is considered to be
unlabelled, i.e. only Di. Major functionality of unsupervised learning is pattern
identifcation.
Some of the tasks under unsupervised learning are:
Clustering: Group all the people wearing white (near white) shirts.
Density Estimation: If points are randomly distributed along an axis, the
regions along the axis with minimum/moderate/maximum number points
need to be estimated.
It can be inferred about the unsupervised learning that there is no target variable
which is to be predicted.
With reference to previous example of patient with ill-health, it can be told that
all the people with a particular symptom of ill-health need to be grouped; however,
disease of the patient need not to be predicted, which is the target variable, with refer-
ence to supervised learning.
Table 1.1 gives the pairs of all possible outcomes with the corresponding real values
mapped.
4 Handbook of Machine Learning
TABLE 1.1
Sample Space and Mapped Values
Pair in the Sample Space Real Value
(1,1) 2
(1,2), (2,1) 3
(1,3),(2,2),(3,1) 4
(1,4),(2,3),(3,2),(4,1) 5
(1,5),(2,4),(3,3),(4,2),(5,1) 6
(1,6),(2,5),(3,4),(4,3),(5,2),(6,1) 7
(2,6),(3,5),(4,4),(5,3),(6,2) 8
(3,6),(4,5),(5,4),(6,3) 9
(4,6),(5,5),(6,4) 10
(5,6),(6,5) 11
(6,6) 12
This X is referred to as random variable taking all the real values as mentioned.
Thus, a random variable can be considered as a rule by which a real value is
assigned to each outcome of the experiment.
If X = f (˜ ) possesses countably infnite range (points in the range are large in
number, but can be counted), X is referred to as discrete random variable (categorical
with reference to machine learning).
On the other hand, if the range of the function is uncountably infnite (large in
number, which can’t be counted), X is referred to as continuous random variable[2,3]
(similar terminology is used with reference to machine learning also).
( i ) 0 ˝ P ( X = n ) ˝ 1 ( ii ) ˜P ( X = n) = 1
n
TABLE 1.2
Probability Distribution
xi ( value taken by X ) 1 2 3 4 –
Probability 1 1 1 1 –
2 4 8 16
6 Handbook of Machine Learning
P ( x1 < X ˝ x 2 ) = FX ( x2 ) − FX ( x1 )
P( x ˙ X < x + ˜ )
f X ( x ) = lim
˜ ˘0 ˜
The certainty or the probability with which X is in the interval ( x , x + ˜ ) is
P( x ° X < x + ˜ ), and the denominator δ is the width (length) of the interval.
Thus, f X ( x ) is the probability normalized by the width of the interval and can be
interpreted as probability divided by width.
From the properties of CDF, P ( x ˝ X < x + ˜ ) = FX ( x + ˜ ) − FX ( x ).
FX ( x + ˜ ) − FX ( x ) d
Hence, f X ( x ) = lim = FX ( x ), i.e. change in CDF is referred
˜ ˇ0 ˜ dx
to as pdf.[2]
f (x) ˛ 0
ˆ
˜ −ˆ
f ( x ) dx = 1
x
FX ( x ) =
˜−ˇ
f (° ) d °
Random Variables in Machine Learning 7
˜a
f ( x ) dx = P ( a < X < b )
The physical signifcance of uniform density is that the random variable X can
1
lie in any interval within the limits (˜ , ° ) with the same probability, i.e. ,
˜ −°
for any confdence interval ( k , k + ˜ ) , where ° < k < ˛ , the confidence level
1
P (˜ < X < x + ° ) = .
˛ −˜
• Any confdence level P(˜ 1 < X < °1 ) for the given confdence interval
°1
1
(˜ , ° ) can be obtained as˜˛1 ° − ˛
. dx
• The CDF of uniformly distributed random variable X (continuous) is
0 for x < ˜
x −˜
FX ( x ) = for ˜ < x ˘ °
° −˜
1 for x > °
• A uniform random variable is said to be symmetric about its mean
˙ ˜ + ° ˘ [6]
ˇˆ = .
2
2˙˜ 2 ˘ 2˜ 2
variance of X, respectively.
Random Variables in Machine Learning 9
• Single fip of a coin, where the possible outcomes are head and tail
• Identifying about the survival of a person in an accident: survived or not
only two values either 0 or 1. It can assume any one value at a time. p and q ( = 1 − p )
are the probability of success and failure, respectively. Success is the required whose
probability is to be computed.
For Example: when an unbiased coin is tossed, if it is required to compute the
1 1−1
˝ 1ˇ ˝ 1ˇ 1
probability of getting a head (represented as m = 1), then P ( X = 1) = ˆ ˆ = .
˙ 2˘ ˙ 2˘ 2
Here, success is getting a head.
˘0 for x < 0
The CDF of Bernoulli random variable is FX ( x ) = q for 0 ˙ x < 1[6]
p + q = 1 for x ˇ 1
• When the variable to be predicted is binary valued, i.e. the test data is to be
categorized into any one of the two classes available, such classifcation is
binary classifcation and the performance of such algorithms can be anal-
ysed using Bernoulli process.[5]
respectively, such that p + q = 1. Here, X is the number of times of having the success
in the experiment.
Here, X is the binomial random variable, since the above probability is the coef-
fcient of kth term in the binomial expansion of ( p + q ) .
r
If it is required to fnd the probability for tail to occur four times, when a fair coin
4 6
˝ 1ˇ ˝ 1ˇ
is tossed ten times, then such probability is p ( X = 4 ) = 10 4 ˆ ˆ .
˙ 2˘ ˙ 2˘
The probability of having success for ‘k’ times in ‘r’ number of trials of the exper-
ˆrck ( p ) ( q )
r−k
˝ k
for k = 0,1,2,r
iment in a random order is p ( X = k ) = ˙ , and this
ˆ0
ˇ otherwise
is the pmf of X.
m
TABLE 1.3
Variable Encoding
Outlet Size Small Medium Big
1 Big 0 0 1
2 Big 0 0 1
3 Medium 0 1 0
4 Small 1 0 0
5 Medium 0 1 0
Random Variables in Machine Learning 13
˜( s − si )
^ 2
i
2 1 2 1
is defned. Its value for the above X is ( 2 − 4 ) + ( 4 − 6 ) = 4. Its positive square
2 2
root is 2, which is the value required.
Thus, E ˝˙( X − m )2 ˆˇ is an indication of the average amount of variation of the val-
ues taken by the random variable with reference to its mean, and hence, its variance
(σ 2), and standard deviation (σ) is its positive square root.
The third central moment of X is E ˝˙( X − m )3 ˆˇ and is referred to as its skew. The
E ˙( X − m )3 ˇ˘
normalized skew, i.e. coeffcient of skewness, is given as ˆ and is a mea-
˜3
sure of symmetry of the density of X.
A random variable with symmetric density about its mean will have zero coef-
fcient of skewness.
If more values taken by the random variable are to the right of its mean, the
corresponding density function is said to be right-skewed and the respective
coeffcient of skewness will be positive (>0). Similarly, the left-skewed density
function can also be specifed and the respective coeffcient of skewness will be
negative (<0).[1]
** Identically distributed random variables will have identical moments.
TABLE 1.4
Data in Nonuniform Scaling
Loan amount EMI Income
(Rs. in lakhs) (Rs. in thousands) (Rs. in hundreds)
30 50 1600
40 40 1200
50 30 800
16 Handbook of Machine Learning
TABLE 1.5
Statistical Parameters of the Data
Standard deviation (σ) (amount by which value
Variable Mean (m) taken by the variable differs from its mean)
Loan amount 40 10
EMI 40 10
Income 1200 400
TABLE 1.6
Scaled Data
Loan Amount EMI Income
(Rs. in lakhs) (Rs. in thousands) (Rs. in hundreds)
30 − 40 50 − 40 1600 − 1200
= −1 =1 =1
10 10 400
40 − 40 40 − 40 1200 − 1200
=0 =0 =0
10 −10 400
50 − 40 30 − 40 800 − 1200
=1 = −1 = −1
10 −10 400
• All these data are subjected to scaling to make them to be on the same scale,
such that they are comparable. Thus, feature pre-processing requires scal-
ing of the variables.
• Standard scaling is a scaling method, where the scaling of variables is done
X−m
as per the formula X ° = .
˜
• Table 1.5 gives the computations of mean and standard deviation for differ-
ent variables of Table 1.4.
• Table 1.6 represents the above data subjected to scaling.
• It appears that all the scaled variables appear on the same reference scale.
• Thus, the concept of standardized random variable is used in feature
scaling.
1.6.1.1.1 Properties
1. FXY ( −˝, −˝ ) = 0
2. FXY ( ˛, ˛ ) = 1
3. FXY ( −˝, y ) = 0
4. FXY ( x, −˝ ) = 0
+ FXY ( x1 , y1 )
8. 0 ˛ FXY ( x, y ) ˛ 1
ˆ ˆ
1.
˜ ˜
−ˆ −ˆ
f XY ( x, y ) dx dy = 1
ˆ
2.
˜ y=−ˆ
f XY ( x, y ) dy = f ( x ) , which is the marginal density of X
ˆ
3.
˜ x=−ˆ
f XY ( x, y ) dx = f ( y ) , which is the marginal density of Y
x y
4.
˜ ˜
−ˆ −ˆ
f XY ( x , y ) dx dy = FXY ( x , y )
x2 y2
5. P ( x1 < X < x 2 , y1 < Y < y2 ) =
˜ ˜
x1 y1
f XY ( x , y ) dx dy
18 Handbook of Machine Learning
x ˆ
6. FX ( x ) =
˜ ˜
−ˆ y=−ˆ
f XY ( x , y ) dx dy
ˆ y
7. FY ( y ) =
˜ ˜
x=−ˆ −ˆ
f XY ( x , y ) dx dy [12]
XY ˜1 ˜2 … ˜n
°1 p (° 1 , ˜1 ) p (° 1 , ˜ 2 ) … p (° 1 , ˜ n )
P ( X ,Y ) = °2 p (° 2 , ˜1 ) ° … p (° 2 , ˜ n )
° ° ° … °
°n p (° n , ˜1 ) p (° n , ˜ 2 ) … p (° n , ˜ n )
3. ˜p(° , ˛ ) = p( ˛ )
i
i j j
4. ˜p(° , ˛ ) = p(° )
j
i j i
• FXY ( x, y ) = FX ( x ) FY ( y )
• f XY ( x , y ) = f X ( x ) fY ( y )
ˆ ˆ
f Z ( z ) = f X ( x ) * fY ( y ) =
˜ −ˆ
f X ( x ) fY ( z − x ) dx =
˜
−ˆ
f X ( z − y ) fY ( y ) dy, which is
convolution of their individual density functions.
This principle can be extended to multiple number of independent random vari-
ables also.
For two random variables X and Y, there will be three number of second-order
joint moments.
They are
X − mx Y − m y E ( X − mx )(Y − m y ) σ
ρ XY = E [ X ′Y ′ ] = E = = XY
σ x σ y σ xσ y σ xσ y
• Thus, correlation coeffcient can be used as the metric for the measure
of the linear relation between the variables.
• Variance and standard deviation are the measures of the spread of the data
set around its mean and are one-dimensional measure. When dealing with
data sets with two dimensions, the relation between these two variables
in the data set (e.g. number of hours spent by a tailor in stitching and the
number of shirts stitched can be the variables) i.e. the statistical analysis
between the two variables will be studied using covariance. Covariance for
one random variable is nothing but its variance. In the case of n variables,
covariance matrix [ n X n ] is used for the statistical analysis of all the pos-
sible pairs of the variables.[9,13,14]
( )
other event being the condition is denoted as p A B =
( ˜ B) = p( A, B) , which
p A
P ( B) p( B)
is the probability of event A, under the occurrence of the event B.
( )
1. f X x y is always non-negative
Y
( )
ˆ
2.
˜
−ˆ
f X x y dx = 1
Y
Similar properties hold good for discrete variables also, but defned under discrete
summation.[15]
( )
p B A p( A)
( )
• Baye’s theorem is stated as p A B =
p( B)
.
• Let the data set consist of various symptoms leading to corona/malaria/
typhoid.
Random Variables in Machine Learning 23
( )
• Then, the probability p having a specific disease Symptom S , i.e. having a
specifc symptom S, the probability of suffering from a specifc disease is
the conditional probability.
• If all the hypotheses are of equal a priori probability, then the above condi-
tional probability can be obtained from the probability of having those symp-
(
toms, knowing the disease, i.e. p Symptom S having the specific disease . )
This probability is referred to as maximum likelihood (ML) of the specifc
hypothesis.[7,9]
• Then, the required conditional probability is
(
p having a specific disease Symptom S )
=
( )
p Symptom S having the specific disease ˝ p ( having the specific disease )
p ( Symptom S )
f ( y) =
i
˜ dx
f ( x ) | x = xi . i [3]
dy
TABLE 1.7
Linear Regression Transformation of Variables[10]
Nonlinear
Relations Reduced to Linear Law
˜ = p° n log ( ˜ ) = log ( p ) + n ˇ log (° ) ˘ Y = nX + C , with Y = log ( ˜ ) , X = log (° ) ,C = log ( x )
˜ = m° n + C Y = mX + C, with X = ˜ n , Y = °
˜ = p° n + q.log (° ) ˜ °n
Y = aX + b, with Y = ,X = , a = p, b = q
log (° ) log (° )
˜ = pe q˛ Y = m˜ + c, with Y = log ( ° ) , m = q ˇ log ( e ) ,
c = log ( p )
1.8 CONCLUSION
Thus, random variables are playing a vital role in the felds of machine learning and
artifcial intelligence. Since prediction about the future values of a variable involves
some amount of uncertainty, theory of probability and random variables are essen-
tial constituent building blocks of the algorithms used to teach a machine to perform
certain tasks that are dealing with the principles of learning based on the experience.
These random variables are very much specifc in the theory of signal estimation too.[11]
REFERENCES
1. Bhagwandas P. Lathi and Zhi Ding – Modern Digital and Analog Communication
Systems, Oxford University Press, New York, International Fourth Edition, 2010.
2. Scott L. Miller and Donald G. Childers – Probability and Random Processes with
Applications to Signal Processing and Communications, Academic Press, Elsevier
Inc., Boston, MA, 2004.
3. Henry Stark and John W. Woods – Probability and Random Processes with Applications
to Signal Processing, Pearson, Upper Saddle River, NJ, Third Edition, 2002.
4. Kevin P. Murphy – Machine Learning: A Probabilistic Perspective, The MIT Press,
Cambridge, MA, 2012.
5. Jose Unpingco – Python for Probability, Statistics, and Machine Learning, Springer,
Cham, 2016.
6. Steven M. Kay – Intuitive Probability and Random Processes using MATLAB, Springer,
New York, 2006.
7. Peter D. Hoff – A First Course in Bayesian Statistical Methods, Springer, New York,
2009.
8. Shai Shalev-Shwartz and Shai Ben-David – Understanding Machine Learning: From
Theory to Algorithms, Cambridge University Press, New York, 2014.
9. Bernard C. Levy – Principles of Signal Detection and Parameter Estimation, Springer,
Cham, 2008.
Random Variables in Machine Learning 25
10. Michael Paluszek and Stephanie Thomas – MATLAB Machine Learning, Apress,
New York, 2017.
11. Rober M. Gray and Lee D. Davisson – An Introduction to Statistical Signal Processing,
Cambridge University Press, Cambridge, 2004.
12. Friedrich Liese and Klaus-J. Miescke – Statistical Decision Theory – Estimation,
Testing and Selection-Springer Series in Statisitcs, Springer, New York, 2008.
13. James O. Berger – Statistical Decision Theory and Bayesian Analysis, Springer-Verlag,
New York Inc., New York, Second Edition, 2013.
14. Ruise He and Zhiguo Ding (Eds.) – Applications of Machine Learning in Wireless com-
munications, IET Telecommunication Series 81, IET The Institution of Engineering
and Technology, London, 2019.
15. Robert M. Fano – Transmission of Information: A Statistical Theory of Communications,
The MIT Press, Cambridge, MA, 1961.
2 Analysis of EMG
Signals using Extreme
Learning Machine with
Nature Inspired Feature
Selection Techniques
A. Anitha
D.G. Vaishnav College
A. Bakiya
MIT Campus, Anna University
CONTENTS
2.1 Introduction..................................................................................................... 27
2.2 Data Set........................................................................................................... 30
2.3 Feature Extraction........................................................................................... 30
2.4 Nature Inspired Feature Selection Methods.................................................... 32
2.4.1 Particle Swarm Optimization Algorithm (PSO) ................................. 32
2.4.2 Genetic Algorithm (GA) ..................................................................... 33
2.4.3 Fire-Fly Optimization Algorithm (FA) ...............................................34
2.4.4 Bat Algorithm (BA) ............................................................................ 36
2.4.5 Whale Optimization Algorithm (WOA) ............................................. 37
2.4.5.1 Exploitation Phase ............................................................... 37
2.4.5.2 Exploration Phase ................................................................ 38
2.5 Extreme Learning Machine (ELM) ................................................................ 39
2.6 Results and Discussion ................................................................................... 41
2.7 Conclusion ...................................................................................................... 45
References................................................................................................................ 47
2.1 INTRODUCTION
Neuromuscular Impairment (NMI) is an ailment that affects the neuromuscular
system by breaking in the communication path between the muscles and the ner-
vous system (Reed et al. 2017). The symptoms of NMI include muscle numbness,
fatigue muscles, abnormal pain sensation, atrophy in muscles, and fasciculation in
DOI: 10.1201/9781003138020-2 27
28 Handbook of Machine Learning
The subset evaluation is a pivotal step in estimating the effectiveness of the feature
subset using evaluation criteria. Examining the evaluation of features, FS techniques
can be categorized as flter methods (Guyon & Elisseeff, 2003), wrapper methods
(Chandrashekar & Sahin 2014), and embedded methods (Stańczyk 2015). Filter
model depends on the common data characteristics, and learning model is not used
for evaluating feature subsets. In flter model, features are ranked, and it is evaluated
independently (univariate) or in batch (multivariate). The chosen feature subsets con-
tain features with highest scores. The features with highest score are grouped as fea-
ture subset. Filter model takes an advantage of less computational complexity (Duch
2006); alternatively during classifcation task, it excludes the performance of selected
features. Wrapper model utilizes a learning algorithm (Chandrashekar & Sahin 2014)
to evaluate the effciency of selected feature subset and thus subjugating the disadvan-
tage of flter model. General framework of wrapper model in FS is shown in Figure
2.2. Embedded model merges FS and classifcation into one process (Stańczyk 2015).
[775] We find it in the Nancy globe of about 1540 (see Vol. IV. p.
81); in the Mercator gores of 1541 (Vol. II. p. 177); and in the
Ruscelli map of 1544 (Vol. II. p. 432), where Greenland
(Grotlandia) is simply a neck connecting Europe with America;
and in Gastaldi “Carta Marina,” in the Italian Ptolemy of 1548,
where it is a protuberance on a similar neck (see Vol. II. 435; IV.
43; and Nordenskjöld’s Studien, 43). The Rotz map of 1542
seems to be based on the same material used by Mercator in his
gores, but he adds a new confusion in calling Greenland the
“Cost of Labrador.” Cf. Winsor’s Kohl Maps, no. 104. The
“Grutlandia” of the Vopellio map of 1556 is also continuous with
Labrador (see Vol. II. 436; IV. 90).
[780] For Mercator’s map, see Vol. II. 452; IV. 94, 373. Ortelius’
separate map of Scandia is much the same. It is the same with
the map of Phillipus Gallæus, dated 1574, but published at
Antwerp in 1585 in the Theatri orbis terrarum Enchiridion.
Gilbert’s map in 1576 omits the “Grocland” (Vol. III. 203). Both
features, however, are preserved in the Judæis of 1593 (Vol. IV.
97), in the Wytfliet of 1597 (Vol. II. 459), in Wolfe’s Linschoten in
1598 (Vol. III. 101), and in Quadus in 1600 (Vol. IV. 101). In the
Zaltière map of 1566 (Vol. II. 451; IV. 93), in the Porcacchi map
of 1572 (Vol. II. 96, 453; IV. 96), and in that of Johannes
Martines of 1578, the features are too indefinite for recognition.
Lelewel (i. pl. 7) gives a Spanish mappemonde of 1573.
[788] Bancroft (v. 199) gives references to those writers who have
discussed this question of giants. Bandelier’s references are more
in detail (Arch. Tour, p. 201). Short (p. 233) borrows largely the
list in Bancroft. The enumeration includes nearly all the old
writers. Acosta finds confirmation in bones of incredible
largeness, often found in his day, and then supposed to be
human. Modern zoölogists say they were those of the Mastodon.
Howarth, Mammoth and the Flood, 297.
[793] Hist. Nations Civilisées, i. 37, 150, etc. Popul Vuh, introd., sec.
v. Bancroft relates the Votan myth, with references, in Nat.
Races, iii. 450. Brasseur identifies the Votanites with the Colhuas,
as the builders of Palenqué, the founders of Xibalba, and thinks a
branch of them wandered south to Peru. There are some stories
of even pre-Votan days, under Igh and Imox. Cf. H. De
Charency’s “Myth d’Imos,” in the Annales de philosophie
Chrétienne, 1872-73, and references in Bancroft, v. 164, 231.
[797] Tylor, Anahuac, 189, and his Early Hist. Mankind, 184. Orozco
y Berra, Geog., 124. Bancroft, v. 169, note. The word Maya was
first heard by Columbus in his fourth voyage, 1503-4. We
sometimes find it written Mayab. It is usual to class the people of
Yucatan, and even the Quiché-Cakchiquels of Guatemala and
those of Nicaragua, under the comprehensive term of Maya, as
distinct from the Nahua people farther north.
[799] Brinton, with his view of myths, speaks of the attempt of the
Abbé Brasseur to make Xibalba an ancient kingdom, with
Palenqué as its capital, as utterly unsupported and wildly
hypothetical (Myths, 251).
[803] Short (p. 248) points out that the linguistic researches of
Orozco y Berra (Geografía de las Lenguas de México, 1-76) seem
to confirm this.
[807] Bancroft, ii. 97. Brasseur, Nat. Civ., i. ch. 4, and his Palenqué
ch. 3.
[820] Bancroft (v. 305) cites the diverse views; so does Short to
some extent (pp. 246, 258, etc.). Cf. Brinton’s Address on “Where
was Aztlan?” p. 6; Short, 486, 490; Nadaillac, 284; Wilson’s
Prehistoric Man, i. 327.
Brinton (Myths of the New World, etc., 89; Amer. Hero. Myths,
92) holds that Aztlan is a name wholly of mythical purport, which
it would be vain to seek on the terrestrial globe. This cradle
region of the Nahuas sometimes appears as the Seven Caves
(Chicomoztoc), and Duran places them “in Teoculuacan,
otherwise called Aztlan, a country toward the north and
connected with Florida.” The Seven Caves were explained by
Sahagún as a valley, by Clavigero as a city, by Schoolcraft and
others as simply seven boats in which the first comers came from
Asia; Brasseur makes them and Aztlan the same; others find
them to be the seven cities of Cibola,—so enumerates Brinton
(Myths, 227), who thinks that the seven divisions of the Nahuas
sprung from the belief in the Seven Caves, and had in reality no
existence.
Gallatin has followed out the series of migrations in the Amer.
Ethnol. Soc. Trans., i. 162. Dawson, Fossil Men (ch. 3), gives his
comprehensive views of the main directions of these early
migrations. Brasseur follows the Nahuas (Popul Vuh, introd., sect.
ix.). Winchell (Pre-Adamites) thinks the general tendency was
from north to south. Morgan finds the origin of the Mexican tribes
in New Mexico and in the San Juan Valley (Peabody Mus. Rept.,
xii. 553. Cf. his article in the North Am. Rev., Oct., 1869).
Humboldt (Views of Nature, 207) touches the Aztec wanderings.
There are two well-known Aztec migration maps, first
published in F. G. Carreri’s Giro del Mondo; in English as “Voyage
round the world,” in Churchill’s Voyages, vol. iv., concerning which
see Bancroft, ii. 543; iii. 68, 69; Short, 262, 431, 433; Prescott,
iii. 364, 382. Orozco y Berra (Hist. Antiq. de Mexico, iii. 61) says
that these maps follow one another, and are not different records
of the same progress. Humboldt (Vues, etc., ii. 176) gives an
interpretation of them in accordance with Sigüenza’s views, which
is the one usually followed, and Bancroft (v. 324) epitomizes it.
Ramirez says that the copies reproduced in Humboldt, Clavigero,
and Kingsborough are not so correct as the engraving given in
Garcia y Cubas’s Atlas geogrâfico, estadistico e histórico de la
Republica Mejicana (April, 1858). Bancroft (ii. 544) gives it as
reproduced by Ramirez. It is also in the Mexican edition of
Prescott, and in Schoolcraft’s Indian Tribes. Cf. Delafield’s Inquiry
(N. Y., 1839) and Léon de Rosny’s Les doc. écrits de l’antiq. Amér.
(Paris, 1882). The original is preserved in the Museo Nacional of
Mexico. A palm-tree on the map, near Aztlan, has pointed some
of the arguments in favor of a southern position for that place,
but Ramirez says it is but a part of a hieroglyphic name, and has
no reference to the climate of Aztlan (Short, p. 266). F. Von
Hellwald printed a paper on “American migrations,” with notes by
Professor Henry, in the Smithsonian Report, 1866, pp. 328-345.
Short defines as “altogether the most enlightened treatment of
the subject” the paper of John H. Becker, “Migrations des
Nahuas,” in the Compte rendu, Congrès des Américanistes
(Luxembourg, 1877), i. 325. This paper finds an identification of
the Tulan Zuiva of the Quichés, the Huehue-Tlapallan of the
Toltecs, the Amaquemecan of the Chichimecs, and the Oztotlan
(Aztlan) of the Aztecs in The valleys of the Rio Grande del Norte
and Rio Colorado, as was Morgan’s view. Short (p. 249)
summarizes his paper. Bancroft (v. 289) shows the diversity of
views respecting Amaquemecan.
[824] Charnay (Eng. trans., ch. 8 and 9) calls it a rival city of Tula or
Tollan, rebuilt by the Chichimecs on the ruins of a Toltec city.
[825] If one wants the details of all this, he can read it in Veytia,
Brasseur (Nat. Civilisées and Palenqué, ch. viii.), and Bancroft,
the latter giving references (v. 285).
[828] Bancroft (v. 287) says: “It is probable that the name Toltec, a
title of distinction rather than a national name, was never applied
at all to the common people.”
[832] Archæol. Tour, 191. The fact that the names which we
associate with the Toltecs are Nahua, only means that Nahua
writers have transmitted them, as Bandelier thinks. Cf. also
Bandelier’s citation in the Peabody Mus. Reports, vol. ii. 388,
where he speaks of our information regarding the Toltecs as
“limited and obscure.” He thinks it beyond question that they
were Nahuas; and the fact that their division of time corresponds
with the system found in Yucatan, Guatemala, etc., with other
evidences of myths and legends, leads him to believe that the
aborigines of more southern regions were, if not descendants, at
least of the same stock with the Toltecs, and that we are justified
in studying them to learn what the Toltecs were. He finds that
Veytia, in his account of the Toltecs, beside depending on
Sahagún and Torquemada, finds a chief source in Ixtlilxochitl, and
locates Huehue-Tlapallan in the north; and Veytia’s statements
reappear in Clavigero.
The best narratives of the Toltec history are those in Veytia,
Historia Antigua de Méjico (Mexico, 1806); Brasseur’s Hist.
Nations Civilisées (vol. i.), and his introduction to his Popul Vuh;
and Bancroft (v. ch. 3 and 4): but we must look to Ixtlilxochitl,
Torquemada, Sahagún, and the others, if we wish to study the
sources. In such a study we shall encounter vexatious problems
enough. It is practically impossible to arrange chronologically
what Ixtlilxochitl says that he got from the picture-writings which
he interpreted. Bancroft (v. 209) does the best he can to give it a
forced perspicuity. Wilson (Prehisoric Man, i. 245) not inaptly
says: “The history of the Toltecs and their ruined edifices stands
on the border line of romance and fable, like that of the ruined
builders of Carnac and Avebury.”
[833] Short (page 255) points out that Bancroft unadvisedly looks
upon these Chichimecs as of Nahua stock, according to the
common belief. Short thinks that Pimentel (Lenguas indigenas de
México, published in 1862) has conclusively shown that the
Chichimecs did not originally speak the Nahua tongue, but
subsequently adopted it. Short (page 256) thinks, after collating
the evidence, that it is impossible to determine whence or how
they came to Anáhuac.
[834] Bancroft, v. 292, gives the different views. Cf. Kirk in Prescott,
i. 16.
[837] Prescott, i., introduction ch. 6, tells the story of their golden
age.
[838] Cf. the map in Lucien Biart’s Les Aztèques (Paris, 1885).
Prescott says the maps in Clavigero, Lopez, and Robertson defy
“equally topography and history.” Cf. note on plans of the city and
valley in Vol. II. pp. 364, 369, 374, to which may be added, as
showing diversified views, those in Stevens’s Herrera (London,
1740), vol. ii.; Bordone’s Libro (1528); Icazbalceta’s Coll. de
docs., i. 390; and the Eng. translation of Cortes’ despatches, 333.
[839] This is placed a.d. 1325. Cf. references in Bancroft (v. 346).
[843] Cf. Brasseur’s Nations Civ. ii. 457, on Tezcuco in its palmy
days.
[847] Bancroft (v. 466) enumerates the great variety of such proofs
of disaster, and gives references (p. 469). Cf. Prescott, i. p. 309.
[849] Brinton’s Amer. Hero Myths, 139, etc. See, on the prevalence
of the idea of the return at some time of the hero-god, Brinton’s
Myths of the New World, p. 160. “We must remember,” he says,
“that a fiction built on an idea is infinitely more tenacious of life
than a story founded on fact.” Brinton (Myths, 188) gathers from
Gomara, Cogolludo, Villagutierre, and others, instances to show
how prevalent in America was the presentiment of the arrival and
domination of a white race,—a belief still prevailing among their
descendants of the middle regions of America who watch for the
coming of Montezuma (Ibid. p. 190). Brinton does not seem to
recognize the view held by many that the Montezuma of the
Aztecs was quite a different being from the demi-god of the
Pueblas of New Mexico.
[863] See Vol. II. p. 417. Cf. Prescott’s Mexico, i. 50; Bancroft (Nat.
Races, ii. ch. 14) epitomizes the information on the laws and
courts of the Nahua; Bandelier (Peabody Mus. Repts., ii. 446),
referring to Zurita’s Report, which he characterizes as marked for
perspicacity, deep knowledge, and honest judgment, speaks of it
as embodying the experience of nearly twenty years,—eleven of
which were passed in Mexico,—and in which the author gave
answers to inquiries put by the king. “If we could obtain,” says
Bandelier, “all the answers given to these questions from all parts
of Spanish America, and all as elaborate and truthful as those of
Zurita, Palacio, and Ondegardo, our knowledge of the aboriginal
history and ethnology of Spanish America would be much
advanced.” Zurita’s Report in a French translation is in Ternaux-
Compans’ Collection; the original is in Pacheco’s Docs. inéditos,
but in a mutilated text.
[869] Anales del Museo Nacional, iii. 4, 120; Brinton’s Am. Hero
Myths, 78. Bandelier, in N. Y. Hist. Soc. Proc., November, 1879,
used a portion of the MS. as printed by Sir Thomas Phillipps
(Amer. Antiq. Soc. Proc., i. 115) under the title of Historia de los
Yndios Mexicanos, por Juan de Tovar; Cura et impensis Dni
Thomæ Phillipps, Bart. (privately printed at Middle Hill, 1860. See
Squier Catalogue, no. 1417). The document is translated by
Henry Phillipps, Jr., in the Proc. Amer. Philosophical Soc. (Philad.),
xxi. 616.
[874] See Vol. II. p. 397. Cf. Prescott, ii. 95. The first part of the
Historia is on the religious rites of the natives; the second on
their conversion to Christianity; the third on their chronology, etc.
[880] Those who used him most, like Clavigero and Brasseur de
Bourbourg, complain of this. Torquemada, says Bandelier
(Peabody Mus. Repts. ii. 119), “notwithstanding his
unquestionable credulity, is extremely important on all questions
of Mexican antiquities.”
[882] Cf. Vol. II. 417; Prescott, i. 13, 163, 193, 196; Bancroft, Nat.
Races, v. 147; Wilson’s Prehistoric Man, i. 325. It must be
confessed that with no more authority than the old Mexican
paintings, interpreted through the understanding of old men and
their traditions, Ixtlilxochitl has not the firmest ground to walk on.
Aubin thinks that Ixtlilxochitl’s confusion and contradictions arise
from his want of patience in studying his documents; and some
part of it may doubtless have arisen from his habit, as Brasseur
says (Annales de Philosophie Chrétienne, May, 1855, p. 329), of
altering his authorities to magnify the glories of his genealogic
line. Max Müller (Chips from a German Workshop, i. 322) says of
his works: “Though we must not expect to find in them what we
are accustomed to call history, they are nevertheless of great
historical interest, as supplying the vague outlines of a distant
past, filled with migrations, wars, dynasties and revolutions, such
as were cherished in the memory of the Greeks in the time of
Solon.” In addition to his Historia Chichimeca and his Relaciones,
(both of which are given by Kingsborough, while Ternaux has
translated portions,)—the MS. of the Relaciones being in the
Mexican archives,—Ixtlilxochitl left a large mass of his manuscript
studies of the antiquities, often repetitionary in substance. Some
are found in the compilation made in Mexico by Figueroa in 1792,
by order of the Spanish government (Prescott, i. 193). Some
were in the Ramirez collection. Quaritch (MS. Collections, Jan.,
1888, no. 136) held one from that collection, dated about 1680,
at £16, called Sumaria Relacion, which concerned the ancient
Chichimecs. Those which are best known are a Historia de la
Nueva España, or Historia del Reyno de Tezcuco, and a Historia
de Nuestra Señora de Guadalupe, if this last is by him.
[884] In his Quatre Lettres, p. 24, he calls it the sacred book of the
Toltecs. “C’est le Livre divin lui-même, c’est le Teoamoxtli.”
[890] Giro del mondo, 1699, vol. vi. Cf. Kingsborough, vol. iv.
Robertson attacked Carreri’s character for honesty, and claimed it
was a received opinion that he had never been out of Italy.
Clavigero defended Carreri. Humboldt thinks Carreri’s local
coloring shows he must have been in Mexico.
[891] Cf. the bibliog., in Vol. II., p. 425, of his Storia Antica del
Messico.
[894] See Vol. II. p. 418. Brasseur de Bourbourg’s Hist. des Nations
Civilisées, p. xxxii. Clavigero had described it.
[899] Prescott, i. 24. Harrisse, Bib. Am. Vet., calls Veytia’s the best
history of the ancient period yet (1866) written.
[900] A second ed. (Mexico, 1832) was augmented with notes and
a life of the author, by Carlos Maria de Bustamante; Field, Ind.
Bibliog., no. 909; Brasseur’s Bibl. Mex.-Guat., p. 68.
[906] Hist. des Nations Civilisées (i. pp. xxxi, lxxvi, etc.; cf. Müller’s
Chips, i. 317, 320, 323). Brasseur in the same place describes his
own collection; and it may be further followed in his Bibl. Mex.-
Guat., and in the Pinart Catalogue. Dr. Brinton says that we owe
much for the preservation during late years of Maya MSS. to Don
Juan Pio Perez, and that the best existing collection of them is
that of Canon Crescencio Carrillo y Ancona. José F. Ramirez (see
Vol. II. p. 398) is another recent Mexican collector, and his MSS.
have been in one place and another in the market of late years.
Quaritch’s recent catalogues reveal a number of them, including
his own MS. Catálogo de Colecciones (Jan., 1888, no. 171), and
some of his unpublished notes on Prescott, not included in those
“notas y ecclarecimientos” appended to Navarro’s translation of
the Conquest of Mexico (Catal., 1885, no. 28,502). The several
publications of Léon de Rosny point us to scattered specimens. In
his Doc. écrits de l’Antiquité Amér. he gives the fac-simile of a
colored Aztec map. A MS. in the collection of the Corps Legislatif,
in Paris, and that of the Codex Indiæ Meridionalis are figured in
his Essai sur le déchiffrement, etc. (pl. ix, x). In the Archives de
la Soc. Amér. de France, n. s., vol. i., etc., we find plates of the
Mappe Tlotzin, and a paper of Madier de Montjau, “sur quelques
manuscrits figuratifs de l’Ancien Méxique.” Cf. also Anales del
Museo, viii.
Cf. for further mention of collections the Revue Orientale et
Américaine; Cyrus Thomas in the Am. Antiquarian, May, 1884
(vol. vi.); and the more comprehensive enumeration in the
introduction to Domenech’s Manuscrit pictographique. Orozco y
Berra, in the introduction to his Geografia de las Lenguas y Carta
Etnográfica (Mexico, 1864), speaks of the assistance he obtained
from the collections of Ramirez and of Icazbalceta.
[908] See Vol. II. p. 418. Bandelier calls this French version “utterly
unreliable.”
[918] Brasseur, Bib. Mex.-Guat., p. 30. See Vol. II. p. 429. The
Spanish title is Relacion de las Cosas de Yucatan.
[920] Cf. Bandelier in Am. Antiq. Soc. Proc., n. s., vol. i. p. 88.
[928] H. H. Bancroft, Nat. Races, ii. 115; iii., ch. 2, and v. 170, 547,
gives a convenient condensation of the book, and says that
Müller misconceives in some parts of his summary, and that
Baldwin in his Ancient America, p. 191, follows Müller. Helps,
Spanish Conquest, iv. App., gives a brief synopsis,—the first one
done in English.
[929] Max Müller dissents from this. Chips, i. 326. Müller reminds
us, if we are suspicious of the disjointed manner of what has
come down to us as the Popul Vuh, that “consecutive history is
altogether a modern idea, of which few only of the ancient
nations had any conception. If we had the exact words of the
Popul Vuh, we should probably find no more history there than
we find in the Quiché MS. as it now stands.”
[931] The names of the gods in the Kiché Myths of Central America
(Philad., 1881), from the Proc. Amer. Philos. Soc. He gives his
reasons (p. 4) for the spelling Kiché.
[932] Cf. Am. Antiq. Soc. Proc., n. s., vol. i. 109; and his paper, “On
the Sources of the Aboriginal Hist. of Spanish America,” in the
Am. Asso. Adv. Sci. Proc., xxvii. 328 (Aug., 1878). In the Peabody
Mus. Eleventh Report, p. 391, he says of it that “it appears to be
for the first chapters an evident fabrication, or at least
accommodation of Indian mythology to Christian notions,—a
pious fraud; but the bulk is an equally evident collection of
original traditions of the Indians of Guatemala, and as such the
most valuable work for the aboriginal history and ethnology of
Central America.”
[933] Hist. Nat. Civ., i. 47. S’il existe des sources de l’histoire
primitive du Méxique dans les monuments égyptiens et de
l’histoire primitive de l’ancien monde dans les monuments
Américains? (1864), which is an extract from his Landa’s Relation.
Cf. Bollaert, in the Royal Soc. of Lit. Trans., 1863. Brasseur (Bib.
Mex.-Guat., p. 45; Pinart, no. 231) also speaks of another Quiché
document, of which his MS. copy is entitled Titulo de los Señores
de Totonicapan, escrito en lengua Quiché, el año de 1554, y
traducido al Castellano el año de 1834, por el Padre Dionisio José
Chonay, indígena, which tells the story of the Quiché race
somewhat differently from the Popul Vuh.
[936] See Vol. II. 419; Bancroft, Nat. Races, v. 564; Bandelier in
Am. Antiq. Soc. Proc., i. 105. Bandelier (Peabody Mus. Repts., ii.
391) says that it is now acknowledged that the Recordacion
florida of Fuentes y Guzman is “full of exaggerations and
misstatements.” Brasseur (Bib. Mex.-Guat., pp. 65, 87), in
speaking of Fuentes’ Noticia histórica de los indios de Guatemala
(of which manuscript he had a copy), says that he had access to
a great number of native documents, but profited little by them,
either because he could not read them, or his translators
deceived him. Brasseur adds that Fuentes’ account of the Quiché
rulers is “un mauvais roman qui n’a pas le sens commun.” This
last is a manuscript used by Domingo Juarros in his Compendio
de la historia de la ciudad de Guatemala (Guatemala, 1808-1818,
in two vols.—become rare), but reprinted in the Museo
Guatemalteco, 1857. The English translation, by John Baily, a
merchant living in Guatemala, was published as a Statistical and
Commercial History of Guatemala (Lond., 1823). Cf. Vol. II. p.
419. Francisco Vazquez depended largely on native writers in his
Crónica de la Provincia de Guatemala (Guatemala, 1714-16). (See
Vol. II. p. 419.)
[938] Vol. II. 419. Helps (iii. 300), speaking of Remesal, says: “He
had access to the archives of Guatemala early in the seventeenth
century, and he is one of those excellent writers so dear to the
students of history, who is not prone to declamation, or rhetoric,
or picturesque writing, but indulges us largely by the introduction
everywhere of most important historical documents, copied boldly
into the text.”
[941] E. G. Squier printed in 1860 (see Vol. II. p. vii.) Diego Garcia
de Palacio’s Carta dirigida al Rey de España, año 1576, under the
English title of Description of the ancient Provinces of
Guazacupan, Izalco, Cuscatlan, and Chiquimula in Guatemala,
which is also included in Pacheco’s Coleccion, vol. vi. Bandelier
refers to Estevan Aviles’ Historia de Guatemala desde los tiempos
de los Indios (Guatemala, 1663). A good reputation belongs to a
modern work, Francisco de Paula Garcia Pelaez’s Memorias para
la Historia del antiguo reyno de Guatemala (Guatemala, 1851-53,
in three vols.).