Introduction to Algorithms for Data Mining and Machine Learning 1st edition - eBook PDF 2024 Scribd Download
Introduction to Algorithms for Data Mining and Machine Learning 1st edition - eBook PDF 2024 Scribd Download
com
https://ebookluna.com/download/introduction-to-algorithms-
for-data-mining-and-machine-learning-ebook-pdf/
OR CLICK BUTTON
DOWNLOAD NOW
https://ebookluna.com/download/big-data-analytics-introduction-to-
hadoop-spark-and-machine-learning-ebook-pdf/
ebookluna.com
https://ebookluna.com/product/ebook-pdf-introduction-to-machine-
learning-with-python-a-guide-for-data-scientists/
ebookluna.com
https://ebookluna.com/product/ebook-pdf-introduction-to-business-data-
mining-1st-edition/
ebookluna.com
(eBook PDF) Machine Learning Refined: Foundations,
Algorithms, and Applications
https://ebookluna.com/product/ebook-pdf-machine-learning-refined-
foundations-algorithms-and-applications/
ebookluna.com
https://ebookluna.com/product/ebook-pdf-introduction-to-data-mining-
global-edition-2nd-edition/
ebookluna.com
Big Data Mining for Climate Change 1st edition - eBook PDF
https://ebookluna.com/download/big-data-mining-for-climate-change-
ebook-pdf/
ebookluna.com
https://ebookluna.com/product/ebook-pdf-introduction-to-data-
mining-2nd-edition-by-pang-ning-tan/
ebookluna.com
https://ebookluna.com/download/machine-learning-for-planetary-science-
ebook-pdf/
ebookluna.com
Xin-She Yang
Introduction to
Algorithms for Data Mining
and Machine Learning
Introduction to Algorithms for Data Mining and
Machine Learning
This page intentionally left blank
Introduction to
Algorithms for Data
Mining and Machine
Learning
Xin-She Yang
Middlesex University
School of Science and Technology
London, United Kingdom
Academic Press is an imprint of Elsevier
125 London Wall, London EC2Y 5AS, United Kingdom
525 B Street, Suite 1650, San Diego, CA 92101, United States
50 Hampshire Street, 5th Floor, Cambridge, MA 02139, United States
The Boulevard, Langford Lane, Kidlington, Oxford OX5 1GB, United Kingdom
Copyright © 2019 Elsevier Inc. All rights reserved.
No part of this publication may be reproduced or transmitted in any form or by any means, electronic or
mechanical, including photocopying, recording, or any information storage and retrieval system, without
permission in writing from the publisher. Details on how to seek permission, further information about the
Publisher’s permissions policies and our arrangements with organizations such as the Copyright Clearance Center
and the Copyright Licensing Agency, can be found at our website: www.elsevier.com/permissions.
This book and the individual contributions contained in it are protected under copyright by the Publisher (other
than as may be noted herein).
Notices
Knowledge and best practice in this field are constantly changing. As new research and experience broaden our
understanding, changes in research methods, professional practices, or medical treatment may become necessary.
Practitioners and researchers must always rely on their own experience and knowledge in evaluating and using
any information, methods, compounds, or experiments described herein. In using such information or methods
they should be mindful of their own safety and the safety of others, including parties for whom they have a
professional responsibility.
To the fullest extent of the law, neither the Publisher nor the authors, contributors, or editors, assume any liability
for any injury and/or damage to persons or property as a matter of products liability, negligence or otherwise, or
from any use or operation of any methods, products, instructions, or ideas contained in the material herein.
ISBN: 978-0-12-817216-2
1 Introduction to optimization 1
1.1 Algorithms 1
1.1.1 Essence of an algorithm 1
1.1.2 Issues with algorithms 3
1.1.3 Types of algorithms 3
1.2 Optimization 4
1.2.1 A simple example 4
1.2.2 General formulation of optimization 7
1.2.3 Feasible solution 9
1.2.4 Optimality criteria 10
1.3 Unconstrained optimization 10
1.3.1 Univariate functions 11
1.3.2 Multivariate functions 12
1.4 Nonlinear constrained optimization 14
1.4.1 Penalty method 15
1.4.2 Lagrange multipliers 16
1.4.3 Karush–Kuhn–Tucker conditions 17
1.5 Notes on software 18
2 Mathematical foundations 19
2.1 Convexity 20
2.1.1 Linear and affine functions 20
2.1.2 Convex functions 21
2.1.3 Mathematical operations on convex functions 22
2.2 Computational complexity 22
2.2.1 Time and space complexity 24
2.2.2 Complexity of algorithms 25
2.3 Norms and regularization 26
2.3.1 Norms 26
2.3.2 Regularization 28
2.4 Probability distributions 29
2.4.1 Random variables 29
2.4.2 Probability distributions 30
vi Contents
3 Optimization algorithms 45
3.1 Gradient-based methods 45
3.1.1 Newton’s method 45
3.1.2 Newton’s method for multivariate functions 47
3.1.3 Line search 48
3.2 Variants of gradient-based methods 49
3.2.1 Stochastic gradient descent 50
3.2.2 Subgradient method 51
3.2.3 Conjugate gradient method 52
3.3 Optimizers in deep learning 53
3.4 Gradient-free methods 56
3.5 Evolutionary algorithms and swarm intelligence 58
3.5.1 Genetic algorithm 58
3.5.2 Differential evolution 60
3.5.3 Particle swarm optimization 61
3.5.4 Bat algorithm 61
3.5.5 Firefly algorithm 62
3.5.6 Cuckoo search 62
3.5.7 Flower pollination algorithm 63
3.6 Notes on software 64
Bibliography 163
Index 171
About the author
Xin-She Yang obtained his PhD in Applied Mathematics from the University of Ox-
ford. He then worked at Cambridge University and National Physical Laboratory (UK)
as a Senior Research Scientist. Now he is Reader at Middlesex University London, and
an elected Bye-Fellow at Cambridge University.
He is also the IEEE Computer Intelligence Society (CIS) Chair for the Task Force
on Business Intelligence and Knowledge Management, Director of the International
Consortium for Optimization and Modelling in Science and Industry (iCOMSI), and
an Editor of Springer’s Book Series Springer Tracts in Nature-Inspired Computing
(STNIC).
With more than 20 years of research and teaching experience, he has authored
10 books and edited more than 15 books. He published more than 200 research pa-
pers in international peer-reviewed journals and conference proceedings with more
than 36 800 citations. He has been on the prestigious lists of Clarivate Analytics and
Web of Science highly cited researchers in 2016, 2017, and 2018. He serves on the
Editorial Boards of many international journals including International Journal of
Bio-Inspired Computation, Elsevier’s Journal of Computational Science (JoCS), In-
ternational Journal of Parallel, Emergent and Distributed Systems, and International
Journal of Computer Mathematics. He is also the Editor-in-Chief of the International
Journal of Mathematical Modelling and Numerical Optimisation.
This page intentionally left blank
Preface
Both data mining and machine learning are becoming popular subjects for university
courses and industrial applications. This popularity is partly driven by the Internet and
social media because they generate a huge amount of data every day, and the under-
standing of such big data requires sophisticated data mining techniques. In addition,
many applications such as facial recognition and robotics have extensively used ma-
chine learning algorithms, leading to the increasing popularity of artificial intelligence.
From a more general perspective, both data mining and machine learning are closely
related to optimization. After all, in many applications, we have to minimize costs,
errors, energy consumption, and environment impact and to maximize sustainabil-
ity, productivity, and efficiency. Many problems in data mining and machine learning
are usually formulated as optimization problems so that they can be solved by opti-
mization algorithms. Therefore, optimization techniques are closely related to many
techniques in data mining and machine learning.
Courses on data mining, machine learning, and optimization are often compulsory
for students, studying computer science, management science, engineering design, op-
erations research, data science, finance, and economics. All students have to develop
a certain level of data modeling skills so that they can process and interpret data for
classification, clustering, curve-fitting, and predictions. They should also be familiar
with machine learning techniques that are closely related to data mining so as to carry
out problem solving in many real-world applications. This book provides an introduc-
tion to all the major topics for such courses, covering the essential ideas of all key
algorithms and techniques for data mining, machine learning, and optimization.
Though there are over a dozen good books on such topics, most of these books are
either too specialized with specific readership or too lengthy (often over 500 pages).
This book fills in the gap with a compact and concise approach by focusing on the key
concepts, algorithms, and techniques at an introductory level. The main approach of
this book is informal, theorem-free, and practical. By using an informal approach all
fundamental topics required for data mining and machine learning are covered, and
the readers can gain such basic knowledge of all important algorithms with a focus
on their key ideas, without worrying about any tedious, rigorous mathematical proofs.
In addition, the practical approach provides about 30 worked examples in this book
so that the readers can see how each step of the algorithms and techniques works.
Thus, the readers can build their understanding and confidence gradually and in a
step-by-step manner. Furthermore, with the minimal requirements of basic high school
mathematics and some basic calculus, such an informal and practical style can also
enable the readers to learn the contents by self-study and at their own pace.
This book is suitable for undergraduates and graduates to rapidly develop all the
fundamental knowledge of data mining, machine learning, and optimization. It can
xii Preface
also be used by students and researchers as a reference to review and refresh their
knowledge in data mining, machine learning, optimization, computer science, and data
science.
Xin-She Yang
January 2019 in London
Acknowledgments
I would like to thank all my students and colleagues who have given valuable feedback
and comments on some of the contents and examples of this book. I also would like to
thank my editors, J. Scott Bentley and Michael Lutz, and the staff at Elsevier for their
professionalism. Last but not least, I thank my family for all the help and support.
Xin-She Yang
January 2019
This page intentionally left blank
Introduction to optimization
Contents
1.1 Algorithms
1 1
1.1.1 Essence of an algorithm 1
1.1.2 Issues with algorithms 3
1.1.3 Types of algorithms 3
1.2 Optimization 4
1.2.1 A simple example 4
1.2.2 General formulation of optimization 7
1.2.3 Feasible solution 9
1.2.4 Optimality criteria 10
1.3 Unconstrained optimization 10
1.3.1 Univariate functions 11
1.3.2 Multivariate functions 12
1.4 Nonlinear constrained optimization 14
1.4.1 Penalty method 15
1.4.2 Lagrange multipliers 16
1.4.3 Karush–Kuhn–Tucker conditions 17
1.5 Notes on software 18
This book introduces the most fundamentals and algorithms related to optimization,
data mining, and machine learning. The main requirement is some understanding of
high-school mathematics and basic calculus; however, we will review and introduce
some of the mathematical foundations in the first two chapters.
1.1 Algorithms
An algorithm is an iterative, step-by-step procedure for computation. The detailed
procedure can be a simple description, an equation, or a series of descriptions in
combination with equations. Finding the roots of a polynomial, checking if a natu-
ral number is a prime number, and generating random numbers are all algorithms.
Example 1
As an example, if x0 = 1 and a = 4, then we have
1 4
x1 = (1 + ) = 2.5. (1.2)
2 1
Similarly, we have
1 4 1 4
x2 = (2.5 + ) = 2.05, x3 = (2.05 + ) ≈ 2.0061, (1.3)
2 2.5 2 2.05
x4 ≈ 2.00000927, (1.4)
√
which is very close to the true value of 4 = 2. The accuracy of this iterative formula or algorithm
is high because it achieves the accuracy of five decimal places after four iterations.
The convergence is very quick if we start from different initial values such as
x0 = 10 and even x0 = 100. However, for an obvious reason, we cannot start with
x0 = 0 due to division by
√zero.
Find the root of x = a is equivalent to solving the equation
f (x) = x 2 − a = 0, (1.5)
which is again equivalent to finding the roots of a polynomial f (x). We know that
Newton’s root-finding algorithm can be written as
f (xk )
xk+1 = xk − , (1.6)
f (xk )
where f (x) is the first derivative or gradient of f (x). In this case, we have
f (x) = 2x. Thus, Newton’s formula becomes
(xk2 − a)
xk+1 = xk − , (1.7)
2xk
1.2 Optimization
V = πr 2 h. (1.12)
There are only two design variables r and h and one objective function S to be min-
imized. Obviously, if there is no capacity constraint, then we can choose not to build
the container, and then the cost of materials is zero for r = 0 and h = 0. However,
Introduction to optimization 5
the constraint requirement means that we have to build a container with fixed volume
V0 = πr 2 h = 10 m3 . Therefore, this optimization problem can be written as
πr 2 h = V0 = 10. (1.14)
To solve this problem, we can first try to use the equality constraint to reduce the
number of design variables by solving h. So we have
V0
h= . (1.15)
πr 2
Substituting it into (1.13), we get
S = 2πr 2 + 2πrh
V0 2V0
= 2πr 2 + 2πr 2 = 2πr 2 + . (1.16)
πr r
This is a univariate function. From basic calculus we know that the minimum or max-
imum can occur at the stationary point, where the first derivative is zero, that is,
dS 2V0
= 4πr − 2 = 0, (1.17)
dr r
which gives
V0 3 V0
r3 = , or r = . (1.18)
2π 2π
Thus, the height is
h V0 /(πr 2 ) V0
= = 3 = 2. (1.19)
r r πr
6 Introduction to Algorithms for Data Mining and Machine Learning
This means that the height is twice the radius: h = 2r. Thus, the minimum surface is
It is worth pointing out that this optimal solution is based on the assumption or re-
quirement to design a cylindrical container. If we decide to use a sphere with radius R,
we know that its volume and surface area is
4π 3
V0 = R , S = 4πR 2 . (1.21)
3
We can solve R directly
3V0 3 3V0
R =
3
, or R = , (1.22)
4π 4π
which gives the surface area
3V 2/3 √
0 4π 3 9 2/3
S = 4π =√ 3
V0 . (1.23)
4π 16π 2
√3 √ √ 3
Since 6π/ 4π 2 ≈ 5.5358 and 4π 3 9/ 16π 2 ≈ 4.83598, we have S < S∗ , that is, the
surface area of a sphere is smaller than the minimum surface area of a cylinder with
the same volume. In fact, for the same V0 = 10, we have
√
4π 3 9 2/3
S(sphere) = √ 3
V0 ≈ 22.47, (1.24)
16π 2
which is smaller than S∗ = 25.69 for a cylinder.
This highlights the importance of the choice of design type (here in terms of shape)
before we can do any truly useful optimization. Obviously, there are many other fac-
tors that can influence the choice of design, including the manufacturability of the
design, stability of the structure, ease of installation, space availability, and so on. For
a container, in most applications, a cylinder may be much easier to produce than a
sphere, and thus the overall cost may be lower in practice. Though there are so many
factors to be considered in engineering design, for the purpose of optimization, here
we will only focus on the improvement and optimization of a design with well-posed
mathematical formulations.
Introduction to optimization 7
where f (x), φj (x), and ψk (x) are scalar functions of the design vector x. Here the
components xi of x = (x1 , . . . , xD )T are called design or decision variables, and they
can be either continuous, discrete, or a mixture of these two. The vector x is often
called the decision vector, which varies in a D-dimensional space RD .
It is worth pointing out that we use a column vector here for x (thus with trans-
pose T ). We can also use a row vector x = (x1 , . . . , xD ) and the results will be the
same. Different textbooks may use slightly different formulations. Once we are aware
of such minor variations, it should cause no difficulty or confusion.
In addition, the function f (x) is called the objective function or cost function,
φj (x) are constraints in terms of M equalities, and ψk (x) are constraints written as
N inequalities. So there are M + N constraints in total. The optimization problem
formulated here is a nonlinear constrained problem. Here the inequalities ψk (x) ≤ 0
are written as “less than”, and they can also be written as “greater than” via a simple
transformation by multiplying both sides by −1.
The space spanned by the decision variables is called the search space RD , whereas
the space formed by the values of the objective function is called the objective or
response space, and sometimes the landscape. The optimization problem essentially
maps the domain RD or the space of decision variables into the solution space R (or
the real axis in general).
The objective function f (x) can be either linear or nonlinear. If the constraints φj
and ψk are all linear, it becomes a linearly constrained problem. Furthermore, when
φj , ψk , and the objective function f (x) are all linear, then it becomes a linear pro-
gramming problem [35]. If the objective is at most quadratic with linear constraints,
then it is called a quadratic programming problem. If all the values of the decision
variables can be only integers, then this type of linear programming is called integer
programming or integer linear programming.
On the other hand, if no constraints are specified and thus xi can take any values
in the real axis (or any integers), then the optimization problem is referred to as an
unconstrained optimization problem.
As a very simple example of optimization problems without any constraints, we
discuss the search of the maxima or minima of a univariate function.
8 Introduction to Algorithms for Data Mining and Machine Learning
2
Figure 1.2 A simple multimodal function f (x) = x 2 e−x .
Example 2
For example, to find the maximum of a univariate function f (x)
f (x) = x 2 e−x ,
2
−∞ < x < ∞, (1.26)
is a simple unconstrained problem, whereas the following problem is a simple constrained mini-
mization problem:
subject to
x1 ≥ 1, x2 − 2 = 0. (1.28)
It is worth pointing out that the objectives are explicitly known in all the optimiza-
tion problems to be discussed in this book. However, in reality, it is often difficult to
quantify what we want to achieve, but we still try to optimize certain things such as the
degree of enjoyment or service quality on holiday. In other cases, it may be impossible
to write the objective function in any explicit form mathematically.
From basic calculus we know that, for a given curve described by f (x), its gradient
f (x) describes the rate of change. When f (x) = 0, the curve has a horizontal tangent
at that particular point. This means that it becomes a point of special interest. In fact,
the maximum or minimum of a curve occurs at
f (x∗ ) = 0, (1.29)
Example 3
To find the minimum of f (x) = x 2 e−x (see Fig. 1.2), we have the stationary condition
2
f (x) = 0 or
Figure 1.3 (a) Feasible domain with nonlinear inequality constraints ψ1 (x) and ψ2 (x) (left) and linear
inequality constraint ψ3 (x). (b) An example with an objective of f (x) = x 2 subject to x ≥ 2 (right).
f (x) = 2e−x (1 − 5x 2 + 2x 4 ),
2
two maxima that occur at x∗ = ±1 with fmax = e−1 . At x = 0, we have f (0) = 2 > 0, thus
the minimum of f (x) occurs at x∗ = 0 with fmin (0) = 0.
Whatever the objective is, we have to evaluate it many times. In most cases, the
evaluations of the objective functions consume a substantial amount of computational
power (which costs money) and design time. Any efficient algorithm that can reduce
the number of objective evaluations saves both time and money.
In mathematical programming, there are many important concepts, and we will
first introduce a few related concepts: feasible solutions, optimality criteria, the strong
local optimum, and weak local optimum.
f (x∗ ) = f (0) = 0.
In fact, f (x) = x 3 has a saddle point x∗ = 0 because f (0) = 0 but f changes sign
from f (0+) > 0 to f (0−) < 0 as x moves from positive to negative.
Example 4
For example, to find the maximum or minimum of a univariate function
we first have to find its stationary points x∗ when the first derivative f (x) is zero, that is,
x∗ = −1, x∗ = 2, x∗ = 0.
From the basic calculus we know that the maximum requires f (x∗ ) ≤ 0 whereas the minimum
requires f (x∗ ) ≥ 0.
At x∗ = −1, we have
f (x∗ ) = −23.
can be converted to the minimization of −f (x). For this reason, the optimization
problems can be expressed as either minimization or maximization depending on the
context and convenience of formulations.
In fact, in the optimization literature, some books formulate all the optimization
problems in terms of maximization, whereas others write these problems in terms of
minimization, though they are in essence dealing with the same problems.
∂f ∂f
= 2x + 0 = 0, = 0 + 2y = 0. (1.32)
∂x ∂y
Since
∂ 2f ∂ 2f
= , (1.34)
∂x∂y ∂y∂x
we can conclude that the Hessian matrix is always symmetric. In the case of f (x, y) =
x 2 + y 2 , it is easy to check that the Hessian matrix is
2 0
H= . (1.35)
0 2
At the stationary point (x∗ , y∗ ), if > 0 and fxx > 0, then (x∗ , y∗ ) is a local mini-
mum. If > 0 but fxx < 0, then it is a local maximum. If = 0, then it is inconclu-
sive, and we have to use other information such as higher-order derivatives. However,
if < 0, then it is a saddle point. A saddle point is a special point where a local
minimum occurs along one direction, whereas the maximum occurs along another
(orthogonal) direction.
Example 5
To minimize f (x, y) = (x − 1)2 + x 2 y 2 , we have
∂f ∂f
= 2(x − 1) + 2xy 2 = 0, = 0 + 2x 2 y = 0. (1.37)
∂x ∂y
The second condition gives y = 0 or x = 0. Substituting y = 0 into the first condition, we have
x = 1. However, x = 0 does not satisfy the first condition. Therefore, we have a solution x∗ = 1
and y∗ = 0.
For our example with f = (x − 1)2 + x 2 y 2 , we have
∂ 2f 2 2 2
2 + 2, ∂ f = 4xy, ∂ f = 4xy, ∂ f = 2x 2 ,
= 2y (1.38)
∂x 2 ∂x∂y ∂y∂x ∂y 2
14 Introduction to Algorithms for Data Mining and Machine Learning
At the stationary point (x∗ , y∗ ) = (1, 0), the Hessian matrix becomes
2 0
H= ,
0 2
which is positive definite because its double eigenvalues 2 are positive. Alternatively, we have
= 4 > 0 and fxx = 2 > 0. Therefore, (1, 0) is a local minimum.
M
N
(x, μi , νj ) = f (x) + μi φi2 (x) + νj max{0, ψj (x)}2 , (1.43)
i=1 j =1
where μi 1 and νj ≥ 0.
For example, let us solve the following minimization problem:
where a is a given value. Obviously, without this constraint, the minimum value occurs
at x = 1 with fmin = 0. If a < 1, then the constraint will not affect the result. However,
if a > 1, then the minimum should occur at the boundary x = a (which can be obtained
by inspecting or visualizing the objective function and the constraint). Now we can
define a penalty function (x) using a penalty parameter μ 1. We have
After that the child given to Satan died, says Tabari, Adam had
another son, and he called him Seth, and Seth was prophet in the
room of his father, after the death of Adam.
Adam had many more children; every time that Eve bore, she bare
twins, whereof one was male, the other female, and the twins were
given to one another as husband and wife.
Now Adam sought to give to Abel the twin sister of Cain, when she
was old enough to be married, but Cain (Kabil, in Arabic) was
dissatisfied.[105] Adam said to the brothers, Cain and Abel, “Go, my
sons, and sacrifice to the Lord; and he whose sacrifice is accepted,
shall have the young girl. Take each of you offerings in your hand
and go, sacrifice to the Lord, and He shall decide.”
Abel was a shepherd, and he took the fattest of the sheep, and bore
it to the place of sacrifice; but Cain, who was a tiller of the soil, took a
sheaf of corn, the poorest he could find, and placed it on the altar.
Then fire descended from heaven and consumed the offering of
Abel, so that not even the cinders remained; but the sheaf of Cain
was left untouched.
Adam gave the maiden to Abel, and Cain was sore vexed.
One day, Abel was asleep on a mountain. Cain took a stone and
crushed his head. Then he threw the corpse on his back, and carried
it about, not knowing what to do with it; but he saw two crows
fighting, and one killed the other; then the crow that survived dug a
hole in the earth with his beak, and buried the dead bird. Cain said, “I
have not the sense of this bird. I too will lay my brother in the
ground.” And he did so.
When Adam learned the death of his son, he set out in search of
Cain, but could not find him; then he recited the following lines:—
This is how the story is told in the Midrash:[107] Cain and Abel could
not agree, for, what one had, the other wanted; then Abel devised a
scheme that they should make a division of property, and thus
remove the possibility of contention. The proposition pleased Cain.
So Cain took the earth, and all that is stationary, and Abel took all
that is moveable.
But the envy which lay in the heart of Cain gave him no rest. One
day he said to his brother, “Remove thy foot, thou standest on my
property; the plain is mine.”
Then Abel ran upon the hills, but Cain cried, “Away, the hills are
mine!” Then he climbed the mountains, but still Cain followed him,
calling, “Away! the stony mountains are mine.”
In the Book of Jasher the cause of quarrel is differently stated. One
day the flock of Abel ran over the ground Cain had been ploughing;
Cain rushed furiously upon him and bade him leave the spot. “Not,”
said Abel, “till you have paid me for the skins of my sheep and wool
of their fleeces used for your clothing.” Then Cain took the coulter
from his plough, and with it slew his brother.[108]
The Targum of Jerusalem says, the subject of contention was that
Cain denied a Judgment to come and Eternal Life; and Abel argued
for both.[109] The Rabbi Menachem, however, asserts that the point
on which they strove was whether a word was written zizit or zizis in
the Parascha.[110]
“And when they were in the field together, the brothers quarrelled,
saying, ‘Let us divide the world.’ One said, ‘The earth you stand on is
my soil.’ The other said, ‘You are standing on my earth.’ One said,
‘The Holy Temple shall stand on my lot;’ the other said, ‘It shall stand
on my lot.’ So they quarrelled. Now there were born with Abel two
daughters, his sisters. Then said Cain, ‘I will take the one I choose, I
am the eldest;’ Abel said, ‘They were born with me, and I will have
them both to wife.’ And when they fought, Abel flung Cain down and
was above him; and he lay on Cain. Then Cain said to Abel, ‘Are we
not both sons of one father; why wilt thou kill me?’ And Abel had
compassion, and let Cain get up. And so Cain fell on him and killed
him. From this we learn not to render good to the evil, for, because
Abel showed mercy to Cain, Cain took advantage of it to slay
Abel.”[111]
S. Methodius the Younger refers to this tradition. He says: “Be it
known that Adam and Eve when they left Paradise were virgins. But
the third year after the expulsion from Eden, they had Cain, their
first-born, and his sister Calmana; and after this, next year, they had
Abel and his sister Deborah. But in the three hundredth year of
Adam’s life, Cain slew his brother, and Adam and Eve wailed over
him a hundred years.”[112]
Eutychius, Patriarch of Alexandria, says, “When Adam and Eve
rebelled against God, He expelled them from Paradise at the ninth
hour on Friday to a certain mountain in India, and He bade them
produce children to increase and multiply upon the earth. Adam and
Eve therefore became parents, first of a boy named Cain, and of a
girl named Azrun, who were twins; then of another boy named Abel,
and of a twin sister named Owain, or in Greek Laphura.
“Now, when the children were grown up, Adam said to Eve, ‘Let Cain
marry Owain, who was born with Abel, and let Abel have Azrun, who
was born with Cain.’ But Cain said to his mother, ‘I will marry my own
twin sister, and Abel shall marry his.’ For Azrun was prettier than
Owain. But when Adam heard this, he said, ‘It is contrary to the
precept that thou shouldst marry thy twin sister.’
“Now Cain was a tiller of the ground, but Abel was a pastor of sheep.
Adam said to them, ‘Take of the fruits of the earth, and of the young
of the sheep, and ascend the top of this holy mountain, and offer
there the best and choicest to God.’ Abel offered of the best and
fattest of the first-born of the flock. Now as they were ascending the
summit of the mountain, Satan put it into the head of Cain to kill his
brother, so as to get Azrun. For that reason his oblation was not
accepted by God. Therefore he was the more inflamed with rage
against Abel, and as they were going down the mount, he rushed
upon him and beat him about the head with a stone and killed him.
Adam and Eve bewailed Abel a hundred years with the greatest
grief.... And God cast out Cain whilst he was still unmarried into the
land of Nod. But Cain carried off with him his sister Azrun.”[113]
The Rabbi Zadok said, “This was the reason why Cain slew Abel.
His twin sister and wife was not at all good-looking. Then he said, ‘I
will kill my brother Abel, and carry off his wife.’”[114]
Gregory Abulfaraj gives this account of the strife: “According to the
opinion of Mar Theodosius, thirty years after he was expelled from
Paradise, Adam knew his wife Eve, and she bore twins, Cain and his
sister Climia; and after thirty more years she bore Abel and his twin
sister Lebuda. Then, seventy years after when Adam wanted to
marry one of the brothers with the twin sister of the other, Cain
refused, asking to have his own twin sister.”[115]
The Pseudo-Athanasius says, “Up to this time no man had died so
that Cain should know how to kill. The devil instructed him in this in a
dream.”[116]
Leonhard Marius on Genesis iv. says, “As to what instrument Cain
used, Scripture is silent. Chrysostom calls it a sword; Prudentius, a
spade; Irenæus, an axe; Isidore says simply, steel; but artists
generally paint a club, and Abulensis thinks he was killed with
stones.” Reuchlin thinks, as iron was not discovered till the times of
Tubal-cain, the weapon must have been made of wood, and he
points out how much more this completes the type of Christ.[117]
Cain and Abel had been born and had lived with Adam in the land of
Adamah; but after Cain slew his brother, he was cast out into the
land Erez, and wherever he went, swords sounded and flashed as
though thirsting to smite him. And he fled that land and came to
Acra, where he had children, and his descendants who live there to
this day have two heads.[118]
Before Cain slew his brother, says the Targum of Jerusalem, the
earth brought forth fruits as the fruits of Eden; but from the day that
blood was spilt upon it, thistles and thorns sprang up; for the face of
earth grew sad, its joy was gone, the stain was on its brow.
Abel’s offering had been of the fattest of his sheep, the Targum adds,
but Cain offered flax.[119]
Abel’s offering, say certain Rabbis, was not perfect; for he offered
the chief part to God, but the remainder he dedicated to the Devil,
and Cain offered the chief part to Satan, and only the remainder to
God.[120]
The Rabbi Johanan said, Cain exclaimed when accused by God of
the murder, “My iniquity is greater than I can bear,” and this is
supposed to mean, “My iniquity is too great to be atoned for, except
by my brother rising from the earth and slaying me.” What did the
Holy One then? He took one letter of the twenty-two which are in the
Law, and He wrote it on the arm of Cain, as it is written, “He put a
mark upon him.”[121]
After Abel was slain, the dog which had kept his sheep guarded his
body, says the Midrash. Adam and Eve sat beside it and wept, and
knew not what to do. Then said a raven whose friend was dead, “I
will teach Adam a lesson,” and he dug a hole in the soil and laid his
friend there and covered him up. And when Adam saw this, he said
to Eve, “We will do the same with Abel.” God rewarded the raven for
this by promising that none should ever injure his young, that he
should always have meat in abundance, and that his prayer for rain
should be immediately answered.[122]
But the Rabbi Johanan taught that Cain buried his brother to hide
what he had done from the eye of God, not knowing that God can
see even the most secret things.[123]
According to some Rabbis, all good souls are derived from Abel and
all bad souls from Cain. Cain’s soul was derived from Satan, his
body alone was from Eve; for the Evil Spirit Sammael, according to
some, Satan, according to others, deceived Eve, and thus Cain was
the son of the Evil One.[124] All the children of Cain also became
demons of darkness and nightmares, and therefore it is, say the
Cabbalists, that there is no mention in Genesis of the death of any of
Cain’s offspring.[125]
When Cain had slain his brother, we are told in Scripture that he fled.
Certain Rabbis give the reason:—He feared lest Satan should kill
him: now Satan has no power over any one whose face he does not
see, thus he had none over Lot’s wife till she turned her face towards
Sodom, and he could see it; and Cain fled, to keep his face from
being seen by the Evil One, and thus give him an opportunity of
taking his life.[126]
With regard to the mark put upon Cain, there is great diverging of
opinion. Some say that his tongue turned white; others, that he was
given a peculiar dress; others, that his face became black; but the
most prevalent opinion is that he became covered with hair, and a
horn grew in the midst of his forehead.
The Little Genesis says, Cain was born when Adam was aged
seventy, and Abel when he was seventy-seven.
The book of the penitence of Adam gives us some curious details.
When Cain had killed his brother, he was filled with terror, for he saw
the earth quivering. He cast the body into a hole and covered it with
dust, but the earth threw the body out. Then he dug another hole and
heaped earth on his brother’s corpse, but again the earth rejected it.
When God appeared before him, Cain trembled in all his limbs, and
God said to him, “Thou tremblest and art in fear; this shall be thy
sign.” And from that moment he quaked with a perpetual ague.
The Rabbis give another mark as having been placed on Cain. They
say that a horn grew out of the midst of his forehead. He was killed
by a son of Lamech, who, being shortsighted, mistook him for a wild
beast; but in the Little Genesis it is said that he was killed by the fall
of his house, in the year 930, the same day that Adam died.
According to the same authority, Adam and Eve bewailed Abel
twenty-eight years.
The Talmud relates the following beautiful incident.
God had cursed Cain, and he was doomed to a bitter punishment;
but moved, at last, by Cain’s contrition, He placed on his brow the
symbol of pardon.
Adam met Cain, and looked with wonder on the seal or token, and
asked,—
“How hast thou turned away the wrath of the Almighty?”
“By confession of sin and repentance,” answered the fratricide.
“Woe is me!” cried Adam, smiting his brow; “is the virtue of
repentance so great, and I knew it not! And by repentance I might
have altered my lot!”[127]
Tabari says that Cain was the first worshipper of fire. Eblis (Satan)
appeared to him and told him that the reason of the acceptance of
Abel’s sacrifice was, that he had invoked the fire that fell on it and
consumed it; Cain had not done this, and therefore fire had not come
down on his oblation. Cain believed this, and adored fire, and taught
his children to do the same.[128]
Cain, says Josephus, having wandered over the earth with his wife,
settled in the land of Nod. But his punishment, so far from proving of
advantage to him, proved only a stimulus to his violence and
passion; and he increased his wealth by rapine, and he encouraged
his children and friends to live by robbery and in luxury. He also
corrupted the primitive simplicity in which men lived, by the
introduction amongst them of weights and measures, by placing
boundaries, and walling cities.[129]
John Malala says the same: “Cain was a tiller of the ground till he
committed the crime of slaying his brother; after that, he lived by
violence, his hand being against every man, and he invented and
taught men the use of weights, measures, and boundaries.”[130]
Welcome to our website – the ideal destination for book lovers and
knowledge seekers. With a mission to inspire endlessly, we offer a
vast collection of books, ranging from classic literary works to
specialized publications, self-development books, and children's
literature. Each book is a new journey of discovery, expanding
knowledge and enriching the soul of the reade
Our website is not just a platform for buying books, but a bridge
connecting readers to the timeless values of culture and wisdom. With
an elegant, user-friendly interface and an intelligent search system,
we are committed to providing a quick and convenient shopping
experience. Additionally, our special promotions and home delivery
services ensure that you save time and fully enjoy the joy of reading.
ebookluna.com