Rough Fuzzy Image Analysis Foundations and Methodologies 1st Edition Sankar K. Pal 2024 scribd download
Rough Fuzzy Image Analysis Foundations and Methodologies 1st Edition Sankar K. Pal 2024 scribd download
https://ebookultra.com
https://ebookultra.com/download/rough-fuzzy-image-
analysis-foundations-and-methodologies-1st-
edition-sankar-k-pal/
https://ebookultra.com/download/image-analysis-and-modeling-in-
ophthalmology-1st-edition-eddie-y-k-ng/
ebookultra.com
https://ebookultra.com/download/foundations-of-fuzzy-logic-and-
semantic-web-languages-1st-edition-umberto-straccia/
ebookultra.com
https://ebookultra.com/download/introduction-to-fuzzy-logic-1st-
edition-james-k-peckol/
ebookultra.com
https://ebookultra.com/download/dermoscopy-image-analysis-1st-edition-
celebi/
ebookultra.com
https://ebookultra.com/download/the-jazz-image-k-heather-pinson/
ebookultra.com
https://ebookultra.com/download/statistical-modeling-analysis-and-
management-of-fuzzy-data-1st-edition-irwin-r-goodman/
ebookultra.com
https://ebookultra.com/download/still-image-and-video-compression-
with-matlab-1st-edition-k-s-thyagarajan/
ebookultra.com
https://ebookultra.com/download/pal-benko-my-life-games-and-
compositions-1st-edition-pal-benko-jeremy-silman-john-l-watson/
ebookultra.com
Rough Fuzzy Image Analysis Foundations and
Methodologies 1st Edition Sankar K. Pal Digital Instant
Download
Author(s): Sankar K. Pal, James F. Peters
ISBN(s): 1439803293
Edition: 1
File Details: PDF, 8.65 MB
Year: 2010
Language: english
Rough Fuzzy
Image Analysis
Foundations and Methodologies
Series Editors
This series aims to capture new developments and summarize what is known over the whole
spectrum of mathematical and computational imaging sciences. It seeks to encourage the
integration of mathematical, statistical and computational methods in image acquisition and
processing by publishing a broad range of textbooks, reference works and handbooks. The
titles included in the series are meant to appeal to students, researchers and professionals
in the mathematical, statistical and computational sciences, application areas, as well as
interdisciplinary researchers involved in the field. The inclusion of concrete examples and
applications, and programming code and examples, is highly encouraged.
Proposals for the series should be submitted to the series editors above or directly to:
CRC Press, Taylor & Francis Group
4th, Floor, Albert House
1-4 Singer Street
London EC2A 4BQ
UK
Rough Fuzzy
Image Analysis
Foundations and Methodologies
Edited by
Sankar K. Pal
James F. Peters
This book contains information obtained from authentic and highly regarded sources. Reasonable efforts have been
made to publish reliable data and information, but the author and publisher cannot assume responsibility for the valid-
ity of all materials or the consequences of their use. The authors and publishers have attempted to trace the copyright
holders of all material reproduced in this publication and apologize to copyright holders if permission to publish in this
form has not been obtained. If any copyright material has not been acknowledged please write and let us know so we may
rectify in any future reprint.
Except as permitted under U.S. Copyright Law, no part of this book may be reprinted, reproduced, transmitted, or uti-
lized in any form by any electronic, mechanical, or other means, now known or hereafter invented, including photocopy-
ing, microfilming, and recording, or in any information storage or retrieval system, without written permission from the
publishers.
For permission to photocopy or use material electronically from this work, please access www.copyright.com (http://
www.copyright.com/) or contact the Copyright Clearance Center, Inc. (CCC), 222 Rosewood Drive, Danvers, MA 01923,
978-750-8400. CCC is a not-for-profit organization that provides licenses and registration for a variety of users. For
organizations that have been granted a photocopy license by the CCC, a separate system of payment has been arranged.
Trademark Notice: Product or corporate names may be trademarks or registered trademarks, and are used only for
identification and explanation without intent to infringe.
Rough fuzzy image analysis : foundations and methodologies / editors, Sankar K. Pal, James F. Peters.
p. cm.
“A CRC title.”
Includes bibliographical references and index.
ISBN 978-1-4398-0329-5 (hardcover : alk. paper)
1. Image analysis. 2. Fuzzy sets. I. Pal, Sankar K. II. Peters, James F. III. Title.
TA1637.R68 2010
621.36’7--dc22 2009053741
This book introduces the foundations and applications in the state-of-art of rough-
fuzzy image analysis. Fuzzy sets* and rough sets** as well as a generalization of
rough sets called near sets*** provide important as well as useful stepping stones
in various approaches to image analysis that are given in the chapters of this book.
These three types of sets and various hybridizations provide powerful frameworks
for image analysis.
Image analysis focuses on the extraction of meaningful information from digital
images. This subject has its roots in studies of space and the senses by J.H. Poincaré
during the early 1900s, studies of visual perception and the topology of the brain
by E.C. Zeeman and picture processing by A.P. Rosenfeld**** . The basic picture
processing approach pioneered by A.P. Rosenfeld was to extract meaningful patterns
in given digital images representing real scenes as opposed to images synthesized by
the computer. Underlying picture processing is an interest in filtering a picture to
detect given patterns embedded in digital images and approximating a given image
with simpler, similar images with lower information content (this, of course, is at the
heart of the near set-based approach to image analysis). This book calls attention
to the utility that fuzzy sets, near sets and rough sets have in image analysis. One
of the earliest fuzzy set-based image analysis studies was published in 1982 by S.K.
Pal***** . The spectrum of fuzzy set-oriented image analysis studies includes edge
ambiguity, scene analysis, image enhancement using smoothing, image description,
motion frame analysis, medical imaging, remote sensing, thresholding and image
frame analysis.
The application of rough sets in image analysis was launched in a seminal paper
published in 1993 by A. Mrózek and L. Plonka****** . Near sets are a recent gener-
alization of rough sets that have proven to be useful in image analysis and pattern
* See, e.g., Zadeh, L.A., Fuzzy sets. Information and Control (1965), 8 (3) 338-353; Zadeh, L.A.,
Toward a theory of fuzzy granulation and its centrality in human reasoning and fuzzy logic, Fuzzy
Sets and Systems 90 (1997), 111-127. See, also, Rosenfeld, A., Fuzzy digital topology, in Bezdek,
J.C., Pal, S.K., Eds., Fuzzy Models for Pattern Recognition, IEEE Press, 1991, 331-339; Banerjee,
M., Kundu, M.K., Maji, P., Content-based image retrieval using visually significant point features,
Fuzzy Sets and Systems 160, 1 (2009), 3323-3341; http://en.wikipedia.org/wiki/Fuzzy set
** See, e.g., Peters, J.F., Skowron, A.: Zdzislaw Pawlak: Life and Work, Transactions on Rough
Sets V, (2006), 1-24; Pawlak, Z., Skowron, A.: Rudiments of rough sets, Information Sciences 177
(2007) 3-27; Pawlak, Z., Skowron, A.: Rough sets: Some extensions, Information Sciences 177
(2007) 28-40; Pawlak, Z., Skowron, A.: Rough sets and Boolean reasoning, Information Sciences
177 (2007) 41-73.; http://en.wikipedia.org/wiki/Rough set
*** See, e.g., Peters, J.F., Puzio, L., Image analysis with anisotropic wavelet-based near-
ness measures, Int. J. of Computational Intelligence Systems 79, 3-4 (2009), 1-17; Peters,
J.F., Wasilewski, P., Foundations of near sets, Information Sciences 179, 2009, 3091-3109;
http://en.wikipedia.org/wiki/Near sets. See, also, http://wren.ee.umanitoba.ca
**** See, e.g., Rosenfeld, A.P., Picture processing by computer, ACM Computing Surveys 1, 3
(1969), 147-176
***** Pal, S.K., A note on the quantitative measure of image enhancement through fuzziness, IEEE
recognition******* .
This volume fully reflects the diversity and richness of rough fuzzy image analysis
both in terms of its underlying set theories as well as its diverse methods and appli-
cations. From the lead chapter by J.F. Peters and S.K. Pal, it can be observed that
fuzzy sets, near sets and rough sets are, in fact, instances of different incarnations
of Cantor sets. These three types of Cantor sets provide a foundation for what
A. Rosenfeld points to as the stages in pictorial pattern recognition, i.e., image
transformation, feature extraction and classification. The chapters by P. Maji and
S.K. Pal on rough-fuzzy clustering, D. Malyszko and J. Stepaniuk on rough-fuzzy
measures, and by A.E. Hassanien, H. Al-Qaheri, A. Abraham on rough-fuzzy clus-
tering for segmentation point to the utility of hybrid approaches that combine fuzzy
sets and rough sets in image analysis. The chapters by D. Sen, S.K. Pal on rough
set-based image thresholding, H. Fashandi, J.F. Peters on rough set-based mathe-
matical morphology as well as an image partition topology and M.M. Mushrif, A.K.
Ray on image segmentation illustrate how image analysis can be carried out with
rough sets by themselves. Tolerance spaces and a perceptual approach in image
analysis can be found in the chapters by C. Henry, A.H. Meghdadi, J.F. Peters, S.
Shahfar, and S. Ramanna (these papers carry forward the work on visual perception
by J.H. Poincaré and E.C. Zeeman). A rich harvest of applications of rough fuzzy
image analysis can be found in the chapters by A.E. Hassanien, H. Al-Qaheri, A.
Abraham, W. Tarnawski, G. Schaefer, T. Nakashima, L. Miroslaw, C. Henry, S.
Shahfar, A.H. Meghdadi and S. Ramanna. Finally, a complete, downloadable im-
plementation of near sets in image analysis called NEAR is presented by C. Henry.
The Editors of this volume extend their profound gratitude to the many reviewers
for their generosity and many helpful comments concerning the chapters in this
volume. Every chapter was extensively reviewed and revised before final acceptance.
We also received many helpful suggestions from the reveiwers of the original proposal
for this CRC Press book. In addition, we are very grateful for the help that we
have received from S. Kumar, A. Rodriguez, R.B. Stern, S.K. White, J. Vakili and
others at CRC Press during the preparation of this volume.
The editors of this volume have been supported by the Natural Sciences and En-
gineering Research Council of Canada (NSERC) research grant 185986, Manitoba
Centre of Excellence Fund (MCEF) grant, Canadian Network of Excellence (NCE)
and Canadian Arthritis Network (CAN) grant SRI-BIO-05, and the J.C. Bose Fel-
lowship of the Government of India.
******* See,e.g.,
Gupta, S., Patnik, S., Enhancing performance of face recognition by using the near
set approach for selecting facial features, J. Theor. Appl. Inform. Technol. 4, 5 (2008), 433-441;
Henry, C., Peters, J.F., Perception-based image analysis, Int. J. Bio-Inspired Comp. 2, 2 (2009),
in press; Peters, J.F., Tolerance near sets and image correspondence, Int. J. of Bio-Inspired
Computation 1(4) (2009), 239-245; Peters, J.F., Corrigenda and addenda: Tolerance near sets and
image correspondence, Int. J. of Bio-Inspired Computation 2(5) (2010), in press; Ramanna, S.,
Perceptually near Pawlak partitions, Transactions on Rough Sets XII, 2010, in press, Ramanna,
S., Meghdadi, A., Measuring resemblances between swarm behaviours: A perceptual tolerance
near set approach, Fundamenta Informaticae 95(4), 2009, 533-552.
0-3
Table of Contents
1 Cantor, Fuzzy, Near, and Rough Sets in Image Analysis
James F. Peters and Sankar K. Pal . . . . . . . . . . . . . . . . . . . . . 1-1
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . I-1
1
Cantor, Fuzzy, Near, and Rough Sets
in Image Analysis
1.1 Introduction
The chapters in this book consider how one might utilize fuzzy sets, near sets, and rough sets, taken
separately or taken together in hybridizations, in solving a variety of problems in image analysis. A
brief consideration of Cantor sets (Cantor, 1883, 1932) provides a backdrop for an understanding
of several recent types of sets useful in image analysis. Fuzzy, near and rough sets provide a wide
spectrum of practical solutions to solving image analysis problems such as image understanding,
image pattern recognition, image retrieval and image correspondence, mathematical morphology,
perceptual tolerance relations in image analysis and segmentation evaluation. Fuzzy sets result from
the introduction of a membership function that generalizes the traditional characteristic function.
The notion of a fuzzy set was introduced by L. Zadeh in 1965 (Zadeh, 1965). Sixteen years later,
rough sets were introduced by Z. Pawlak in 1981 (Pawlak, 1981a). A set is considered rough
whenever the boundary between its lower and upper approximation is non-empty. Of the three forms
of sets, near sets are newest, introduced in 2007 by J.F. Peters in a perception-based approach to the
study of the nearness of observable objects in a physical continuum (Peters and Henry, 2006; Peters,
2007c,a; Peters, Skowron, and Stepaniuk, 2007; Henry and Peters, 2009b; Peters and Wasilewski,
2009; Peters, 2010).
This chapter highlights a context for three forms of sets that are now part of the computational
intelligence spectrum of tools useful in image analysis and pattern recognition. The principal con-
1–1
1–2 Rough Fuzzy Image Analysis
tribution of this chapter is an overview of the high utility of fuzzy sets, near sets and rough sets
with the emphasis on how these sets can be used in image analysis, especially in classifying parts
of digital images presented in this book.
In this mature interpretation of the notion of a set, G. Cantor points to a property or law that de-
termines elementhood in a set and “unites [the elements] into a whole” (Cantor, 1883), elaborated
in (Cantor, 1932), and commented on in Lavine (1994). In 1851, Bolzano (Bolzano, 1959) writes
that “an aggregate so conceived that is indifferent to the arrangement of its members I call a set”.
At that time, the idea that a set could contain just one element or no elements (null set) was not con-
templated. This is important in the current conception of a near set, since such a set must contain
pairs of perceptual objects with similar descriptions and such a set is never null. That is, a set is a
perceptual near set if, and only if it is never empty and it contains pairs of perceived objects that
have descriptions that are within some tolerance of each other (see Def. 2).
Symbol Interpretation
O, X,Y Set of perceptual objects, X,Y ⊆ O, A ⊂ X, x ∈ X, y ∈ Y ,
F, B Sets of probe functions, B ⊆ F, φi ∈ B,
φi (x) φi : X → ℜ, ith probe function representing feature of x,
φ B (x) (φ1 (x), φ2 (x), . . . , φi (x), . . . , φk (x)),description of x of length k,
ε ε ∈ ℜ (reals) such that ε ≥ 0,
1
· 2 = (∑ki=1 (·i )2 ) 2 , L2 (Euclidean) norm,
∼
=B,ε {(x, y) ∈ O × O : φ (x) − φ (y) 2 ≤ ε }, tolerance relation,
∼
=B shorthand for ∼ =B,ε ,
A⊂ ∼ =B,ε ∀x, y ∈ A, x ∼ =B,ε y (i.e., A∼ ∼
=B,ε is a preclass in =B,ε ),
C∼
=B,ε tolerance class, maximal preclass of =B,ε , ∼
X B,ε Y X resembles (is near) Y ⇐⇒ X ∼ =B,ε Y .
Near sets are disjoint sets that resemble each other (Henry and Peters, 2010). Resemblance
between disjoint sets occurs whenever there are observable similarities between the objects in the
sets. Similarity is determined by comparing lists of object feature values. Each list of feature values
defines an object’s description. Comparison of object descriptions provides a basis for determining
the extent that disjoint sets resemble each other. Objects that are perceived as similar based on
their descriptions are grouped together. These groups of similar objects can provide information
and reveal patterns about objects of interest in the disjoint sets. For example, collections of digital
images viewed as disjoint sets of points provide a rich hunting ground for near sets. For example,
near sets can be found in the favite pentagona coral fragment in Fig. 1.1a from coral reef near Japan.
If we consider the greyscale level, the sets X,Y in Fig. 1.1b are near sets, since there are many pixels
in X with grey levels that are very similar to pixels in Y .
interest in comparing similarities between digital images. Unlike rough sets, the near set approach
does not require set approximation Peters and Wasilewski (2009). Simple examples of near sets can
sometimes be found in tolerance classes in pairs of image coverings, if, for instance, a subimage of
a class in one image has a description that is similar to the description of a subimage in a class in
the second image. In general, near sets are discovered by discerning objects–either within a single
set or across sets–with descriptions that are similar.
From the beginning, the near set approach to perception has had direct links to rough sets in its
approach to the perception of objects (Pawlak, 1981a; Orłowska, 1982) and the classification of ob-
jects (Pawlak, 1981a; Pawlak and Skowron, 2007c,b,a). This is evident in the early work on nearness
of objects and the extension of the approximation space model (see, e.g., (Peters and Henry, 2006;
Peters et al., 2007)). Unlike the focus on the approximation boundary of a set, the study of near sets
focuses on the discovery of affinities between perceptual granules such as digital images viewed as
sets of points. In the context of near sets, the term affinity means close relationship between per-
ceptual granules (particularly images) based on common description. Affinities are discovered by
comparing the descriptions of perceptual granules, e.g., descriptions of objects contained in classes
found in coverings defined by the tolerance relation ∼ =F,ε .
Near sets offer an ideal framework for solving problems based on human perception that arise in
areas such as image processing, computer vision as well as engineering and science problems. In
near set theory, perception is a combination of the view of perception in psychophysics (Hoogs,
Collins, Kaucic, and Mundy, 2003; Bourbakis, 2002) with a view of perception found in Merleau-
Ponty’s work (Merleau-Ponty, 1945, 1965). In the context of psychophysics, perception of an object
(i.e., in effect, our knowledge about an object) depends on sense inputs that are the source of signal
values (stimularions) in the cortex of the brain. In this view of perception, the transmissions of
sensory inputs to cortex cells senses are likened to probe functions defined in terms of mappings
of sets of sensed objects to sets of real-values representing signal values (the magnitude of each
cortex signal value represents a sensation) that are a source of object feature values assimilated by
the mind.
Perception in animals is modelled as a mapping from sensory cells to brain cells. For example,
visual perception is modelled as a mapping from stimulated retina sensory cells to visual cortex cells
(see Fig. 1.2). Such mappings are called probe functions. A probe measures observable physical
characteristics of objects in our environment. In other words, a probe function provides a basis
for what is commonly called feature extraction (Guyon, Gunn, Nikravesh, and Zadeh, 2006). The
sensed physical characteristics of an object are identified with object features. The term feature
is used in S. Watanabe’s sense of the word (Watanabe, 1985), i.e., a feature corresponds to an
observable property of physical objects. Each feature has a 1-to-many relationship to real-valued
functions called probe functions representing the feature. For each feature (such as colour) one or
more probe functions can be introduced to represent the feature (such as grayscale, or RGB values).
Objects and sets of probe functions form the basis of near set theory and are sometimes referred to
as perceptual objects due to the focus on assigning values to perceived object features.
Axiom 1 An object is perceivable if, and only if the object is describable.
In Merleau-Ponty’s view (Merleau-Ponty, 1945, 1965), an object is perceived to the extent that it
can be described. In other words, object description goes hand-in-hand with object perception. It
is our mind that identifies relationships between object descriptions to form perceptions of sensed
objects. It is also the case that near set theory has been proven to be quite successful in finding
solutions to perceptual problems such as measuring image correspondence and segmentation eval-
uation. The notion of a sensation in Poincaré (Poincaré, 1902) and a physical model for a probe
function from near set theory (Peters and Wasilewski, 2009; Peters, 2010) is implicitly explained by
Zeeman (Zeeman, 1962) in terms of visual perception. That is, ‘seeing’ consists of mappings from
1–6 Rough Fuzzy Image Analysis
sense inputs from sensory units in the retina of the eye to cortex cells of the brain stimulated by sense
inputs. A sense input can be represented by a number representing the intensity of the light from the
visual field (i.e., everything in the physical world that causes light to fall on the retina.) impacting
on the retina. The intensity of light from the visual field will determine the level of stimulation of a
cortex cell from retina sensory input. Over time, varying cortex cell stimulation has the appearance
of an electrical signal. The magnitude of cortex cell stimulation is a real-value. The combination
of an activated sensory cell in the retina and resulting retina-originated impulses sent to cortex cells
(visual stimulation) is likened to what Poincaré calls a sensation in his essay on separate sets of
similar sensations leading to a perception of a physical continuum (Poincaré, 1902). This model for
a sensation underlies what is known as a probe function in near set theory (Peters, 2007b; Peters
and Wasilewski, 2009).
In effect, a probe function value φ (x) measures the strength of a feature value extracted from
each sensation. In Poincaré, sets of sensations are grouped together because they are, in some sense,
similar within a specified distance, i.e., tolerance. Implicit in this idea in Poincaré is the perceived
feature value of a particular sensation that makes it possible for us to measure the closeness of an
individual senation to other sensations.
A human sensation modelled as a probe measures observable physical characteristics of objects in
our environment. The sensed physical characteristics of an object are identified with object features.
In Merleau-Ponty’s view, an object is perceived to the extent that it can be described (Merleau-Ponty,
1945, 1965). In other words, object description goes hand-in-hand with object perception. It is our
mind that identifies relationships between object descriptions to form perceptions of sensed objects.
It is also the case that near set theory has been proven to be quite successful in finding solutions to
perceptual problems such as measuring image correspondence and segmentation evaluation.
Axiom 2 Formulate object description to achieve object perception.
In a more recent interpretation of the notion of a near set, the nearness of sets is considered in
the context of perceptual systems (Peters and Wasilewski, 2009). Poincaré’s idea of perception of
objects such as digital images in a physical continuum can be represented by means of perceptual
systems, which is akin to but not the same as what has been called a perceptual information sys-
tem (Peters and Wasilewski, 2009; Peters, 2010). A perceptual system is a pair O, F where O is
a non-empty set of perceptual objects and F is a non-empty, countable set of probe functions (see
Def. 1).
Definition 1 Perceptual System (Peters, 2010)
A perceptual system O, F consists of a sample space O containing a finite, non-empty set of
sensed sample objects and a non-empty, countable set F containing probe functions representing
object features.
The perception of physical objects and their description within a perceptual system facilitates pattern
recognition and the discovery of sets of similar objects. In the near set approach to image analysis,
one starts by identifying a perceptual system and the defining a cover on the sample space with an
appropriate perceptual tolerance relation.
Method 1 Perceptual Tolerance
Cantor, Fuzzy, Near, and Rough Sets in Image Analysis 1–7
1. identify a sample space O and a set F to formulate a perceptual system O, F, and then
2. introduce a tolerance relation τε that defines a cover on O.
represented by what is known as a probe function that maps an object to a real value. Since our
main interest is in detecting similarities between seemingly quite disjoint sets such as subimages in
an image or pairs of classes in coverings on a pair of images, a near set is defined in context of a
tolerance space.
Definition 2 Tolerance Near Sets (Peters, 2010)
Let O, F be a perceptual system. Put ε ∈ ℜ, B ⊂ F. Let X,Y ⊂ O denote disjoint sets with
coverings determined by a tolerance relation ∼
=B,ε . Sets X,Y are tolerance near sets if, and only if
there are preclasses A ⊂ X, B ⊂ Y such that A B,ε B.
Rough sets were introduced by Z. Pawlak in (Pawlak, 1981a) and elaborated in (Pawlak, 1981b;
Pawlak and Skowron, 2007c,b,a). In a rough set approach to classifying sets of objects X, one
considers the size of the boundary region in the approximation of X. By contrast, in a near set
approach to classification, one does not consider the boundary region of a set. In particular, assume
that X is a non-empty set belonging to a universe U and that F is a set of features defined either by
total or partial functions. The lower approximation of X relative to B ⊆ F is denoted by B∗ (X) and
the upper approximation of X is denoted by B ∗ (X), where
B∗ (X) = x/∼ ,
B
x/∼ ⊆X
B
B ∗ (X) = x/∼ .
B
x/∼ ∩X=0/
B
A set X is roughly classified whenever BndB (X) is not empty. In other words, X is a rough set
whenever the boundary region BndB (X) = 0. / In sum, a rough set is a Cantor set if, and only if its
approximation boundary is non-empty. It should also be noted that rough sets differ from near sets,
since near sets are defined without reference to an approximation boundary region. This means, for
example, with near sets the image correspondence problem can be solved without resorting to set
approximation.
1. Let (U, B) denote a sample space (universe) U and set of object features B,
2. Using relation ∼B , partition the universe U,
3. Determine the size of the boundary of a set X.
In other words, if a set X does not equal its lower approximation, then the set X is rough, i.e.,
roughly approximated by the equivalence classes in the quotient set U/∼ .
B
In terms of rough sets and image analysis, it can be observed that A. Mrózek and L. Plonka were
pioneers (Mrózek and Plonka, 1993). For example, he was one of the first to introduce a rough
set approach to image analysis and to view a digital image as a universe viewed as a set of points.
The features of pixels (points) in a digital image are a source of knowledge discovery. Using Z.
Pawlak’s indiscernibility relation, it is then a straightforward task to partition an image and to con-
sider set approximation relative to interesting objects contained in subsets of an image. This work
on digital images by A. Mrózek and L. Plonka appeared six or more years before the publication
of papers on approximate mathematical morphology by Lech Polkowski (Polkowski, 1999) (see,
also, (Polkowski, 1993; Polkowski and Skowron, 1994)) and connections between mathematical
morphology and rough sets pointed to by Isabelle Bloch (Bloch, 2000). The early work on the use
of rough sets in image analysis has been followed by a number of articles by S.K. Pal and others
(see, e.g., (Pal and Mitra, 2002; Pal, UmaShankar, and Mitra, 2005; Peters and Borkowski, 2004;
Borkowski and Peters, 2006; Borkowski, 2007; Maji and Pal, 2008; Mushrif and Ray, 2008; Sen
and Pal, 2009)).
From set composition Law 3, it can be observed that rough sets are Cantor sets.
1.6 Conclusion
In sum, fuzzy sets, near sets and rough sets are particular forms of Cantor sets. In addition, each of
these sets in the computational intelligence spectrum offer very useful approaches in image analysis,
especially in classifying objects.
Acknowledgements
This research by James Peters has been supported by the Natural Sciences and Engineering Re-
search Council of Canada (NSERC) grant 185986, Manitoba Centre of Excellence Fund (MCEF)
grant, Canadian Centre of Excellence (NCE) and Canadian Arthritis Network grant SRI-BIO-05,
and Manitoba Hydro grant T277 and that of Sankar Pal has been supported by the J.C. Bose Fel-
lowship of the Govt. of India.
1–12 Rough Fuzzy Image Analysis
Bibliography
Bloch, L. 2000. On links between mathematical morphology and rough sets. Pattern Recog-
nition 33(9):1487–1496.
Bolzano, B. 1959. Paradoxien des unendlichen (paradoxes of the infinite), trans. by d.a. steele.
London: Routledge and Kegan Paul.
Borkowski, M., and J.F. Peters. 2006. Matching 2d image segments with genetic algorithms
and approximation spaces. Transactions on Rough Sets V(LNAI 4100):63–101.
Bourbakis, N. G. 2002. Emulating human visual perception for measuring difference in images
using an spn graph approach. IEEE Transactions on Systems, Man, and Cybernetics, Part B
32(2):191–201.
Deng, T.Q., and H.J.A.M. Heijmans. 2002. Grey-scale morphology based on fuzzy logic. J.
Math. Imag. Vis. 16:155–171.
Fashandi, H., J.F. Peters, and S. Ramanna. 2009. L2 norm length-based image similarity mea-
sures: Concrescence of image feature histogram distances. In Signal and image processing,
int. assoc. of science & technology for development, 178–185. Honolulu, Hawaii.
Gupta, S., and K.S. Patnaik. 2008. Enhancing performance of face recognition systems by
using near set approach for selecting facial features. J. Theoretical and Applied Information
Technology 4(5):433–441.
Guyon, I., S. Gunn, M. Nikravesh, and L.A. Zadeh. 2006. Feature extraction. foundations and
applications. Berlin: Springer.
Hassanien, A.E., A. Abraham, J.F. Peters, G. Schaefer, and C. Henry. 2009. Rough sets and
near sets in medical imaging: A review. IEEE Trans. Info. Tech. in Biomedicine 13(6):
955–968. Digital object identifier: 10.1109/TITB.2009.2017017.
Henry, C., and J.F. Peters. 2007, 475-482. Image pattern recognition using approximation
spaces and near sets. In Proc. 11th int. conf. on rough sets, fuzzy sets, data mining and
granular computing (rsfdgrc 2007), joint rough set symposium (jrs 2007), lecture notes in
artificial intelligence 4482. Heidelberg, Germany.
———. 2008, 1-6. Near set image segmentation quality index. In Geobia 2008 pixels, objects,
intelligence. geographic object based image analysis for the 21st century. University of
Calgary, Alberta.
———. 2009a. Near set evaluation and recognition (near) system. Tech. Rep., Computationa
Intelligence Laboratory, University of Manitoba. UM CI Laboratory Technical Report No.
TR-2009-015.
Cantor, Fuzzy, Near, and Rough Sets in Image Analysis 1–13
Hoogs, A., R. Collins, R. Kaucic, and J. Mundy. 2003. A common set of perceptual observ-
ables for grouping, figure-ground discrimination, and texture classification. IEEE Transac-
tions on Pattern Analysis and Machine Intelligence 25(4):458–474.
Lavine, S. 1994. Understanding the infinite. Cambridge, MA: Harward University Press.
Maji, P., and S.K. Pal. 2008. Maximum class separability for rough-fuzzy c-means based brain
mr image segmentation. Transactions on Rough Sets IX, LNCS-5390:114134.
Martino, F.D., S. Sessa, and H. Nobuhara. 2008. Eigen fuzzy sets and image informa-
tion retrieval. In Handbook of granular computing, ed. W. Pedrycz, A. Skowron, and
V. Kreinovich, 863–872. West Sussex, England: John Wiley & Sons, Ltd.
Meghdadi, A.H., J.F. Peters, and S. Ramanna. 2009. Tolerance classes in measuring image
resemblance. Intelligent Analysis of Images & Videos, KES 2009, Part II, Knowledge-Based
and Intelligent Information and Engineering Systems, LNAI 5712 127–134. ISBN 978-3-
64-04591-2, doi 10.1007/978-3-642-04592-9 16.
Merleau-Ponty, Maurice. 1945, 1965. Phenomenology of perception. Paris and New York:
Smith, Gallimard, Paris and Routledge & Kegan Paul. Trans. by Colin Smith.
Mrózek, A., and L. Plonka. 1993. Rough sets in image analysis. Foundations of Computing
and Decision Sciences 18(3-4):268–273.
Mushrif, M., and A.K. Ray. 2008. Color image segmentation: Rough-set theoretic approach.
Pattern Recognition Letters 29(4):483493.
Nachtegael, M., and E.E. Kerre. 2001. Connections between binary, grayscale and fuzzy
mathematical morphologies. Fuzzy Sets and Systems 124:73–85.
Orłowska, E. 1982. Semantics of vague concepts. applications of rough sets. Polish Academy
of Sciences 469. In G.Dorn, P. Weingartner (Eds.), Foundations of Logic and Linguistics.
Problems and Solutions, Plenum Press, London/NY, 1985, 465-482.
Pal, S.K. 1982. A note on the quantitative measure of image enhancement through fuzziness.
IEEE Trans. Pattern Anal. Machine Intell. PAMI-4(2):204–208.
———. 1986. A measure of edge ambiguity using fuzzy sets. Pattern Recognition Letters
4(1):51–56.
———. 1992. Fuzziness, image information and scene analysis. In An introduction to fuzzy
logic applications in intelligent systems, ed. R.R. Yager and L.A. Zadeh, 147–183. Dor-
drecht: Kluwer Academic Publishers.
Pal, S.K., and R.A. King. 1980. Image enhancement with fuzzy set. Electronics Letters 16(10):
376–378.
———. 1981. Image enhancement using smoothing with fuzzy set. IEEE Trans. Syst. Man
and Cyberns. SMC-11(7):495–501.
1–14 Rough Fuzzy Image Analysis
Pal, S.K., R.A. King, and A.A. Hashim. 1983. Image description and primitive extraction
using fuzzy sets. IEEE Trans. Syst. Man and Cyberns. SMC-13(1):94–100.
Pal, S.K., and A.B. Leigh. 1995. Motion frame analysis and scene abstraction: Discrimination
ability of fuzziness measures. J. Intelligent & Fuzzy Systems 3:247–256.
Pal, S.K., and P. Mitra. 2002. Multispectral image segmentation using rough set initialized em
algorithm. IEEE Transactions on Geoscience and Remote Sensing 11:24952501.
Pal, S.K., and S. Mitra. 1996. Noisy fingerprint classification using multi layered perceptron
with fuzzy geometrical and textual features. Fuzzy Sets and Systems 80(2):121–132.
Pal, S.K., B. UmaShankar, and P. Mitra. 2005. Granular computing, rough entropy and object
extraction. Pattern Recognition Letters 26(16):401–416.
Pavel, M. 1993. Fundamentals of pattern recognition. 2nd ed. N.Y., U.S.A.: Marcel Dekker,
Inc.
Pawlak, Z., and A. Skowron. 2007a. Rough sets and boolean reasoning. Information Sciences
177:41–73.
Peters, J.F. 2007a. Near sets. general theory about nearness of objects. Applied Mathematical
Sciences 1(53):2609–2029.
———. 2007b. Near sets. general theory about nearness of objects. Applied Mathematical
Sciences 1(53):2609–2029.
———. 2007c. Near sets. special theory about nearness of objects. Fundamenta Informaticae
75(1-4):407–433.
———. 2009b. Tolerance near sets and image correspondence. Int. J. of Bio-Inspired Com-
putation 1(4):239–445.
———. 2010. Corrigenda and addenda: Tolerance near sets and image correspondence. Int.
J. Bio-Inspired Computation 2(5). in press.
Peters, J.F., and M. Borkowski. 2004. K-means indiscernibility relation over pixels. In Lecture
notes in computer science 3066, ed. S. Tsumoto, R. Slowinski, K. Komorowski, and J.W.
Gryzmala-Busse, 580–585. Berlin: Springer. Doi 10.1007/b97961.
Peters, J.F., and C. Henry. 2006. Reinforcement learning with approximation spaces. Funda-
menta Informaticae 71:323–349.
Cantor, Fuzzy, Near, and Rough Sets in Image Analysis 1–15
Peters, J.F., and L. Puzio. 2009. Image analysis with anisotropic wavelet-based nearness mea-
sures. International Journal of Computational Intelligence Systems 3(2):1–17.
Peters, J.F., and S. Ramanna. 2009. Affinities between perceptual granules: Foundations and
perspectives. In Human-centric information processing through granular modelling sci 182,
ed. A. Bargiela and W. Pedrycz, 49–66. Berlin: Springer-Verlag.
Peters, J.F., A. Skowron, and J. Stepaniuk. 2007. Nearness of objects: Extension of approxi-
mation space model. Fundamenta Informaticae 79(3-4):497–512.
Peters, J.F., and P. Wasilewski. 2009. Foundations of near sets. Information Sciences. An Inter-
national Journal 179:3091–3109. Digital object identifier: doi:10.1016/j.ins.2009.04.018.
Poincaré, H. 1902. La science et l’hypothèse. Paris: Ernerst Flammarion. Later ed,,̇ Champs
sciences, Flammarion, 1968 & Science and Hypothesis, trans. by J. Larmor, Walter Scott
Publishing, London, 1905.
Polkowski, L. 1993. Mathematical morphology of rough sets. Bull. Polish Acad. Ser. Sci.Math,
Warsaw: Polish Academy of Sciences.
———. 1999. Approximate mathematical morphology. rough set approach. Rough and Fuzzy
Sets in Soft Computing, Berlin: Springer - Verlag.
Sen, D., and S. K. Pal. 2009. Histogram thrsholding using fuzzy and rough means of associa-
tion error. IEEE Trans. Image Processing 18(4):879–888.
Sussner, P., and M.E. Valle. 2008. Fuzzy associative memories and their relationship to math-
ematical morphology. In Handbook of granular computing, ed. W. Pedrycz, A. Skowron,
and V. Kreinovich, 733–753. West Sussex, England: John Wiley & Sons, Ltd.
Watanabe, S. 1985. Pattern recognition: Human and mechanical. John Wiley & Sons: Chich-
ester.
Zeeman, E.C. 1962. The topology of the brain and the visual perception. New Jersey: Prentice
Hall. In K.M. Fort, Ed., Topology of 3-manifolds and Selected Topics, 240-256.
2
Rough-Fuzzy Clustering Algorithm
for Segmentation of Brain MR
Images
2.1 Introduction
Segmentation is a process of partitioning an image space into some non-overlapping mean-
ingful homogeneous regions. The success of an image analysis system depends on the quality
of segmentation (Rosenfeld and Kak, 1982). A segmentation method is supposed to find
those sets that correspond to distinct anatomical structures or regions of interest in the
image. In the analysis of medical images for computer-aided diagnosis and therapy, seg-
mentation is often required as a preliminary stage. However, medical image segmentation is
a complex and challenging task due to intrinsic nature of the images. The brain has a par-
ticularly complicated structure and its precise segmentation is very important for detecting
tumors, edema, and necrotic tissues, in order to prescribe appropriate therapy (Suetens,
2002).
In medical imaging technology, a number of complementary diagnostic tools such as x-
ray computer tomography (CT), magnetic resonance imaging (MRI), and position emission
tomography (PET) are available. MRI is an important diagnostic imaging technique for
the early detection of abnormal changes in tissues and organs. Its unique advantage over
other modalities is that it can provide multispectral images of tissues with a variety of
2–1
2–2 Rough Fuzzy Image Analysis
contrasts based on the three MR parameters ρ, T1, and T2. Therefore, majority of research
in medical image segmentation concerns MR images (Suetens, 2002).
Conventionally, the brain MR images are interpreted visually and qualitatively by ra-
diologists. Advanced research requires quantitative information, such as the size of the
brain ventricles after a traumatic brain injury or the relative volume of ventricles to brain.
Fully automatic methods sometimes fail, producing incorrect results and requiring the in-
tervention of a human operator. This is often true due to restrictions imposed by image
acquisition, pathology and biological variation. So, it is important to have a faithful method
to measure various structures in the brain. One of such methods is the segmentation of
images to isolate objects and regions of interest.
Many image processing techniques have been proposed for MR image segmentation, most
notably thresholding (Lee, Hun, Ketter, and Unser, 1998; Maji, Kundu, and Chanda, 2008),
region-growing (Manousakes, Undrill, and Cameron, 1998), edge detection (Singleton and
Pohost, 1997), pixel classification (Pal and Pal, 1993; Rajapakse, Giedd, and Rapoport,
1997) and clustering (Bezdek, 1981; Leemput, Maes, Vandermeulen, and Suetens, 1999;
Wells III, Grimson, Kikinis, and Jolesz, 1996). Some algorithms using the neural network
approach have also been investigated in the MR image segmentation problems (Cagnoni,
Coppini, Rucci, Caramella, and Valli, 1993; Hall, Bensaid, Clarke, Velthuizen, Silbiger, and
Bezdek, 1992). One of the main problems in medical image segmentation is uncertainty.
Some of the sources of this uncertainty include imprecision in computations and vagueness
in class definitions. In this background, the possibility concept introduced by the fuzzy
set theory (Zadeh, 1965) and rough set theory (Pawlak, 1991) have gained popularity in
modeling and propagating uncertainty. Both fuzzy set and rough set provide a mathematical
framework to capture uncertainties associated with human cognition process (Dubois and
H.Prade, 1990; Maji and Pal, 2007b; Pal, Mitra, and Mitra, 2003). The segmentation of MR
images using fuzzy c-means has been reported in (Bezdek, 1981; Brandt, Bohan, Kramer,
and Fletcher, 1994; Hall et al., 1992; Li, Goldgof, and Hall, 1993; Xiao, Ho, and Hassanien,
2008). Image segmentation using rough sets has also been done (Mushrif and Ray, 2008;
Pal and Mitra, 2002; Widz, Revett, and Slezak, 2005a,b; Widz and Slezak, 2007; Hassanien,
2007).
In this chapter, a hybrid algorithm called rough-fuzzy c-means (RFCM) algorithm is pre-
sented for segmentation of brain MR images. Details of this algorithm have been reported
in (Maji and Pal, 2007a,c). The RFCM algorithm is based on both rough sets and fuzzy
sets. While the membership function of fuzzy sets enables efficient handling of overlapping
partitions, the concept of lower and upper approximations of rough sets deals with uncer-
tainty, vagueness, and incompleteness in class definition. Each partition is represented by
a cluster prototype (centroid), a crisp lower approximation, and a fuzzy boundary. The
lower approximation influences the fuzziness of the final partition. The cluster prototype
(centroid) depends on the weighting average of the crisp lower approximation and fuzzy
boundary. However, an important issue of the RFCM based brain MR image segmentation
method is how to select initial prototypes of different classes or categories. The concept of
discriminant analysis, based on the maximization of class separability, is used to circumvent
the initialization and local minima problems of the RFCM, and enables efficient segmenta-
tion of brain MR images (Maji and Pal, 2008). The effectiveness of the RFCM algorithm,
along with a comparison with other c-means algorithms, is demonstrated on a set of brain
MR images using some standard validity indices.
The chapter is organized as follows: Section 2.2 briefly introduces the necessary notions
of fuzzy c-means and rough sets. In Section 2.3, the RFCM algorithm is described based on
the theory of rough sets and fuzzy c-means. While Section 2.4 deals with pixel classification
problem, Section 2.5 gives an overview of the feature extraction techniques employed in seg-
Rough-Fuzzy Clustering Algorithm for Segmentation of Brain MR Images 2–3
mentation of brain MR images along with the initialization method of c-means algorithm
based on the maximization of class separability. Implementation details, experimental re-
sults, and a comparison among different c-means are presented in Section 2.6. Concluding
remarks are given in Section 2.7.
where 1 ≤ ḿ < ∞ is the fuzzification factor, vi is the ith centroid corresponding to cluster
βi , µij ∈ [0, 1] is the fuzzy membership of the pattern xj to cluster βi , and ||.|| is the distance
norm, such that
1X X
n n
vi = (µij )ḿ xj ; where ni = (µij )ḿ (2.2)
ni j=1 j=1
and
X
c
dij ḿ−1
)−1 ; where d2ij = ||xj − vi ||2
2
µij = ( ( ) (2.3)
dkj
k=1
subject to
X
c X
n
µij = 1, ∀j, and 0 < µij < n, ∀i.
i=1 j=1
The process begins by randomly choosing c objects as the centroids (means) of the c
clusters. The memberships are calculated based on the relative distance of the object xj to
the centroids by Equation 2.3. After computing memberships of all the objects, the new
centroids of the clusters are calculated as per Equation 2.2. The process stops when the
centroids stabilize. That is, the centroids from the previous iteration are identical to those
generated in the current iteration. The basic steps are outlined as follows:
1. Assign initial means vi , i = 1, 2, · · · , c. Choose values for ḿ and threshold ǫ. Set
iteration counter t = 1.
2. Compute memberships µij by Equation 2.3 for c clusters and n objects.
3. Update mean (centroid) vi by Equation 2.2.
4. Repeat steps 2 to 4, by incrementing t, until |µij (t) − µij (t − 1)| > ǫ.
Although fuzzy c-means is a very useful clustering method, the resulting memberships
values do not always correspond well to the degrees of belonging of the data, and it may
2–4 Rough Fuzzy Image Analysis
U/R = {X1 , X2 , · · · , Xm }
That is, the lower approximation R(X) is the union of all the elementary sets which are
subsets of X, and the upper approximation R(X) is the union of all the elementary sets
which have a non-empty intersection with X. The interval [R(X), R(X)] is the representa-
tion of an ordinary set X in the approximation space < U, R > or simply called the rough
set of X. The lower (resp., upper) approximation R(X) (resp., R(X)) is interpreted as the
collection of those elements of U that definitely (resp., possibly) belong to X. Further,
• a set X ∈ 2U is said to be definable (or exact) in < U, R > iff R(X) = R(X).
• for any X, Y ∈ 2U , X is said to be roughly included in Y , denoted by X ⊂Y
˜ , iff
R(X) ⊆ R(Y ) and R(X) ⊆ R(Y ).
• X and Y is said to be roughly equal, denoted by X ≃R Y , in < U, R > iff
R(X) = R(Y ) and R(X) = R(Y ).
|R(X)|
αR (X) =
|R(X)|
Rough-Fuzzy Clustering Algorithm for Segmentation of Brain MR Images 2–5
X
c X X
c X
A1 = ||xj − vi ||2 B1 = (µij )ḿ ||xj − vi ||2
i=1 xj ∈A(βi ) i=1 xj ∈B(βi )
vi represents the centroid of the ith cluster βi , the parameter w and w̃ correspond to the
relative importance of lower bound and boundary region, and w + w̃ = 1. Note that, µij
has the same meaning of membership as that in fuzzy c-means.
In the RFCM, each cluster is represented by a centroid, a crisp lower approximation,
and a fuzzy boundary (Fig. 2.1). The lower approximation influences the fuzziness of final
partition. According to the definitions of lower approximations and boundary of rough sets,
if an object xj ∈ A(βi ), then xj ∈/ A(βk ), ∀k 6= i, and xj ∈
/ B(βi ), ∀i. That is, the object
xj is contained in βi definitely. Thus, the weights of the objects in lower approximation of
a cluster should be independent of other centroids and clusters, and should not be coupled
with their similarity with respect to other centroids. Also, the objects in lower approxima-
tion of a cluster should have similar influence on the corresponding centroid and cluster.
Whereas, if xj ∈ B(βi ), then the object xj possibly belongs to βi and potentially belongs
to another cluster. Hence, the objects in boundary regions should have different influence
on the centroids and clusters. So, in the RFCM, the membership values of objects in lower
approximation are µij = 1, while those in boundary region are the same as fuzzy c-means
(Equation 2.3). In other word, the RFCM algorithm first partitions the data into two classes
- lower approximation and boundary. Only the objects in boundary are fuzzified.
2–6 Rough Fuzzy Image Analysis
Cluster βi
Fuzzy Boundary B( βi )
with µ ij [0, 1]
FIGURE 2.1 RFCM: cluster βi is represented by crisp lower bound and fuzzy boundary
1 X
C1 = xj ; where |A(βi )| represents the cardinality of A(βi )
|A(βi )|
xj ∈A(βi )
1 X X
and D1 = (µij )ḿ xj ; where ni = (µij )ḿ
ni
xj ∈B(βi ) xj ∈B(βi )
Thus, the cluster prototypes (centroids) depend on the parameters w and w̃, and fuzzifica-
tion factor ḿ rule their relative influence. The correlated influence of these parameters and
fuzzification factor, makes it somewhat difficult to determine their optimal values. Since
the objects lying in lower approximation definitely belong to a cluster, they are assigned
a higher weight w compared to w̃ of the objects lying in boundary region. Hence, for the
RFCM, the values are given by 0 < w̃ < w < 1.
From the above discussions, the following properties of the RFCM algorithm can be
derived.
S
1. A(βi ) = U , U be the set of objects of concern.
2. A(βi ) ∩ A(βk ) = ∅, ∀i 6= k.
3. A(βi ) ∩ B(βi ) = ∅, ∀i.
4. ∃i, k, B(βi ) ∩ B(βk ) 6= ∅.
5. µij = 1, ∀xj ∈ A(βi ).
6. µij ∈ [0, 1], ∀xj ∈ B(βi ).
Let us briefly comment on some properties of the RFCM. The property 2 says that if an
object xj ∈ A(βi ) ⇒ xj ∈
/ A(βk ), ∀k 6= i. That is, the object xj is contained in βi definitely.
The property 3 establishes the fact that if xj ∈ A(βi ) ⇒ xj ∈ / B(βi ), - that is, an object
may not be in both lower and boundary region of a cluster βi . The property 4 says that
Rough-Fuzzy Clustering Algorithm for Segmentation of Brain MR Images 2–7
The performance of the RFCM depends on the value of δ, which determines the class
labels of all the objects. In other word, the RFCM partitions the data set into two classes
- lower approximation and boundary, based on the value of δ. In the present work, the
following definition is used:
1X
n
δ= (µij − µkj ) (2.6)
n j=1
where n is the total number of objects, µij and µkj are the highest and second highest
memberships of xj . That is, the value of δ represents the average difference of two highest
memberships of all the objects in the data set. A good clustering procedure should make
the value of δ as high as possible. The value of δ is, therefore, data dependent.
In this section, we present the results of different c-means algorithms on pixel classification
of brain MR images, that is, the results of clustering based on only gray value of pixels.
Above 100 MR images with different sizes and 16 bit gray levels are tested with different
c-means algorithms. All the brain MR images are collected from Advanced Medicare and
Research Institute, Salt Lake, Kolkata, India. The comparative performance of different c-
means is reported with respect to DB, and Dunn index, as well as the β index (Pal, Ghosh,
and Sankar, 2000), which are reported next.
for 1 ≤ i, k ≤ c. The DB index minimizes the within-cluster distance S(vi ) and maximizes
the between-cluster separation d(vi , vk ). Therefore, for a given data set and c value, the
higher the similarity values within the clusters and the between-cluster separation, the lower
would be the DB index value. A good clustering procedure should make the value of DB
index as low as possible.
Dunn Index:
Dunn index (Bezdek and Pal, 1988) is also designed to identify sets of clusters that are
compact and well separated. Dunn index maximizes
d(vi , vk )
Dunn = mini mini6=k for 1 ≤ i, k, l ≤ c.
maxl S(vl )
A good clustering procedure should make the value of Dunn index as high as possible.
β Index:
The β-index of Pal et al. (Pal et al., 2000) is defined as the ratio of the total variation and
within-cluster variation, and is given by
N Xc X ni Xc X ni Xc
β= ; where N = ||xij − v||2 ; M = ||xij − vi ||2 ; ni = n;
M i=1 j=1 i=1 j=1 i=1
ni is the number of objects in the ith cluster (i = 1, 2, · · · , c), n is the total number of
objects, xij is the jth object in cluster i, vi is the mean or centroid of ith cluster, and v is
the mean of n objects. For a given image and c value, the higher the homogeneity within
the segmented regions, the higher would be the β value. The value of β increases with c.
Consider the image of Fig. 2.3 as an example, which represents an MR image (I-20497774)
of size 256×180 with 16 bit gray levels. So, the number of objects in the data set of IMAGE-
20497774 is 46080. Table 2.1 depicts the values of DB index, Dunn index, and β index of
FCM and RFCM for different values of c on the data set of I-20497774 considering only
gray value of pixel. The results reported here with respect to DB and Dunn index confirm
that both FCM and RFCM achieve their best results for c = 4 (background, gray matter,
white matter, and cerebro-spinal fluid). Also, the value of β index, as expected, increases
Rough-Fuzzy Clustering Algorithm for Segmentation of Brain MR Images 2–9
TABLE 2.1 Performance of FCM and RFCM on
I-20497774
Value DB Index Dunn Index β Index
of c FCM RFCM FCM RFCM FCM RFCM
2 0.51 0.21 2.30 6.17 2.15 2.19
3 0.25 0.17 1.11 1.62 3.55 3.74
4 0.16 0.15 1.50 1.64 9.08 9.68
5 0.39 0.17 0.10 0.64 10.45 10.82
6 0.20 0.19 0.66 1.10 16.93 17.14
7 0.23 0.27 0.98 0.12 21.63 22.73
8 0.34 0.27 0.09 0.31 25.82 26.38
9 0.32 0.28 0.12 0.13 31.75 32.65
10 0.30 0.24 0.08 0.12 38.04 39.31
with increase in the value of c. For a particular value of c, the performance of RFCM is
better than that of FCM.
Fig. 2.2 shows the scatter plots of the highest and second highest memberships of all the
objects in the data set of I-20497774 at first and final iterations respectively, considering
w = 0.95, ḿ1 = 2.0, and c = 4. The diagonal line represents the zone where two highest
memberships of objects are equal. From Fig. 2.2, it is observed that though the average
difference between two highest memberships of the objects are very low at first iteration
(δ = 0.145), they become ultimately very high at the final iteration (δ = 0.652).
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
Highest Membership Value Highest Membership Value
FIGURE 2.2 Scatter plots of two highest membership values of all objects in data set I-20497774
Table 2.2 compares the performance of different c-means algorithms on some brain MR
images with respect to DB, Dunn, and β index considering c = 4 (back-ground, gray
matter, white matter, and CSF). All the results reported in Table 2.2 confirm that the
RFCM algorithm produces pixel clusters more promising than do the conventional methods.
Some of the existing algorithms like PCM and FPCM have failed to produce multiple
clusters as they generate coincident clusters even when they have been initialized with the
final prototypes of FCM. Also, the values of DB, Dunn, and β index of RFCM are better
compared to other c-means algorithms.
In this section, the feature extraction methodology for segmentation of brain MR images
is first described. Next, the methodology to select initial centroids for different c-means
algorithms is provided based on the concept of maximization of class separability (Maji and
Pal, 2008).
Homogeneity
If H is the homogeneity of a pixel Im,n within 3 × 3 neighborhood, then
1
H=1− {|Im−1,n−1 + Im+1,n+1 − Im−1,n+1 − Im+1,n−1 | +
6(Imax − Imin )
|Im−1,n−1 + 2Im,n−1 + Im+1,n−1 − Im−1,n+1 − 2Im,n+1 − Im+1,n+1 |}
where Imax and Imin represent the maximum and minimum gray values of the image. The
region that is entirely within an organ will have a high H value. On the other hand, the
regions that contain more than one organ will have lower H values (Maji and Pal, 2008).
Rough-Fuzzy Clustering Algorithm for Segmentation of Brain MR Images 2–11
Edge Value
In MR imaging, the histogram of the given image is in general unimodal. One side of the
peak may display a shoulder or slope change, or one side may be less steep than the other,
reflecting the presence of two peaks that are close together or that differ greatly in height.
The histogram may also contain a third, usually smaller, population corresponding to points
on the object-background border. These points have gray levels intermediate between those
of the object and background; their presence raises the level of the valley floor between the
two peaks, or if the peaks are already close together, makes it harder to detect the fact that
they are not a single peak.
As the histogram peaks are close together and very unequal in size, it may be difficult
to detect the valley between them. In determining how each point of the image should
contribute to the segmentation method, the current method takes into account the rate of
change of gray level at the point, as well as the point’s gray level (edge value); that is, the
maximum of differences of average gray levels in pairs of horizontally and vertically adjacent
2 × 2 neighborhoods (Maji et al., 2008; Weszka and Rosenfeld, 1979). If ∆ is the edge value
at a given point Im,n , then
1
∆= max{|Im−1,n + Im−1,n+1 + Im,n + Im,n+1 − Im+1,n − Im+1,n+1 − Im+2,n − Im+2,n+1 |,
4
|Im,n−1 + Im,n + Im+1,n−1 + Im+1,n − Im,n+1 − Im,n+2 − Im+1,n+1 − Im+1,n+2 |}
According to the image model, points interior to the object and background should gen-
erally have low edge values, since they are highly correlated with their neighbors, while
those on the object-background border should have high edge values (Maji et al., 2008).
entropy, difference variance, and difference entropy, are computed for each window. For
four angular directions, a set of four values is obtained for each of ten measures. The mean
of each of the ten measures, averaged over four values, along with gray value, homogeneity,
and edge value of the pixel, comprise the set of 13 features which is used as feature vector
of the corresponding pixel.
Here, L is the total number of discrete values ranging between [0, L − 1], T is the threshold
value, which maximizes J(T), and h(z) represents the percentage of data having feature
value z over the total number of discrete values of the corresponding feature. To maximize
J(T), the means of the two classes should be as well separated as possible and the variances
in both classes should be as small as possible.
Based on the concept of maximization of class separability, the method for selecting initial
centroids is described next. The main steps of this method proceeds as follows.
1. The data set X = {x1 , · · · , xj , · · · , xn } with xj ∈ ℜm are first discretized to
facilitate class separation method. Suppose, the possible value range of a feature
fm in the data set is (fm,min , fm,max ), and the real value that the data element xj
takes at fm is fmj , then the discretized value of fmj is
fmj − fm,min
Discretized(fmj ) = (L − 1) × (2.8)
fm,max − fm,min
where L is the total number of discrete values ranging between [0, L − 1].
2. For each feature fm , calculate h(z) for 0 ≤ z < L.
Rough-Fuzzy Clustering Algorithm for Segmentation of Brain MR Images 2–13
3. Calculate the threshold value Tm for the feature fm , which maximizes class sep-
arability along that feature.
4. Based on the threshold Tm , discretize the corresponding feature fm of the data
element xj as follows
1, if Discretized(fmj ) ≥ Tm
f mj =
0, Otherwise
5. Repeat steps 2 to 4 for all the features and generate the set of discretized objects
X = {x1 , · · · , xj , · · · , xn }.
6. Calculate total number of similar discretized objects N(xi ) and mean of similar
objects v(xi ) of xi as
X
n
1 X
n
N(xi ) = δj and v(xi ) = δj × xj
j=1
N(xi ) j=1
1 if xj = xi
where δj =
0 Otherwise
7. Sort n objects according to their values of N(xi ) such that N(x1 ) > N(x2 ) >
· · · > N(xn ).
8. If xi = xj , then N(xi ) = N(xj ) and v(xj ) should not be considered as a centroid
(mean), resulting in a reduced set of objects to be considered for initial centroids.
9. Let there be ń objects in the reduced set having N(xi ) values such that N(x1 ) >
N(x2 ) > · · · > N(xń ). A heuristic threshold function can be defined as follows
(Banerjee, Mitra, and Pal, 1998):
R X ń
1
Tr = ; where R =
ǫ̃ i=1
N(xi ) − N(xi+1 )
where ǫ̃ is a constant (= 0.5, say), so that all the means v(xi ) of the objects in
reduced set having N(xi ) value higher than it are regarded as the candidates for
initial centroids (means).
The value of Tr is high if most of the N(xi )’s are large and close to each other. The above
condition occurs when a small number of large clusters are present. On the other hand, if
the N(xi )’s have wide variation among them, then the number of clusters with smaller size
increases. Accordingly, Tr attains a lower value automatically. Note that the main motive
of introducing this threshold function lies in reducing the number of centroids. Actually,
it attempts to eliminate noisy centroids (data representatives having lower values of N(xi ))
from the whole data set. The whole approach is, therefore, data dependent.
FIGURE 2.3 I-20497774: original and segmented images of HCM, FCM, RCM, and RFCM
RFCM for different values of c on the data set of I-20497774, considering w = 0.95 and
ḿ = 2.0. The results reported here with respect to DB and Dunn index confirm that both
FCM and RFCM achieve their best results for c = 4. Also, the value of β index, as expected,
increases with increase in the value of c. For a particular value of c, the performance of
RFCM is better than that of FCM.
Threshold Values:
Gray value = 1959, Homogeneity = 0.17, Edge value = 0.37
Angular second moment = 0.06, Contrast = 0.12
Correlation = 0.57, Inverse difference moment = 0.18
Sum average = 0.17, Sum variance = 0.14, Sum entropy = 0.87
Entropy = 0.88, Difference variance = 0.07, Difference entropy = 0.79
Finally, Table 2.5 provides the comparative results of different c-means algorithms on
I-20497774 with respect to the values of DB index, Dunn index, and β index. The cor-
responding segmented images along with the original one are presented in Fig. 2.3. The
results reported in Fig. 2.3 and Table 2.5 confirm that the RFCM algorithm produces seg-
mented image more promising than do the conventional c-means algorithms. Some of the
existing algorithms like PCM and FPCM fail to produce multiple segments as they generate
coincident clusters even when they are initialized with final prototypes of the FCM.
Rough-Fuzzy Clustering Algorithm for Segmentation of Brain MR Images 2–15
TABLE 2.4 Performance of FCM and RFCM on
I-20497774 data set
Value DB Index Dunn Index β Index
of c FCM RFCM FCM RFCM FCM RFCM
2 0.38 0.19 2.17 3.43 3.62 4.23
3 0.22 0.16 1.20 1.78 7.04 7.64
4 0.15 0.13 1.54 1.80 11.16 13.01
5 0.29 0.19 0.95 1.04 11.88 14.83
6 0.24 0.23 0.98 1.11 19.15 19.59
7 0.23 0.21 1.07 0.86 24.07 27.80
8 0.31 0.21 0.46 0.95 29.00 33.02
9 0.30 0.24 0.73 0.74 35.06 40.07
10 0.30 0.22 0.81 0.29 41.12 44.27
Section 2.5.2 for the data sets I-20497761, I-20497763, and I-20497777 (Fig. 2.4). The
Table 2.8 compares the performance of different c-means algorithms on some brain MR
images with respect to DB, Dunn, and β index. The segmented versions of different c-
Rough-Fuzzy Clustering Algorithm for Segmentation of Brain MR Images 2–17
means are shown in Figs. 2.5-2.7. All the results reported in Table 2.8 and Figs. 2.5-2.7
confirm that although each c-means algorithm, except PCM and FPCM, generates good
segmented images, the values of DB, Dunn, and β index of the RFCM are better compared
to other c-means algorithms. Both PCM and FPCM fail to produce multiple segments of
the brain MR images as they generate coincident clusters even when they are initialized
with the final prototypes of other c-means algorithms.
Table 2.8 also provides execution time (in milli sec.) of different c-means. The execution
time required for the RFCM is significantly lesser compared to FCM. For the HCM and
RCM, although the execution time is less, the performance is considerably poorer than that
of RFCM. Following conclusions can be drawn from the results reported in this chapter:
FIGURE 2.5 I-20497761: segmented versions of HCM, FCM, RCM, and RFCM
FIGURE 2.6 I-20497763: segmented versions of HCM, FCM, RCM, and RFCM
FIGURE 2.7 I-20497777: segmented versions of HCM, FCM, RCM, and RFCM
to FCM. But, the performance of RFCM with respect to DB, Dunn, and β is
significantly better than all other c-means. The performance of FCM and RCM
is intermediate between RFCM and HCM.
2. The discriminant analysis based initialization is found to improve the values of
DB, Dunn, and β as well as reduce the time requirement substantially for all
c-means algorithms.
3. Two features proposed in (Maji and Pal, 2008) are as important as Haralick’s
ten features for clustering based segmentation of brain MR images.
4. Use of rough sets and fuzzy memberships adds a small computational load to
HCM algorithm; however the corresponding integrated method (RFCM) shows
a definite increase in Dunn index and decrease in DB index.
The best performance of the segmentation method in terms of DB, Dunn, and β is
achieved due to the following reasons:
1. the discriminant analysis based initialization of centroids enables the algorithm
to converge to an optimum or near optimum solutions;
2. membership of the RFCM handles efficiently overlapping partitions; and
3. the concept of crisp lower bound and fuzzy boundary of the RFCM algorithm
deals with uncertainty, vagueness, and incompleteness in class definition.
In effect, promising segmented brain MR images are obtained using the RFCM algorithm.
2.7 Conclusion
A robust segmentation technique is presented in this chapter, integrating the merits of
rough sets, fuzzy sets, and c-means algorithm, for brain MR images. Some new measures
are reported, based on the local properties of MR images, for accurate segmentation. The
method, based on the concept of maximization of class separability, is found to be suc-
cessful in effectively circumventing the initialization and local minima problems of iterative
refinement clustering algorithms like c-means. The effectiveness of the algorithm, along
with a comparison with other algorithms, is demonstrated on a set of brain MR images.
The extensive experimental results show that the rough-fuzzy c-means algorithm produces
a segmented image more promising than do the conventional algorithms.
Acknowledgments.
The authors thank Advanced Medicare and Research Institute, Kolkata, India, for providing
brain MR images. This work was done when S. K. Pal was a Govt. of India J.C. Bose Fellow.
Rough-Fuzzy Clustering Algorithm for Segmentation of Brain MR Images 2–19
Bibliography
Banerjee, Mohua, Sushmita Mitra, and Sankar K Pal. 1998. Rough Fuzzy MLP: Knowl-
edge Encoding and Classification. IEEE Transactions on Neural Networks 9(6):
1203–1216.
Bezdek, J. C., and N. R. Pal. 1988. Some New Indexes for Cluster Validity. IEEE
Transactions on System, Man, and Cybernetics, Part B 28:301–315.
Cagnoni, S., G. Coppini, M. Rucci, D. Caramella, and G. Valli. 1993. Neural Network
Segmentation of Magnetic Resonance Spin Echo Images of the Brain. Journal of
Biomedical Engineering 15(5):355–362.
Dubois, D., and H.Prade. 1990. Rough Fuzzy Sets and Fuzzy Rough Sets. International
Journal of General Systems 17:191–209.
Dunn, J. C. 1974. A Fuzzy Relative of the ISODATA Process and its Use in Detecting
Compact, Well-Separated Clusters. Journal of Cybernetics 3:32–57.
Haralick, R. M., K. Shanmugam, and I. Dinstein. 1973. Textural Features for Image
Classification. IEEE Transactions on Systems, Man and Cybernetics SMC-3(6):
610–621.
Hassanien, Aboul Ella. 2007. Fuzzy Rough Sets Hybrid Scheme for Breast Cancer
Detection. Image Vision Computing 25(2):172–183.
Lee, C., S. Hun, T. A. Ketter, and M. Unser. 1998. Unsupervised Connectivity Based
Thresholding Segmentation of Midsaggital Brain MR Images. Computers in Biology
and Medicine 28:309–338.
3.1 Introduction
Real-life images are inherently embedded with various ambiguities. In order to perceive
the nature of ambiguities in images, let us consider a 1001 × 1001 grayscale image (see
Figure 3.1(a)) that has sinusoidal gray value gradations in horizontal direction. When
an attempt is made to mark the boundary of an arbitrary region in the image, an exact
boundary can not be defined as a consequence of the presence of steadily changing gray
values (gray value gradation). This is evident from Figure 3.1(b) that shows a portion of
the image, where it is known that the pixels in the ‘white’ shaded area uniquely belong
to a region. However, the boundary (on the left and right sides) of this region is vague as
it can lie anywhere in the gray value gradations present in the portion. Value gradation
is a common phenomenon in real-life images and hence it is widely accepted (Pal, 1982;
Pal, King, and Hashim, 1983; Udupa and Saha, 2003) that regions in an image have fuzzy
boundaries.
Moreover, the gray levels at various pixels in grayscale images are considered to be impre-
cise, which means that a gray level resembles other nearby gray levels to certain extents. It
is also true that pixels in a neighborhood with nearby gray levels have limited discernibility
due to the inadequacy of contrast. For example, Figure 3.1(c) shows a 6 × 6 portion cut
from the image in Figure 3.1(a). Although the portion contains gray values separated by
6 gray levels, it appears to be almost homogeneous. The aforementioned ambiguities in
3–1
3–2 Rough Fuzzy Image Analysis
FIGURE 3.1: Ambiguities in a grayscale image with sinusoidal gray value gradations in horizontal
direction
images due to fuzzy boundaries of various regions and rough resemblance of nearby gray
levels is studied and modeled in this chapter. Note that, the aforementioned ambiguities
are related to the indefiniteness in deciding an image pixel as white or black and hence they
can be collectively referred to as the grayness ambiguity (Pal, 1999).
Fuzzy set theory of Lofti Zadeh, is based on the concept of vague boundaries of sets in
the universe of discourse (Klir and Yuan, 2005). Rough set theory of Zdzislaw Pawlak, on
the otherhand, focuses on ambiguity in terms of limited discernibility of sets in the domain
of discourse (Pawlak, 1991). Therefore, fuzzy sets can be used to represent the grayness
ambiguity in images due to the vague definition of region boundaries (fuzzy boundaries)
and rough sets can be used to represent the grayness ambiguity due to the indiscernibility
between individual or groups of gray levels (rough resemblance).
Rough set theory, which was initially developed considering crisp equivalence approx-
imation spaces (Pawlak, 1991), has been generalized by considering fuzzy (Dubois and
Prade, 1990, 1992; Thiele, 1998) and tolerance (Skowron and Stepaniuk, 1996) approxima-
tion spaces. Furthermore, rough set theory, which was initially developed to approximate
crisp sets, has also been generalized to approximate fuzzy sets (Dubois and Prade, 1990,
1992; Thiele, 1998).
In this chapter, we propose the use of the rough set theory and its certain generalizations
to quantify grayness ambiguity in images. Here the generalizations to rough set theory based
on the approximation of crisp and fuzzy sets considering crisp equivalence, fuzzy equiva-
lence, crisp tolerance and fuzzy tolerance approximation spaces in different combinations
are studied. All these combinations give rise to different concepts for modeling vagueness,
which can be quantified using the roughness measure (Pawlak, 1991).
We propose classes of entropy measures which use the roughness measures obtained con-
sidering the aforementioned various concepts for modeling vagueness. We perform rigorous
theoretical analysis of the proposed entropy measures and provide some properties which
they satisfy. We then use the proposed entropy measures to quantify grayness ambiguity in
images, giving an account of the manner in which the grayness ambiguity is captured. We
show that the aforesaid generalizations to rough set theory regarding the approximation of
fuzzy sets can be used to quantify grayness ambiguity due to both fuzzy boundaries and
rough resemblance.
We then propose an image thresholding methodology that employs grayness ambiguity
Exploring the Variety of Random
Documents with Different Content
to serve as the mouthpiece of its author). “Thackeray,” so W. C.
Brownell has asserted in his sympathetic study,
enwraps and embroiders his story with his personal philosophy, charges it with his
personal feeling, draws out with inexhaustible personal zest its typical
suggestiveness, and deals with his material directly instead of dispassionately and
disinterestedly.
III
Yet all its parts are as interdependent as they are numerous and varied. It is
Thackeray’s largest canvas, and it is filled with the greatest ease and to the
borders.... It illustrates manners with an unexampled crowd of characters, the
handling of which, without repetition or confusion, without digression or discord,
exhibits the control of the artist equally with the imaginative and creative faculty of
the poet.
IV
He shrieked out in his cracked voice the parts, and his pupil learned them from
him by rote. He indicated the attitudes, and set and moved those beautiful arms of
hers.... With what indomitable patience and dulness she followed him. She knew
that he made her; and she let herself be made.... She was not grateful, or
ungrateful, or unkind, or ill-humored.
She might not be grateful, but she knew very well who had made
her; she said so simply enough, explaining why she had not earlier
played the more important parts, “I didn’t take the leading business
then; I wasn’t fit for it till Bows taught me.”
So it was that Adrienne Lecouvreur, in the play which Scribe and
Legouvé wrote for Rachel, thanked the little old prompter, Michonnet,
who had taught her, “I was ungrateful in saying I had never had a
teacher. There is a kind-hearted man, a sincere friend, whose
counsels have always sustained me.” And Legouvé has told us that
at one of the rehearsals Rachel suddenly turned from Regnier, who
was the Michonnet, and knelt before Samson, who was the Duc de
Bouillon, and addressed this speech directly to him.
It would be interesting to know whether Thackeray ever saw
‘Adrienne Lecouvreur,’ which was produced in Paris in April, 1849,
six months before ‘Pendennis’ began to appear in monthly parts.
(1920)
IX
MARK TWAIN AND THE THEATER
IX
MARK TWAIN AND THE THEATER
Mark Twain was a born story-teller; he was a born actor; he was not
affrighted by the idea of facing an audience; he was fond of the
theater; he lived in a time when the drama was regaining its proud
position in our literature and when men of letters who had begun as
novelists were turning dramatists. Why is it that he did not leave us
even one play worthy to be set by the side of the ‘Adventures of
Huckleberry Finn’? Why is it that the only piece of his which was
successful on the stage, is a poor thing, not wholly his own? Why is
it that he did not persevere in playwriting as did his fellow humorists,
George Bernard Shaw and George Ade, and his fellow story-tellers,
Barrie and Tarkington?
These are questions which must have occurred to not a few of his
admirers; and they are questions to which it is not easy to find an
immediate answer. Yet there must be an explanation of some sort for
this puzzling fact; and there may be profit in trying to discover it.
Even if the answer shall prove to be incomplete and unsatisfactory,
the inquiry is worth while for its own sake.
II
He was the most consummate public performer I ever saw, and it was an
incomparable pleasure to hear him lecture; on the platform he was the great and
finished actor he probably would not have been on the stage.... When he read his
manuscript to you, it was with a thorough, however involuntary, recognition of its
dramatic qualities.... He was realistic, but he was essentially histrionic; and rightly
so. What we have strongly conceived we ought to make others strongly imagine,
and we ought to use every art to that end.
I had stage-fright then for the first and last time.... After the first agonizing five
minutes, my stage-fright left me, never to return. I know if I was going to be
hanged I could get up and make a good showing—and I intend to.
the greatest of all arts is to write a drama. It is a most difficult thing. It requires the
highest talents possible and the rarest gifts. No, there is another talent that ranks
with it—for anybody can write a drama—I have written about four hundred—but to
get one accepted requires real ability. And I have never had that felicity yet.
III
Altho I have quoted Mark’s assertion that he had never had the
felicity of having a play accepted, he did have two pieces produced
by managers; and a third, written in collaboration with Howells, had a
brief and inglorious career at the expense of its authors. His first
play, made out of one of his novels, drew delighted audiences for
several seasons; the second, written in partnership with Bret Harte,
and the third, written in partnership with Howells, met with so little
success that they sank at once beneath the wave of oblivion, being
almost unknown except in the hazy memories of the few surviving
spectators who chanced to see one or the other during its brief stay
on the stage. No one of the three has ever been published.
After Mark had settled in Hartford he formed a close friendship
with his near neighbor Charles Dudley Warner; and in 1873 they
joined forces in a novel, the ‘Gilded Age.’ They wrote it not so much
in collaboration as in conjunction,—that is to say, each of the writers
was responsible for the chapters he prepared himself; and there was
no integrating co-ordination of their respective contributions. Mark
was the author of more than half of the chapters; and he was the
creator of the one outstanding character, Colonel Mulberry Sellers,
an imaginative reproduction of a man he had known since boyhood,
James Lampton. Mark began by writing the first eleven chapters,
then Warner wrote two, Mark followed with two more; and thus they
worked alternately. They labored, so Mark declared, “in the
superstition that we were writing one coherent yarn, when I suppose,
as a matter of fact, we were writing two incoherent yarns.”
It was not long after the publication of their conjoint work that they
were informed of the performance in San Francisco of a
dramatization by one Gilbert S. Densmore, otherwise unknown to
fame, the character of Colonel Sellers being impersonated by John
T. Raymond. Action was at once taken to put a stop to this
infringement on the copyright of the story. In the end a satisfactory
arrangement was arrived at. Densmore was bought out; Warner,
discovering that his share in the story had been but little drawn upon,
relinquished any claim he might have; Mark made the piece over;
and Raymond continued to play Colonel Sellers, under a contract
which divided the profits between the author and the actor. For a
season or two Mark’s agent travelled with the company and reported
on a postal card every night the author’s share; and Howells has
related how these welcome missives would come about dinnertime
and how Mark would read them aloud in triumph. “One hundred and
fifty dollars—two hundred dollars—three hundred dollars, were the
gay figures which they bore and which he flaunted in the air before
he sat down at table.”
It is difficult now to determine how much of the dramatic skeleton
Densmore had put together to enable Colonel Sellers to exhibit the
facets of his lovable character, survived in the play which drew
crowded houses one long winter in New York. Here Mark himself is
the best witness in his own behalf; and his biographer has quoted
from an unpublished letter a clear-cut statement:
I entirely rewrote the play three separate and distinct times. I had expected to
use little of [Densmore’s] language and but little of his plot. I do not think that there
are now twenty sentences of Mr. Densmore’s in the play, but I used so much of his
plot that I wrote and told him I should pay him about as much more as I had
already paid him, in case the play proved a success.
But the greatest scenes are in the last act, where Colonel Sellers appears as a
witness for the defence of Laura Hawkins. As he mounts the stand he affably
recognizes and shakes hands with several acquaintances among the jury; he
delivers his testimony in the form of a stump speech; he helplessly overrides all
the protests, exceptions, and interruptions of the prosecution; from time to time he
irresistibly turns and addresses the jury and can scarcely be silenced; while the
attorneys are wrangling together he has seized a juryman by the coat-lapel and is
earnestly exhorting him in whisper. The effect is irresistibly ludicrous. It is farce and
not farce, for, however extravagantly impossible the situation is, the man in it is
deliciously true to himself. There is one bit of pathos, where Sellers tells how he
knew Laura as a little girl, and implies that, though she might have killed a man,
she could not have done murder.
IV
A few days later he wrote again, telling his friend that he had
piled up one hundred and fifty-one pages. The first, second and fourth acts are
done, and done to my satisfaction too. Never had so much fun over anything in my
life—such consuming interest and delight.
This piece was intended for Sol Smith Russell. But the theatrical
experts to whom it was submitted did not share its author’s
consuming interest. Dion Boucicault said that it was better than ‘Ah
Sin’; but to say this was saying little. John Brougham wrote that it
was “altogether too diffuse for dramatic representation.” In time
Mark’s own opinion of his play seems to have cooled, and he put his
manuscript aside. Possibly he utilized it more or less many years
later when he wrote ‘Tom Sawyer, Detective’; but this is mere
conjecture.
Then, after a longer interval he asked Howells to collaborate with
him in a sequel to Colonel Sellers; and in ‘My Mark Twain’ Howells
has given a detailed account of their conjoint misadventure. Mark
had a host of suggestions but no story, so Howells supplied one as
best he could; and the two friends spent a hilarious fortnight in
writing the play. Mark had quarrelled with Raymond and did not want
to let him reincarnate Sellers; and yet he had ultimately to recognize
that Raymond was the only actor the public would accept in the
character. So the piece was sent to Raymond, who accepted it,
asking for certain alterations; and then most unexpectedly he
returned the manuscript, refusing to have anything to do with it. After
hawking their play about, the authors arranged to produce it
themselves with Burbank (who was not an actor but an elocutionist-
entertainer) as Sellers,—Burbank playing the part in imitation of
Raymond. At last they had lost confidence in it so completely that
they paid a forfeit rather than undertake the risk of a production in
New York. So it was that the ‘American Claimant, or Mulberry Sellers
Ten Years Later’ was made visible in New York only at a special
matinee in the fall of 1887. It had a few performances in unimportant
out of town theaters; and then it disappeared from the stage. Still, it
had not lived in vain since it supplied material for several chapters in
Mark’s later novel, to which he gave the same title, without the
subtitle.
After this play had been withdrawn from the boards Mark’s
ambition to establish himself as a dramatist did not again manifest
itself. However, it is pleasant to believe that the pain of his own
failure may have been more or less assuaged by the better fortune
of dramatizations of two of his novels.
I have already noted that not long after the publication of the
‘Prince and the Pauper’ Mrs. Clemens had arranged scenes from it
to be acted by members of the family and by their young friends, and
that Mark himself had undertaken the part of Miles Hendon. A little
later a dramatization of the whole story was made by Abby Sage
Richardson; and this was produced in New York in January, 1890. It
achieved instant popularity, as well it might, since the story is
indisputably dramatic and since it has a more direct action than any
other of Mark’s novels. This version, revised by Amélie Rives, was
revived in 1920 by William Faversham, who appeared as Miles
Hendon. The revival met with a reception as warm as that which had
greeted the original production.
In one respect this professional dramatization was inferior to Mrs.
Clemens’s amateur arrangement; it was so devised that one
performer should assume two characters, the little Prince and the
little Pauper; and this necessitated the omission of the culminating
moment in the tale when the Prince and the Pauper stand face to
face. And in both the amateur and the professional performances
these two lads were impersonated by girls. This may have been
necessary, since it is almost impossible to find competent boy actors,
while there are girl actors aplenty; but none the less was it
unfortunate, since a girl is never entirely satisfactory in boy’s clothes.
Very rarely can she conceal from us the fact that she is a girl, doing
her best to be a boy. Curiously enough, boys can act girls’ parts and
make us forget for the moment that they are not what they seem.
Five years after Mrs. Richardson had dramatized the ‘Prince and
the Pauper,’ Frank Mayo made a most effective play out of
‘Pudd’nhead Wilson.’ He arranged the title-part for his own vigorous
and impressive acting. He simplified Mark’s story and he amplified it;
he condensed it and he heightened it; he preserved the ingenious
incidents and the veracious characters; he made his profit out of the
telling dialog; and he was skilful in disentangling the essentially
dramatic elements of Mark’s rather rambling story. He produced it in
New York in the spring of 1895. Mark was then in Europe; but when
he returned he made haste to see the piece. He was discovered by
the audience and called upon for a speech, in which he
congratulated the player-playwright on a “delightful play.” He ended
by saying, “Confidentially I have always had an idea that I was well
equipt to write plays, but I have never encountered a manager who
has agreed with me”—which was not strictly accurate since two
different managers had accepted the ‘Gilded Age’ and ‘Ah Sin.’
When the ‘Gilded Age’ was brought out in New York in the fall of
1874, Mark climbed the eighty steps which led to the editorial offices
of the New York World, then in the control of Manton Marble. He
asked for the city editor and he was shown into the cubicle occupied
by William C. Brownell. He explained that he had come to ask the
editor to puff his play; whereupon Brownell inquired if it was a good
play. “No,” was Mark’s drawling answer, “it isn’t a good play. It’s a
bad play, a damned bad play. I couldn’t write a good play. But it has
a good character. I can write character; and that character is the best
I can do. If it was a good play, I shouldn’t have had to climb up here
to ask you to puff it.”
Here Mark was unconsciously revealing his agreement with
Aristotle, the master of all who know. Aristotle declared that in a
tragedy—and the remark is even more applicable to comedy—plot is
more important than character, since you can have an appealing
drama without character but you cannot have it without plot. Lowell
said the same thing in more detail, in one of his lectures on the ‘Old
English Dramatists.’
In a play we not only expect a succession of scenes, but that each scene should
lead by a logic more or less stringent, if not to the next, at any rate to something
that is to follow and that all should contribute their fraction of impulse to the
inevitable catastrophe. That is to say, the structure should be organic, with a
necessary and harmonious connection and relation of parts, and not merely
mechanical with an arbitrary or haphazard joining of one part to another.