SlideShare ist ein Scribd-Unternehmen logo
1 von 43
Downloaden Sie, um offline zu lesen
ROUGH SET & IT’SVARIANTS:VARIABLE
PRICISION ROUGH SET AND FUZZY ROUGH
SET APPROACHES
presented by-
Rajdeep Chatterjee
PLP, MIU ISI Kolkata
Overview
 Introduction to Rough Set
 Information/Decision Systems
 Indiscernibility
 Set Approximations of Rough Set
 Reducts and Core
 Dependency of Attributes
 Variable Precision rough Set (VPRS)
 Set Approximations (VPRS)
 Fuzzy Rough Set (FRS)
 Set Approximations and Dependency (FRS)
 Observations
R Chatterjee, PLP, MIU ISI Kolkata
Introduction
 Often, information on the surrounding
world is
◦ Imprecise
◦ Incomplete
◦ uncertain.
 We should be able to process uncertain
and/or incomplete information.
R Chatterjee, PLP, MIU ISI Kolkata
Introduction
 “Rough set theory” was developed by
Zdzislaw Pawlak in the early 1980’s.
 Representative Publications:
◦ Z. Pawlak, “Rough Sets”, International Journal of
Computer and Information Sciences,Vol.11,341-
356 (1982).
◦ Z. Pawlak, Rough Sets -Theoretical Aspect of
Reasoning about Data, Kluwer Academic
Pubilishers (1991).
R Chatterjee, PLP, MIU ISI Kolkata
Information Systems
Age LEMS
X1 16-30 50
X2 16-30 0
X3 31-45 1-25
X4 31-45 1-25
X5 46-60 26-49
X6 16-30 26-49
X7 46-60 26-49
 IS is a pair (U,A)
 U is a non-empty
finite set of objects.
 A is a non-empty
finite set of
attributes such that
for every
 is called the value
set of a.
aV
R Chatterjee, PLP, MIU ISI Kolkata
Decision Systems
Age LEMS Walk
X1 16-30 50 Yes
X2 16-30 0 No
X3 31-45 1-25 No
X4 31-45 1-25 Yes
X5 46-60 26-49 No
X6 16-30 26-49 Yes
X7 46-60 26-49 No
 DS:
 is the decision
attribute (instead of
one we can consider
more decision
attributes).
 The elements of A
are called the
condition attributes.
}){,( dAUT 
Ad 
Condition attributes
Decision attribute
R Chatterjee, PLP, MIU ISI Kolkata
Indiscernibility
 The equivalence relation
A binary relation which is
reflexive (xRx for any object x) ,
symmetric (if xRy then yRx), and
transitive (if xRy and yRz then xRz).
 The equivalence class of an element
consists of all objects
such that xRy.
XXR 
Rx][
Xx
Xy
R Chatterjee, PLP, MIU ISI Kolkata
Indiscernibility
 Let IS = (U,A) be an information system, then with
any there is an associated equivalence
relation:
where is called the B-indiscernibility relation.
 If then objects x and x’ are
indiscernible from each other by attributes from B.
 The equivalence classes of the B-indiscernibility
relation are denoted by
AB 
)}'()(,|)',{()( 2
xaxaBaUxxBINDIS 
)(BINDIS
),()',( BINDxx IS
Rx][
R Chatterjee, PLP, MIU ISI Kolkata
An Example of Indiscernibility
Age LEMS Walk
X1 16-30 50 Yes
X2 16-30 0 No
X3 31-45 1-25 No
X4 31-45 1-25 Yes
X5 46-60 26-49 No
X6 16-30 26-49 Yes
X7 46-60 26-49 No
 The non-empty
subsets of the
condition attributes
are {Age}, {LEMS}, and
{Age, LEMS}.
 IND({Age}) =
{{x1,x2,x6}, {x3,x4},
{x5,x7}}
 IND({LEMS}) = {{x1},
{x2}, {x3,x4}, {x5,x6,x7}}
 IND({Age,LEMS}) =
{{x1}, {x2}, {x3,x4},
{x5,x7}, {x6}}.
R Chatterjee, PLP, MIU ISI Kolkata
Set Approximation
 Let T = (U,A) and let and
We can approximate X using only the
information contained in B by
constructing the B-lower and B-upper
approximations of X, denoted and
respectively, where
AB  .UX 
XB
XB
},][|{ XxxXB B 
}.][|{  XxxXB B
R Chatterjee, PLP, MIU ISI Kolkata
Set Approximation
 B-boundary region of X,
consists of those objects that we cannot
decisively classify into X in B.
 B-outside region of X,
consists of those objects that can be with
certainty classified as not belonging to X.
 A set is said to be “rough” if its boundary
region is non-empty,otherwise the set is crisp.
,)( XBXBXBNB 
,XBU 
R Chatterjee, PLP, MIU ISI Kolkata
An Example of Set Approximation
 LetW = {x |Walk(x)
= yes}.
 The decision class,
Walk, is rough since
the boundary region
is not empty
}.7,5,2{
},4,3{)(
},6,4,3,1{
},6,1{
xxxWAU
xxWBN
xxxxWA
xxWA
A




Age LEMS Walk
X1 16-30 50 Yes
X2 16-30 0 No
X3 31-45 1-25 No
X4 31-45 1-25 Yes
X5 46-60 26-49 No
X6 16-30 26-49 Yes
X7 46-60 26-49 No
R Chatterjee, PLP, MIU ISI Kolkata
An Pictorial Depiction of Set Approximation
yes
yes/no
no
{{x1},{x6}}
{{x3,x4}}
{{x2}, {x5,x7}}
WA
AW
R Chatterjee, PLP, MIU ISI Kolkata
Lower & Upper Approximations
}:/{ XYRUYXR  
}:/{  XYRUYXR 
LowerApproximation:
Upper Approximation:
R Chatterjee, PLP, MIU ISI Kolkata
Properties of Approximations
1.
2.
3.
4.
5.
6. YX
YX
YBXBYXB
YBXBYXB
UUBUBBB
XBXXB






)()()(
)()()(
)()(,)()(
)(

)()( YBXB 
)()( YBXB 
implies
implies
R Chatterjee, PLP, MIU ISI Kolkata
Properties of Approximations
7.
8.
9.
10.
11.
12. )())(())((
)())(())((
)()(
)()(
)()()(
)()()(
XBXBBXBB
XBXBBXBB
XBXB
XBXB
YBXBYXB
YBXBYXB






Where, -X denotes U - X.
R Chatterjee, PLP, MIU ISI Kolkata
Four Basic Classes of Rough Sets
 X is roughly B-definable, iff and
 X is internally B-undefinable, iff
and
 X is externally B-undefinable, iff
and
 X is totally B-undefinable, iff
and
)(XB
,)( UXB 
)(XB
,)( UXB 
)(XB
,)( UXB 
)(XB
.)( UXB 
R Chatterjee, PLP, MIU ISI Kolkata
Accuracy of Approximation
where |X| denotes the cardinality of
Obviously
If X is crisp with respect to B.
If X is rough with respect to B.
|)(|
|)(|
)(
XB
XB
XB 
.X
.10  B
,1)( XB
,1)( XB
R Chatterjee, PLP, MIU ISI Kolkata
Issues in the Decision Table
 The same or indiscernible objects may be
represented several times.
 Some of the attributes may be
superfluous (redundant).
That is, their removal cannot worsen the
classification.
R Chatterjee, PLP, MIU ISI Kolkata
Reduct
 Keep only those attributes that preserve
the indiscernibility relation and,
consequently, set approximation.
 There are usually several such subsets of
attributes and those which are minimal
are called reducts.
R Chatterjee, PLP, MIU ISI Kolkata
Dispensable & Indispensable
Attributes
Let
Attribute c is dispensable in T
if , otherwise
attribute c is indispensable in T.
.Cc
)()( }){( DPOSDPOS cCC 
XCDPOS
DUX
C /
)(


The C-positive region of D:
R Chatterjee, PLP, MIU ISI Kolkata
Independent
 T = (U, C, D) is independent
if all are indispensable in TCc
R Chatterjee, PLP, MIU ISI Kolkata
Reduct & Core
 The set of attributes is called a
reduct of C, if T’ = (U, R, D) is independent
and
 The set of all the condition attributes
indispensable in T is denoted by CORE(C).
where RED(C) is the set of all reducts of C.
CR 
).()( DPOSDPOS CR 
)()( CREDCCORE 
R Chatterjee, PLP, MIU ISI Kolkata
Discernibility Matrix
 Let T = (U, C, D) be a decision table, with
By a discernibility matrix of T, denoted M(T),
we will mean matrix defined as:
for i, j = 1,2,…,n
 is the set of all the condition attributes
that classify objects ui and uj into different
classes
}.,...,,{ 21 nuuuU 
ijc
R Chatterjee, PLP, MIU ISI Kolkata
Discernibility Function
 A discernibility function for an information
system IS is a boolean function om m boolean
variables (corresponding to the attributes
a1,a2,…,am) defined as follows.
 Where .The set of all prime
implicants of determines the set of all reduct
of IS.
R Chatterjee, PLP, MIU ISI Kolkata
Examples of Discernibility Matrix
a b c d
u1 a0 b1 c1 Y
u2 a1 b1 c0 n
u3 a0 b2 c1 n
u4 a1 b1 c1 Y
 In order to discern equivalence
classes of the decision attribute d,
to preserve conditions described
by the discernibility matrix for
this table
u1 u2 u3
u2
u3
u4
a,c
b
c a,b

C = {a, b, c}
D = {d}
Reduct = {b, c}
cb
bacbca

 )()(
R Chatterjee, PLP, MIU ISI Kolkata
Dependency of Attributes
 Set of attribute D depends totally on a set
of attributes C, denoted if all
values of attributes from D are uniquely
determined by values of attributes from C.
,DC 
R Chatterjee, PLP, MIU ISI Kolkata
Dependency of Attributes
 Let D and C be subsets of A. We will say
that D depends on C in a degree k
denoted by if
where called C-
positive region of D.
),10(  k ,DC k
||
|)(|
),(
U
DPOS
DCk C
 
),()(
/
XCDPOS
DUX
C

 
R Chatterjee, PLP, MIU ISI Kolkata
Dependency of Attributes
 Obviously
 If k = 1 we say that D depends totally on
C.
 If k < 1 we say that D depends partially
(in a degree k) on C.
.
||
|)(|
),(
/


DUX U
XC
DCk 
R Chatterjee, PLP, MIU ISI Kolkata
Variable Precision Rough Set
 A generalized model of rough sets called
variable precision model (VPRS) aimed at
modeling classification problems involving
uncertain or imprecise information, is
presented by Wojceich Ziarko in 1993.
 This extended rough set model able to
allow some degree of misclassification in
the largely correct classification.
R Chatterjee, PLP, MIU ISI Kolkata
Variable Precision Rough Set
 c(X,Y) of the relative degree of misclassification of the
set X with respect to set Y defined as
where card denotes set cardinality.
 The quantity c(X,Y) will be referred to as the relative
classification error.
 The actual number of misclassified elements is given
by the product c(X,Y)*card(X) which is referred to as
an absolute classification error.
R Chatterjee, PLP, MIU ISI Kolkata
-majority (VPRS)
 is known as admissible classification
error must be within the range .
 More than *100 elements of X should be
common with Y. then it is called –majority
relations.
 Let X1={x1,x2,x3,x4}
X2={x1,x2,x5}
Y={x1,x2,x3,x8}
XY


XY 1
25.0
 XY 2
33.0

R Chatterjee, PLP, MIU ISI Kolkata
Set Approximations inVPRS
 Let A=(U,R) which consists of a non-
empty, finite universe U and of the
equivalence relation R on U. The
equivalence relation R, referred to as an
indiscernibility relation, corresponds to a
partitioning of the universe U into a
collection of equivalence classes or
elementary sets R*={E1,E2,…,En}
R Chatterjee, PLP, MIU ISI Kolkata
Lower & Upper Approximation (VPRS)
 Lower approximation:
or, equivalently,
 Upper approximation:
EX

R
R
R

R Chatterjee, PLP, MIU ISI Kolkata
Boundary & Negative Region (VPRS)
 Boundary region:
 Negative region:
R Chatterjee, PLP, MIU ISI Kolkata
Theoretical aspect of Approximation
(VPRS)
 The lower approximation of the set X can
be interpreted as the collection of all
those elements of U which can be
classified into X with the classification
error not greater than b.
 The negative region of the set X is the
collection of all those elements of U
which can be classified into the
complement of X, -X with the
classification error not greater than b.
R Chatterjee, PLP, MIU ISI Kolkata
Theoretical aspect of Approximation
(VPRS)
 The boundary region of the set X cannot
be classified either into X or –X with the
classification error not greater than b.
 The upper approximation of the set X
includes all those elements of U which
cannot be classified into -X with the error
not greater than b.
R Chatterjee, PLP, MIU ISI Kolkata
Fuzzy-Rough Sets (FRS)
 One particular use of RST is that of attribute reduction
in datasets. Given dataset with discretized attribute
values, it is possible to find a subset of the original
attributes that are the most informative (termed as
Reduct).
 However, most often the case that the values of
attributes may be real-valued and cannot be handled by
traditional rough set.
 Some discretization is possible which in turn gives you
loss of information.
 To deal with vagueness and noisy data in the dataset,
Fuzzy Rough Set was introduced by Richard Jensen.
R Chatterjee, PLP, MIU ISI Kolkata
Fuzzification for Conditional Features
a b c q
1 -0.4 -0.3 -0.5 No
2 -0.4 0.2 -0.1 Yes
3 -0.3 -0.4 -0.3 No
4 0.3 -0.3 0 Yes
5 0.2 -0.3 0 Yes
6 0.2 0 0 no
 Fuzzy-rough set is defined by two
fuzzy sets: fuzzy lower and upper
approximations, obtained by
extending the corresponding
crisp rough set notions.
 In crisp case, elements that
belong to the lower
approximation (i.e., have
membership of 1) are said to
belong to the approximated set
with absolute certainty. In fuzzy-
rough case, elements may have a
membership in the range [0,1],
allowing greater flexibility in
handling uncertainty.
R Chatterjee, PLP, MIU ISI Kolkata
Membership values from MFs of
linguistic labels
a b c q
Na Za Nb Zb Nc Zc {1,3,6} {2,4,5}
1 0.8 0.2 0.6 0.4 1.0 0.0 1.0 0.0
2 0.8 0.2 0.0 0.6 0.2 0.8 0.0 1.0
3 0.6 0.4 0.8 0.2 0.6 0.4 1.0 0.0
4 0.0 0.4 0.6 0.4 0.0 1.0 0.0 1.0
5 0.0 0.6 0.6 0.4 0.0 1.0 0.0 1.0
6 0.0 0.6 0.0 1.0 0.0 1.0 1.0 0.0
R Chatterjee, PLP, MIU ISI Kolkata
Lower & Upper Approximations (FRS)
 Lower approximation:
 Upper approximation:
sup
/ PUF
infUy
sup
Uy
sup
/ PUF
P
P
R Chatterjee, PLP, MIU ISI Kolkata
Positive Region & Dependency Measure (FRS)
 The membership of an object , belonging to the fuzzy
positive region can be defined by
 Using the definition of the fuzzy positive region, the new
dependency function can be defined as follows:
P
sup
/ QUX
R Chatterjee, PLP, MIU ISI Kolkata
Observations
 Evaluation of importance of particular attributes and
elimination of redundant attributes from the decision
table.
 Construction of a minimal subset of independent
attributes ensuring the same quality of classification as
the whole set, i.e. reducts of the set of attributes.
 Intersection of these reducts giving a core of attributes,
which cannot be eliminated without disturbing the
ability of approximating the classification and
 Generation of logical rules from the reduced decision
table.
R Chatterjee, PLP, MIU ISI Kolkata

Weitere ähnliche Inhalte

Was ist angesagt?

Image segmentation ppt
Image segmentation pptImage segmentation ppt
Image segmentation ppt
Gichelle Amon
 

Was ist angesagt? (20)

random forest regression
random forest regressionrandom forest regression
random forest regression
 
Data discretization
Data discretizationData discretization
Data discretization
 
Morphological image processing
Morphological image processingMorphological image processing
Morphological image processing
 
Image segmentation
Image segmentationImage segmentation
Image segmentation
 
Decision Trees
Decision TreesDecision Trees
Decision Trees
 
Support vector machine
Support vector machineSupport vector machine
Support vector machine
 
Birch Algorithm With Solved Example
Birch Algorithm With Solved ExampleBirch Algorithm With Solved Example
Birch Algorithm With Solved Example
 
Feature Reduction Techniques
Feature Reduction TechniquesFeature Reduction Techniques
Feature Reduction Techniques
 
K Nearest Neighbors
K Nearest NeighborsK Nearest Neighbors
K Nearest Neighbors
 
K-Nearest Neighbor Classifier
K-Nearest Neighbor ClassifierK-Nearest Neighbor Classifier
K-Nearest Neighbor Classifier
 
Inductive bias
Inductive biasInductive bias
Inductive bias
 
Image segmentation ppt
Image segmentation pptImage segmentation ppt
Image segmentation ppt
 
My8clst
My8clstMy8clst
My8clst
 
Outlier Detection
Outlier DetectionOutlier Detection
Outlier Detection
 
08 clustering
08 clustering08 clustering
08 clustering
 
support vector regression
support vector regressionsupport vector regression
support vector regression
 
Decision Tree Learning
Decision Tree LearningDecision Tree Learning
Decision Tree Learning
 
Classification Using Decision tree
Classification Using Decision treeClassification Using Decision tree
Classification Using Decision tree
 
Control Strategies in AI
Control Strategies in AIControl Strategies in AI
Control Strategies in AI
 
You Only Look Once: Unified, Real-Time Object Detection
You Only Look Once: Unified, Real-Time Object DetectionYou Only Look Once: Unified, Real-Time Object Detection
You Only Look Once: Unified, Real-Time Object Detection
 

Andere mochten auch (6)

Text categorization using Rough Set
Text categorization using Rough SetText categorization using Rough Set
Text categorization using Rough Set
 
Fuzzy logic application (aircraft landing)
Fuzzy logic application (aircraft landing)Fuzzy logic application (aircraft landing)
Fuzzy logic application (aircraft landing)
 
Rough set based decision tree for identifying vulnerable and food insecure ho...
Rough set based decision tree for identifying vulnerable and food insecure ho...Rough set based decision tree for identifying vulnerable and food insecure ho...
Rough set based decision tree for identifying vulnerable and food insecure ho...
 
Thesis ppt
Thesis pptThesis ppt
Thesis ppt
 
Decision trees
Decision treesDecision trees
Decision trees
 
Application of fuzzy logic
Application of fuzzy logicApplication of fuzzy logic
Application of fuzzy logic
 

Ähnlich wie Roughset & it’s variants

Descriptive Statistics with R
Descriptive Statistics with RDescriptive Statistics with R
Descriptive Statistics with R
Kazuki Yoshida
 
Co clustering by-block_value_decomposition
Co clustering by-block_value_decompositionCo clustering by-block_value_decomposition
Co clustering by-block_value_decomposition
AllenWu
 
[ABDO] Logic As A Database Language
[ABDO] Logic As A Database Language[ABDO] Logic As A Database Language
[ABDO] Logic As A Database Language
Carles Farré
 

Ähnlich wie Roughset & it’s variants (20)

Improving circuit miniaturization and its efficiency using Rough Set Theory( ...
Improving circuit miniaturization and its efficiency using Rough Set Theory( ...Improving circuit miniaturization and its efficiency using Rough Set Theory( ...
Improving circuit miniaturization and its efficiency using Rough Set Theory( ...
 
Chapter 5
Chapter 5Chapter 5
Chapter 5
 
Descriptive Statistics with R
Descriptive Statistics with RDescriptive Statistics with R
Descriptive Statistics with R
 
Simultaneous State and Actuator Fault Estimation With Fuzzy Descriptor PMID a...
Simultaneous State and Actuator Fault Estimation With Fuzzy Descriptor PMID a...Simultaneous State and Actuator Fault Estimation With Fuzzy Descriptor PMID a...
Simultaneous State and Actuator Fault Estimation With Fuzzy Descriptor PMID a...
 
E5-roughsets unit-V.pdf
E5-roughsets unit-V.pdfE5-roughsets unit-V.pdf
E5-roughsets unit-V.pdf
 
Bayesian_Decision_Theory-3.pdf
Bayesian_Decision_Theory-3.pdfBayesian_Decision_Theory-3.pdf
Bayesian_Decision_Theory-3.pdf
 
Co clustering by-block_value_decomposition
Co clustering by-block_value_decompositionCo clustering by-block_value_decomposition
Co clustering by-block_value_decomposition
 
C023014030
C023014030C023014030
C023014030
 
C023014030
C023014030C023014030
C023014030
 
Spike sorting: What is it? Why do we need it? Where does it come from? How is...
Spike sorting: What is it? Why do we need it? Where does it come from? How is...Spike sorting: What is it? Why do we need it? Where does it come from? How is...
Spike sorting: What is it? Why do we need it? Where does it come from? How is...
 
On observer design methods for a
On observer design methods for aOn observer design methods for a
On observer design methods for a
 
K-algebras on quadripartitioned single valued neutrosophic sets
K-algebras on quadripartitioned single valued neutrosophic setsK-algebras on quadripartitioned single valued neutrosophic sets
K-algebras on quadripartitioned single valued neutrosophic sets
 
Comparison of Different Methods for Fusion of Multimodal Medical Images
Comparison of Different Methods for Fusion of Multimodal Medical ImagesComparison of Different Methods for Fusion of Multimodal Medical Images
Comparison of Different Methods for Fusion of Multimodal Medical Images
 
Calibrating the Lee-Carter and the Poisson Lee-Carter models via Neural Netw...
Calibrating the Lee-Carter and the Poisson Lee-Carter models  via Neural Netw...Calibrating the Lee-Carter and the Poisson Lee-Carter models  via Neural Netw...
Calibrating the Lee-Carter and the Poisson Lee-Carter models via Neural Netw...
 
Reduct generation for the incremental data using rough set theory
Reduct generation for the incremental data using rough set theoryReduct generation for the incremental data using rough set theory
Reduct generation for the incremental data using rough set theory
 
MVPA with SpaceNet: sparse structured priors
MVPA with SpaceNet: sparse structured priorsMVPA with SpaceNet: sparse structured priors
MVPA with SpaceNet: sparse structured priors
 
Lesson 26
Lesson 26Lesson 26
Lesson 26
 
AI Lesson 26
AI Lesson 26AI Lesson 26
AI Lesson 26
 
[ABDO] Logic As A Database Language
[ABDO] Logic As A Database Language[ABDO] Logic As A Database Language
[ABDO] Logic As A Database Language
 
Observer design for descriptor linear systems
Observer design for descriptor linear systemsObserver design for descriptor linear systems
Observer design for descriptor linear systems
 

Mehr von Dr. Rajdeep Chatterjee

Mehr von Dr. Rajdeep Chatterjee (8)

Artificial Neural Networks (ANN)
Artificial Neural Networks (ANN)Artificial Neural Networks (ANN)
Artificial Neural Networks (ANN)
 
Genetic Algorithm (GA)
Genetic Algorithm (GA)Genetic Algorithm (GA)
Genetic Algorithm (GA)
 
Derivative free optimizations
Derivative free optimizationsDerivative free optimizations
Derivative free optimizations
 
EEG based Motor Imagery Classification using SVM and MLP
EEG based Motor Imagery Classification using SVM and MLPEEG based Motor Imagery Classification using SVM and MLP
EEG based Motor Imagery Classification using SVM and MLP
 
Data Structure: Algorithm and analysis
Data Structure: Algorithm and analysisData Structure: Algorithm and analysis
Data Structure: Algorithm and analysis
 
IPR: Introduction to Trademark
IPR: Introduction to TrademarkIPR: Introduction to Trademark
IPR: Introduction to Trademark
 
Lecture 1 introduction to computer and its organization
Lecture 1 introduction to computer and its organizationLecture 1 introduction to computer and its organization
Lecture 1 introduction to computer and its organization
 
Newtonian Law Inspired Optimization Techniques Based on Gravitational Search ...
Newtonian Law Inspired Optimization Techniques Based on Gravitational Search ...Newtonian Law Inspired Optimization Techniques Based on Gravitational Search ...
Newtonian Law Inspired Optimization Techniques Based on Gravitational Search ...
 

Kürzlich hochgeladen

Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Christo Ananth
 
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
dharasingh5698
 
AKTU Computer Networks notes --- Unit 3.pdf
AKTU Computer Networks notes ---  Unit 3.pdfAKTU Computer Networks notes ---  Unit 3.pdf
AKTU Computer Networks notes --- Unit 3.pdf
ankushspencer015
 

Kürzlich hochgeladen (20)

MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINEMANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
 
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 
Porous Ceramics seminar and technical writing
Porous Ceramics seminar and technical writingPorous Ceramics seminar and technical writing
Porous Ceramics seminar and technical writing
 
KubeKraft presentation @CloudNativeHooghly
KubeKraft presentation @CloudNativeHooghlyKubeKraft presentation @CloudNativeHooghly
KubeKraft presentation @CloudNativeHooghly
 
Online banking management system project.pdf
Online banking management system project.pdfOnline banking management system project.pdf
Online banking management system project.pdf
 
Russian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur Escorts
Russian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur EscortsRussian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur Escorts
Russian Call Girls in Nagpur Grishma Call 7001035870 Meet With Nagpur Escorts
 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
 
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptxBSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
 
The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...
The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...
The Most Attractive Pune Call Girls Manchar 8250192130 Will You Miss This Cha...
 
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
 
Water Industry Process Automation & Control Monthly - April 2024
Water Industry Process Automation & Control Monthly - April 2024Water Industry Process Automation & Control Monthly - April 2024
Water Industry Process Automation & Control Monthly - April 2024
 
Introduction and different types of Ethernet.pptx
Introduction and different types of Ethernet.pptxIntroduction and different types of Ethernet.pptx
Introduction and different types of Ethernet.pptx
 
Extrusion Processes and Their Limitations
Extrusion Processes and Their LimitationsExtrusion Processes and Their Limitations
Extrusion Processes and Their Limitations
 
University management System project report..pdf
University management System project report..pdfUniversity management System project report..pdf
University management System project report..pdf
 
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
 
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
 
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
 
AKTU Computer Networks notes --- Unit 3.pdf
AKTU Computer Networks notes ---  Unit 3.pdfAKTU Computer Networks notes ---  Unit 3.pdf
AKTU Computer Networks notes --- Unit 3.pdf
 
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete RecordCCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
 

Roughset & it’s variants

  • 1. ROUGH SET & IT’SVARIANTS:VARIABLE PRICISION ROUGH SET AND FUZZY ROUGH SET APPROACHES presented by- Rajdeep Chatterjee PLP, MIU ISI Kolkata
  • 2. Overview  Introduction to Rough Set  Information/Decision Systems  Indiscernibility  Set Approximations of Rough Set  Reducts and Core  Dependency of Attributes  Variable Precision rough Set (VPRS)  Set Approximations (VPRS)  Fuzzy Rough Set (FRS)  Set Approximations and Dependency (FRS)  Observations R Chatterjee, PLP, MIU ISI Kolkata
  • 3. Introduction  Often, information on the surrounding world is ◦ Imprecise ◦ Incomplete ◦ uncertain.  We should be able to process uncertain and/or incomplete information. R Chatterjee, PLP, MIU ISI Kolkata
  • 4. Introduction  “Rough set theory” was developed by Zdzislaw Pawlak in the early 1980’s.  Representative Publications: ◦ Z. Pawlak, “Rough Sets”, International Journal of Computer and Information Sciences,Vol.11,341- 356 (1982). ◦ Z. Pawlak, Rough Sets -Theoretical Aspect of Reasoning about Data, Kluwer Academic Pubilishers (1991). R Chatterjee, PLP, MIU ISI Kolkata
  • 5. Information Systems Age LEMS X1 16-30 50 X2 16-30 0 X3 31-45 1-25 X4 31-45 1-25 X5 46-60 26-49 X6 16-30 26-49 X7 46-60 26-49  IS is a pair (U,A)  U is a non-empty finite set of objects.  A is a non-empty finite set of attributes such that for every  is called the value set of a. aV R Chatterjee, PLP, MIU ISI Kolkata
  • 6. Decision Systems Age LEMS Walk X1 16-30 50 Yes X2 16-30 0 No X3 31-45 1-25 No X4 31-45 1-25 Yes X5 46-60 26-49 No X6 16-30 26-49 Yes X7 46-60 26-49 No  DS:  is the decision attribute (instead of one we can consider more decision attributes).  The elements of A are called the condition attributes. }){,( dAUT  Ad  Condition attributes Decision attribute R Chatterjee, PLP, MIU ISI Kolkata
  • 7. Indiscernibility  The equivalence relation A binary relation which is reflexive (xRx for any object x) , symmetric (if xRy then yRx), and transitive (if xRy and yRz then xRz).  The equivalence class of an element consists of all objects such that xRy. XXR  Rx][ Xx Xy R Chatterjee, PLP, MIU ISI Kolkata
  • 8. Indiscernibility  Let IS = (U,A) be an information system, then with any there is an associated equivalence relation: where is called the B-indiscernibility relation.  If then objects x and x’ are indiscernible from each other by attributes from B.  The equivalence classes of the B-indiscernibility relation are denoted by AB  )}'()(,|)',{()( 2 xaxaBaUxxBINDIS  )(BINDIS ),()',( BINDxx IS Rx][ R Chatterjee, PLP, MIU ISI Kolkata
  • 9. An Example of Indiscernibility Age LEMS Walk X1 16-30 50 Yes X2 16-30 0 No X3 31-45 1-25 No X4 31-45 1-25 Yes X5 46-60 26-49 No X6 16-30 26-49 Yes X7 46-60 26-49 No  The non-empty subsets of the condition attributes are {Age}, {LEMS}, and {Age, LEMS}.  IND({Age}) = {{x1,x2,x6}, {x3,x4}, {x5,x7}}  IND({LEMS}) = {{x1}, {x2}, {x3,x4}, {x5,x6,x7}}  IND({Age,LEMS}) = {{x1}, {x2}, {x3,x4}, {x5,x7}, {x6}}. R Chatterjee, PLP, MIU ISI Kolkata
  • 10. Set Approximation  Let T = (U,A) and let and We can approximate X using only the information contained in B by constructing the B-lower and B-upper approximations of X, denoted and respectively, where AB  .UX  XB XB },][|{ XxxXB B  }.][|{  XxxXB B R Chatterjee, PLP, MIU ISI Kolkata
  • 11. Set Approximation  B-boundary region of X, consists of those objects that we cannot decisively classify into X in B.  B-outside region of X, consists of those objects that can be with certainty classified as not belonging to X.  A set is said to be “rough” if its boundary region is non-empty,otherwise the set is crisp. ,)( XBXBXBNB  ,XBU  R Chatterjee, PLP, MIU ISI Kolkata
  • 12. An Example of Set Approximation  LetW = {x |Walk(x) = yes}.  The decision class, Walk, is rough since the boundary region is not empty }.7,5,2{ },4,3{)( },6,4,3,1{ },6,1{ xxxWAU xxWBN xxxxWA xxWA A     Age LEMS Walk X1 16-30 50 Yes X2 16-30 0 No X3 31-45 1-25 No X4 31-45 1-25 Yes X5 46-60 26-49 No X6 16-30 26-49 Yes X7 46-60 26-49 No R Chatterjee, PLP, MIU ISI Kolkata
  • 13. An Pictorial Depiction of Set Approximation yes yes/no no {{x1},{x6}} {{x3,x4}} {{x2}, {x5,x7}} WA AW R Chatterjee, PLP, MIU ISI Kolkata
  • 14. Lower & Upper Approximations }:/{ XYRUYXR   }:/{  XYRUYXR  LowerApproximation: Upper Approximation: R Chatterjee, PLP, MIU ISI Kolkata
  • 15. Properties of Approximations 1. 2. 3. 4. 5. 6. YX YX YBXBYXB YBXBYXB UUBUBBB XBXXB       )()()( )()()( )()(,)()( )(  )()( YBXB  )()( YBXB  implies implies R Chatterjee, PLP, MIU ISI Kolkata
  • 16. Properties of Approximations 7. 8. 9. 10. 11. 12. )())(())(( )())(())(( )()( )()( )()()( )()()( XBXBBXBB XBXBBXBB XBXB XBXB YBXBYXB YBXBYXB       Where, -X denotes U - X. R Chatterjee, PLP, MIU ISI Kolkata
  • 17. Four Basic Classes of Rough Sets  X is roughly B-definable, iff and  X is internally B-undefinable, iff and  X is externally B-undefinable, iff and  X is totally B-undefinable, iff and )(XB ,)( UXB  )(XB ,)( UXB  )(XB ,)( UXB  )(XB .)( UXB  R Chatterjee, PLP, MIU ISI Kolkata
  • 18. Accuracy of Approximation where |X| denotes the cardinality of Obviously If X is crisp with respect to B. If X is rough with respect to B. |)(| |)(| )( XB XB XB  .X .10  B ,1)( XB ,1)( XB R Chatterjee, PLP, MIU ISI Kolkata
  • 19. Issues in the Decision Table  The same or indiscernible objects may be represented several times.  Some of the attributes may be superfluous (redundant). That is, their removal cannot worsen the classification. R Chatterjee, PLP, MIU ISI Kolkata
  • 20. Reduct  Keep only those attributes that preserve the indiscernibility relation and, consequently, set approximation.  There are usually several such subsets of attributes and those which are minimal are called reducts. R Chatterjee, PLP, MIU ISI Kolkata
  • 21. Dispensable & Indispensable Attributes Let Attribute c is dispensable in T if , otherwise attribute c is indispensable in T. .Cc )()( }){( DPOSDPOS cCC  XCDPOS DUX C / )(   The C-positive region of D: R Chatterjee, PLP, MIU ISI Kolkata
  • 22. Independent  T = (U, C, D) is independent if all are indispensable in TCc R Chatterjee, PLP, MIU ISI Kolkata
  • 23. Reduct & Core  The set of attributes is called a reduct of C, if T’ = (U, R, D) is independent and  The set of all the condition attributes indispensable in T is denoted by CORE(C). where RED(C) is the set of all reducts of C. CR  ).()( DPOSDPOS CR  )()( CREDCCORE  R Chatterjee, PLP, MIU ISI Kolkata
  • 24. Discernibility Matrix  Let T = (U, C, D) be a decision table, with By a discernibility matrix of T, denoted M(T), we will mean matrix defined as: for i, j = 1,2,…,n  is the set of all the condition attributes that classify objects ui and uj into different classes }.,...,,{ 21 nuuuU  ijc R Chatterjee, PLP, MIU ISI Kolkata
  • 25. Discernibility Function  A discernibility function for an information system IS is a boolean function om m boolean variables (corresponding to the attributes a1,a2,…,am) defined as follows.  Where .The set of all prime implicants of determines the set of all reduct of IS. R Chatterjee, PLP, MIU ISI Kolkata
  • 26. Examples of Discernibility Matrix a b c d u1 a0 b1 c1 Y u2 a1 b1 c0 n u3 a0 b2 c1 n u4 a1 b1 c1 Y  In order to discern equivalence classes of the decision attribute d, to preserve conditions described by the discernibility matrix for this table u1 u2 u3 u2 u3 u4 a,c b c a,b  C = {a, b, c} D = {d} Reduct = {b, c} cb bacbca   )()( R Chatterjee, PLP, MIU ISI Kolkata
  • 27. Dependency of Attributes  Set of attribute D depends totally on a set of attributes C, denoted if all values of attributes from D are uniquely determined by values of attributes from C. ,DC  R Chatterjee, PLP, MIU ISI Kolkata
  • 28. Dependency of Attributes  Let D and C be subsets of A. We will say that D depends on C in a degree k denoted by if where called C- positive region of D. ),10(  k ,DC k || |)(| ),( U DPOS DCk C   ),()( / XCDPOS DUX C    R Chatterjee, PLP, MIU ISI Kolkata
  • 29. Dependency of Attributes  Obviously  If k = 1 we say that D depends totally on C.  If k < 1 we say that D depends partially (in a degree k) on C. . || |)(| ),( /   DUX U XC DCk  R Chatterjee, PLP, MIU ISI Kolkata
  • 30. Variable Precision Rough Set  A generalized model of rough sets called variable precision model (VPRS) aimed at modeling classification problems involving uncertain or imprecise information, is presented by Wojceich Ziarko in 1993.  This extended rough set model able to allow some degree of misclassification in the largely correct classification. R Chatterjee, PLP, MIU ISI Kolkata
  • 31. Variable Precision Rough Set  c(X,Y) of the relative degree of misclassification of the set X with respect to set Y defined as where card denotes set cardinality.  The quantity c(X,Y) will be referred to as the relative classification error.  The actual number of misclassified elements is given by the product c(X,Y)*card(X) which is referred to as an absolute classification error. R Chatterjee, PLP, MIU ISI Kolkata
  • 32. -majority (VPRS)  is known as admissible classification error must be within the range .  More than *100 elements of X should be common with Y. then it is called –majority relations.  Let X1={x1,x2,x3,x4} X2={x1,x2,x5} Y={x1,x2,x3,x8} XY   XY 1 25.0  XY 2 33.0  R Chatterjee, PLP, MIU ISI Kolkata
  • 33. Set Approximations inVPRS  Let A=(U,R) which consists of a non- empty, finite universe U and of the equivalence relation R on U. The equivalence relation R, referred to as an indiscernibility relation, corresponds to a partitioning of the universe U into a collection of equivalence classes or elementary sets R*={E1,E2,…,En} R Chatterjee, PLP, MIU ISI Kolkata
  • 34. Lower & Upper Approximation (VPRS)  Lower approximation: or, equivalently,  Upper approximation: EX  R R R  R Chatterjee, PLP, MIU ISI Kolkata
  • 35. Boundary & Negative Region (VPRS)  Boundary region:  Negative region: R Chatterjee, PLP, MIU ISI Kolkata
  • 36. Theoretical aspect of Approximation (VPRS)  The lower approximation of the set X can be interpreted as the collection of all those elements of U which can be classified into X with the classification error not greater than b.  The negative region of the set X is the collection of all those elements of U which can be classified into the complement of X, -X with the classification error not greater than b. R Chatterjee, PLP, MIU ISI Kolkata
  • 37. Theoretical aspect of Approximation (VPRS)  The boundary region of the set X cannot be classified either into X or –X with the classification error not greater than b.  The upper approximation of the set X includes all those elements of U which cannot be classified into -X with the error not greater than b. R Chatterjee, PLP, MIU ISI Kolkata
  • 38. Fuzzy-Rough Sets (FRS)  One particular use of RST is that of attribute reduction in datasets. Given dataset with discretized attribute values, it is possible to find a subset of the original attributes that are the most informative (termed as Reduct).  However, most often the case that the values of attributes may be real-valued and cannot be handled by traditional rough set.  Some discretization is possible which in turn gives you loss of information.  To deal with vagueness and noisy data in the dataset, Fuzzy Rough Set was introduced by Richard Jensen. R Chatterjee, PLP, MIU ISI Kolkata
  • 39. Fuzzification for Conditional Features a b c q 1 -0.4 -0.3 -0.5 No 2 -0.4 0.2 -0.1 Yes 3 -0.3 -0.4 -0.3 No 4 0.3 -0.3 0 Yes 5 0.2 -0.3 0 Yes 6 0.2 0 0 no  Fuzzy-rough set is defined by two fuzzy sets: fuzzy lower and upper approximations, obtained by extending the corresponding crisp rough set notions.  In crisp case, elements that belong to the lower approximation (i.e., have membership of 1) are said to belong to the approximated set with absolute certainty. In fuzzy- rough case, elements may have a membership in the range [0,1], allowing greater flexibility in handling uncertainty. R Chatterjee, PLP, MIU ISI Kolkata
  • 40. Membership values from MFs of linguistic labels a b c q Na Za Nb Zb Nc Zc {1,3,6} {2,4,5} 1 0.8 0.2 0.6 0.4 1.0 0.0 1.0 0.0 2 0.8 0.2 0.0 0.6 0.2 0.8 0.0 1.0 3 0.6 0.4 0.8 0.2 0.6 0.4 1.0 0.0 4 0.0 0.4 0.6 0.4 0.0 1.0 0.0 1.0 5 0.0 0.6 0.6 0.4 0.0 1.0 0.0 1.0 6 0.0 0.6 0.0 1.0 0.0 1.0 1.0 0.0 R Chatterjee, PLP, MIU ISI Kolkata
  • 41. Lower & Upper Approximations (FRS)  Lower approximation:  Upper approximation: sup / PUF infUy sup Uy sup / PUF P P R Chatterjee, PLP, MIU ISI Kolkata
  • 42. Positive Region & Dependency Measure (FRS)  The membership of an object , belonging to the fuzzy positive region can be defined by  Using the definition of the fuzzy positive region, the new dependency function can be defined as follows: P sup / QUX R Chatterjee, PLP, MIU ISI Kolkata
  • 43. Observations  Evaluation of importance of particular attributes and elimination of redundant attributes from the decision table.  Construction of a minimal subset of independent attributes ensuring the same quality of classification as the whole set, i.e. reducts of the set of attributes.  Intersection of these reducts giving a core of attributes, which cannot be eliminated without disturbing the ability of approximating the classification and  Generation of logical rules from the reduced decision table. R Chatterjee, PLP, MIU ISI Kolkata