cellular automata evolution for pattern recognitionpmaji/index_files/pmajippt.pdf · cellular...
TRANSCRIPT
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Cellular Automata Evolutionfor Pattern Recognition
Pradipta Maji Center for Soft Computing Research
Indian Statistical Institute, Kolkata, 700 108, INDIA
Under the supervision of
Prof. P Pal Chaudhuri Prof. Debesh K DasProfessor Emeritus Professor
Dept. of Comp. Sci. & Tech. Dept. of Comp. Sci. & Engg.Bengal Engineering College (DU), Shibpur, INDIA Jadavpur University, INDIA
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Introduction
• Cellular Automata (CA) – promising research areaArtificial Intelligence (AI) Artificial Life (ALife)
• Considerable research in –modeling toolimage processinglanguage recognitionpattern recognitionVLSI testing
• Cellular Automata (CA) –learns association from a set of examples apply this knowledge-base to handle unseen cases such associations effective for classifying patterns
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Contribution of the Thesis
• Analysis and synthesis of linear boolean CA (MACA) – CA with only XOR logicapplication of MACA in pattern recognition• data mining• image compression• fault diagnosis of electronic circuit
• Analysis and synthesis of non-linear boolean CA (GMACA) – CA with all possible logicapplication of non-linear CA in pattern recognition
• Analysis and synthesis offuzzy CA (FMACA) – CA with fuzzy logicapplication of fuzzy CA in pattern recognition
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Cellular Automata (CA)
• A special type of computing model –50’s - J Von Neumann80’s - S. Wolfram
• A CA displays three basic characteristicsSimplicity: Basic unit of CA – cell – is simpleVast parallelism: CA achieves parallelism on a scale larger than massively parallel computersLocality: CA – characterized by local connectivity of its cell – all interactions take place on a purely local basis – a cell can only communicate with its neighboring cells – interconnection links usually carry only a small amount of information – no cell has a global view of the entire system
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Cellular Automata (CA)• A computational model with discrete cells updated synchronously• Uniform CA, hybrid / non-uniform CA, null boundary CA, periodic boundary CA
outputInput
Combinational Logic
Clock
From left neighbor
From right neighbor
0/1
2 - state 3-neighborhood CA cell
………..
Each cell can have 256 different rules
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
CA – State Transition0 0 1 1
0 1 1 1
98 230 226 107
0 0 1 0
98 230 226 107
3
7
For 2nd Cell
Rule 230
PS NS111 1110 1101 1100 0011 0010 1001 1000 0 2
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Different Types of CA
• Linear CA– Based on XOR logic– Total 7 rules (60, 90, 102, 150, 170, 204, 240) – Can be expressed through matrix (T), characteristic polynomial– Next state of the CA cell P(t+1) = T. P(t)
• Additive CA– Based on XOR and XNOR logic– Total 14 rules (linear rules + 195,165,153,105,85,51,15)– Can be expressed through matrix, inversion vector, characteristic polynomial– The next state of the CA cell P(t+1) = T. P(t) + F
60 102 150 2041 0 0 00 1 1 00 1 1 10 0 0 1
T =
60 153 105 204 0 1 1 0F =
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Additive Cellular AutomataXOR Logic XNOR Logic
Rule 60 : qI(t+1) = qI-1(t) ⊕ qI(t) Rule 195 : qI(t+1) = qI-1(t) ⊕ qI(t)
Rule 90 : qI(t+1) = qI-1(t) ⊕ qI+1(t) Rule 165 : qI(t+1) = qI-1(t) ⊕ qI+1(t)
Rule 102 : qI(t+1) = qI(t) ⊕ qI+1(t) Rule 153 : qI(t+1) = qI(t) ⊕ qI+1(t)
Rule 150 : qI(t+1) = qI-1(t) ⊕ qI(t) ⊕ qI-1(t) Rule 105 : qI(t+1) = qI-1(t) ⊕ qI(t) ⊕ qI-1(t)
Rule 170 : qI(t+1) = qI-1(t) Rule 85 : qI(t+1) = qI-1(t)
Rule 204 : qI(t+1) = qI (t) Rule 51 : qI(t+1) = qI (t)
Rule 240 : qI(t+1) = qI+1(t) Rule 15 : qI(t+1) = qI+1(t)
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
CA - State Transition Diagram
9 15
613
7 12
3 1411
52 8
1 410
0
Group CA 5
15
10
0
4
14
11
1
2
7
13
8
3
6
12
9
Non-group CA
Associative Memory
Non-group Cellular Automata
Linear Non-linear Fuzzy
MACA GMACA FMACA Perform pattern recognition task
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Pattern Recognition
• Pattern Recognition/Classificationmost important foundation stone of knowledge extraction methodologydemands automatic identification of patterns of interest (objects, images) from its background (shapes, forms, outlines, etc)conventional approach – machine compares given input pattern with each of stored patterns – identifies the closest matchtime to recognize the closest match – O(k) – recognition slow
• Associative Memory Entire state space - divided into some pivotal points
Transient
Transient
Transient
States close to pivot -associated with that pivot
Time to recognize a pattern -Independent of number of stored patterns1. MACA (linear) 2. GMACA (non-linear) 3. FMACA (fuzzy)
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Multiple Attractor CA (MACA)Employs linear CA rules State with self loop – attractorTransient states and attractor – form attractor basinBehaves as an associative memory Forms natural clusters
10001 01001
1100010000 01000
10010 10011
1101101010 01011
00001 00000 11001 00010 00011 11010
10100 10101
1110001100 01101
10110 10111
1111101110 01111
00101 00100 11101 00110 00111 11110
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Multiple Attractor CA (MACA)
• Next state: P(t+1) = T. P(t)
• Characteristic Polynomial: X (n-m) (1+X)m where m=log2(k)• n denotes number of CA cell• k denotes number of attractor basins
• Depth d of MACA –• number of edges between a non-reachable state and an attractor state
• Attractor of a basin: P(t+d) = Td P(t)
• m-bit positions pseudo-exhaustive: extract PEF (pseudo-exhaustive field) from attractor state
• Problem:
Complexity of identification of attractor basin is O(n3)Exponential search spaceRedundant solutions
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Chromosome of Genetic Algorithm
1 1 0 0 01 1 0 0 00 0 1 0 00 0 0 1 00 0 0 1 0
T =Matrix
1 1 1 1 0 … … .. 1 0
102 60 204 204 240
Characteristic polynomial x3(1+x)2 Elementary divisorsx3(1+x)2 ……… x2 (1+x) (1+x) x
Rule vector
x2 (1+x) ( x1+x)
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Dependency Vector/Dependency String
1 1 0 0 01 1 0 0 00 0 1 0 00 0 0 1 00 0 0 1 0
T =
Characteristic polynomial x3(1+x)2
Matrix T is obtained from T1 and T2 by Block Diagonal Method
1 1 01 1 00 0 1
T1 =
1 01 0
T2 =
Characteristic polynomial x2(1+x)
Characteristic polynomial x(1+x)
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Dependency Vector/Dependency String
0 0 0 0 00 0 0 0 00 0 1 0 00 0 0 1 00 0 0 1 0
Td =
Dependency Vector DV1 = < 0 0 1 >
0 0 00 0 00 0 1
T1d =
1 01 0
T2d =
Dependency Vector DV1 = < 1 0 >
Dependency String DS = < 0 0 1 > < 1 0 >
Dependency String DS = < 0 0 1 2 0 >
0 0 1 2 0
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Dependency Vector/Dependency String
0 00 11 01 1
Zero basin of T1 + Zero basin of T2 -----Zero basin of T1 + Non-zero basin of T2 ---
Non-zero basin of T1 + Zero basin of T2 ----Non-zero basin of T1 + Non-zero basin of T2 ---
PEF Bits
DV1 contributes 1st PEF Bits DV2 contributes 2nd PEF BitsPEF = [PEF1] [PEF2] = [DS.P] = [DV1.P1] [DV2.P2]
P = [ 1 1 1 1 1 ] DS = [ 0 0 1 2 0 ]
PEF = [PEF1][PEF2] = [<0 0 1><1 1 1>][<1 0><1 1>] = 1 1
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Matrix/Rule from Dependency String
• ‘0’ in DV T1 = [ 0 ]
• ’11’ in DV T3 = [T1] 1
0 0
[T1] 1 0
0 1 1
0 1 1
• ‘1’ in DV T2 = [ 1 ]
• ’101’ in DV T4 =
DV = < 1 1 1 >
T2 =1 1 0
0 0 1
0 0 0
DS = < 1 0 1 1 2 2 2>
T =
1 1 0 0 0 0 0
0 1 1 0 0 0 0
0 1 1 1 0 0 0
0 0 0 0 0 0 0
0 0 0 0 1 1 0
0 0 0 0 0 0 1
0 0 0 0 0 0 0
CA Rule Vector
<102, 102, 150, 0, 102, 170, 0>
DV = < 1 0 1 1 >
T1 =
1 1 0 0
0 1 1 0
0 1 1 1
0 0 0 0
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Dependency Vector/Dependency String
Characteristic polynomial x4(1+x)
0 1 0 0 00 0 0 0 00 1 1 1 00 0 0 0 00 0 0 1 0
T1 =
0 1 0 0 00 0 1 0 00 1 1 1 00 0 1 0 00 0 0 1 0
T2 =
0 1 0 0 00 0 1 0 00 1 1 1 00 0 1 0 00 0 0 0 0
T3 =
<170, 0, 150, 0, 240> <170, 0, 150, 0, 240> <170, 0, 150, 0, 240>
Dependency Vector <0 1 1 1 0>Identification of attractor basins in O(n)
Reduction of search space
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Image CompressionBlock diagram of codebook generation scheme
Training Images
8 X 8 Set
4 X 4 Set
16 X 16 Set
TSVQ
8 X 8 Codebook
4 X 4 Codebook
16 X 16 Codebook
Spatial Domain
High Compression ratio
Acceptable image quality
Applications - Human Portraits
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Tree-Structured Vector Quantization
N X N Set
Cluster 1 Cluster 2
Centroid 1 Centroid 2
S1, S2, S3, S4
S1, S2 S3, S4
S1 S2 S3 S4
Clusters and centroids generation using Tree-Structured Vector Quantization (TSVQ)
Logical structure of multi-class classifier equivalent to PTSVQ
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
MACA Based Two Stage Classifier
OutputLayer
Cla
ssif
ier
1
Cla
ssif
ier
2 Classifier 1: n-bit DS consists of m DVs
Classifier 2: m-bit DV
HiddenLayerInput
Layer
0.0280.00915
0.9790.9195300
0.0190.00615
0.9690.8845200
0.0100.00315
0.9420.7955100
Memory Size RatioSoftware Hardware
Value of PEF (m)
No of Bits (n)
MSR (software) = (n+m) / (n+2m)
MSR (hardware) = (3n+3m-4) / (3n-2+2m)
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Image CompressionOriginal Decompressed
compression 96.43%, PSNR 32.81
compression 95.66%, PSNR 34.27
0.041020.131920.194116 X 16
0.013670.033120.04738 X 8
0.005620.008240.01214 X 4
CATSVQFull Search
Block Size
Execution Time (in milli seconds)
High compression
Acceptable image quality
Higher speed
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
MACA based Tree-Structured Classifier
I
MACA1
IV
IVII
III
MACA2 MACA3
I II III
MACA4
I II IV
Selection of MACA:
Diversity of ith attractor basin (node): Mi = max{Nij } / ∑j Nijwhere Nij - number of tuples of class j covered by ith attractor basin
Mi ≈ 1, ith attractor indicates class attractor indicates class j j for whichfor which Nij is maximumFigure of Merit: FM = 1/k ∑i Mi
where k denotes number of attractor basins
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Fault Diagnosis of Digital CircuitFault
Injection Diagnosis of an example CUT ‘EC’
0.442695.1299.9410S6669(55,6358)
0.930879.5497.058S3271(14,2585)
0.750489.6392.4310S4863(16,4123)
0.223898.8198.966C7552(108,7053)
0.210499.3399.726C6288(32,7648)
0.910696.0398.836C1908(25,1801)
MemoryDictionaryMACA# PartitionCUT (n,p)
Module 1
Module 2
CUT ‘EC’Set of
Test Vectors
Fault Injection
S
ASignature
SetPattern
Classifier MACA
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Fault Diagnosis of Analog Circuit
99.87756035610C2
99.83954215430C1
99.88654245430OTA2
99.91889628970OTA1
SRNot detectedDetected# SamplesComponent
OTA2OTA1
C2C1Vin
Vout
OTA2OTA1
Vin
X1
OTA3
X2
X1: Output of BPFX2: Output of LPF
97.6410242194321C2
97.3011641814297C1
99.88544364441OTA3
99.651543064321OTA2
99.89871937201OTA1
SRNot detectedDetected# SamplesComponent
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Performance on STATLOG Dataset
Memory Overhead (Kbyte)Classification Accuracy (%)
89.594.494.696.7Segment
78.779.368.572.9Vehicle
84.467.286.687.4Letter
94.199.999.999.9Shuttle
77.586.285.285.4Satimage
86.680.779.380.1Heart
74.667.167.466.8German
87.991.493.390.3DNA
75.975.374.272.9Diabetes
86.584.785.883.4Australian
MACAMLPC4.5BayesianDataset
24.512.74370.42121.89
29.831.3272.1447.02
354.112.571299.28766
55.170.711513.571500
222.7411.33709.72669
3.171.5219.469.8
17.4213.6499.3149
50.8637.221067.961000
9.960.4427.1514
8.041.9337.8519
MACAMLPC4.5Bayesian
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
MACATree VS C4.5 on Statlog Dataset
24.51370.4289.594.6Segment
29.8372.1478.768.5Vehicle
354.111299.2884.486.6Letter
55.171513.5794.199.9Shuttle
222.74709.7277.585.2Satimage
3.1719.4686.679.3Heart
17.4299.3174.667.4German
50.861067.9687.993.3DNA
9.9627.1575.974.2Diabetes
8.0437.8586.585.8Australian
Memory OverheadC4.5 MACA
Classifn. AccuracyC4.5 MACA
STATLOGDataset
58250839.782
28196350.8139
563314352524.62107
6633398547.249
49432255199.2433
1718198.633
19196749.8134
494655124.3127
10173934.239
446127.335
Retrieval Time(ms)C4.5 MACA
No of NodesC4.5 MACA
Comparable classification accuracy
Low memory overhead
Lesser number of intermediate nodes
Lesser retrieval time
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Conclusion
• Advantages:– Explore computational capability of MACA– Introduction of Dependency Vector (DV)/String (DS) to characterize MACA– Reduction of complexity to identify attractor basins from O(n3) to O(n)– Elegant evolutionary algorithm – combination of DV/DS and GA– MACA based tree-structured pattern classifier– Application of MACA in
• Classification• image compression• fault diagnosis of electronic circuits• Codon to amino acid mapping, S-box of AES
• Problems:– Linear MACA – employs only XOR logic, functionally incomplete– Distribution of each attractor basin is even – Can handle only binary patterns
• Solutions:– Nonlinear MACA (GMACA) – Fuzzy MACA (FMACA)
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Generalized MACA (GMACA)
• Employs non-linear hybrid rules with all possible logic• Cycle or attractor length greater than 1• Can perform pattern recognition task• Behaves as an associative memory
0100 1000
1010 0001
0101
0011
0010
0000
11010111
1100 1001
10110110
1110
1111
P1 attractor-1
P2 attractor-2
Rule vector:<202,168,218,42>
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Basins of Attraction (Theoretical)
n = 50
k = 10Error correcting capability
at multiple bit noise
Error correcting capability at single bit noise
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Distribution of CA Rule (Theoretical)
More homogeneous – less probability of occurrence
Degree of Homogeneity
DH = | 1- r/4 | where r = number of 1’s of a rule
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Synthesis of GMACA
0100 1000
0001
0010 0000
Basin 1
Basin
1110 11011011
0111
1111
Basin 2
Next State
01110111011111111111
Present State Next StatePhase I: Random Generation of a directed sub-graph
0100 00011 1000 0001Phase II: State transition table
from sub-graph 0001 00000000 00000010 0000
Phase III: GMACA rule vector from State transition table
g1For 2nd Cell:-
111 1110 1101 1100 0011 1 010 0001 0000 0
Rule 232
Basin Present State
1110
2 1011110101111111
g2
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Genetic AlgorithmResolution of Collision:
if n0 = n1; next state is either `0’ or `1’
if n0 > n1; next state is `0’
if n0 < n1; next state is `1’
where n0 = Occurrence of state `0’ for a configuration
n1 = Occurrence of state `1’
g1 g2 ….. gkExample chromosome format – each gx a basin of a pattern Px
k numbers of genes in a chromosome
Each gene - a single cycle directed sub-graph with p number of nodes, where p = 1 + n
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Maximum Permissible Noise/Height
Minimum value of maximum permissible height hmax = 2
Minimum value of maximum permissible noise rmax = 1
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Performance Analysis of GMACA
Higher memorizing capacity than Hopfield network
Cost of computation is constant –depends on transient length of CA
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Basins of Attraction (Experimental)
n = 50 k = 10
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Distribution of CA Rule (Experimental)
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Conclusion
• Advantages:– Explore computational capability of non-linear MACA– Characterization of basins of attraction of GMACA– Fundamental results to characterize GMACA rules– Reverse engineering method to synthesize GMACA– Combination of reverse engineering method and GA– Higher memorizing capacity than Hopfield network
• Problems:– Can handle only binary patterns
• Solutions:– Fuzzy MACA (FMACA)
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Fuzzy Cellular Automata (FCA)
• A linear array of cells• Each cell assumes a state - a rational value in [0, 1]• Combines both fuzzy logic and Cellular Automata• Out of 256 rules, 16 rules are OR and NOR rules (including 0 and 255)
Boolean Function Opeartion FCA Operation
OR (a + b) min{1, (a + b)}
AND (a.b) (a.b)
NOT (~a) (1 – a)
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Fuzzy Cellular Automata (FCA)OR Logic NOR Logic
Rule 170 : qI(t+1) = qI-1(t) Rule 85 : qI(t+1) = qI-1(t)
Rule 204 : qI(t+1) = qI (t) Rule 51 : qI(t+1) = qI (t)
Rule 238 : qI(t+1) = qI (t) + qI+1(t) Rule 17 : qI(t+1) = qI (t) + qI+1(t)
Rule 240 : qI(t+1) = qI+1(t) Rule 15 : qI(t+1) = qI+1(t)
Rule 250 : qI(t+1) = qI-1(t) + qI+1(t) Rule 5 : qI(t+1) = qI-1(t) + qI+1(t)
Rule 252 : qI(t+1) = qI-1(t) + qI (t) Rule 3 : qI(t+1) = qI-1(t) + qI (t)
Rule 254 : qI(t+1) = qI-1(t) + qI (t) + qI+1(t) Rule 1 : qI(t+1) = qI-1(t) + qI (t) + qI+1(t)
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Fuzzy Cellular Automata (FCA)
• 16 OR and NOR rules can be represented by n x n matrix T and an ndimensional binary vector F
• Si(t) represents the state of ith cell at tth time instantSi(t+1) = | Fi - min{1, Σj Tij.Sj(t)}|
where Tij = 1 if next state of ith cell dependents on jth cell
0 otherwise F = Inversion vector, contains 1 where NOR rule is applied
4-cell null boundary hybrid FCA
<238,1,238,3>T =
1 1 0 0
1 1 1 0
0 0 1 1
0 0 1 1
F =
0
1
0
1
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Fuzzy Multiple Attractor CA (FMACA)
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Fuzzy Multiple Attractor CA (FMACA)
• Dependency Vector (DV) – corresponding to matrix
• Derived Complement Vector (DCV) – corresponding to inversion vector
• Pivot cell (PC) – represents an attractor basin uniquely
• State of Pivot Cell (PC) of attractor of the basin where a statebelongs
qm = min {1, Σj |DCVj - DVj.Sj(t)|}
• Size of attractor basins – equal as well as unequal
• Matrix, inversion vector from DV/DCV
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Fuzzy Multiple Attractor CA (FMACA)
1.0 0.0 1.0
1.0 0.5 1.00.5 0.5 1.0
0.5 1.0 1.0
1.0 1.0 1.0
0.0 1.0 1.0
1.0 1.0 0.0
0.0 0.5 0.50.5 0.0 0.5
0.5 0.5 0.0
0.5 0.0 1.0
0.0 0.5 1.0 0.5 1.0 0.0 0.0 1.0 0.0
0.5 0.0 0.0
0.0 0.5 0.0
1.0 0.0 0.0
0.0 0.0 0.5
0.0 0.0 0.0
0.0 0.0 1.0
1.0 0.5 0.0
1.0 1.0 0.50.5 0.5 0.5
0.0 1.0 0.5
0.5 1.0 0.5
1.0 0.5 0.5
1.0 0.0 0.5
T = 0 0 1
1 1 0
0 0 0
0
0
0
1
1
1
0
0
0 F = DV = DCV =
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Fuzzy Multiple Attractor CA (FMACA)
1.0 0.0 0.0
1.0 0.5 0.00.5 0.5 0.0
0.5 1.0 0.0
1.0 1.0 0.0
0.0 1.0 0.0
1.0 1.0 1.0
0.0 0.5 0.50.5 0.0 0.5
0.5 0.5 1.0
0.5 0.0 0.0
0.0 0.5 0.0 0.5 1.0 1.0 0.0 1.0 1.0
0.5 0.0 1.0
0.0 0.5 1.0
1.0 0.0 1.0
0.0 0.0 0.5
0.0 0.0 1.0
0.0 0.0 0.0
1.0 0.5 1.0
1.0 1.0 0.50.5 0.5 0.5
0.0 1.0 0.5
0.5 1.0 0.5
1.0 0.5 0.5
1.0 0.0 0.5
T = 0 0 1
1 1 0
0 0 0
0
1
1
1
1
1
0
0
1 F = DV = DCV =
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
FMACA based Tree-Structured Classifier
I
FMACA1
IV
IVII
III
FMACA2 FMACA3
I II III
FMACA4
I II IV
• FMACA based tree-structured pattern classifier
• Can handle binary as well as real valued datasets
• Provides equal and unequal size of attractor basins
• Combination of GA and DV/DCV
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Experimental Setup
• Randomly generate K number of centroids• Around each centroid, generate t number of tuples• 50 % patterns are taken for training• 50 % patterns are taken for testing
A2A1
AK
dmax
Dmin
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Performance Analysis of FMACA
Dataset Depth TrainingAccuracy
TestingAccuracy
Breadth
(n=5,K=2,t=4000) 12
16.981.8
15.680.9
121
34
97.798.4
91.692.4
359Can generalize dataset irrespective of classes,
tuples, attributes
• Generalization of FMACA tree
– Depth: Number of layers from root to leaf
– Breadth: Number of intermediate nodes
Attributes (n) Size (t) No of Classes FMACA C4.5
6 6000 48
96.6889.26
93.1081.60
8 10000 610
85.9185.61
79.6073.93
10 10000 48
83.8274.01
77.1066.90
Higher classification accuracy compared to C4.5
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Performance Analysis of FMACA
High generation time - but, one time cost
Lower retrieval time compared to C4.5
Generation Time (ms) Retrieval Time (ms)Dataset
FMACA C4.5 FMACA C4.5(n=5,k=2,t=2000) 14215 273 3 306
(n=5,k=2,t=20000) 52557 756 80 812
(n=6,k=2,t=2000) 722725 162 4 259
(n=6,k=2,t=20000) 252458 791 35 874
Attributes (n) Size (t) No of Classes FMACA C4.5
6 6000 48
13621331
51096932
8 8000 68
12611532
95198943
10 10000 610
13641481
77277984
Lower memory overhead compared to C4.5
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Performance on STATLOG Dataset
Classification Accuracy (%)
85.193.181.988. 1
FMACA
84.4Letter94.1Shuttle77.5Satimage87.9DNA
MACADataset
16
936180
FMACA
1008
567540180
MACANumber of CA cells
661.8
51.8161.6122.1
FMACA
524.6
47.2199.2124.3MACA
No of Nodes of Tree
261.0818.29189.6251.22
FMACA
354.11Letter55.17Shuttle222.74Satimage50.86DNAMACADataset
Memory Overhead
1382
19822419491
FMACA
5633
66334943494
MACARetrieval Time (ms)
Comparable accuracy
Lesser CA cells
Lesser memory overhead
Lesser retrieval time
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Conclusion
Introduction of fuzzy CA in pattern recognition
New mathematical tools –• Dependency matrix, Dependency vector• Complement vector, Derived complement vector
Reduction of complexity to identify attractors from O(n3) to O(n)
Both equal and unequal size of attractor basins
Movement of patterns from one to another basin
Reduction of search space
Elegant evolutionary algorithm – combination of DV/DCV and GA
FMACA based tree-structured pattern classifier
Department of Computer Science & Technology,
Bengal Engineering College (DU), Shibpur, Howrah, W. B. India
Future Extensions
Applications in pattern clustering, mix-mode learning
Theoretical analysis of memorizing capacity of non-linear CA
Combination of fuzzy set and fuzzy CA 1-D CA to 2-D CA
Development of hybrid systems using CA• CA + neural network + fuzzy set• CA + fuzzy set + rough set
Boolean CA to multi-valued / hierarchical CA
Application of CA in• Bioinformatics• Medical Image Analysis• Image Compression• Data Mining