Two Modifications of CNN
Two Modifications of CNN
necessarily subsume any other column of Ai since duplicate Two Modifications of CNN
columns are removed in Step 3-b.3. However, at some point in IVAN TOMEK
the reduction of P' to I', P' is reduced to some I'yj*, P' Iyj* c
and, as shown above, there exists an FPI Q such that Q is the Abstract-The condensed nearest-neighbor (CNN) method chooses
disjunction of Q'yj* and possibly Xi and/or Xi. Q' may also have samples randomly. This results in a) retention of unnecessary samples
been reduced to Q" by Step 3-b, Q' c Q". So now, and b) occasional retention of internal rather than boundary samples.
Two modifications of CNN are presented which remove these dis-
col col 2 advantages by considering only points close to the boundary. Per-
formance is illustrated by an example.
I' Q"
Ai = Xixi [..
INTRODUCTION
*
yj* *
yj *
-*
The condensed nearest-neighbor (CNN) method [1 ] is a
method of preprocessing of the design set for pattern recognition.
where col 1 is derived from P and col 2 is derived from Q. Since It is based on the nearest-neighbor (NN) rule [2]. Its purpose is
we areinterested in generating I'XiXi, we will consider two cases. to reduce the size of the original design set D (set of samples
Case 1: F' ' Q". If P contains XiXi, then an e is placed with known membership) by the elimination of certain samples
below col 1 and XiXi' = I is added to A. However, P = without affecting significantly the performance of NN classifica-
XiXiP ' XiXiyi*I' C XiXjI' 1. Therefore, P c I, which
tion: The NN rule used with E (the new design set E c D),
is a contradiction since P is an FPI. Therefore, P cannot contain should give almost the same result as NN used with D.
Xi Xi. So an e is placed below col 1, and 1 is not added to A. CNN works as follows:
Case 2: I' = Q". As shown above, P does not contain XiXi.
If Q contains XiXi, then an e is placed below col 1 and col 2, a) pass v- 1,
and XiXiI' = I is added to A. However, Q = XLXiyJ*Q` C b) choose x E D randomly, D(1) = D - {x}, E = {x},
XiXiy-*Q` XiXiQ = XiXI' = 1. Therefore, Q c I, which is c) D (pass + 1) = 0, count v- 0,
a contradiction since Q is an FPI. Therefore, Q cannot contain d) choose x E D (pass) randomly, classify x by NN using E,
XiXi. So an e is placed below col 1 and col 2, and I is not added e) if classification found in d) agrees with actual membership
to A. of x
Therefore, any nonessential FPI of F will not appear in the then D(pass + 1) = D(pass + 1) U {x}
function resulting from the algorithm of this paper. Q.E.D. else E = F u {x }, count v- count + 1,
f) D(pass)= D(pass) - {x},
V. CONCLUSIONS g) if D(pass) 0 go to d),
h) if count = 0
In this paper we have described an algorithm which finds then end of algorithm
precisely the set of essential fuzzy prime implicants of a fuzzy else pass v- pass + 1, go to b).
function in a sum of products form. At present, this is the only
known fuzzy minimization method for disjunctive functions It is clear that CNN has the following properties. It generates a
which does so without the construction of a cover table. design set E 1) which is a subset of the original design set and
This algorithm is less cumbersome than existing methods of 2) which classifies (NN rule) all samples in D correctly. Property
generating all fuzzy prime implicants in that it does not include 1) usually means that E is much smaller than D and thus com-
an intermediate expansion of the function. Reduction takes place putationally much better suited for NN classification: it requires
only on those phrases which are in the original representation of less storage and computation. Property 2) indicates that NN
the function. For this reason we feel that this is a significant classification with E is very similar (although not necessarily
algorithm, introducing concepts which may prove useful in other identical) to NN classification with D. This is especially true
areas of fuzzy set theory. when D is "representative" (by this we mean that the number of
samples and their distribution is such that the approximation
REFERENCES of the "true" underlying probability distribution by relative
[1] S. R. Das, "An approach for simplifying switching functions by
frequency of samples is "good").
utilizing cover table representation," IEEE Trans. Computers (Short The disadvantage of CNN is that it processes samples from D
Notes), vol. C-20, pp. 355-359, 1971. randomly moves them from D to E quite randomly at the
[2] S. R. Das and N. S. Khabra, "Clause-column table approach for
generating all the prime implicants of switching functions," IEEE beginning and less so later on (when it tends to take samples
Trans. Computers, vol. C-21, pp. 1239-1246, 1972. closer to the boundary). This means that E contains a) interior
[3] A. Kandel, "On minimization of fuzzy functions," IEEE Trans.
Computers, vol. C-22, pp. 826-832, 1973. samples which could be eliminated completely without change in
[41 ,"On the minimization of incompletely specified fuzzy functions," the performance and b) samples which define a boundary on E
Information and Control, pp. 141-153, 1974.
[5] ,"On the Properties of fuzzy switching functions," Journal of but not on D (i.e., samples not essential in D become boundary
Cybernetics, pp. 119- 126, 1974.
[6] -, "A note on the simplification of fuzzy switching functions," points in E). Point a) implies that E is larger than necessary,
Computer Science Report 139, New Mexico Institute of Mining and point b) causes an undesirable shift between boundaries.
Technology, Socorro, New Mexico, 1975.
[71 ,"Inexact switching logic," IEEE Trans. Systems, Man, and The ideal method of reduction of D would work essentially as
Cybernetics, vol. SMC-6, pp. 215-219, 1976. CNN but would only use points close to the decision boundary
[8] A. Kandel and H. A. Davis, "The first fuzzv decade (Bibliography on
fuzzy sets and their applications)," Computer Science Report 140, to generate E. Unfortunately, the true decision boundary is
New Mexico Institute of Mining and Technology, Socorro, New
Mexico, 1976. unknown by definition. The next best is to use only those points
[9] T. P. Neff and A. Kandel, "Simplification of fuzzy switching func- which generate the piecewise-linear decision boundary in D (as
tions," to appear in the International Journal of Computer and In-
formation Sciences, 1977. given by the application of the NN rule). Even this is difficult.
[10] S. M. Rickman and A. Kandel, "Column table approach for the
minimization of fuzzy functions," Computer Science Report 137,
New Mexico Institute of Mining and Technology, Socorro, New
Mexico, 1975. Manuscript received March 12, 1976; revised June 9, 1976.
[11] L. A. Zadeh, "Fuzzy sets," Information and Control, vol. 8, pp. 338- The author is with the Department of Computer Science, Acadia
353. 1965. University, Wolfville, NS, Canada BOP IXO.
770 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS, NOVEMBER 1976
0
.
.
.
0
U
0
0
%4
_
ii
U 0 *
I
U
U
00
n 0C
0
0
0
a
0
0
0
0 lI lI
61 ~~~~1
0
111
0 0 0
0 a
0
0 0 0 a ts
0 0 a o
0 o
0 0~~~~~
a0
Fig. 2. Points represented by full symbols should be chosen for the final
I I
design set. They are not retained by Method 1.
0
*yU ~~~II I
g
I
A .
x *
__0J
a
.
~~~ I
Fig. 3. Step e) in Method 1. y is the nearest neighbor of x (y E F). N
i) I I I I
_..- _m-
qx ~ I
0
YOOk a
a
0 X .
a
0
0
a
0 U
0
Fig. 5. Example of application of Method 2: The upper pair (X,Y) is
accepted as a member of C, the lower pair is not.
(a) (b)
(c) (d)
Fig. 6. Comparison of proposed methods with two others. Example taken from [4] with two classes with uniform distributions
separated by the indicated boundary. Original set (400 samples) is not shown. Results of processing: (a) Method 1 (only subset
C is shown). (b) Method 2. (c) Method from [4]. (d) CNN.
772 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS, NOVEMBER 1976
else classify x by F,
if classification agrees
then E = E {x}, F = F- {x}
u
y
else find z nno(x), z D(pass) and assign it to F: F
=
Fig. 7. Illustration for proof in Appendix.
F u {z }, next find u such that dist (u,z) min E A dist
= v