3 Adaptation of Fuzzy Inference System Using Neural Learning
3 Adaptation of Fuzzy Inference System Using Neural Learning
3 Adaptation of Fuzzy Inference System Using Neural Learning
A. Abraham
The integration of neural networks and fuzzy inference systems could be for-
mulated into three main categories: cooperative, concurrent and integrated
neuro-fuzzy models. We present three different types of cooperative neuro-
fuzzy models namely fuzzy associative memories, fuzzy rule extraction using
self-organizing maps and systems capable of learning fuzzy set parameters.
Different Mamdani and Takagi-Sugeno type integrated neuro-fuzzy systems
are further introduced with a focus on some of the salient features and ad-
vantages of the different types of integrated neuro-fuzzy models that have
been evolved during the last decade. Some discussions and conclusions are
also provided towards the end of the chapter.
3.1 Introduction
Hayashi et al. [21] showed that a feedforward neural network could approxi-
mate any fuzzy rule based system and any feedforward neural network may
be approximated by a rule based fuzzy inference system [31]. Fusion of Ar-
tificial Neural Networks (ANN) and Fuzzy Inference Systems (FIS) have at-
tracted the growing interest of researchers in various scientific and engineer-
ing areas due to the growing need of adaptive intelligent systems to solve
the real world problems [5, 6, 10, 11, 12, 13, 17, 19, 20, 22, 33, 37]. A
neural network learns from scratch by adjusting the interconnections be-
tween layers. Fuzzy inference system is a popular computing framework
based on the concept of fuzzy set theory, fuzzy if-then rules, and fuzzy
reasoning. The advantages of a combination of neural networks and fuzzy
inference systems are obvious [12, 32]. An analysis reveals that the draw-
backs pertaining to these approaches seem complementary and therefore it
is natural to consider building an integrated system combining the con-
cepts [37]. While the learning capability is an advantage from the view-
point of fuzzy inference system, the automatic formation of linguistic rule
base will be advantage from the viewpoint of neural network. There are
A. Abraham: Adaptation of Fuzzy Inference System Using Neural Learning, StudFuzz 181, 53–
83 (2005)
www.springerlink.com
c Springer-Verlag Berlin Heidelberg 2005
54 A. Abraham
several works related to the integration of neural networks and fuzzy inference
systems [1, 2, 3, 15, 17, 23, 28, 30, 32, 34, 43, 45, 49, 52].
Fuzzy rules
Data Output
Fuzzy sets
its weights. The main disadvantage of FAM is the weighting of rules. Just
because certain rules, does not have much influence does not mean that they
are very unimportant. Hence, the reliability of FAMs for certain applications
is questionable. Due to implementation simplicity, FAMs are used in many
applications.
Pedryz et al. [46] used self-organizing maps with a planar competition layer to
cluster training data, and they provide means to interpret the learning results.
The learning results could show whether two input vectors are similar to each
other or belong to the same class. However, in the case of high-dimensional
input vectors, the structure of the learning problem can rarely be detected
in the two dimensional map. A procedure is provided for interpreting the
learning results using linguistic variables.
After the learning process, the weight matrix W represents the weight of
each feature of the input patterns to the output. Such a matrix defines a map
for a single feature only. For each feature of the input patterns, fuzzy sets
are specified by a linguistic description B (one fuzzy set for each variable).
They are applied to the weight matrix W to obtain a number of transformed
matrices. Each combination of linguistic terms is a possible description of a
pattern subset or cluster. To check a linguistic description B for validity, the
transformed maps are intersected and a matrix D is obtained. Matrix D deter-
mines the compatibility of the learning result with the linguistic description
B. D(B) is a fuzzy relation, and D(B) is interpreted as the degree of support
of B. By describing D(B) by its α-cuts DαB one obtains subsets of output
nodes, whose degree of membership is at least α such that the confidence of
all patterns Xα belong to the class described by Bvanishes with decreasing
α. Each B is a valid description of a cluster if D(B) has a non-empty α-cut
DαB . If the features are separated into input and output features according to
56 A. Abraham
Data output
Table 3.1. Comparison between neural networks and fuzzy inference systems
Table 3.1 summarizes the comparison between neural networks and fuzzy
inference system. To a large extent, the drawbacks pertaining to these two ap-
proaches seem complementary. Therefore, it seems natural to consider building
an integrated system combining the concepts of FIS and ANN modeling. A
common way to apply a learning algorithm to a fuzzy system is to represent it
in a special neural network like architecture. However the conventional neural
network learning algorithms (gradient descent) cannot be applied directly to
such a system as the functions used in the inference process are usually non
differentiable. This problem can be tackled by using differentiable functions in
the inference system or by not using the standard neural learning algorithm.
In Sects. 3.4.1 and 3.4.2, we will discuss how to model integrated neuro-fuzzy
systems implementing Mamdani [36] and Takagi-Sugeno FIS [47].
Layer 5
rule inference and defuzzification layer
Layer 4
rule consequent layer
Layer 3
R1 R2 R3 rule antecedent layer
Layer 2
(fuzzification layer)
Layer 1
(input layer)
x1 x2
Layer-5 (rule consequent layer): Every node i in this layer is with a node
function
wi fi = wi (pi x1 + qi x2 + ri ) (3.2)
where wi is the output of layer 4, and {pi , qi , ri } is the parameter set.
A well-established way is to determine the consequent parameters using the
least means squares algorithm.
Layer-6 (rule inference layer) The single node in this layer computes the
overall output as the summation of all incoming signals
wi fi
Overall output = wi fi = i (3.3)
i i wi
ANFIS is perhaps the first integrated hybrid neuro-fuzzy model [24] and the
architecture is very similar to Fig. 3.4. A modified version of ANFIS as shown
in Fig. 3.5 is capable of implementing the Tsukamoto fuzzy inference system
[24, 51] as depicted in Fig. 3.6. In the Tsukamoto FIS, the overall output is
the weighted average of each rule’s crisp output induced by the rule’s firing
strength (the product or minimum of the degrees of match with the premise
part) and output membership functions. The output membership functions
used in this scheme must be monotonically non-decreasing. The first hidden
layer is for fuzzification of the input variables and T-norm operators are de-
ployed in the second hidden layer to compute the rule antecedent part. The
third hidden layer normalizes the rule strengths followed by the fourth hidden
layer where the consequent parameters of the rule are determined. Output
layer computes the overall input as the summation of all incoming signals.
3 Adaptation of Fuzzy Inference System Using Neural Learning 61
y
Layer 6
rule inference layer
Layer 5
rule consequent layer
x1 x2
Layer 4
rule strength normalization
Layer 3
R1 R2 R3 rule antecedent layer
Layer 2
(fuzzification layer)
Layer 1
(input layer)
x1 x2
FALCON [32] has a five-layered architecture as shown in Fig. 3.7 and imple-
ments a Mamdani type FIS. There are two linguistic nodes for each output
variable. One is for training data (desired output) and the other is for the
actual output of FALCON. The first hidden layer is responsible for the fuzzi-
fication of each input variable. Each node can be a single node representing
a simple membership function (MF) or composed of multilayer nodes that
y1 y'1 ym y'm
R1 R1 R2 R3
x1 xn
Fig. 3.7. Architecture of FALCON
3 Adaptation of Fuzzy Inference System Using Neural Learning 63
compute a complex MF. The Second hidden layer defines the preconditions
of the rule followed by rule consequents in the third hidden layer. FALCON
uses a hybrid-learning algorithm comprising of unsupervised learning and a
gradient descent learning to optimally adjust the parameters to produce the
desired outputs. The hybrid learning occurs in two different phases. In the ini-
tial phase, the centers and width of the membership functions are determined
by self-organized learning techniques analogous to statistical clustering tech-
niques. Once the initial parameters are determined, it is easy to formulate the
rule antecedents. A competitive learning algorithm is used to determine the
correct rule consequent links of each rule node. After the fuzzy rule base is es-
tablished, the whole network structure is established. The network then enters
the second learning phase to adjust the parameters of the (input and output)
membership functions optimally. The backpropagation algorithm is used for
the supervised learning. Hence FALCON algorithm provides a framework for
structure and parameter adaptation for designing neuro-fuzzy systems [32].
R1
ξ1
R2
η
R3
ξ2
R4
five layers. The first hidden layer stores the linguistic values of all the input
variables. Each input unit is only connected to those units of the first hidden
layer, which represent its associated linguistic values. The second hidden layer
represents the fuzzy rules nodes, which determine the degree of fulfillment of a
rule using a softmin operation. The third hidden layer represents the linguistic
values of the control output variable η. Conclusions of the rule are computed
depending on the strength of the rule antecedents computed by the rule node
layer. GARIC makes use of local mean-of-maximum method for computing
the rule outputs. This method needs a crisp output value from each rule.
Therefore, the conclusions must be defuzzified before they are accumulated
to the final output value of the controller. The learning algorithm of the AEN
of GARIC is equivalent to that of its predecessor ARIC. However, the ASN
learning procedure is different from the procedure used in ARIC. GARIC
uses a mixture of gradient descent and reinforcement learning to fine-tune the
node parameters. The hybrid learning stops if the output of the AEN ceases
to change. The interpretation of GARIC is improved compared to GARIC.
The relatively complex learning procedure and the architecture of GARIC can
be seen as a main disadvantage of GARIC.
The learning algorithm defined for NEFCON is able to learn fuzzy sets as well
as fuzzy rules implementing a Mamdani type FIS [40]. This method can be
considered as an extension to GARIC that also use reinforcement learning but
need a previously defined rule base. Figure 3.9 illustrates the basic NEFCON
architecture with 2 inputs and five fuzzy rules [40]. The inner nodes R1 , . . . , R5
represent the rules, the nodes ξ 1 , ξ 2 , and η the input and output values, and
µr , Vr the fuzzy sets describing the antecedents and consequents. In contrast
to neural networks, the connections in NEFCON are weighted with fuzzy sets
instead of real numbers. Rules with the same antecedent use so-called shared
weights, which are represented by ellipses drawn around the connections as
shown in the figure. They ensure the integrity of the rule base. The knowl-
edge base of the fuzzy system is implicitly given by the network structure.
The input units assume the task of fuzzification interface, the inference logic
is represented by the propagation functions, and the output unit is the defuzzi-
fication interface. The learning process of the NEFCON model can be divided
into two main phases. The first phase is designed to learn the rule base and the
second phase optimizes the rules by shifting or modifying the fuzzy sets of the
rules. Two methods are available for learning the rule base. Incremental rule
learning is used when the correct out put is not known and rules are created
based on estimated output values. As the learning progresses, more rules are
added according to the requirement. For decremental rule learning, initially
rules are created due to fuzzy partitions of process variables and unnecessary
rules are eliminated in the course of learning. Decremental rule learning is
less efficient compared to incremental approach. However it can be applied to
3 Adaptation of Fuzzy Inference System Using Neural Learning 65
Defuzzification layer
V1
V2
V3 Rule base
layer
R1 R2 R3 R4 R5
(1) (2)
µ µ
3 1
Fuzzification layer
Defuzzification layer
c1 c2
1 1 1 1 1 Rule
base layer
R1 R2 R3 R4 R5
(1) (2)
µ µ
3 1
x1 Fuzzification layer x2
Defuzzification layer
y1 y2
Rule
base layer
R1 R2 R3 R4 R5
(1) (2)
µ µ
3 1
x1 Fuzzification layer x2
and the parameters in the precondition part are adjusted by the backpropaga-
tion algorithm. SONFIN can be used for normal operation at anytime during
the learning process without repeated training on the input-output pattern
when online operation is required. In SONFIN rule base is dynamically created
as the learning progresses by performing the following learning processes:
• Input-output space partitioning
The way the input space is partitioned determines the number of rules ex-
tracted from the training data as well as the number of fuzzy sets on the
universal of discourse of each input variable. For each incoming pattern x
the strength a rule is fired can be interpreted as the degree the incoming
pattern belongs to the corresponding cluster. The center and width of the
corresponding membership functions (of the newly formed fuzzy rules) are
assigned according to the first-neighbor heuristic. For each rule generated,
the next step is to decompose the multidimensional membership function to
corresponding 1 − D membership function for each input variable. For the
output space partitioning, almost a similar measure is adopted. Performance
of SONFIN can be enhanced by incorporating a transformation matrix R into
the structure, which accommodates all the a priori knowledge of the data set.
• Construction of fuzzy rule base
Generation of new input cluster corresponds to the generation of a new fuzzy
rule, with its precondition part constructed by the learning algorithm in
3 Adaptation of Fuzzy Inference System Using Neural Learning 71
process. At the same time we have to decide the consequent part of the gen-
erated rule. This is done using a algorithm based on the fact that different
preconditions of rules may be mapped to the same consequent fuzzy set. Since
only the center of each output membership function is used for defuzzifica-
tion, the consequent part of each rule may simply be regarded as a singleton.
Compared to the general fuzzy rule based models with singleton output where
each rule has its own singleton value, fewer parameters are needed in the con-
sequent part of the SONFIN, especially for complicated systems with a large
number of rules.
TSK model can model a sophisticated system with a few rules. In SONFIN, in-
stead of using the linear combination of all input variables as the consequent
part, only the most significant input variables are used as the consequent
terms of the SONFIN. The significant terms will be chosen and added to the
network incrementally any time when the parameter learning cannot improve
the network output accuracy anymore during the online learning process. The
consequent structure identification scheme in SONFIN is a kind of node grow-
ing method in ANNs. When the effect of the parameter learning diminished
(output error is not decreasing), additional terms are added to the consequent
part.
• Parameter identification
After the network structure is adjusted according to the current training pat-
tern, the network then enters the parameter identification phase to adjust the
parameters of the network optimally based on the same training pattern. Pa-
rameter learning is performed on the whole network after structure learning,
no matter whether the nodes (links) are newly added or are existent originally.
Backpropagation algorithm is used for this supervised learning. SONFIN is
perhaps one of the most computational expensive among all neuro-fuzzy mod-
els. The network is adaptable to the users specification of required accuracy.
Defuzzification
Steering
Fuzzification
Goal Sensor
Input
Fig. 3.14. Architecture of the FUN showing the implementation of a sample rule
fuzzification of the input values. In the second hidden layer, the conjunctions
(fuzzy-AND) are calculated. Membership functions of the output variables
are stored in the third hidden layer. Their activation function is a fuzzy-OR.
Finally, the output neurons contain the output variables and have a defuzzi-
fication activation function. FUN network is depicted in Fig. 3.14.
The rules and the membership functions are used to construct an initial
FUN network. The rule base can then be optimized by changing the struc-
ture of the net or the data in the neurons. To learn the rules, the connections
between the rules and the fuzzy values are changed. To learn the member-
ship functions, the data of the nodes in the first and three hidden layers
are changed. FUN can be trained with the standard neural network training
strategies such as reinforcement or supervised learning.
3 Adaptation of Fuzzy Inference System Using Neural Learning 73
The rules are represented in the net through the connections between the
layers. The learning of the rules is implemented as a stochastic search in the
rule space: a randomly chosen connection is changed and the new network
performance is verified with a cost function. If the performance is worse, the
change is undone, otherwise it is kept and some other changes are tested, until
the desired output is achieved. As the learning algorithm should preserve the
semantic of the rules, it has to be controlled in such a way that no two values
of the same variable appear in the same rule. This is achieved by swapping
connections between the values of the same variable. FUN uses a mixture
of gradient descent and stochastic search for updating the membership func-
tions. A maximum change in a random direction is initially assigned to all
Membership function Descriptors (MFDs). In a random fashion one MFD
of one linguistic variable is selected, and the network performance is tested
with this MFD altered according to the allowable change for this MFD. If
the network performs better according to the given cost function, the new
value is accepted and next time another change is tried in the same direction.
Contrary if the network performs worse, the change is reversed. To guaran-
tee convergence, the changes are reduced after each training step and shrink
asymptotically towards zero according to the learning rate. As evident, FUN
system is initialized by specifying a fixed number of rules and a fixed number
of initial fuzzy sets for each variable and the network learns through a sto-
chastic procedure that randomly changes parameters of membership functions
and connections within the network structure Since no formal neural network
learning technique is used it is questionable to call FUN a neuro-fuzzy system.
EFuNNs [26] and dmEFuNNs [27] are based on the ECOS (Evolving COnnec-
tionist Systems) framework for adaptive intelligent systems formed because of
evolution and incremental, hybrid (supervised/unsupervised), online learning.
They can accommodate new input data, including new features, new classes,
and etc. through local element tuning.
In EFuNNs all nodes are created during learning. EFuNN has a five-layer
architecture as shown in Fig. 3.15. The input layer is a buffer layer representing
the input variables. The second layer of nodes represents fuzzy quantification
of each input variable space. Each input variable is represented here by a group
of spatially arranged neurons to represent a fuzzy quantization of this variable.
The nodes representing membership functions (triangular, Gaussian, etc) can
be modified during learning. The third layer contains rule nodes that evolve
through hybrid supervised/unsupervised learning. The rule nodes represent
prototypes of input-output data associations, graphically represented as an
association of hyper-spheres from the fuzzy input and fuzzy output spaces.
Each rule node ris defined by two vectors of connection weights: W1 (r) and
74 A. Abraham
W2 (r), the latter being adjusted through supervised learning based on the
output error, and the former being adjusted through unsupervised learning
based on similarity measure within a local area of the input problem space.
The fourth layer of neurons represents fuzzy quantification for the output
variables. The fifth layer represents the real values for the output variables.
In the case of “one-of-n” EFuNNs, the maximum activation of the rule node
is propagated to the next level. In the case of “many-of-n” mode, all the
activation values of rule nodes that are above an activation threshold are
propagated further in the connectionist structure.
fuzzy rule will be formed using the weighted least square estimator. The last
layer calculates the output of dmEFuNN.
The number m of activated nodes used to calculate the output values for
a dmEFuNN is not less than the number of the input nodes plus one. Like
the EFuNNs, the dmEFuNNs can be used for both offline learning and online
learning thus optimising global generalization error, or a local generalization
error. In dmEFuNNs, for a new input vector (for which the output vector is
not known), a subspace consisted of m rule nodes are found and a first order
TakagiSugeno fuzzy rule is formed using the least square estimator method.
This rule is used to calculate the dmEFuNN output value. In this way, a dmE-
FuNN acts as a universal function approximator using m linear functions in a
small m dimensional node subspace. The accuracy of approximation depends
on the size of the node subspaces, the smaller the subspace is, the higher
the accuracy. It means that if there are sufficient training data vectors and
sufficient rule nodes are created, a satisfying accuracy can be obtained.
3.5 Discussions
As evident, both cooperative and concurrent models are not fully interpretable
due to the presence of neural network (black box concept). Whereas an in-
tegrated neuro-fuzzy model is interpretable and capable of learning in a su-
pervised mode (or even reinforcement learning like NEFCON). In FALCON,
76 A. Abraham
GARIC, ANFIS, NEFCON, SONFIN, FINEST and FUN the learning process
is only concerned with parameter level adaptation within fixed structures.
For large-scale problems, it will be too complicated to determine the opti-
mal premise-consequent structures, rule numbers etc. User has to provide the
architecture details (type and quantity of MF’s for input and output vari-
ables), type of fuzzy operators etc. FINEST provides a mechanism based on
the improved generalized modus ponens for fine tuning of fuzzy predicates and
combination functions and tuning of an implication function. An important
feature of EFuNN and dmEFuNN is the one pass (epoch) training, which is
highly capable of online learning. Table 3.2 provides a comparative perfor-
mance of some neuro fuzzy systems for predicting the Mackey-Glass chaotic
time series [35]. Due to unavailability of source codes, we are unable to provide
a comparison with all the models. Training was done using 500 data sets and
the considered NF models were tested with another 500 data sets [1].
Among NF models ANFIS has the lowest Root Mean Square Error
(RMSE) and NEPROX the highest. This is probably due to Takagi-Sugeno
rules implementation in ANFIS compared to the Mamdani-type fuzzy system
in NEFPROX. However, NEFPROX outperformed ANFIS in terms of compu-
tational time. Due to fewer numbers of rules SONFIN, EFuNN and dmEFuNN
are also able to perform faster than ANFIS. Hence, there is a tradeoff between
interpretability and accuracy. Takagi Sugeno type inference systems are more
accurate but require more computational effort. While Mamdani type infer-
ence, systems are more interpretable and required less computational load but
often with a compromise on accuracy.
As the problem become, more complicated manual definition of NF ar-
chitecture/parameters becomes complicated. The following questions remain
unanswered:
• For input/output variables, what are the optimal number of membership
functions and shape?
• What is the optimal structure (rule base) and fuzzy operators?
• What are the optimal learning parameters?
• Which fuzzy inference system (example. Takagi-Sugeno, Mamdani etc.) will
work the best for a given problem?
3 Adaptation of Fuzzy Inference System Using Neural Learning 77
Slow
Search of fuzzy inference system
Fig. 3.18. Chromosome representation of the adaptive fuzzy inference system using
evolutionary computation and neural learning
is more prior knowledge about the fuzzy rule base than the fuzzy operators
then it is better to implement the search for fuzzy rule base at a higher level.
The problem representation (genetic coding) is illustrated in Fig. 3.18. Please
refer [4] for more technical details.
Automatic adaptation of membership functions is popularly known as self
tuning and the genome encodes parameters of trapezoidal, triangle, logistic,
hyperbolic-tangent, Gaussian membership functions etc.
Evolutionary search of fuzzy rules can be carried out using three ap-
proaches. In the first method, (Michigan approach) the fuzzy knowledge base
is adapted because of antagonistic roles of competition and cooperation of
fuzzy rules [14]. Each genotype represents a single fuzzy rule and the entire
population represents a solution. A classifier rule triggers whenever its condi-
tion part matches the current input, in which case the proposed action is sent
to the process to be controlled. The global search algorithm will generate new
classifier rules based on the rule strengths acquired during the entire process.
The fuzzy behavior is created by an activation sequence of mutually collabo-
rating fuzzy rules. The entire knowledge base is build up by a cooperation of
competing multiple fuzzy rules.
The second method (Pittsburgh approach) evolves a population of knowl-
edge bases rather than individual fuzzy rules [14]. Genetic operators serve to
provide a new combination of rules and new rules. In some cases, variable
length rule bases are used; employing modified genetic operators for dealing
with these variable length and position independent genomes. The disadvan-
tage is the increased complexity of search space and additional computational
burden especially for online learning.
3 Adaptation of Fuzzy Inference System Using Neural Learning 79
The third method (iterative rule learning approach) is very much similar
to the first method with each chromosome representing a single rule, but
contrary to the Michigan approach, only the best individual is considered
to form part of the solution, discarding the remaining chromosomes in the
population. The evolutionary learning process builds up the complete rule
base through a iterative learning process [18].
3.6 Conclusions
In this chapter, we presented the different ways to learn fuzzy inference sys-
tems using neural network learning techniques. As a guideline, for neuro-
fuzzy systems to be highly intelligent some of the major requirements are fast
learning (memory based - efficient storage and retrieval capacities), on-line
adaptability (accommodating new features like inputs, outputs, nodes, con-
nections etc), achieve a global error rate and computationally inexpensive.
The data acquisition and preprocessing training data is also quite impor-
tant for the success of neuro-fuzzy systems. Many neuro-fuzzy models use
supervised/unsupervised techniques to learn the different parameters of the
inference system. The success of the learning process is not guaranteed, as the
designed model might not be optimal. Empirical research has shown that gra-
dient descent technique (most commonly used supervised learning algorithm)
is trapped in local optima especially when the error surface is complicated.
Global optimization procedures like evolutionary algorithms, simulated
annealing, tabu search etc. might be useful for adaptive evolution of fuzzy
if-then rules, shape and quantity of membership functions, fuzzy operators
and other node functions, to prevent the network parameters being trapped in
local optima due to reliance on gradient information by most of the supervised
learning techniques. For online learning, global optimization procedures might
sound computational expensive. Fortunately, evolutionary algorithms work
with a population of independent solutions, which makes it easy to distribute
the computational load among several processors using parallel algorithms.
Sugeno-type fuzzy systems are high performers (less RMSE) but often
requires complicated learning procedures and computational expensive. How-
ever, Mamdani-type fuzzy systems can be modeled using faster heuristics but
with a compromise on the performance (accuracy). Hence there is always a
compromise between performance and computational time.
3.7 Acknowledgements
The author wishes to thank the anonymous reviewers for their constructive
comments which helped to improve the presentation of the chapter.
80 A. Abraham
References
1. A. Abraham, Neuro-Fuzzy Systems: State-of-the-Art Modeling Techniques,
Connectionist Models of Neurons, Learning Processes, and Artificial Intelli-
gence, Springer-Verlag Germany, Jose Mira and Alberto Prieto (Eds.), Granada,
Spain, pp. 269–276, 2001. 54, 76
2. A. Abraham, Intelligent Systems: Architectures and Perspectives, Recent Ad-
vances in Intelligent Paradigms and Applications, Abraham A., Jain L. and
Kacprzyk J. (Eds.), Studies in Fuzziness and Soft Computing, Springer Verlag
Germany, Chap. 1, pp. 1–35, 2002. 54
3. A. Abraham and M.R. Khan, Neuro-Fuzzy Paradigms for Intelligent Energy
Management, Innovations in Intelligent Systems: Design, Management and Ap-
plications, Abraham A., Jain L. and Jan van der Zwaag B. (Eds.), Studies in
Fuzziness and Soft Computing, Springer Verlag Germany, Chap. 12, pp. 285–
314, 2003. 54
4. A. Abraham, EvoNF: A Framework for Optimization of Fuzzy Inference Systems
Using Neural Network Learning and Evolutionary Computation, The 17th IEEE
International Symposium on Intelligent Control, ISIC’02, IEEE Press, pp. 327–
332, 2002. 60, 77, 78
5. P. Andlinger and E.R. Reichl, Fuzzy-Neunet: A Non Standard Neural Network,
In Prieto et al., pp. 173–180, 1991. 53
6. M. Arao, T. Fukuda and K. Shimokima, Flexible Intelligent System based on
Fuzzy Neural Networks and Reinforcement Learning, In proceedings of IEEE
International Conference on Fuzzy Systems, Vol 5(1), pp. 69–70, 1995. 53
7. H.R. Berenji and P. Khedkar, Fuzzy Rules for Guiding Reinforcement Learning,
In International. Conference on Information Processing and Management of
Uncertainty in Knowledge-Based Systems (IPMU’92), pp. 511–514, 1992. 63
8. H.R. Berenji and P. Khedkar, Learning and Tuning Fuzzy Logic Controllers
through Reinforcements, IEEE Transactions on Neural Networks, Vol (3),
pp. 724–740, 1992. 60, 63
9. J.C. Bezdek and S.K. Pal, Fuzzy Models for Pattern Recognition, IEEE Press,
New York, 1992. 56
10. M. Brown, K. Bossley and D. Mills, High Dimensional Neurofuzzy Systems:
Overcoming the Course of Dimensionality, In Proceedings. IEEE International.
Conference on Fuzzy Systems, pp. 2139–2146, 1995. 53
11. J.J. Buckley and Y. Hayashi, Hybrid neural nets can be fuzzy controllers and
fuzzy expert systems, Fuzzy Sets and Systems, 60: pp. 135–142, 1993. 53
12. H. Bunke and A. Kandel, Neuro-Fuzzy Pattern Recognition, World Scientific
Publishing CO, Singapore, 2000. 53
13. G.A. Carpenter, S. Grossberg, N. Markuzon, J.H. Reynolds, and D.B. Rosen,
Fuzzy ARTMAP: A Neural Network Architecture for Incremental Supervised
Learning of Analog Multidimensional Maps, IEEE Transactions Neural Net-
works, 3(5), pp. 698–712, 1992. 53
14. O. Cordón F. Herrera, F. Hoffmann and L. Magdalena, Genetic Fuzzy Systems:
Evolutionary Tuning and Learning of Fuzzy Knowledge Bases, World Scientific
Publishing Company, Singapore, 2001. 77, 78
15. F. De Souza, M.M.R. Vellasco, M.A.C. Pacheco, The Hierarchical Neuro-Fuzzy
BSP Model: An Application in Electric Load Forecasting, Connectionist Mod-
els of Neurons, Learning Processes and Artificial Intelligence, Jose Mira et al
(Editors), LNCS 2084, Springer Verlag Germany, 2001. 54
3 Adaptation of Fuzzy Inference System Using Neural Learning 81
16. J.C. Feng and L.C. Teng, An Online Self Constructing Neural Fuzzy Inference
Network and its Applications, IEEE Transactions on Fuzzy Systems, Vol 6,
No.1, pp. 12–32, 1998. 60, 69
17. R. Fuller, Introduction to Neuro-Fuzzy Systems, Studies in Fuzziness and Soft
Computing, Springer Verlag, Germany, 2000. 53, 54
18. A. Gonzalez and F. Herrera, Multi-Stage Genetic Fuzzy Systems Based on
the Iterative Rule Learning Approach, Mathware and Soft Computing Vol 4,
pp. 233–249, 1997. 79
19. M.M. Gupta, Fuzzy Neural Computing Systems, In Proceedings of SPIE,
Vol 1710, Science of Artificial Neural Networks, Vol 2, pp. 489–499, 1992. 53
20. S.K. Halgamuge and M. Glesner, Neural Networks in Designing Fuzzy Systems
for Real World Applications, Fuzzy Sets and Systems, 65: pp. 1–12, 1994. 53
21. Y. Hayashi and J.J. Buckley, Approximations Between Fuzzy Expert Systems
and Neural Networks, International Journal of Approximate Reasoning, Vol 10,
pp. 63–73, 1994. 53
22. Y. Hayashi and A. Imura, Fuzzy Neural Expert System with Automated Ex-
traction of Fuzzy If-Then Rules from a Trained Neural Network, In First In-
ternational. Symposium on Uncertainty Modeling and Analysis, pp. 489–494,
1990. 53
23. J. Hollatz, Neuro-Fuzzy in Legal Reasoning, In Proceedings. IEEE International.
Conference on Fuzzy Systems, pp. 655–662, 1995. 54
24. R. Jang, Neuro-Fuzzy Modeling: Architectures, Analyses and Applications,
Ph.D. Thesis, University of California, Berkeley, 1992. 60
25. A. Kandel, Q.Y. Zhang and H. Bunke, A Genetic Fuzzy Neural Network for
Pattern Recognition, In IEEE Transactions on Fuzzy Systems, pp. 75–78, 1997. 60
26. N. Kasabov, Evolving Fuzzy Neural Networks – Algorithms, Applications and
Biological Motivation, In Yamakawa T and Matsumoto G (Eds), Methodologies
for the Conception, Design and Application of Soft Computing, World Scientific,
pp. 271–274, 1998. 60, 73
27. N. Kasabov and S. Qun, Dynamic Evolving Fuzzy Neural Networks with
m-out-of-n Activation Nodes for On-line Adaptive Systems, Technical Re-
port TR99/04, Department of information science, University of Otago, New
Zealand, 1999. 60, 73
28. E. Khan and P. Venkatapuram, Neufuz: Neural Network Based Fuzzy Logic
Design Algorithms, In Proceedings IEEE International Conference on Fuzzy
Systems, pp. 647–654, 1993. 54
29. B. Kosko, Neural Networks and Fuzzy Systems: A Dynamical Systems Approach
to Machine Intelligence, Prentice Hall, Englewood Cliffs, New Jersey, 1992. 54
30. W. Li, Optimization of a Fuzzy Controller Using Neural Network, In Proceedings
IEEE International Conference on Fuzzy Systems, pp. 223–227, 1994. 54
31. X.H. Li and C.L.P. Chen, The Equivalance Between Fuzzy Logic Systems and
Feedforward Neural Networks, IEEE Transactions on Neural Networks, Vol 11,
No. 2, pp. 356–365, 2000. 53
32. C.T. Lin and C.S.G. Lee, Neural Network based Fuzzy Logic Control and
Decision System, IEEE Transactions on Comput. (40(12): pp. 1320–1336,
1991. 53, 54, 60, 62, 63
33. C.T. Lin and C.S.G. Lee, Neural Fuzzy Systems: A Neuro-Fuzzy Synergism to
Intelligent Systems, Prentice Hall Inc, USA, 1996. 53
82 A. Abraham
53. R.R. Yager and D.P. Filev, Adaptive Defuzzification for Fuzzy System Model-
ing, In Proceedings of the Workshop of the North American Fuzzy Information
Processing Society, pp. 135–142, 1992. 56
54. Q.Y. Zhang and A. Kandel, Compensatory Neuro-fuzzy Systems with Fast
Learning Algorithms, IEEE Transactions on Neural Networks, Vol 9, No. 1,
pp. 83–105, 1998. 60