Techniques for Learning and Tuning Fuzzy Rule-Based...

47
Techniques for Learning and Tuning Fuzzy Rule-Based Systems for Linguistic Modeling and their Application * R.Alcal´a ? , J. Casillas ? ,O.Cord´on ? , F. Herrera ? , S. J. I. Zwir ? Department of Computer Science and Artificial Intelligence. E.T.S. de Ingenier´ ıa Inform´atica, University of Granada. 18071 – Granada, Spain e-mails: {alcala,casillas,ocordon,herrera}@decsai.ugr.es Department of Computer Science. University of Buenos Aires, Buenos Aires, Argentina e-mail: [email protected] Abstract Nowadays, Linguistic Modeling is considered to be one of the most important areas of ap- plication for Fuzzy Logic. Linguistic Mamdani-type Fuzzy Rule-Based Systems (FRBSs), the ones used to perform this task, provide a human-readable description of the model in the form of linguistic rules, which is a desirable characteristic in many problems. In this Chapter we are going to accomplish a short revision of the FRBSs where we shall see the different types that currently exist, along with their structures and characteristics, centering our attention on linguistic Mamdani-type FRBS. The performance of a linguistic FRBS depends on its Rule Base and the membership functions associated to the fuzzy partitions. Due to the complexity in the design of these components, a large quantity of automatic techniques has been proposed to put it into effect. Thereafter, we are going to review several learning (when it sets the Rule Base and sometimes the Data Base as well) and tuning (when it only sets the Data Base) methods. These methods are inspired in the three most well known approaches: ad hoc data covering, neural networks, and genetic algorithms. We shall introduce a brief description of these techniques and their synergy with FRBSs. The accuracy of the reviewed methods will be compared when solving two real-world ap- plications. Some interesting conclusions will be obtained about the behavior of the methods, approaches, and techniques. I Introduction At present, the most important application of Fuzzy Set Theory as developed by Zadeh in 1965 [1] are Fuzzy Rule-Based Systems (FRBSs). These kinds of systems constitute an extension of classical Rule-Based Systems, because they deal with fuzzy rules instead of classical logic rules. Thanks to this, they have been successfully applied to a wide range of problems from different areas presenting uncertainty and vagueness in different ways [2, 3, 4, 5]. An FRBS presents two main components: 1) the Inference System, which puts into effect the fuzzy inference process needed to obtain an output from the FRBS when an input is specified, and 2) the Knowledge Base (KB) representing the knowledge known about the problem being solved, constituted by a collection of fuzzy rules. * This research has been supported by CICYT TIC96-0778 1

Transcript of Techniques for Learning and Tuning Fuzzy Rule-Based...

Page 1: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Techniques for Learning and Tuning Fuzzy Rule-BasedSystems for Linguistic Modeling and their Application ∗

R. Alcala?, J. Casillas?, O. Cordon?, F. Herrera?, S. J. I. Zwir†

? Department of Computer Science and Artificial Intelligence.

E.T.S. de Ingenierıa Informatica, University of Granada. 18071 – Granada, Spain

e-mails: {alcala,casillas,ocordon,herrera}@decsai.ugr.es

† Department of Computer Science.

University of Buenos Aires, Buenos Aires, Argentina

e-mail: [email protected]

Abstract

Nowadays, Linguistic Modeling is considered to be one of the most important areas of ap-plication for Fuzzy Logic. Linguistic Mamdani-type Fuzzy Rule-Based Systems (FRBSs), theones used to perform this task, provide a human-readable description of the model in the formof linguistic rules, which is a desirable characteristic in many problems.

In this Chapter we are going to accomplish a short revision of the FRBSs where we shall seethe different types that currently exist, along with their structures and characteristics, centeringour attention on linguistic Mamdani-type FRBS. The performance of a linguistic FRBS dependson its Rule Base and the membership functions associated to the fuzzy partitions. Due to thecomplexity in the design of these components, a large quantity of automatic techniques has beenproposed to put it into effect.

Thereafter, we are going to review several learning (when it sets the Rule Base and sometimesthe Data Base as well) and tuning (when it only sets the Data Base) methods. These methodsare inspired in the three most well known approaches: ad hoc data covering, neural networks,and genetic algorithms. We shall introduce a brief description of these techniques and theirsynergy with FRBSs.

The accuracy of the reviewed methods will be compared when solving two real-world ap-plications. Some interesting conclusions will be obtained about the behavior of the methods,approaches, and techniques.

I Introduction

At present, the most important application of Fuzzy Set Theory as developed by Zadeh in 1965 [1]are Fuzzy Rule-Based Systems (FRBSs). These kinds of systems constitute an extension of classicalRule-Based Systems, because they deal with fuzzy rules instead of classical logic rules. Thanks tothis, they have been successfully applied to a wide range of problems from different areas presentinguncertainty and vagueness in different ways [2, 3, 4, 5].

An FRBS presents two main components: 1) the Inference System, which puts into effect thefuzzy inference process needed to obtain an output from the FRBS when an input is specified, and2) the Knowledge Base (KB) representing the knowledge known about the problem being solved,constituted by a collection of fuzzy rules.

∗This research has been supported by CICYT TIC96-0778

1

Page 2: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

There are two different kinds of FRBSs in the literature, the Mamdani and TSK ones, dependingon the expression of the consequent of the fuzzy rules composing the KB. While Mamdani-type fuzzyrules consider a linguistic variable in the consequent, TSK fuzzy rules are based on representingthe consequent as a polynomial function of the inputs.

Linguistic Modeling based on Fuzzy Logic is considered as a system model constituting a lin-guistic description, being put into effect by means of a linguistic Mamdani-type FRBS. Therebythe concept of linguistic variable plays a central role. A crucial reason why the linguistic fuzzy rule-based approach is worth considering is that it may remain verbally interpretable. These FRBSshave been widely used and have obtained very good results in many different applications.

The main of this Chapter is to present a short description of linguistic FRBSs and a reviewof different design techniques for them. Many different approaches have been presented takingdifferent learning algorithms as a base, such as, ad hoc methods, neural networks (NNs), geneticalgorithms (GAs), clustering, etc. We shall present the most well known approaches to learn andtune linguistic FRBSs, apply them to some examples, analyze their results, and give finally someguidelines for their application.

In this Chapter, we use the terms learning and tuning of FRBSs refering to the design of anFRBS by means of a learning process, deriving the Rule Base and sometimes defining the DataBase as well, or by a tuning process, which only defines the Data Base. Therefore, we shall use thefirst expressions to make easy the reading of this Chapter.

We arrange this Chapter as follows. Section II presents some preliminaries by briefly introducingthe different types and structures of FRBSs. In Sections III and IV, several linguistic FRBSslearning and tuning methods will be presented, covering the different types of techniques. Section Vshows two experiments developed to evaluate the behavior of the methods. Finally, in Section VI,some concluding remarks are pointed out.

II Fuzzy Rule-Based Systems

In a very wide sense, an FRBS is a Rule-Based System where Fuzzy Logic may be used as a tool forrepresenting different forms of knowledge about the problem being solved, as well as for modelingthe interactions and relationships existing between its variables. Thanks to this, they have beensuccessfully applied to a wide range of problems from different areas presenting uncertainty andvagueness in different ways [2, 3, 4, 5].

In this section, the basic aspects of FRBSs will be introduced: the different existing types,their composition and functioning will be described. Nevertheless, we shall not focus on the basicprinciples of Fuzzy Logic, that are to be found in texts like [6, 7].

II.A Framework: Fuzzy Logic and Fuzzy Systems

As it is known, Rule-Based Systems (production rule systems) have been successfully used tomodel human problem-solving activity and adaptive behavior, where a classical way to representthe human knowledge is the use of “IF-THEN” rules. The satisfaction of the rule antecedent givesrise to the execution of the consequent, i.e., one action is performed. Conventional approachesto knowledge representation are based on bivalent logic, which makes them have a serious short-coming associated: their inability to deal with the issue of uncertainty and imprecision. As aconsequence, conventional approaches do not provide an adequate model for modes of reasoningand all commonsense reasoning fall into this category.

Fuzzy Logic may be viewed as an extension of classical logical systems, providing an effectiveconceptual framework for dealing with the problem of knowledge representation in an environment

2

Page 3: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

of uncertainty and imprecision. Fuzzy Logic, as its name suggests, is the logic underlying modesof reasoning which are approximate rather than exact. The importance of Fuzzy Logic derivesfrom the fact that most modes of human reasoning —and especially commonsense reasoning— areapproximate in nature. Fuzzy Logic is concerned in the main with imprecision and ApproximateReasoning [8].

In a wide interpretation of Fuzzy Logic (in terms of which it is coextensive with Fuzzy SetTheory, that is, classes of objects in which the transition from membership to nonmembership isgradual rather than abrupt), Fuzzy Logic and fuzzy sets have placed modeling into a new andbroader perspective by providing innovative tools to cope with complex and ill-defined systems.The area of fuzzy sets has emerged following some pioneer works of Zadeh [1, 8] where the firstfundamentals of Fuzzy Systems were established.

From a very broad point of view, a Fuzzy System is any Fuzzy Logic-Based System, whereFuzzy Logic may be used either as the basis for the representation of different forms of knowledge,or to model the interactions and relationships among the system variables. Fuzzy Systems haveproven to be an important tool for modeling complex systems in which, due to the complexity orthe imprecision, classical tools are unsuccessful [2, 3, 4, 5].

Concretely, the application of Fuzzy Logic to Rule-Based Systems leads us to FRBSs. Thesesystems consider rules with the “IF-THEN” form whose antecedents and consequents are composedof Fuzzy Logic statements, thus presenting two essential advantages with respect to classical Rule-Based Systems:

• the key features of knowledge captured by fuzzy sets involve handling uncertainty, and

• inference methods become more robust and flexible with approximate reasoning methods ofFuzzy Logic.

Knowledge representation is enhanced with the use of linguistic variables and their linguisticvalues, that are defined by context-dependent fuzzy sets whose meanings are specified by gradedmembership functions [8]. On the other hand, inference methods such as generalized modus ponens,tollens, etc., which are based on Fuzzy Logic, form the basis of Approximate Reasoning with patternmatching scores of similarity [8]. Hence, Fuzzy Logic provides a unique computational frameworkfor inference in Rule-Based Systems. Unlike traditional logical systems, Fuzzy Logic is aimed atproviding modes of reasoning which are approximate and analog rather than exact.

II.B Types of Fuzzy Rule-Based Systems

Two different types of FRBSs are usually distinguished in the specialized literature according tothe form of the fuzzy rules considered and to the types of inputs and outputs used [9]. We shallintroduce them in the next subsections.

II.B.1 Mamdani Fuzzy Rule-Based Systems

This type of FRBS was proposed by Mamdani [10], who was able to translate Zadeh’s preliminaryassumptions to the first FRBS applied to a control problem. These kinds of Fuzzy Systems, theones most used since then, are also referred to as FRBSs with Fuzzifier and Defuzzifier or, morecommonly, as fuzzy logic controllers (name proposed by Mamdani in his pioneer papers [11]), dueto the fact that their main application has been system control historically.

Mamdani FRBSs are composed of four main components: a Knowledge Base (KB), an InferenceSystem and the Fuzzification and Defuzzification Interfaces. The KB —composed of Data Base (DB)

3

Page 4: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

and Rule Base (RB)— stores the available knowledge about the problem in the form of linguistic“IF-THEN” rules. The Inference System puts into effect the inference process on the system inputsmaking use of the information stored in the KB. The generic structure of Mamdani FRBSs is shownin Figure 1.

Figure 1: Basic structure of a Mamdani Fuzzy Rule-Based System

The Mamdani-type FRBS has several very interesting characteristics:

• On the one hand, it may be used in real-world applications due to the fact that it is able todeal with real-valued inputs and outputs.

• On the other hand, it provides a natural framework to include expert knowledge in theform of linguistic rules and allows us to combine, in a very simple way, the latter with rulesautomatically generated from data sets representing the behavior of the system.

• Finally, it has more freedom in the choice of the Fuzzification and Defuzzification Interfacecomponents, as well as the Inference System ones, thus allowing us to design a more suitableFRBS for a specific problem. We shall analyze these aspects in the next subsection.

As we shall see in Section II.C, where each of the components of this system kind will beanalyzed in deep, the Fuzzification Interface establishes a mapping between crisp values inthe input domain U of the system outputs and fuzzy sets defined in the same universe ofdiscourse. On the other hand, the Defuzzification Interface develops the opposite operationby defining a mapping between fuzzy sets defined in the output domain V and crisp valuesdefined in the same universe.

These systems present the maximum description level. The fuzzy rules are composed of inputand output linguistic variables taking values from a linguistic term set with a real-world meaning.Therefore, each rule is a description of a condition-action statement that may be clearly interpretedby human beings. This fact makes Mamdani FRBSs appropriate for Linguistic Modeling, subareaof Fuzzy Logic Modeling in which the main characteristic is the model interpretability [12], andSystem Control [13] problems.

There exists an extension of Mamdani FRBSs based on changing the linguistic rule structuremaking it more flexible. This new type of FRBS makes use of DNF (Disjunctive Normal Form)fuzzy rules with the following form [14, 15, 16, 17]:

IF X1 is ˜A1 and . . . and Xn is ˜An THEN Y is B ,

where each variable Xi has a referential set Ui and takes values in a finite domain (term set) Di,i = 1, . . . , n. The referential set for Y is V and its domain is F . The value of the variable Y is B,

4

Page 5: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

where B ∈ F and the value of the variable Xi is ˜Ai, where ˜Ai ∈ P (Di) and P (Di) denotes the setof subsets of Di. Thus, the complete syntax for the antecedent of the rule is

X1 is ˜A1 = {A11 or . . . or A1l1} and . . . and Xn is ˜An = {An1 or . . . or Anln} .

An example of this kind of rule is shown as follows. Let us suppose we have three input variables,X1, X2, and X3, and one output variable, Y , such that the linguistic term set associated with eachone is

D1 = {A11, A12, A13} , D2 = {A21, A22, A23, A24, A25} , D3 = {A31, A32} , F = {B1, B2, B3} .

In this case, a possible DNF rule may be

IF X1 is {A11 or A13} and X2 is {A23 or A25} and X3 is {A31 or A32} THEN Y is B2 .

Recently a new more flexible KB structure that allows us to improve the accuracy of LinguisticModels without losing their interpretability has been proposed [18]. Now, it is allowed the linguisticrule to have two different consequents associated. This will allow the FRBS to locally improve itsaccuracy in complex space zones, because the final rule output lie in an intermediate zone betweenthe both main labels. This transformation has a linguistic interpretation:

IF X1 is A1 and . . . and Xn is An THEN Y is between B1 and B2 .

On the other hand, in the last few years it has been proposed a new variant of these kindsof systems, in which the system accuracy is more preferable than its interpretability. These sys-tems are usually called approximate Mamdani-type FRBSs [2, 19, 20, 21, 22], the opposite to theones following the classical structure presented that are usually named as linguistic or descriptiveMamdani-type FRBSs. The main application of the former is the area of Fuzzy Modeling [2], wherethe model accuracy is the main requirement instead of its description ability.

The structure of an approximate FRBS is the same that the descriptive one, which was shown inFigure 1. The only difference is related to the type of rule considered in the KB and, consequently,to the composition of this component. In this case, the rules do not use linguistic variables but,directly, fuzzy variables. Therefore, the structure of the fuzzy rules considered in approximatesystems is the following:

IF X1 is A1 and . . . and Xn is An THEN Y is B ,

where Ai and B are fuzzy sets without a direct interpretation instead of linguistic labels.Hence, in these systems a DB is not found storing the existing linguistic terms and the fuzzy

sets associated defining their semantics. In this case, the KB used in FRBSs of the previous types,that was composed of the said DB and of the RB, becomes a single Fuzzy Rule Base formed by aset of rules presenting the latter structure shown, where each individual rule directly contains themeaning describing it. Figure 2 graphically shows this.

II.B.2 Takagi-Sugeno-Kang Fuzzy Rule-Based Systems

Instead of working with linguistic rules of the kind introduced in the previous subsection, Takagi,Sugeno, and Kang [23, 24] proposed a new model based on rules where the antecedent was composedof linguistic variables and the consequent was represented by a function of the input variables. The

5

Page 6: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Figure 2: Graphical comparison between a linguistic KB and an approximate Fuzzy Rule Base

most usual form of these kinds of rules is the one shown in the following, in which the consequentconstitutes a linear combination of the variables involved in the antecedent:

IF X1 is A1 and . . . and Xn is An THEN Y = p1 ·X1 + . . . + pn ·Xn + p0 ,

with Xi being the system input variables, Y being the output variable, and pi being real parameters.As regards Ai, they may be either linguistic labels with a meaning associated in the form of fuzzysets, or fuzzy sets, in the case in which they are directly fuzzy variables. These kinds of rules areusually called TSK fuzzy rules, in reference to their creators.

The output of a TSK FRBS using a KB composed of m rules is obtained as the weighted averageof the individual outputs provided by each rule, Yi, i = 1, . . . , m, as follows:

∑mi=1 hi · Yi∑m

i=1 Yi,

with hi = T (Ai1(x1), . . . , Ai

n(xn)) being the matching degree between the antecedent part of the i-thrule and the current inputs to the system, x = (x1, . . . , xn). T stands for a conjunctive operatormodeled by a t-norm.

Therefore, and as it was enunciated by its creators in [23], this FRBS is based on dividing theinput space in several fuzzy subspaces and defining a linear input-output relationship in each one ofthese subspaces. In the inference process, these partial relationships are combined in the said wayfor obtaining the global input-output relationship, taking into account the dominance of the partialrelationships in their respective areas of application and the conflict existing in the overlappedzones. A graphical representation of this second kind of FRBSs is shown in Figure 3.

The main advantage of these systems is the fact that they present a compact system equationthat allows us to estimate the parameters pi by means of classical methods, which makes its designeasier. Nevertheless, the main drawback relates to the form of the rule consequents as well, whichcauses the system not to constitute a natural framework for representing expert knowledge. Itis possible to integrate expert knowledge in these FRBSs by performing a small modification onthe rule consequent: when a linguistic rule with consequent Y is B is provided by an expert, this

6

Page 7: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Figure 3: Basic structure of a TSK FRBS

consequent is substituted by Y = p0, with p0 standing for a characteristic value of the fuzzy setassociated to the label B. These kinds of rules are usually called simplified TSK rules or zero orderTSK rules.

II.C Components of a Linguistic Mamdani-type Fuzzy Rule-Based System

Summarizing the concepts introduced previously, a Mamdani-type FRBS is composed of the fol-lowing components:

• A Knowledge Base, that contains the linguistic rules which guide the system behavior.

• A Fuzzification Interface, that takes care on transforming the crisp input data in values thatmay be handled in the fuzzy reasoning process, i.e., in fuzzy sets of any kind.

• An Inference System, that makes use of these values and of the information stored in the baseto put into effect the inference process.

• A Defuzzification Interface, that transforms the fuzzy action obtained from the inferenceprocess in a crisp action that constitutes the global output of the FRBS.

We shall analyze in deep each one of these components in the following subsections.

II.C.1 The Knowledge Base

As we have mentioned, the component that stores these rules is the KB that is composed of twodifferent components:

• RB: Formed by a set of linguistic “IF-THEN” rules that, in the case of multiple input-singleoutput FRBSs, present the following structure:

IF X1 is A1 and . . . and Xn is An THEN Y is B ,

with Xi and Y being input and output linguistic variables, respectively, and with Ai and Bbeing linguistic labels. The RB is comprised by a collection of rules of this kind joined by thealso operator, which means that, as we shall see in the following subsection, all of them maybe fired when a specific input is given to the system.

• DB: Containing the definitions of the fuzzy sets associated to the linguistic terms used in therules collected in the RB, that is, the membership function parameters.

7

Page 8: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

II.C.2 The Fuzzification Interface

The Fuzzification Interface is one of the two components that allow Mamdani-type FRBSs to dealwith real inputs and outputs. Its aim is to define a mapping that establishes a correspondencebetween each value in the crisp input space and a fuzzy set defined in the universe of discourseof this input, obtaining the membership function associated to each one of the system inputs.Symbolically, this component works as follows:

A′ = F (x0) ,

with x0 being a crisp input value for the FRBS defined in the universe of discourse U, A′ being afuzzy set defined in the same domain, and F being a fuzzification operator.

There are two main possibilities to choose the operator F :

1. Punctual fuzzification: A′ is built as a punctual fuzzy set (a singleton) with support x0, i.e.,it presents the following membership function:

A′(x) =

{

1, if x = x00, otherwise .

2. Non punctual or approximate fuzzification: In this case, A′(x0) = 1 and the membershipdegree of the remaining values decreases when moving away from x0. This second kind ofoperator allows us to deal with different types of membership functions. For example, in thecase of triangular-shaped fuzzy sets, it is possible to use the following one:

A′(x) =

{

1− |x−x0|σ , if |x− x0| ≤ σ

0, otherwise .

The former is the one most used due to its simplicity.

II.C.3 The Inference System

The Inference System is the component that puts into effect the fuzzy inference process. To doso, it makes use of Fuzzy Logic principles for establishing a mapping between fuzzy sets defined inU = U1 ×U2 × . . .×Un and fuzzy sets defined in V (with U1, . . . ,Un and V being the domainswhere the input variables X1, . . . , Xn, and the output one Y are defined, respectively).

The fuzzy inference process is based on the application of the Generalized Modus Ponens, anextension of the classical Modus Ponens, proposed by Zadeh in the way [8]:

IF X is A THEN Y is BX is A′

Y is B′ .

A fuzzy conditional statement with the form “IF X is A THEN Y is B” represents a fuzzyrelation between A and B defined in U × V. This fuzzy relation is expressed by a fuzzy set Rwhose membership function µR(x, y) is given by:

µR(x, y) = I(µA(x), µB(y)),∀x ∈ U, y ∈ V ,

with µA(x) and µB(y) being the membership functions of the fuzzy sets A and B, respectively, andI being a fuzzy implication operator modeling the existing fuzzy relation.

8

Page 9: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

The membership function of the fuzzy set B′ is obtained as a result of the application of theCompositional Rule of Inference (CRI) introduced by Zadeh in [8] as follows: “If R is a fuzzyrelation defined in U and V and A′ is a fuzzy set defined in U, then the fuzzy set B′, induced byA′, is obtained from the composition of R and A′”, that is:

B′ = A′ ◦R .

Therefore, when the CRI is applied on rules whose antecedent is formed by n input variablesand whose consequent has only one output variable, it presents the following expression:

µB′(y) = Supx∈U{

T ′(µA′(x), I(µA(x), µB(y)))}

,

where µA′(x) = T (µA′1(x), . . . , µA′n(x)), µA(x) = T (µA1(x), . . . , µAn(x)), with T and T ′ being fuzzy

conjunctive operators (t-norms) and with I being a fuzzy implication operator.Since, as we mentioned in the previous subsection, in the great majority of the cases the Fuzzi-

fication Interface transforms the input x0 = (x1, . . . , xn) received by the system in several punctualfuzzy sets A′1, . . . , A

′n, and due to the application of the properties T (1, 1) = 1 and T (x, 1) = x,

satisfied by t-norms [25, 26], the CRI is finally reduced to the form:

µB′(y) = I(µA(x0), µB(y)) .

II.C.4 The Defuzzification Interface

From the operation mode of the linguistic Mamdani-type FRBS Inference System introduced inthe previous subsection, it may be clearly drawn the fact that the fuzzy inference process is appliedat the level of individual rules. Thus, once the CRI has been applied on the m rules composing theKB, m fuzzy sets B′

i are obtained representing the fuzzy actions deduced by the FRBS from theinputs received.

Since the system must give a crisp output, the Defuzzification Interface has to develop the taskof aggregating the information provided by each one of the fuzzy sets and transform it in a singlecrisp value. There are two different operation modes to do this aggregation [27]:

1. Mode A: Aggregation first, defuzzification after: In this first case, the Defuzzification Interfaceperforms the following tasks:

• Aggregates the individual fuzzy sets inferred, B′i, for obtaining a final fuzzy set B′ by

means of a fuzzy aggregation operator, G, which models the also operator that relatesthe rules in the base:

µB′(y) = G{

µB′1(y), µB′2

(y), . . . , µB′n(y)}

.

• By using a defuzzification method, D, transforms the fuzzy set B′ obtained in a crispvalue, y0, that will be given as system global output:

y0 = D(µB′(y)) .

2. Mode B: Defuzzification first, aggregation after: In this second operation mode, the contri-bution of each fuzzy set inferred is individually considered by means of a characteristic value(center of gravity, mean of maximum value, etc.) and the final crisp value is obtained bymeans of an aggregation operator (an average, a weighted average, or the selection of one of

9

Page 10: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

them, among others) performed on a crisp characteristic value of each one of the individualfuzzy sets.

This operation mode constitutes a different approach of the concept represented by the alsooperator.

Historically, the first mode proposed was the Mode A, that was already used by Mamdani inhis first approach to Fuzzy Control [10]. In the last few years, the second approach is becomingmore used [27, 28].

III Learning of Linguistic FRBSs

As we have mentioned, two main tasks have to be performed to design an intelligent system ofthis kind for a specific application: to select the fuzzy operators involved in the Inference System—i.e., to define the way in which the fuzzy inference process will be performed—, and to derive anadequate KB about the problem being solved. The accuracy of the FRBS in solving this problemwill directly depend on both components.

The first design task has been widely analyzed in the specialized literature, and a big amountof theoretical and comparative studies have been carried out in order to deal with the problem ofthe selection of the best possible fuzzy operators in the Inference System [27, 29, 30].

As regards the second design task, it seems to be a more difficult decision because the compo-sition of the KB depends directly on the problem being solved. Due to the complexity of the KBderivation, a large quantity of automatic techniques has been proposed to put it into effect.

Many of these techniques are collected under the name of Soft Computing (SC) [31, 32]. SCis a new field of Computer Science that deals with the integration of problem-solving techniquessuch as Fuzzy Logic, NNs, or GAs. Each of these techniques provides us with complementaryreasoning and search methods to solve complex problems. Among all the possible combinations,we are interested on how NNs [33] or GAs [34] can help us to design an FRBS by defining the KB,acting as it is shown in Figure 4.

First of all, it is interesting to distinguish between tuning and learning problems. In tuningproblems, a predefined RB and a preliminary DB are used and the objective is to find a better setof parameters defining the DB. In learning problems, a process including the learning of the RB orof the whole KB is performed.

In this section we shall focus on the learning of linguistic FRBSs. As we have mentioned in theprevious section, Linguistic Models are characterized by offering a very appropriate interpretabil-ity that have repercussions on an expensive efficiency loss. Therefore, the methods for learningLinguistic Models are adopted in problems that require a high degree of interpretability.

We are going to introduce several learning algorithms for linguistic FRBSs. These are groupedin three approaches according to the three most well know techniques: ad hoc data covering ap-proaches, NNs, and GAs. The methods that we shall see are collected in Table 1.

III.A Ad Hoc Data Covering Methods

With this name, “Ad Hoc Data Covering Methods”, we are collecting those methods that are basedon processes in which the learning of the fuzzy rules is guided by covering criteria of the data in theexample set. Generally, they are characterized by being methods based on a short time-consumingiterative procedure.

10

Page 11: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Figure 4: Design of FRBSs

III.A.1 Wang and Mendel’s Learning Method

The ad hoc data covering RB generation process proposed by Wang and Mendel in [35] has beenwidely known because of its simplicity and good performance. It is based on working with an input-output data set representing the behavior of the problem being solved, using a previous definitionof the DB composed of the input and output primary fuzzy partitions.

The generation of the RB is put into effect by means of the following steps:

1. Consider a fuzzy partition of the input variable spaces: It may be obtained from the expertinformation (if it is available) or by a normalization process. If the latter is the case, performa fuzzy partition of the input variable spaces dividing each universe of discourse into a numberof equal or unequal partitions, select a kind of membership function and assign one fuzzy setto each subspace. In our case, we shall work with symmetrical fuzzy partitions of triangularmembership functions (see Figure 5).

Figure 5: Graphical representation of a uniform fuzzy partition

2. Generate a preliminary linguistic rule set: This set will be formed by the rule best covering

11

Page 12: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Table 1: Learning and Tuning Methods used

Reference Author(s) of the Method Technique used[35] Wang and Mendel Ad hoc Data Covering[36] Nozaki, Ishibuchi, and Tanaka Ad hoc Data Covering[37] Shann and Fu Neural Network[38] Jang (ANFIS) Neural Network[39] Thrift Genetic Algorithm[40] Liska and Melsheimer Genetic Algorithm[15] Gonzalez and Perez (SLAVE) Genetic Algorithm[20] Cordon and Herrera (D-MOGUL) Ad hoc DC + Genetic Algorithma

[18] Cordon and Herrera (WM-ALM) Ad hoc DC + Genetic Algorithma

[38] Jang (ANFIS for tuning) Neural Network[20] Cordon and Herrera (tuning) Genetic Algorithm

a They are composed of two stages: the first (Generation Process) is Ad Hoc and the second (Simplifi-cation Process) is GA-based.

each example (input-output data pair) contained in the input-output data set. The structureof these rules is obtained by taking a specific example, i.e., an n + 1-dimensional real array(n input and 1 output values), and setting each one of the variables to the linguistic label(associated fuzzy set) best covering every array component.

3. Give an importance degree to each rule: Let Rl = IF x1 is A1 and . . . and xn is An THEN yis C be the linguistic rule generated from the example el = (xl

1, . . . , xln, yl). The importance

degree associated to it will be obtained as follows:

G(Rl) = µA1(xl1) · . . . · µAn(xl

n) · µB(yl) .

4. Obtain a final RB from the preliminary fuzzy rule set: The rule with the highest importancedegree is chosen for each combination of antecedents.

III.A.2 Nozaki, Ishibuchi, and Tanaka’s Learning Method

Nozaki et al. try to obtain a Linguistic Model that, maintaining the level of interpretability, presentsa higher accuracy. To achieve it, they design a fuzzy model based on simplified TSK-type rulesto, thereinafter, transform it into a Linguistic Model (having linguistic terms in the consequent)with the same performance. As we have said, the method consists of two phases. In a first step, asimplified TSK KB is learnt. The rule consequents are composed of a single real value that can beconsidered as the independent term of the linear combination in a TSK representation (introducedin Section II.B.2). In the second phase, each TSK rule is associated to two descriptive Mamdani-type rules defined in the same input subspace. To do so, we shall have two RBs: a main and asecondary ones. The method works as follows:

1. First phase: Generation of TSK Rules. To determine the real consequent of each rule, aweight is defined for each training data array as the result of raising the membership functionof the input to the power of α (a parameter that defines a non-linear scaling function) and ofgetting the product of the membership function values of each input. The real consequent will

12

Page 13: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

be obtained as the weighted average of the known output value of each array of input values.If the weight is zero, the rule will not be considered. In this way, we can quickly obtain abase of TSK rules with a linguistic antecedent and whose consequent is a real number.

2. Second phase: Derivation of Mamdani rules from TSK rules. To obtain better results takingthe advantage offered by the TSK representation, the authors use two Mamdani-type RBs:the main and secondary ones. In the first place it will be necessary to accomplish a fuzzypartition of the universe of discourse of the consequent in some linguistic labels. They consideruniformly distributed partitions with triangular membership functions (see Figure 5).

For each TSK rule, the greatest value of the membership function applied to the consequentis searched. The real value (x0) will be substituted for the corresponding label obtaining aMamdani rule. This rule will be inserted in the main RB. The corresponding label to thesecond best value will form the rule that is included in the secondary RB. Figure 6 graphicallyclarifies the operation mode.

Figure 6: Derivation of Mamdani rules from TSK rules in the Nozaki et al.’s Method

To draw profit from the two RBs, the authors present an Inference System capable of obtaininga combined result from them. To do so, a real value between 0 and 1 that acts as a certaintyfactor is associated to each rule. This value is the result of evaluating the TSK consequent on thecorresponding membership functions, according to the RB (in the drawing, it would be h3 and h2).

III.B Neural Learning FRBS

A short introduction to Neural Networks and Neuro-Fuzzy Systems is presented in Appendix A. Inthe next subsections we introduce two specific methods with different characteristics: the first onelearns fuzzy rules through the weights setting while the second one learns the membership functionshapes.

III.B.1 Shann and Fu’s Learning Method

The method proposed by Shann and Fu [37] use a layer-structured Neuro-Fuzzy System (NFS)for learning fuzzy rules by considering weights. The learning is produced adjusting the certaintyfactors of each rule.

The method is a two-phase learning procedure. The first phase is an error-backpropagation(EBP) training, and the second one is a rule-pruning algorithm.

1. First phase: EBP training. The NFS is composed of five layers (as shown in Figure 7).

13

Page 14: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Figure 7: Neural Network Structure proposed by Shann and Fu

• Layer 1 (input layer): Each node in this layer represents an input linguistic variable ofthe network and it is used as a buffer to broadcast the input to the next layer. Thedomain of each input linguistic variable is determined by the application and it does notneed to be constrained within [0,1].

• Layer 2 (membership-function layer): Each node in this layer represents the membershipfunction of a linguistic value associated to an input linguistic variable. Therefore, amembership function node is a fuzzifier. We can use any shape of membership functionsuch as the triangle, trapezoid or bell-shaped ones. The function of the nodes in thislayer are determined and formulated by applications.

• Layer 3 (AND layer): Each node in layer 3 represents a possible IF-part for the fuzzyrules. In fuzzy set theory, there are many different operators for fuzzy intersection[6, 7, 27, 29, 30]. The authors choose the most commonly used one, i.e., the min-operator, which is simple and effective and has strong characteristics of competition.

• Layer 4 (OR layer): Each node in layer 4 represents a possible THEN-part for the fuzzyrules. The operation performed by an OR node is to combine fuzzy rules with the sameconsequent. In this first phase, the links between layers 3 and 4 are fully connected sothat all the possible fuzzy rules are embedded in the structure of the network. As in theprevious layer, the authors choose the most commonly used one, i.e., the max-operatorsuggested by Zadeh [1], as the function of an OR node.The links of this layer have weights associated. For example, the weight wkj of an inputlink in layer 4 represents the certainty factor of a fuzzy rule, which comprises the ANDnode j in layer 3 as the IF-part and the OR node k in layer 4 as the THEN-part. Hence,these weights are adjustable while learning the knowledge of fuzzy rules and they haveto be nonnegative real numbers.

14

Page 15: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

• Layer 5 (defuzzification layer): Each node in this layer represents an output linguisticvariable and performs defuzzification, taking into account the effects of all membershipfunctions of the output linguistic values. Suppose that the correlation-product inferenceand the fuzzy centroid defuzzification scheme [41] are used. Then, the function of nodel in layer 5 is defined as follows:

zl =∑

k∈Pl(xlkalkclk)

k∈Pl(xlkalk)

,

with Pl being the set of indices of the nodes in layer 4 that have an output link connectedto node l, xlk = zk, and with alk and clk being the area and centroid of the membershipfunction of the output linguistic value represented by node k in layer 4, respectively.Since it is assumed that the membership functions of the output linguistic values areknown, the areas and centroids can be calculated before learning.

2. Second phase: Pruning. The RB directly obtained right after the training phase will containall the rules with nonzero weight, i.e., certainty factor. This causes the size of the RB to bevery large in most cases. With this second phase, a concise RB is obtained. At the meantime,the structure of the NFS will be reduced as well. The technique involves calculating thecentroids among the rules with the same antecedent. The linguistic variable where eachcentroid lies will determine the rule selected for each set of rules.

III.B.2 Jang’s Learning Method (ANFIS)

In [38], Jang presents a new algorithm called ANFIS (Adaptive-Network-Based Fuzzy InferenceSystem) that defines the composition of the DB. The inference system is implemented in an NNthat uses an hybrid method to adjust the parameters in its nodes, a gradient method and the leastsquares estimate (LSE) to identify the parameters.

The proposed ANFIS can construct an input-output mapping based on both expert knowledge(in the form of linguistic rules) and specified input-output data pairs.

In spite of the fact that the author talks about several types of inference systems, he focuses ona TSK system to describe ANFIS, but he notes that it is possible to implement a Mamdani systemas well. In the following, it will be first explained the TSK system and then the modifications totransform to Mamdani.

An adaptive network is a multilayer feedforward network in which each node performs a partic-ular function (node function) on the incoming signals as well as a set of parameters pertaining tothis node of which its output depends. These parameters can be fixed or variable, and it is throughthe change of the last ones that the network is tuned. ANFIS has nodes with variable parameters,called square nodes, which will represent the membership functions of the antecedents, and themembership function for the Mamdani-type consequent or the linear functions for the TSK-typeconsequent. The nodes in the intermediate layers connect the antecedents with the consequent.Their parameters are fixed and they are called circular nodes.

The node functions in the same layer belong the same function family as described below:Layer 1: Parameters in this layer are referred to as premise parameters. In fact, any differen-

tiable function, such as bell and triangular-shaped membership functions are valid for the nodesin this layer. Every node i in this layer is a square node with a node function. From now on, weconsider that bell-shaped membership functions are used:

O1i = µAi(x) , µAi(x) =

11 + [(x−ci

ai)2]bi

.

15

Page 16: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Bell-shaped membership functions are used.Layer 2: Each node represents the firing strength of a rule ωi through a conjunction operator.

The function considered is the minimum t-norm. They are circular nodes with a node function:

ωi =

{

µAi(x), if µAi(x) ≤ µBi(x)

µBi(x), if µAi(x) > µBi(x), i = 1, 2 .

Layer 3: It calculates the ratio of ith rule’s firing strength with respect to the sum of all therule’s firing strengths.

ωi =ωi

ω1 + ω2, i = 1, 2

Layer 4: Every node i in this layer is a square node with a node function:

O4i = ωifi = ωi (pix + qiy + ri) .

Layer 5: The single node in this layer computes the overall output as the sum of all incomingsignals.

O5i =

i

ωifi =∑

i ωifi∑

i ωi

The initial values of the antecedent and consequent parameters are set through a partition ofthe input universe of each variable.

After the network parameters are set with their initial values, the consequent is adjusted throughthe LSE [42]. This method is used because the structure of the TSK consequent, a linear combina-tion of the inputs, makes it suitable to learn the weights composing the consequent. On the otherhand, this is preferred to the strict gradient descent due to the former is quicker than the latter.

In order to achieve a desired input-output mapping, the network parameters are updated ac-cording to the given training data and the gradient descent method [43], that propagates the errorrates from the output end towards the input end as a linear combination of the error in the nodesof the following layer which is multiplied by the derivative from the function of these nodes withrespect to the output of this. By using this measure, it will be able to calculate the error thatis produced by each parameter and to correct it. Let us suppose an adaptive network with Llayers and #(k) nodes in layer k, we can denote by (k, i) the node at position i in layer k, and itsmembership function by

Oki = Ok

i (Ok−1i , . . . , Ok−1

#(k−1), a, b, c, . . .) ,

with a, b, c, . . . being the parameters pertaining to this node.Assuming that the given training data set has P entries, taking Tm,p as the mth component of

the pth target output vector and Olm,p as the mth component of the actual output vector producedby the presentation of the pth input vector, the error measure for the pth entry of training dataentry will be:

Ep =#(L)∑

m=1

(Tm,p −OLm,p)

2 =⇒ E =P

p=1

Ep .

The error rate for the output node at (L, i) can be calculated as

∂Ep

∂OLi,p

= −2(Ti,p −OLi,p) ,

16

Page 17: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

and for the internal node (k, i), the error rate can be derived by the chain rule as

∂Ep

∂OLi,p

=#(k+1)

m=1

∂Ep

∂Ok+1m,p

∂Ok+1m,p

∂Oki,p

.

Therefore, by means of these two expressions we can find ∂Ep

∂Oki,p

for each node in the network.

Then, the derivative of the overall error measure E with respect to α is

∂E∂α

=P

p=1

∂Ep

∂α.

Thus, the update formula for the generic parameter α is

4α = −η∂E∂α

,

in which η is a learning rate which can be further expresed as

η =s

α (∂E∂α )2

,

with s being the step size, the length of each gradient transition in the parameter space. Usually,we can change the value of s to vary the speed of convergence.

Some modifications have been made over the original system proposed by Jang in order totransform it in a Mamdani system and to consider triangular-shaped membership functions. Thesechanges let the method vary the number of rules and tune the rules learnt previously.

In order to construct a Mamdani system, the expression of node function in the layer 4 mustbe changed: Every node in this layer calculates the rule consequent multiplied by the output of itscorresponding node in layer 3. The consequent of the rule is obtained through the center of gravityweighted by the matching degree,

O4i = ωifi = ωi (

V y · µCi(y) · dy∫

V µCi(y) · dy) ,

whith Ci being the consequent membership function for the i rule.The network architecture for a Mamdani system in ANFIS is shown in Figure 8.It will not be permitted the use of the LSE technique in a Mamdani system with linguistic

variables in the consequent. In Mamdani systems, the complete network will be adjusted by thegradient method.

To provide well formed triangular-shaped membership functions using the gradient method,each triangle must be represented by three parameters, its center, the square root of the distanceto its left extreme, and the square root of the distance to its extreme right, allowing any type oftriangle to be obtained. Thus, if the updated parameter has a negative value the triangle will stillbe correct.

Fuzzy partitions considering triangular-shaped membership functions do not necessarily haveto satisfy the epsilon-completeness, which means that given a value x of one of the inputs in theuniverse of discourse, a linguistic label A can always be found so that µA(x) ≥ ε, being

µAi(x) =

x−ab−a , if a ≤ x ≤ bc−xc−b , if b > x ≥ c

0, if x < a or x > c .

17

Page 18: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Figure 8: Structure of a Mamdani-type ANFIS, Jang’s Method

Using the gradient method, the membership functions could fall apart from the domain causingthe corresponding rules to be eliminated. If too many rules are eliminated, the system will have awrong behavior. The solution is to insert in the universe of discourse, according to a probabilitydefined beforehand, some of those membership functions, so that they can be tuned again. This isdone for the antecedents as well as for the consequent.

III.C Genetic Learning FRBS

A short introduction to GAs and Genetic Fuzzy Systems is presented in Appendix B. In the followingwe shall review several methods based on GAs under the Pittsburgh and IRL approaches.

III.C.1 Thrift’s Learning Method

This method is based on encoding all the cells of the complete decision table in the chromosomes.In this way, Thrift [39] establishes a mapping between the label set associated to the system outputvariable and an ordered integer set (containing one more element and taking 0 as its first element)representing the allele set. An example is shown to clarify the concept. Let {NB, NS, ZR, PS, PB}be the term set associated to the output variable, and let us note the absence of value for the outputvariable by the symbol “–”. The complete set formed joining this symbol to the term set is mappedinto the set {0, 1, 2, 3, 4, 5}. Hence the label NB is associated with the value 0, NS with 1, . . . ,PB with 4 and the blank symbol “–” with 5.

Therefore, the GA employs an integer coding. Each one of the chromosomes is constituted byjoining the partial coding associated to each one of the linguistic labels contained in the decisiontable cells. A gene presenting the alelle “–” will represent the absence of the fuzzy rule containedin the corresponding cell in the RB.

The GA proposed employs an elitist selection scheme and the genetic operators used are ofdifferent nature. While the crossover operator is the standard two-point crossover, the mutationoperator is specifically designed for the process. When it is applied over an allele different fromthe blank symbol, changes it one level either up or down or to the blank code. When the previousgene value is the blank symbol, it selects a new value at random.

Finally, the fitness function is based on an application specific measure. The fitness of anindividual is determined by applying the FRBS considering the RB coded in its genotype to thecontrolled system with several different starting points and computing the convergence to thedesired equilibrium point.

18

Page 19: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

III.C.2 Liska and Melsheimer’s Learning Method

Liska and Melsheimer [40] present a general method that covers all the areas of KB learning:number of fuzzy rules, structure of the rules, and membership function parameters. The powerof their method lies on learning optimizing all the three parts simultaneously, against doing itseparately, which could result in a suboptimal solution from the authors’ point of view.

The KB is represented as a chromosome composed of three substrings:

• The first substring of real numbers encodes the membership functions associated to all systemvariables. Each membership function is represented by two parameters —center and width—that allow to use bell-shaped or isosceles triangular-shaped membership functions.

• The second substring of integer numbers encodes the structure of each rule in the RB so thatone integer number represents one membership function in the space of an input variable.Membership functions are numbered in ascending order according to their centers, i.e., anumber ‘1’ refers to the membership function with the lowest value of the membership functioncenter in a particular input variable. The value ‘0’ in the second substring indicates the “null”membership function, i.e., the input variable is not involved in the rule.

• The third substring of integer numbers encodes membership functions in the rule consequents.The integer numbers refer to membership functions in the same way as in the second substring.A value ‘0’ in the third substring means that the rule is deleted from the FRBS RB.

The inclusion of ‘0’ in the second and third substrings allows both the number of input variablesinvolved in each rule and the number of rules to change dynamically during the GA search. Thenumber of rules is constrained by an upper limit specified by the designer.

As regards the genetic operators (crossover and mutation), they act in a different way dependingon what substring of the chromosome they are applied on.

On the one hand, if we are working with the first substring it is necesary that the operatoracts in the way that the relative order of membership functions in each variable is preserved, e.g.,a center of the membership function with a linguistic meaning “low” can take any value in thevariable range but always lower than a center of the membership function “high”.

On the other hand, with the second and the third substring we shall use three specific operatorsthat are not applied at the level of genes but of fields in the second substring. In this case, a fieldis considered to be the antecedents of a rule.

• Uniform crossover creates two offspring from two parents by deciding randomly, field by field,which offspring receives the field from which parent.

• Mutation randomly replaces selected fields in a parent by a random value between the mini-mum and maximum allowed values.

• Creep creates one offspring from one parent by randomly altering its field within a specifiedrange.

The performance of each chromosome in a population is evaluated by an exponential rankingtechnique based on the error function. This technique assigns the highest fitness fitness(1) = 1000to the chromosome with the lowest value of the error function, the chromosome with the nextlowest value of the error function receives fitness(2) = α ∗ fitness(1), α ∈ [0, 1], or generally,fitness(N + 1) = α ∗ fitness(N) with the exception that no string was given a fitness lesser than

19

Page 20: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

1. The higher the fitness a string receives, the higher the probability it has to be selected forreproduction. Liska and Melsheimer obtained the best results when α was set constant to 0.96. Intheir GA implementation they also do not allow duplicates, i.e., a new offspring is introduced intothe current population only if it differs from every member of the current population at least inone field.

Once the genetic learning process has finished, the method applies a second phase of tuning ofmembership functions for ensuring the optimum adjustment. The objective is to avoid the stagna-tion of the GA. This fine-tuning is developed with a multilayer feedforward network [40]. In contrastto the feedforward NNs, the tuned parameters in this realization are the centers and widths of mem-bership functions in the fuzzification and defuzzification interfaces, and all the interconnections areassigned a constant weight 1.

III.C.3 Gonzalez and Perez’s Learning Method (SLAVE)

SLAVE (Structural Learning Algorithms in Vague Environments) is a inductive learning algorithm,that was initially proposed in [14] and later developed in [15, 44]. The basic element of the SLAVElearning algorithm is its model of rule

IF X1 is ˜A1 and . . . and Xn is ˜An THEN Y is B ,

where each variable Xi has a referential set Ui and takes values in a finite domain (term set) Di,i = 1, . . . , n. The referential set for Y is V and its domain is F . The value of the variable Y isB, where B ∈ F and the value of the variable Xi is ˜Ai, where ˜Ai ∈ P (Di) and P (Di) denotes theset of subsets of Di. As we saw in Section II.B.1, this type of FRBS are called DNF (DisjunctiveNormal Form). It has been used by other authors [16, 17] as well.

The key to this rule model is that each variable can take as a value an element or a subset ofelements from its domain. The concept may be clarified with the following example:

Let Xi be a variable whose domain is shown in Figure 5(a). An antecedent like

. . . and Xi = {ZR, PS, PM} and . . . ,

is equivalent to. . . and {Xi is ZR or Xi is PS or Xi is PM} and . . .

Using the previous model of rule, the set of all possible rules is

∆ = P (D1)× P (D2)× . . .× P (Dn)× F .

Another important characteristic of this learning algorithm is that it uses the genetic IRLapproach for generating the rule set (see Genetic Fuzzy Systems in Appendix B). The iterativeapproach is presented as an alternative to the well-known Michigan and Pittsburgh approachesfor genetic learning. This approach involves including a GA in a iterative scheme similar to thefollowing [45]:

1. Use a GA to obtain a rule for the system.

2. Incorporate the rule into the final set of rules.

3. Penalize this rule.

4. If the set of rules obtained is sufficient to represent the examples in the training set, thesystem will return this set of rules as the solution. Otherwise return to step 1.

20

Page 21: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

The GA obtains a rule that is a partial solution of the learning problem in each step. The truesolution is obtained by appending each rule to the rule set. In the recent literature we may finddifferent algorithms that use this new approach, such as [14, 20, 46]. A study about the problemsgenerated by the use of the IRL approaches can be seen in [45].

The SLAVE parameters are described in [15]. For measuring the goodness of the rule are intro-duced two definitions that are based on the classical consistency and completeness conditions. Theseconditions provide the logical foundation of the algorithms for concept learning from examples.

Definition 1. The completeness condition states that every example in some class must verify somerule from this class.

Definition 2. The consistency condition states that if an example satisfies a description of some class,then it cannot be a member of a training set of any other class.

These definitions are associated to the whole set of rules, but SLAVE obtains the set of rulesthat describes the system by extracting one rule in each iteration of the learning process. For thisreason, the authors need to define these concepts on each rule. Moreover, they are not interestedin proposing hard definitions on fuzzy problems, thus, they propose a degree of completeness anda degree of consistency, both definitions using the concepts proposed in [15].

Definition 3. The degree of completeness of a rule RB(A) is defined as

Λ(RB(A)) =n+(RB(A))

nB,

where nB is the number of examples of the B class and n+(RB(A)) is the number of positive examplescovered by RB(A).

The soft consistency degree is based on the possibility of admitting some noise in the rules.Thus, in order to define the soft consistency degree the authors use the following set:

∆k = {RB(A)/n−(RB(A)) < k n+(RB(A))} ,

that represents the set of rules having a number of negative examples (n−(RB(A))) strictly lesserthan a percentage (depending on k) of the positive examples (n+(RB(A))).

Definition 4. The degree in which a rule R satisfies the soft consistency condition is

Γk1k2(R) =

1 if R ∈ ∆k1

k2n+(R)−n−(R)n+(R)(k2−k1) if R /∈ ∆k1 y R ∈ ∆k2

0 otherwise

where k1, k2 ∈ [0, 1] and k1 < k2, and n−(R), n+(R) are the number of positive and negative examplesto the rule R.

This definition uses two parameters, k1 is a lower bound of the noisy threshold and k2 is anupper bound of the noisy threshold.

These definitions are based on the use of the cardinality of two fuzzy sets (the positive andnegative example sets). We can say that an example is positive for a rule when it matches theantecedent and consequent of the rule. On the contrary, we consider it as a negative example whenit matches the antecedent but not the consequent. In any other case, we can say that it has noinfluence on this rule.

21

Page 22: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

SLAVE selects the rule that simultaneously verifies the completeness and soft consistency con-ditions to a high degree. Therefore, the rule selection in SLAVE can be accomplished by solvingthe following optimization problem:

maxA∈D

{ Λ(RB(A)) Γk1k2(RB(A))} ,

where D = P (D1)× P (D2)× . . .× P (Dn).Figure 9 shows the different elements of the SLAVE learning algorithm and the relations among

them.

Figure 9: Description of SLAVE Learning Process

III.C.4 Cordon and Herrera’s Learning Method (D-MOGUL)

In this subsection we are going to introduce the Descriptive-MOGUL approach [20] based in theMOGUL paradigm presented in [47]. Making use of this method it will be possible to automaticallygenerate a complete KB when a example training set is available. It consists of the following threesteps:

1. An iterative RB generation process of desirable fuzzy rules able to include the completeknowledge of the set of examples.

2. A genetic simplification process, which finds the final RB able to approximate the input-outputbehavior of the real system. It is based on eliminating some unnecessary rules from the ruleset obtained in the previous stage, avoiding thus the possible overlearning, by selecting thesubset of rules best cooperating.

3. A genetic tuning process of the DB used in order to improve as far as possible the accuracyof the final KB.

22

Page 23: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

The Rule Base Generation Process

The first stage consists of an increasing RB generation process based on an iterative explorationof the problem search space. Apart from the training data set Ep, a previously defined DB,constituted by uniform fuzzy partitions with triangular membership functions crossing at height0.5, is considered. The FRBS designer can specify the number of linguistic terms forming each oneof them in order to obtain the desired granularity level. Figure 5 shows the generic structure of afuzzy partition with seven linguistic labels.

This component allows us to obtain a set of Mamdani-type fuzzy rules Bg describing the systembehavior. In order to do that, it is necessary to establish a condition for it. This is the requirementof covering all possible situation-action pairs, el ∈ Ep, the completeness property [15, 48]. Thismay be formalized for a constant τ ∈ [0, 1], it requires the non-zero union of fuzzy sets Ai(·), Bi(·),i = 1, ..., T , T = |Bg|, and is formulated by the following expressions:

CR(el) =⋃

i=1..T

Ri(el) ≥ τ , l = 1, ..., p

Ri : If x1 is Ai1 and ... and xn is Ain then y is B

el = (exl1, . . . , ex

ln, eyl)

Ri(el) = ∗(Ai(exl), Bi(eyl))

Ai(exl) = ∗(Ai1(exl1), ..., Ain(exl

n)) ;

where ∗ is a t-norm, and Ri(el) is the compatibility degree between the rule Ri and the example el.Given a set of rules R, the covering value of an example el is defined as

CVR(el) =T

i=1

Ri(el) ,

and we require the following condition

CVR(el) ≥ ε , l = 1, ..., p .

A good set of rules must satisfy both the conditions presented above, to verify the completenessproperty and to have an adequate final covering value.

The RB is derived rule by rule, selecting the most accurate one at each step in the algorithm.Once this rule is obtained, its covering over the training set examples is taken into account. Thoseexamples covered in a degree higher than a value ε specified by the FRBS designer are removedfrom the training set. Hence, the increasing example covering guides the search to other promisingspace zones at each step.

Each time the best rule has to be selected in the generation process, the accuracy of thecandidates is measured by using a multicriteria fitness function. This function is designed takinginto account the following three criteria allowing us to ensure the completeness and consistency ofthe final KB generated:

23

Page 24: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

a) High frequency value [48]

The frequency of a fuzzy rule, Ri, through the set of examples, Ep, is defined as:

ΨEp(Ri) =∑p

l=1 Ri(el)p

.

b) High average covering degree over positive examples [48]

The set of positive examples to Ri with compatibility degree greater than or equal to ω isdefined as:

E+ω (Ri) = {el ∈ Ep/Ri(el) ≥ ω} .

The average covering degree on E+ω (Ri) can be defined as:

Gω(Ri) =∑

el∈E+ω (Ri)

Ri(el)/n+ω (Ri) .

where n+ω (Ri) = |E+

ω (Ri)|.

c) Small negative examples set [15]

The set of the negative examples to Ri is defined as:

E−(Ri) = {el ∈ Ep/Ri(el) = 0 and Ai(exl) > 0} .

An example is considered negative for a rule when it better matches some another rule that hasthe same antecedent but a different consequent. The negative examples are always consideredover the complete training set.

With n−Ri= |E−(Ri)| being the number of negative examples, the penalty function on the

negative examples set will be:

gn(Ri−) =

1 if n−Ri≤ k · n+

ω (Ri)1

n−Ri−kn+

ω (Ri)+exp(1)otherwise ,

where we permit up to a percentage of the number of positive examples, k ·n+ω (Ri), of negative

examples per rule without any penalty. This percentage is determined by the parameterk ∈ [0, 1].

Therefore these three criteria are combined into a fitness function using any aggregation functionincreasing in the three variables. The authors propose the product:

F (Ri) = ΨEp(Ri) ·Gω(Ri) · gn(Ri−) .

Rules getting a higher value in this function will be more accurate.Taking the previous comments into account, the generation method is developed in the following

steps:

1. Initialization:

(a) To introduce the k, ω, and ε parameter values.

24

Page 25: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

(b) To set the example covering degree CV [l] ← 0, l = 1, ..., p.

(c) To initialize the final Rule Base Bg to empty.

2. To initialize the candidate fuzzy rule set Bc to empty.

3. For every el ∈ Ep, generate the fuzzy rule Rc best covering it by taking the linguistic labelmatching best with the el component value for each variable. If Rc 6∈ Bc, add it to Bc.

4. To evaluate all the fuzzy rules contained in Bc and to select the one getting a higher value inthe fitness function, Rr.

5. To introduce Rr into the set of rules Bg.

6. For every el ∈ Ep do

(a) CV [l] ← CV [l] + Rr(el).

(b) If CV [l] ≥ ε then remove el from Ep.

7. If Ep = ∅ then Stop else return to Step 2.

The Genetic Simplification Process

Due to the iterative nature of the generation process, an overlearning phenomenon may appear.This occurs when some examples are covered at a higher degree than the desired one and it makesthe RB obtained perform worse. In order to solve this problem and improve its accuracy, it isnecessary to simplify the rule set obtained from the previous process, removing the redundant rulesand selecting the rule subset with best cooperation for deriving the final RB solving the problem.

The simplification process used was proposed in [48]. It is based on a binary coded GA, inwhich the selection of the individuals is developed using the stochastic universal sampling procedureproposed by Baker in [49] together with an elitist selection scheme, and the recombination is putinto effect by using the classical binary multipoint crossover (performed at two points) and uniformmutation operators.

The coding scheme generates fixed-length chromosomes. Considering the rules contained inthe rule set derived from the previous step counted from 1 to m, an m-bit string C = (c1, ..., cm)represents a subset of candidate rules to form the RB finally obtained as this stage output, Bs,such that,

If ci = 1 then Ri ∈ Bs else Ri 6∈ Bs .

The initial population is generated by introducing a chromosome representing the completepreviously obtained rule set Bg, that is, with every ci = 1. The remaining chromosomes areselected at random.

As regards to the fitness function, E(·) it is based on an application specific measure usuallyemployed in the design of FRBSs, the mean square error over a training data set, ETDS , which isrepresented by the following expression:

E(Cj) =1

2|ETDS |∑

el∈ETDS

(eyl − S(exl))2 ,

whith S(exl) being the output value obtained from the FRBS using the RB coded in Cj , R(Cj)when the input variable values are exl, and eyl is the known desired value.

25

Page 26: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Anyway, there is a need to keep the control rule completeness property considered in the previousstage. An FRBS must always be able to infer a proper output for every system input. We shallensure this condition by forcing every example contained in the training set to be covered by theencoded RB in a degree greater than or equal to τ ,

CR(Cj)(el) =⋃

j=1..T

Rj(el) ≥ τ , ∀el ∈ ETDS and Rj ∈ R(Cj) ,

where τ is the minimal training set completeness degree acepted in the simplification process.Usually, τ is less than or equal to ω, the compatibility degree used in the generation process.

Therefore, we define a training set completeness degree of R(Cj) over the set of examples ETDSas

TSCD(R(Cj), ETDS) =⋂

el∈ETDS

CR(Cj)(el) ,

and the final fitness function penalizing the lack of the completeness property is:

F (Cj) =

{

E(Cj) if TSCD(R(Cj), ETDS) ≥ τ12

el∈ETDS(eyl)2 otherwise

.

The Genetic Tuning Process

Finally, a modified version of the genetic tuning method presented in [50] is applied by theauthors. Since we are going to use this third phase as a independent algorithm to adjust the KBpreviously learned with some of the reviewed methods, it will be presented in Section IV.B.

III.C.5 Cordon and Herrera’s Learning Method (WM-ALM)

In [18], the authors propose a methodology to design Linguistic Models with high accuracy and agood description level, called Accurate Linguistic Modeling (ALM). In the same paper, two differentlearning methods based on the paradigm are proposed. The WM-ALM method is composed of twostages, which will be described in the following:

1. A linguistic rule generation method from examples based on a modification made on a ad hocdata covering fuzzy rule generation processes, the Wang and Mendel’s method [35] (showed inSection III.A.1). The modification involves generating the two most important consequentsfor each combination of antecedents (instead of only the most important one, as this kindof methods usually do). The authors remark that an input subspace will only have twoconsequents associated (and thus, two linguistic rules) when there is a need to do so, i.e.,when there is any data in it or when two different consequents may be generated.

2. A rule selection genetic process, that removes the redundant or unnecesary rules from thefuzzy rule set generated in the previous step to select the subset of them cooperating best.This process will be implemented by means of a binary-coded Genetic Algorithm.

As mentioned, the Wang and Mendel’s method is modified to allow to have (if is necessary)two consequent for a specific combination of antecedents. The fourth step (see Section III.A.1)is the only one that is different from the original Wang and Mendel’s algorithm. Whilst in thatmethod the rule with the highest importance degree is the only one chosen for each combination ofantecedents, now two different rules, the two most important ones in each input subspace (if theyexist), are allowed to form part of the RB.

26

Page 27: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Of course, a combination of antecedents may not have rules associated (if there are no examplesin that input subspace) or only one rule (if all the examples in that subspace generated the samerule). Therefore, the generation of rules with double consequent is only addressed when the problemcomplexity, represented by the example set, shows that it is necessary.

The operation mode of the generation method means that, in each input subspace, the rules arecreated individually from the examples in the input-output data set without taking into accountthe cooperation existing among them to give the final model output. That is, no information aboutthe neighbor rules is considered in order to generate them.

Because of this, the generated RB may present redundant or unnecessary rules making themodel using this KB less accurate. In order to avoid this fact and to achieve that more thana single rule is used only in those zones where it is really necessary, the authors use a geneticrule selection process with the aim of simplifying the initial linguistic rule set by removing theunnecessary rules from it and generating a KB with good cooperation.

The selection of the subset of linguistic rules best cooperating is a combinatorial optimizationproblem [12, 51]. Since the number of variables involved in it, i.e., the number of preliminary rules,may be very large, they consider an approximate algorithm to solve it, a GA.

This selection algorithm was presented in the second subsection of the Section III.C.4 as asimplification process.

IV Tuning of Linguistic FRBSs

The performance of an FRBS depends on its RB and on the membership functions of associated tothe fuzzy partition, i.e., the DB. Hence, it is very important to tune these parameters to the processto be modeled. The tuning methods fit the membership functions of the fuzzy rules obtaining high-performance membership functions by minimizing an error function defined by means of the systembehavior or the evaluation of a training example set.

Recent work has centered on the use of mathematical and heuristic optimization techniquesas gradient descent, GAs, etc. In the following, we shall present two approaches for tuning theparameters of the membership functions, the first one that represents the KB by means of an NN,being tuned by the EBP, and the second one that uses GAs.

IV.A Jang’s Tuning Method

The method proposed by Jang [38] is more a tuning method than a learning method. Jang initiallydetermines the labels of the input variables through a uniform partition of the universe of eachvariable, and the parameters of the consequent are initiated to zero. A fixed rule base is createdfrom the one that it begins to adjust. A first assumption for using this approach as a tuning methodis that with a finer initialization, the method will work better. We shall allow it to start from aKB already learnt through another method. If the initial KB is simplified, the nodes that do notparticipate are disconnected. In this sense will be used ANFIS as a tuning method. The operationof the method can be consulted in Section III.B.2.

IV.B Cordon and Herrera’s Tuning Method

This process, introduced in [20], is based on the existence of a previous complete Mamdani KB,that is, an initial DB definition and an RB constituted by m Mamdani-type rules.

Each chromosome forming the genetic population will encode a complete DB definition thatwill be combined with the existing RB in order to evaluate the individual adaptation.

27

Page 28: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

The GA designed for the tuning process presents real coding issue [52], uses the stochasticuniversal sampling [49] as a selection procedure and Michalewicz’s non-uniform mutation operator.As regards to the crossover operator, the Max-Min-Arithmetical [50], which makes use of fuzzytools in order to improve the GA behavior, is employed.

As it has been commented in Section III.C.4, the fuzzy partitions are triangular-shaped (seeFigure 5). Thus, each one of the membership functions has an associated parametric representationbased on a 3-tuple of real values and a primary fuzzy partition can be represented by an arraycomposed by 3 · N real values, with N being the number of terms forming the linguistic variableterm set. The DB is encoded into a fixed length real coded chromosome Cr built by joining thepartial representations of each one of the variable fuzzy partitions as is shown in the following:

Cri = (ai1, bi1, ci1, . . . , aiNi , biNi , ciNi)Cr = Cr1 Cr2 ... Crm ,

where Cri codes the fuzzy partition of the i-th variable.The initial gene pool is created making use of the initial DB definition. This is encoded directly

in a chromosome, denoted as C1. The remaining individuals are generated by associating an intervalof performance, [cl

h, crh] to every gene ch in C1, h = 1 . . .

∑mi=1 Ni · 3. Each interval of performance

will be the interval of adjustment for the corresponding gene, ch ∈ [clh, cr

h].If (t mod 3) = 1 then ct is the left value of the support of a fuzzy number. The fuzzy number is

defined by the three parameters (ct, ct+1, ct+2) and the intervals of performance are the following:

ct ∈ [clt, c

rt ] = [ct − ct+1−ct

2 , ct + ct+1−ct2 ] ,

ct+1 ∈ [clt+1, c

rt+1] = [ct+1 − ct+1−ct

2 , ct+1 + ct+2−ct+12 ] ,

ct+2 ∈ [clt+2, c

rt+2] = [ct+2 − ct+2−ct+1

2 , ct+2 + ct+2−ct+12 ] .

Figure 10 shows a graphical representation of these intervals.

Figure 10: Intervals of performance

Therefore the authors create a population of chromosomes containing C1 as its first individ-ual and the remaining ones initiated randomly, with each gene being in its respective interval ofperformance.

The fitness function E(·) presented in the genetic simplification process introduced in Sec-tion III.C.4 is used for evaluating the adaptation of each individual of the population in the genetictuning process.

28

Page 29: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

V Examples of Application: Experiments Developed and ResultsObtained

To illustrate the performance of the seen methods, two applications with different features have beenchosen: the problem of rice taste evaluation [36] and a real-world Spanish electrical distributionnetwork problem [53, 54, 55].

We are going to develop a comparative study between the eleven methods for each problem.Afterwards, in Section V.C, we shall carry out a global analysis where we shall look at the generalbehavior of the different methods.

With the aim of making easier the identification of the methods, the abbreviations shown inTable 2 will be considered.

Table 2: Notations considered in this section

Section Author(s) of the Method Abbreviation-NameIII.A.1 Wang and Mendel WMIII.A.2 Nozaki, Ishibuchi, and Tanaka NITIII.B.1 Shann and Fu SFIII.B.2 Jang (ANFIS) J-ANFISIII.C.1 Thrift TIII.C.2 Liska and Melsheimer LMIII.C.3 Gonzalez and Perez (SLAVE) GP -SLAV EIII.C.4 Cordon and Herrera (D-MOGUL) CH-D-MOGULIII.C.5 Cordon and Herrera (WM-ALM) CH-WM -ALMIV.A Jang (ANFIS for tuning) J-ANFIS-tuningIV.B Cordon and Herrera (tuning) CH-tuning

We have considered the parameters showed in Table 3 for every method in every experimentdeveloped. With these values we have tried to select standard parameters that work well in mostcases instead of searching very specific values for each problem.

Each table of results will present three groups of data associated to each method: the resultsobtained by the method itself (noted by Generation) and two columns showing those obtained afterthe application of each one of the introduced tuning processes (noted by J-ANFIS-tuning andCH-tuning respectively).

The NIT method uses a scale factor (α) that deforms the membership function shape, whichcan make it to be considered as a tuning process. For this reason, we do not apply any additionaltuning to this method. In the case of the CH-D-MOGUL method, the generation results showedare the ones obtained after the application of the Simplification Process (second stage).

To evaluate the quality of the results we are going to use an error function called mean squareerror (MSE). We have defined MSE as

MSE =1

2 ·N

N∑

i=1

(y′i − yi)2 .

with y′ being the output obtained from the FRBS and with y being the known desired output. Themore closer to zero the measure is, the greater the efficiency is.

29

Page 30: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Table 3: Parameters values considered for the methods

Method Parameters DecisionWM Without Parameters

NIT Parameter α 5

SF Maximum number of epochs 1000Learning rate β 0.01

J-ANFIS Maximum number of epochs 1000and Step size 0.1

J-ANFIS-tuning Decrease rate of step size 0.9Increase rate of step size 1.1

Label insertion percentage 60%

T Genetic Parameters explained lower down

LM Creep probability 0.1Maximum number of rules Max allow

Genetic Parameters explained lower down

GP -SLAV E Maximum number of iterations 1000Parameter λ 0.8

Parameter K1 0Parameter K2 1

Adaptation Degree 0

CH-D-MOGUL Generation:Example covering value ε 1.5Positive example degree ω 0.05

k-consistency property parameter 0.1Simplification:

Completeness property parameter τ 0.1Number of simplified KBs to generate 3

Niche radius r 10% of initial KB rule no.Power factor β 0.5

Genetic Parameters explained lower down

CH-WM -ALM Genetic Parameters explained lower downbut with 500 generations

CH-tuning Non-uniform mutation parameter b 5Max-min-arithmetical parameter a 0.35Completeness property parameter τ 0.1

Genetic Parameters explained lower down

Genetic Parameters Maximum number of generations 1000Population size 61

Crossover probability 0.6Mutation probability 0.1

V.A Rice Taste Evaluation

V.A.1 The Rice Taste Evaluation Problem

Subjective qualification of food taste is a very important but difficult problem. In the case of therice taste qualification, it is usually put into effect by means of a subjective evaluation called thesensory test. In this test, a group of experts, usually composed of 24 persons, evaluate the riceaccording to a set of characteristics associated to it. These factors are: flavor, appearance, taste,stickiness, and toughness [36].

Because of the large quantity of relevant variables, the problem of rice taste analysis becomesvery complex, thus leading to solve it by means of modeling techniques capable of obtaining amodel representing the non-linear relationships existing in it. Moreover, the problem-solving goalis not only to obtain an accurate model, but to obtain a user-interpretable model as well, capableof putting some light on the reasoning process performed by the expert for evaluating a kind of

30

Page 31: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

rice in a specific way. Due to all these reasons, in this section we deal with obtaining a linguisticmodel to solve the said problem.

In order to do so, we are going to use the data set presented in [36]. This set is composed of105 data arrays collecting subjective evaluations of the six variables in question (the five mentionedand the overall evaluation of the kind of rice), made up by experts on this number of kinds of ricegrown in Japan (for example, Sasanishiki, Akita-Komachi, etc.). The six variables are normalized,thus taking values in the real interval [0, 1].

With the aim of not biasing the learning, we have randomly obtained ten different partitionsof the set mentioned, composed by 75 pieces of data in the training set and 30 in the test one,for generating ten qualitative models in each experiment. We have worked with fuzzy partitionsobtained from a normalization process in which the universe of discourse of each one of the sixvariables has been equally divided into 2 parts, and a triangular fuzzy set has been associated toeach one of them. Figure 5 shows an example of a fuzzy partition with seven linguistic labels.

V.A.2 Experiments and Results in the Rice Taste Evaluation Problem

The results obtained in the experiments developed are collected in Table 4. The values shownin columns MSEtra and MSEtst have been computed as an average of the MSE values obtainedin the approximation of the training and test data sets, respectively, by the ten linguistic modelsgenerated in each case. The column #R stands for the average of the number of linguistic rules inthe KBs of the models generated from each process.

Table 4: Results obtained in the rice taste evaluation problem

Generation J-ANFIS-tuning CH-tuningMethod #R MSEtra MSEtst #R MSEtra MSEtst MSEtra MSEtst

WM 15 0.01328 0.01312 15 0.00363 0.00372 0.00111 0.00214NIT 64 0.00300 0.00352 – — — — —SF 32 0.01940 0.02137 32 0.00397 0.00481 0.00183 0.00331

J-ANFIS 32 0.00503 0.00563 – — — 0.00277 0.00387T 15.9 0.00495 0.00600 15.9 0.00348 0.00483 0.00115 0.00293

LM 30.6 0.00128 0.00236 30.6 0.00128 0.00236 0.00081 0.00234GP -SLAV E 24 0.01803 0.02049 24 0.01771 0.02290 0.00791 0.01325

CH-D-MOGUL 6 0.00486 0.00370 4 0.01018 0.01015 0.00108 0.00233CH-WM -ALM 5 0.00341 0.00398 5 0.00247 0.00323 0.00103 0.00274

It is very remarkable the excellent treatment of the complexity by CH-D-MOGUL and CH-WM -ALM that obtain bases with a very small number of rules (an average of 6 and 5 respectively).The number of rules is extremely important in problems that need to be as highly interpretable asthis. Therefore, the work carried out in the phase of Simplification (used in both methods) is veryvaluable.

On the other hand, LM obtains results of excellent accuracy but at the expense of to generatea large number of rules that, as said, is not desirable in this case.

As regards the tuning processes, the adjustment of CH-tuning improves all the results andequalizes the efficiency grades of the methods. The adjustment of J-ANFIS-tuning also improvesthe results but less significantly than CH-tuning. When the models obtained by CH-D-MOGULare tuned, the results makes worse a lot. This is due to the fact that, upon to have a low number of

31

Page 32: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

rules, when J-ANFIS-tuning eliminates some of them, it increases the error. This newly indicatesthat the selection realized in CH-D-MOGUL is very appropriate and the number of rules is optimal.

Although NIT generate good values, all of the methods produce less rules than it. We mustnot forget that in this problem the complexity is very important. Furthermore, if CH-tuning isapplied, all except J-ANFIS and GP -SLAV E obtain also better accuracy.

V.B Electrical Distribution Networks

V.B.1 The Electrical Distribution Networks Problem

Sometimes, there is a need to measure the amount of electricity lines that an electric company owns.This measurement may be useful for several aspects such us the estimation of the maintenance costsof the network, which was the main goal of the problem presented here in Spain [53, 54, 55]. Highand medium voltage lines can be easily measured, but low voltage line is contained in cities andvillages, and it would be very expensive to measure it. This kind of line used to be very convolutedand, in some cases, one company may serve more than 10,000 small nuclei. An indirect method fordetermining the length of line is needed.

The problem involves finding a model that relates the total length of low voltage line installedin a rural town with the number of inhabitants in the town and the mean of the distances from thecenter of the town to the three furthest clients in it [54, 55]. This model will be used to estimatethe total length of line being maintained.

We shall limit ourselves to the estimation of the length of line in a town, given the inputsmentioned before. Hence, our objective is to relate the first variable (line length) with the othertwo ones (population, radius of village).

In this problem, it would be preferable that the solutions obtained verify the following require-ment: they have not only to be numerically accurate in the problem-solving, but must be able toexplain how a specific value is computed for a certain village. That is, it is interesting that thesesolutions are interpretable by human beings to some degree.

To compare the methods, we have randomly divided the sample, composed of 495 pieces of realdata obtained from direct measures in this number of villages [53], into two sets comprising 386and 99 samples, labeled training and test. In this case, the linguistic variable fuzzy partitions aredivided into 7 fuzzy sets in the experiments developed.

V.B.2 Experiments and Results in the Electrical Distribution Networks Problem

The results obtained with the different linguistic modeling methods considered are shown in Table5.

The significant differences existing between WM and CH-WM -ALM are because of the factthat the RB obtained with WM is insufficient since in this application is necessary to deep in certainzones. CH-WM -ALM allows us to use two different rules defined in the same input subspace butpresenting a different consequent in the case in which it is necessary.

In the generation phase, the genetic methods stand out against the remaining ones. Especially,LM offers a good result since it generates a model accurate which is in both learning and predictionthus not presenting overlearning.

SF obtains very bad results because of it has a big dependency of the value assigned to theweights to accomplish the learning process. Let us suppose that an important rule begins with alow weight associated, its influence on the NN evolution will be nonvalued and it will be eliminatedin the pruning phase. However, the authors propose to randomly initialize these weights, provokingwith this action the convergence to a local optimum.

32

Page 33: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Table 5: Results obtained in the electrical application problem

Generation J-ANFIS-tuning CH-tuningMethod #R MSEtra MSEtst #R MSEtra MSEtst MSEtra MSEtst

WM 24 222623 240018 24 161889 160370 144510 173167NIT 64 173230 190808 – — — — —SF 45 1281547 1067993 45 247696 296587 233133 252569

J-ANFIS 49 256605 268451 – — — 244286 263940T 47 185204 168060 47 184804 168318 169689 175985

LM 49 167014 167383 49 165650 296804 150224 290461GP -SLAV E 61 346921 379076 39 184650 256420 176145 191721

CH-D-MOGUL 35 167621 207598 35 165633 208693 142503 181288CH-WM -ALM 20 155866 178601 17 209259 256890 136437 210236

The adjustment of CH-tuning only causes overlearning precisely in the three methods thatobtained the best value of prediction in the generation phase, LM , T , and CH-WM -ALM . Thisalso occurs when J-ANFIS-tuning are applied. The adjustment of J-ANFIS-tuning works wellwith the WM method, which has wrong results in the generation phase, but generally it does notobtain significant improvement.

V.C Global Analysis

In the following, some conclusions about the methods’ behavior and about the different techniquesused are presented.

V.C.1 Learning Methods

We should underline the good behavior presented by the methods based on GAs on the generationphase.

The bad results of J-ANFIS are due to the lack of the LSE technique in the Mamdani approach.The power of this method is based on this technique, but it is only applicable when the consequentsare not linguistic terms (TSK approach). This also affects to the tuning method.

As regards the number of rules in the KBs, the models designed by means of process with GeneticSimplification Phase (CH-D-MOGUL and CH-WM -ALM) should be highlighted as well, due tothe fact that they always present less rules than the ones in the models generated from remainingmethods.

The number of rules used by GP -SLAV E is only 2 and 9 for the rice and the electrical problemsrespectively if we represent them in DNF form. This aspect provides a good interpretation to theRB generated and, although its accuracy is worse than most methods, it is very interesting inspecific problems. For example, one of the RBs derived by GP -SLAV E in the rice test problem isas simple as:

IF Taste is Bad and Stickiness is Not sticky THEN Overall Evaluation is LowIF Taste is Good THEN Overall Evaluation is High

33

Page 34: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

V.C.2 Tuning Methods

We should remark the good performance of the tuning processes (especially the genetic approach,CH-tuning), that broadly improve the accuracy of the preliminary RB definitions. However, someproblems of high complexity appear: the tuning applied to initial KBs with good accuracy doesnot obtain beneficial results since it increases the overlearning and makes worse the generalizationability. This is due to that the tuning tends to overfit on the training data in these cases.

A clear inconvenient of J-ANFIS-tuning is that it is more restrictive than the GA-based oneregarding the structure of the KBs to adapt. It does not allow initial KBs with several rules withthe same antecedents and different consequents. In this case, we must consider only one of themin each input subspace, but this will provoke the loss of part of the input information.

V.C.3 Techniques Used

Next, we remark some comments on the techniques used by the different methods:

• The ad hoc approaches are generally based in data covering, i.e., in an attempt to coverproperly the example set. This implies a big dependency on the data training that may beimprudent.

• The principal advantages of NNs is that they can learn by generalization and many NNsare resistive against noise or partial damage after learning. As disadvantages we found thatthe convergence of the learning process cannot be guaranteed but has to be gained by often-extensive experimentation. The NNs have a serious problem when the number of variablesor the number of labels is increased, since the complexity grows geometrically. This is due tothe way considered to represent the rules by means of connections between nodes. Anotherimportant weakness is that NNs are not flexible with the KB representation.

• Whereas GAs have a guaranteed local convergence and are easy to implement, there is noguarantee to find the optimal solution, and the high amount of individuals to be processedmakes the algorithm relatively time consuming. Pittsburgh GA approaches, as NNs, alsohave problems when they work with a lot of variables since the search space will be huge.A good advantage is that GAs can represent different types of fuzzy rules as approximateMamdani or DNF ones.

VI Concluding Remarks

In this Chapter we have accomplished a short revision of FRBSs and we have seen the differenttypes that currently exist.

Thereinafter, we have focused on linguistic Mamdani FRBSs, which obtain better legibility inexchange for offering worse accuracy. We have seen ad hoc learning methods and two techniques,NNs and GAs, that can be applied to the learning of FRBSs. Several forms of facing the learninghave been presented.

Finally, through the application of the reviewed methods, two real applications have beenconsidered to analyze their behavior. In this study, we have obtained some interesting conclusions.Among other things, we have proven that the GAs work best and that the tuning phase improvessignificantly the accuracy. We have also been able to see the limitations associated with thelinguistic representation when the problem is complex.

34

Page 35: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Appendix A: Neural Networks

A.1. Brief Description of Neural Networks

Artificial Neural Networks (ANNs) are computational models that emerged as mathematicalformalization attempts of the structure of the brain [56, 57, 58, 59]. They are outlined as calculationmodels for those which exist algorithms that permit to develop cognition tasks such as learning,classification or optimization. Furthermore, they are characterized by a massively parallel operationthat makes them very efficient. The ANN model can be understood as another way of representingknowledge. It differs of the high-level mappings (rules, semantic networks, frames, . . . ) in whichworks at a subsymbolic level.

An ANN is formed by a set of elemental processing units that are communicated mutually beingshipped signals through weighted connections. The main characteristics of an NN model are:

• A set of processing units, called neurons, cell process elements, etc.

• Each unit, i, receives an input series (from the environment or from other units), that cantake values in various domains. If the set is {0, 1}, the neuron is called binary. If values outletin {−1, 1} the name of the neuron is bipolar. But also there are very used the cases in whichthe set is a continuous interval as [0, 1] or [−1, 1].

• The neurons are connected among them through links, also called connections or synapses,that tend to be one-way, though there are models with symmetrical connections as well.Each connection communicates two units and have associated a weighting or weight, wij ,that determines the efficiency with which the signals of the neuron i affects the j one.

• An aggregation rule of the inputs, that determines the excitement level of a unit i.

• An activation function, f , that indicates the state or activation level, ai, according to theexcitement level, that can be propagated as input to other neurons.

• In many cases, an external input or trend for each unit, θi, is employed as well. Its performanceis equivalent to a threshold that the total signal that arrives to the neuron must exceed toenable it.

Processing units. Each unit accomplishes a relatively simple role: it takes the inputs of theneurons connected to it and the external one and obtains its new state or excitement level fromthem. From a new state, and applying the activation function, the output value of the neuronis obtained to be transmitted to the neurons connected to its output. This output value is theactivation state.

We can build networks by connecting neurons. In them, we distinguish three types of neurons:input neurons, by which the network downloads from the environment; output neurons that providethe result of the processing to the environment and hidden neurons whose input and output signalsremain within the system.

The network is inherently parallel in the sense of the fact that many units can accomplishtheir calculations at the same time. During the processing, the units can update their state ofsynchronous or asynchronous mode. In the synchronous mode, all the neurons are updated simul-taneously. In the asynchronous one, each unit has a variable probability, time-sensitive, of beingupdated and usually only one unit is updated in a given moment. In some instances, the secondmode is more advantageous.

35

Page 36: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Connections among the units and aggregation rule. In most cases, it is supposed thatthe contributions of the input units are additive. Thus, it is established that the excitement of theunit j is the weighted sum of the inputs, result to which the threshold is added:

uj(t) =∑

i

wij(t)xi(t) + θj(t) .

When a connection has a positive weight wij , it is said that the connection is exciting and whenit is negative, it is said it is inhibiting.

Activation rules and output rules. The effect of the excitement level on the activation stateof the neuron is established through the activation function, f . This function takes the currentstate of the neuron and the excitement level as inputs and returns the new activation state:

aj(t + 1) = f(aj(t), uj(t)) .

Usually, the activation function is a non-diminishing function of the excitement level of theneuron. The more employed functions are the step ones: the sign function, the linear function orS elongated functions such as the logistics and the hyperbolic tangent.

In some instances, the output of a unit can be a stochastic function of the inputs. In this case,the new activation state is a dependent random variable of the input.

Connection topologies. A relevant aspect of an ANN is the mode in which the units areconnected. According to the type of connection topology, the models of ANN are split into:

• Feed-forward networks. The neurons are arranged in sets called layers. The layers are ordered,beginning in the input layer (the one which contains the input units) and terminating in theoutput layer, so that the links only exist among the neurons of a layer and those of theimmediately next one and only in that sense. There are no links among neurons of a samelayer neither among neurons of non-consecutive layers. In this way, the signals flow from theinput to the output.

• Feed-back networks. Connections toward back and intralayer are allowed, disappearing basi-cally the layer idea as a functional neurons group.

This distinction is important because the analysis of the network operation (above all, itsdynamic behavior) is very different according to if it spreads toward forward or with feedback.

Once the neurons which a given one is connected to are established, it can be represented theset of connections with a weight vector. The ordered arrangement of the arrays permits us to builda matrix that is designated weight matrix of the network. When there is no connection betweentwo neurons, it is understood that the associated weight is zero.

Training of an ANN. An ANN must be trained so that the application of a set of inputsproduces, well directly or through a learning process, the waited output. The learning involvesmodifying the free parameters of the network, i.e., the weights of the connections, so that theassignment target error to be made zero.

There are several methods to accomplish this. If there is enough knowledge, the weights canbe beforehand calculated. But the more usual method of training the network is based on showingexamples to it and modifying its weights in view of the error produced according to some learningrule.

The learning methods are classified in two groups:

36

Page 37: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

• Supervised. Examples that include the desired input and output are shown to the network.

• Unsupervised. The network is trained with examples including only the inputs (the waitedoutput is not indicated to it). The network must be capable of discovering statistically themain characteristics of the input samples and to employ them to classify the input patterns.It does not exist any set of categories in which the patterns should be classified a priori; thesystem must develop its own mapping of the input stimuli.

Learning rules. All the learning methods give the setting of the weights of the connectionsamong units as a result according to a rule.

Nearly all these learning rules are variations of the Hebb rule [60]. The basic idea of this rule isthat if two units i and j are activated simultaneously, their interconnections should be reinforced.If i is connected to the input of j, the simplest formulation of the Hebb rule prescribes the followingmodification:

∆wij = γaiaj ,

with γ being a constant of proportion that represents the learning rate. Another common ruleemploys the discrepancy among the current and the desired output instead of the output of theunit j:

∆wij = γai(dj − aj) ,

with dj being the value expected for the activation level, supplied by a teacher. This rule is knownas the delta rule [61].

Both the Hebb and the delta rules are adapted for the training of feed-forward propagationnetworks in which there are no hidden neurons. However, these kinds of networks have a verylimited area of application since they are only capable of solving linearly separable problems. Hence,simple problems as the representation of or-exclusive (XOR) function can not be solved employingthem. This limitation was demonstrated by Minsky and Papert [62], who also proved that it couldbe solved employing hidden neurons. The problem was to find appropriate training procedures.Rumelhart et al. [63] settled the problem with the presentation in 1986 of the error-backpropagationalgorithm that has turned out to be fruitful and very used.

The error-backpropagation algorithm (EBP). The EBP is a generalization of the deltarule. The main idea is again to quantify the existing error through the difference among the systemoutput and the real output. The difficulty lies in that there is now a great number of connectionsthat increase the error, which do not only relate to the neurons of the output layer. Particularly,since there is not information on the role that the hidden units play, their contribution to the totalerror can not be directly calculated and it is not known how to correctly modify the weights of theconnections associated to them.

The problem can be tackled if it is admitted that the error observed in the output layer shouldbe primarily due to the action of the hidden units located in the immediately previous layer, that atthe same time are seen as influenced by the elements of the previous layer, and so on. The adaptiveprocess remains defined in two stages that are going to be repeated until the learning is consideredto be accomplished. In the first the activity level of all the neurons of the network is evaluatedmaintaining fixed the values of the connections, which permits to determine the magnitude of theexisting error. In a second stage this error is propagated back layer to layer, successfully modifyingthe weights that in the next stage will serve to calculate the new error. This process is the onewhich gives name to the learning algorithm.

37

Page 38: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Figure 11: Multilayer Neural Network

We consider a network with n input neurons, h hidden neurons and m output neurons (seeFigure 11). The activation function is a differential function of the actual input to the neuron:

apj = Fj(e

pj ) ,

whereepj =

i

wijapi + θj ,

that is the actual input to the j-th unit for the p-th example.After the first epoch in which the assignment error by the network is evaluated, each weight is

adjusted according to:∆pwij = γδp

j api .

The parameter δpj adopts different values for the case of output and internal units. It is started

making the setting for the weights that bind the units of the hidden layer and the output ones. Inthis case:

δpj = (dp

j − apj )F

′j(e

pj )

for any output unit j. If a hidden unit is considered, the values of δpj are calculated according to:

δpj = F ′

j(epj )

No∑

h=1

δphwhj .

The latter two equations give us a recursive procedure to calculate the deltas for all the unitsin the network. This procedure constitutes the generalized delta rule or the EBP.

The expressions of the algorithm for the case where the activation function is the logistics isdefined as:

F (epj ) =

11 + exp(−ep

j ),

and its derivative is:F ′(ep

j ) = apj (1− ap

j ) .

The deltas for the output units result:

δpj = (dp

j − apj )a

pj (1− ap

j ) ,

38

Page 39: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

and for the units of the hidden layer:

δpj = ap

j (1− apj )

No∑

h=1

δphwhj .

The gradient descent algorithm requires that the setting is accomplished through minimal steps,but for practical effects it is interesting to consider a large learning rate γ that permits a quicklearning. The problem is that oscillations can appear. A way to avoid the oscillations for largevalues is to accomplish changes in the dependent weights of the previous changes appending a termmoment:

∆wij(t + 1) = γδpi a

pj + α∆wij(t) ,

with α being a constant of the moment, measuring the importance of the previous setting in thecurrent one.

A.2. Neuro-Fuzzy Systems

Fuzzy Logic and NNs are two disciplines that efficiently deal with two different areas of infor-mation processing. Fuzzy Logic deals with various aspects of uncertain knowledge representationand processing, and it allows approximate reasoning in the field of knowledge based systems. NNsare efficient computation structures capable of learning and adapting from examples.

Both techniques have their advantages and their weaknesses. In some sense, they become com-plementary since some features lacking in one approach are dominant in the other. Fuzzy Systemsare not powerful in learning, adaptation, and parallel computation, where NNs offer a good per-formance in these aspects. NNs lacks in flexibility, human interaction or knowledge representation,where Fuzzy Logic is a powerful tool. In the last few years, a large research effort has been madeto synthesize Fuzzy Logic and NNs to produce hybrid Fuzzy-Neural systems.

We prefer to call Neuro-Fuzzy Systems (NFS) to these approaches where NNs are used to provideinputs for a Fuzzy System, or to change the output of a Fuzzy System to remark that the parametersof a Fuzzy System are not changed by a learning process in these approaches. If the creation of anNN is the main target, it is possible to apply fuzzy techniques to speed up the learning process, orto fuzzify an NN by the extension principle to be able to process fuzzy inputs. These approachescould be called Fuzzy Neural Networks to stress that fuzzy techniques are used to create or enhanceNNs. We are interested in the first approach.

An NFS has the following five characteristics as in [64] is described:

• An NFS is a Fuzzy System that is trained by a learning algorithm usually derived fromNN theory. The learning procedure operates on local information, and causes only localmodifications in the underlying Fuzzy System. The learning process is not knowledge-based,but data driven.

• It can be viewed as a special 3-layer feedforward NN. The units in this network use t-norms ort-conorms instead of the activation functions common in NNs. The first layer represents inputvariables, the middle or hidden layer represents fuzzy rules, and the third layer representsoutput variables. Fuzzy sets are encoded as connection weights. This view of a Fuzzy Systemillustrates the data flow within the system, and its parallel nature. However, this neuralnetwork view is not a prerequisite for applying a learning procedure, but merely a convenience.

• An NFS can always —i.e., before, during, and after the learning— be interpreted as a FRBS.It is both possible to create the system out of training data from scratch, and to initialize itby prior knowledge in the form of fuzzy rules.

39

Page 40: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

• The learning procedure of an NFS takes the semantic properties of the underlying FuzzySystem into account. This results in constraints on the possible modifications applicable tothe system parameters.

• An NFS approximates an n-dimensional unknown function that is partially given by thetraining data. An NFS should not be seen as a kind of fuzzy expert system, and it hasnothing to do with Fuzzy Logic in the narrow sense.

NFSs can be considered as a technique to derive a Fuzzy System from data, or to enhance it bylearning from examples. It is possible to use an NN to learn certain parameters of a Fuzzy System,like using a self-organizing feature map to find fuzzy rules [65] (cooperative models), or to view aFuzzy System as a special NN, and directly apply a learning algorithm [66] (hybrid models).

For more information, refer to [41, 64, 67, 68].

Appendix B: Genetic Algorithms

B.1. Brief Description of Genetic Algorithms

GAs are general-purpose search algorithms that use principles inspired by natural populationgenetics to evolve solutions to problems [69]. The basic idea is to maintain a population of knowl-edge structures that evolves over time through a process of competition and controlled variation.Each structure in the population represents a candidate solution to the specific problem and has anassociated fitness to determine which structures are used to form new ones in the process of com-petition. The new individuals are created using genetic operators such as crossover and mutation.GAs have had a great measure of success in search and optimization problems. The reason of greatpart of this success is their ability to exploit accumulative information about an initially unknownsearch space in order to bias subsequent search into useful subspaces, i.e., their robustness. Thisis their key feature, especially in large, complex and poorly understood search spaces, where theclassical search tools (enumerative, heuristic, . . . ) are inappropriate, offering a valid approach toproblems requiring efficient and effective search.

A GA starts with a population of randomly generated solutions, chromosomes, and advancestoward better solutions by applying genetic operators, modeled on the genetic processes occurringin nature. As mentioned in these algorithms we maintain a population of solutions for a givenproblem; this population undergoes evolution in a form of natural selection. In each generation,relatively good solutions reproduce to give offspring that replace the relatively bad solutions, whichdie. An evaluation or fitness function plays the role of the environment to distinguish between goodand bad solutions. The process of going from the current population to the next one constitutesone generation in the execution of a genetic algorithm.

Although there are many possible variants of the basic GA, the fundamental underlying mech-anism involves three operations:

(1) evaluation of individual fitness,

(2) formation of a gene pool (intermediate population), and

(3) recombination and mutation.

Figure 12 shows the structure of a simple GA.A fitness function must be devised for each problem to be solved. Given a particular chro-

mosome, a solution, the fitness function returns a single numerical fitness that is supposed to beproportional to the utility or adaptation of the solution represented by this chromosome.

40

Page 41: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Procedure Genetic Algorithmbegin (1)

t = 0;initialize P (t);evaluate P (t);While (Not termination-condition) dobegin (2)

t = t + 1;select P (t) from P (t− 1);recombine P (t);evaluate P (t);

end (2)end (1)

Figure 12: Structure of a GA

There are a number of ways to do selection. We might view the population as a mapping ontoa roulette wheel, where each individual is represented by a space that proportionally correspondsto its fitness. By repeatedly spinning the roulette wheel, individuals are chosen using stochasticsampling with replacement to fill the intermediate population. The selection procedure proposedby Baker [49], called stochastic universal sampling, is one of the efficient proposals for avoiding thegenetic drift [70]. The number of offspring of any structure is bound by the floor and ceiling of theexpected number of offspring.

After selection has been carried out, the construction of the intermediate population is com-pleted and recombination and mutation can occur.

The crossover operator combines the features of two parent structures to form two similaroffspring. Classically, it is applied at a random position with a probability of performance, thecrossover probability, Pc. The mutation operator arbitrarily alters one or more components ofa selected structure so as to increase the structural variability of the population. Each positionof each solution vector in the population undergoes a random change according to a probabilitydefined by a mutation rate, the mutation probability, Pm.

Figures 13, 14, and 15 illustrate the basic operations: reproduction, crossover, and mutation.

Figure 13: Evaluation and contribution to the gene pool

It is generally accepted that a GA must take into account the five following components to solvea problem:

41

Page 42: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

Figure 14: Recombination. One-point crossover

Figure 15: Mutation

1. A genetic representation of solutions to the problem,

2. a way to create an initial population of solutions,

3. an evaluation function which gives the fitness of each individual,

4. genetic operators that alter the genetic composition of children during reproduction, and

5. values for the parameters that the GA uses (population size, probabilities of applying geneticoperators, etc.).

The basic principles of the GAs were first laid down rigorously by Holland [69] and are welldescribed in many texts as [70, 71].

B.2. Genetic Fuzzy Systems

Although GAs are not learning algorithms, they may offer a powerful and domain-independentsearch method for a variety of learning tasks. In fact, there has been a good deal of interest inusing GAs for machine learning problems [72].

Three alternative approaches, in which GAs have been applied to learning processes, havebeen proposed, the Michigan [73], the Pittsburgh [74], and the Iterative Rule Learning (IRL) [46]approaches. In the first one, the chromosomes correspond to classifier rules that are evolved as awhole, whereas in the Pittsburgh approach, each chromosome encodes a complete set of classifiers.In the IRL approach each chromosome represents only one rule, but contrary to the first, onlythe best individual is considered as the solution, discarding the remaining chromosomes in thepopulation.

Recently, numerous papers and applications combining fuzzy concepts and GAs have appeared,and there is an increasing concern about the integration of these two topics. In particular, a greatnumber of publications explore the use of GAs for designing Fuzzy Systems. These approachesreceive the general name of Genetic Fuzzy Systems (GFSs) [19, 34].

The automatic design of Fuzzy Systems can be considered in many cases as an optimization orsearch process on the space of potential solutions. GAs are the best known and most widely usedglobal search technique with an ability to explore and exploit a given operating space using availableperformance measures. A priori knowledge in the form of linguistic variables, fuzzy membershipfunction parameters, fuzzy rules, number of rules, etc., may be easily incorporated into the geneticdesign process. The generic code structure and independent performance features of GAs makethem suitable candidates for incorporating a priori knowledge. Over the last few years, these

42

Page 43: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

advantages have extended the use of GAs in the development of a wide range of approaches fordesigning Fuzzy Systems.

As in the general case of Fuzzy Systems, the main application area of GFSs is system modelingand control. Regardless the kind of optimization problem, i.e., given a system to be modeled orcontrolled, the involved learning or tuning process will be based on evolution. Three points are thekeys to a genetic process: the population of potential solutions, the pair evolution operators/code,and the performance index.

A GFS combines the main aspects of the system to be obtained, a Fuzzy System, and the designtechnique used to obtain it, a GA, with the aim of improving as far as possible the accuracy ofthe final Fuzzy System generated. One of the most interesting features of a Fuzzy System is theinterpolated reasoning that develops. This characteristic plays a key role in the high performanceof Fuzzy Systems and is a consequence of the cooperation among the fuzzy rules composing the KB.As is known, the output obtained from a Fuzzy System is not usually due to a single fuzzy rule butto the cooperative action of several fuzzy rules that have been fired because they match the inputto the system to some degree.

On the other hand, the main feature of a GA is the competition among members of the populationrepresenting possible solutions to the problem being solved. In this case, this characteristic is dueto the mechanisms of natural selection on which the GA is based.

Therefore, since a GFS combines both aforementioned features, it works by inducing compe-tition to get the best possible cooperation. This seems to be a very interesting way to solve theproblem of designing a Fuzzy System, because the different members of the population competewith one another to provide a final solution presenting the best cooperation among the fuzzy rulescomposing it. The problem is to obtain the best possible way to put this way of working intoeffect. This is referred to as Cooperation vs. Competition Problem [75]. The difficulty of solvingthe introduced problem depends directly on the genetic learning approach followed by the GFS(Michigan, Pittsburgh, or IRL).

Different GFS proposals can be found in [34, 76, 77, 78].

References

[1] Zadeh, L. A. (1965). Fuzzy sets. Information and Control 8, pp. 338–353.

[2] Bardossy, A., and Duckstein, L. (1995). “Fuzzy rule-based modeling with application to geophysical,biological and engineering systems.” CRC Press.

[3] Chi, Z., Yan, H., and Pham, T. (1996). “Fuzzy algorithms: with applications to image processing andpattern recognition.” World Scientific.

[4] Hirota, K. (Ed.) (1993). “Industrial applications of fuzzy technology.” Springer-Verlag.

[5] Pedrycz, W. (Ed.) (1996). “Fuzzy Modelling. Paradigms and Practice.” Kluwer Academic Press.

[6] Klir, G. J., and Yuan, B. (1995). “Fuzzy sets and fuzzy logic.” Prentice-Hall.

[7] Zimmermann, H. J. (1996). “Fuzzy sets theory and its applications.” Kluwer Academic Press.

[8] Zadeh, L. A. (1973). Outline of a new approach to the analysis of complex systems and decisionprocesses. IEEE Transactions on Systems, Man, and Cybernetics 3, pp. 28–44.

[9] Wang, L. X. (1994). “Adaptive fuzzy systems and control.” Prentice-Hall.

[10] Mamdani, E. H. (1974). Applications of fuzzy algorithm for control a simple dynamic plant. Proc. ofthe IEEE 121, pp. 1585–1588.

43

Page 44: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

[11] Mamdani, E. H., and Assilian, S. (1975). An experiment in linguistic synthesis with a fuzzy logiccontroller. Int. Journal of Man-Machine Studies 7, pp. 1–13.

[12] Sugeno, M., and Yasukawa, T. (1993). A fuzzy-logic-based approach to qualitative modeling. IEEETransactions on Fuzzy Systems 1, pp. 7–31.

[13] Lee, C. C. (1990). Fuzzy logic in control systems: Fuzzy logic controller – Parts I and II. IEEE Trans-actions on Systems, Man, and Cybernetics 20, pp. 404–435.

[14] Gonzalez, A., Perez, R., and Verdegay, J. L. (1993). Learning the structure of a fuzzy rule: A geneticapproach. Proc. of the 1st European Congress on Fuzzy and Intelligent Technologies, Aachen (Germany),pp. 814–819.

[15] Gonzalez, A., and Perez, R. (1998). Completeness and consistency conditions for learning fuzzy rules.Fuzzy Sets and Systems 96, pp. 37–51.

[16] Magdalena, L., and Monasterio, F. (1997). A fuzzy logic controller with learning through the evolutionof its knowledge base. Int. Journal of Approximate Reasoning 16, pp. 335–358.

[17] Magdalena, L. (1997). Adapting the gain of an FLC with genetic algorithms. Int. Journal of Approxi-mate Reasoning 17, pp. 327–349.

[18] Cordon, O., and Herrera, F. (1998). A proposal for improving the accuracy of linguistic modeling.Dept. of Computer Science and Artificial Intelligence, University of Granada, Spain. Technical Report#DECSAI-98113.

[19] Cordon, O., and Herrera, F. (1995). A general study on genetic fuzzy systems. In “Genetic algorithmsin engineering and computer science” (J. Periaux, G. Winter, M. Galan, and P. Cuesta, Eds.), pp.33–57. John Wiley & Sons.

[20] Cordon, O., and Herrera, F. (1997). A three-stage evolutionary process for learning descriptive andapproximative fuzzy logic controller knowledge bases from examples. Int. Journal of Approximate Rea-soning 17, pp. 369–407.

[21] Koczy, L. (1996). Fuzzy if ... then rule models and their transformation into one another. IEEE Trans-actions on Systems, Man, and Cybernetics 26, pp. 621–637.

[22] Cordon, O., Herrera, F. (to appear). Hibridizing genetic algorithms with sharing scheme and evolutionstrategies for designing approximate fuzzy rule-based systems. Fuzzy Sets and Systems.

[23] Takagi, T., and Sugeno, M. (1985). Fuzzy identification of systems and its application to modeling andcontrol. IEEE Transactions on Systems, Man, and Cybernetics 15(1), pp. 116–132.

[24] Sugeno, M., and Kang, G. T. (1988). Structure identification of fuzzy model. Fuzzy Sets and Systems28, pp. 15–33.

[25] Gupta, M. M., and Qi, J. (1991). Design of fuzzy logic controllers based on generalized T-operators.Fuzzy Sets and Systems 40, pp. 473–489.

[26] Trillas, E., and Valverde, L. (1985). On implication and indistinguishability in the setting of fuzzy logic.In “Management Decision Support Systems Using Fuzzy Logic and Possibility Theory” (J. Kacpryzkand R. R. Yager, Eds.), pp. 198–212. Verlag TUV Rheinland.

[27] Cordon, O., Herrera, F., and Peregrın, A. (1997). Applicability of the fuzzy operators in the design offuzzy logic controllers. Fuzzy Sets and Systems 86, pp. 15–41.

[28] Driankov, D., Hellendoorn, H., and Reinfrank, M. (1993). “An introduction to fuzzy control.” Springer-Verlag.

44

Page 45: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

[29] Cao, Z., and Kandel, A. (1989). Applicability of some fuzzy implication operators. Fuzzy Sets andSystems 31, pp. 151–186.

[30] Kiszka, J., Kochanska, M., and Sliwinska, D. (1985). The influence of some fuzzy implication operatorson the accuracy of a fuzzy model — Parts I and II. Fuzzy Sets and Systems 15, pp. 111–128 and223–240.

[31] Zadeh, L. A. (1994). Fuzzy logic and soft computing: Issues, contentions and perspectives. Proc. of the3rd Int. Conf. on Fuzzy Logic, Neural Nets and Soft Computing, Iizuka (Japan), pp. 1–2.

[32] Bonissone, P.P. (1997). Soft computing: The convergence of emerging reasoning technologies. SoftComputing. A Fusion of Foundations, Methodologies, and Applications 1(1), pp. 6–18.

[33] Takagi, T. (1990). Fusion technology of fuzzy theory and neural networks — survey and future direc-tions. Int. Conf. on Fuzzy Logic and Neural Networks, Iizuka (Japan), pp. 13–26.

[34] Cordon, O., Herrera, F., Hoffmann, F., Magdalena, L. (in preparation). Genetic fuzzy systems: Evolu-tionary tuning and learning of fuzzy knowledge bases. World Scientific.

[35] Wang, L. X., and Mendel, J. M. (1992). Generating fuzzy rules by learning from examples. IEEETransactions on Systems, Man, and Cybernetics 22(6), pp. 1414-1427.

[36] Nozaki, K., Ishibuchi, H., and Tanaka, H. (1997). A simple but powerful heuristic method for generatingfuzzy rules from numerical data. Fuzzy Sets and Systems 86, pp. 251–270.

[37] Shann, J. J., and Fu, H. C. (1995). A fuzzy neural network for rule acquiring on fuzzy control systems.Fuzzy Sets and Systems 71, pp. 345–357.

[38] Jang, J.-S.R. (1993). ANFIS: Adaptive-Network-Based Fuzzy Inference System. IEEE Transactions onSystems, Man, and Cybernetics 23(3), pp. 665–685.

[39] Thrift, P. (1991). Fuzzy logic synthesis with genetic algorithms. Proc. of the 4th Int. Conf. on GeneticAlgorithms, pp. 509–513.

[40] Liska, J., and Melsheimer, S. S. (1994). Complete design of fuzzy logic systems using genetic algorithms.Proc. of the 3rd IEEE Int. Conf. on Fuzzy Systems, Vol. 2, pp. 1377–1382.

[41] Kosko, B. (1992). “Neural networks and fuzzy systems: A dynamic systems approach to machineintelligence.” Prentice-Hall.

[42] Jang, J.-S.R. (1991). Fuzzy modeling using generalized neural networks and kalman filter algorithm.Proc. of the 9th Nat. Conf. Artificial Intelligence, pp. 762–767.

[43] Watrous, R. L. (1991). Learning algorithms for connectionist network: Applied gradient methods ofnonlinear optimization. Proc. of the IEEE Int. Conf. Neural Networks, pp. 619–627.

[44] Gonzalez, A., and Perez, R. (1995). Structural learning of fuzzy rules from noisy examples. Proc. of the4th IEEE Int. Conf. on Fuzzy Systems/IFES’95, Yokohama, Vol. 3, pp. 1323–1330.

[45] Gonzalez, A., and Herrera, F. (1996). Multi-stage genetic fuzzy systems based on the iterative rulelearning approach. Mathware and Soft Computing 4, pp. 233–249.

[46] Venturini, G. (1993). SIA: A supervised inductive algorithm with genetic search for learning attributebased concepts. Proc. of the European Conf. on Machine Learning, Vienna, pp. 280–296.

[47] Cordon, O., del Jesus, M. J., Herrera, F., and Lozano, M. (1999). MOGUL: A methodology to obtaingenetic fuzzy rule-based systems under the iterative rule learning. Int. Journal of Intelligent Systems,to appear.

45

Page 46: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

[48] Herrera, F., Lozano, M., and Verdegay, J. L. (1998). A learning process for fuzzy control rules usinggenetic algorithms. Fuzzy Sets and Systems 100, pp. 143–158.

[49] Baker, J. E. (1987). Reducing bias and inefficiency in the selection algorithm. Proc. of the 2nd Int.Conf. on Genetic Algorithms, Lawrence Erlbaum (Hillsdale, NJ), pp. 14–21.

[50] Herrera, F., Lozano, M., and Verdegay, J. L. (1995). Tuning fuzzy controllers by genetic algorithms.Int. Journal of Approximate Reasoning 12, pp. 299–315.

[51] Ishibuchi, H., Nozaki, K., Yamamoto, N., and Tanaka, H. (1993). Selecting fuzzy if-then rules forclassification problems using genetic algorithms. IEEE Trans. on Fuzzy Systems 3(3), pp. 260–270.

[52] Herrera, F., Lozano, M., and Verdegay, J. L. (1998). Tackling real-coded genetic algorithms: Operatorsand tools for behavioural analysis. Artificial Intelligence Review 12, pp. 265–319.

[53] Sanchez, L. (1997). Interval-valued GA-P algorithms. Technical Report, Dept. of Computer Science,University of Oviedo, Oviedo (Spain).

[54] Sanchez, L. (1997). Study of the Asturias rural and urban low voltage network. Technical Report,Hidroelectrica del Cantabrico Research and Development Department (in spanish), Asturias (Spain).

[55] Cordon, O., Herrera, F., and Sanchez, L. (1999). Solving electrical distribution problems using hybridevolutionary data analysis techniques. Applied Intelligence 10, pp. 5–24.

[56] Haykin, S. (1994). “Neural Networks.” Macmillan.

[57] Krose, B. J. A., and van der Smagt, P. D. (1993). “An introduction to neural networks.” University ofAmsterdam.

[58] Muller, B., and Reinhardt, J. (1990). “Neural networks. An introduction.” Springer-Verlag.

[59] Sim, K. P. (1989). “Artificial Neural Systems.” Pergamon Press.

[60] Hebb, D. (1949). “Organization of Behaviour.” John Wiley & Sons.

[61] Widrow, B., and Hoff, M. (1960). Adaptive switching circuits. WESCON Convention Record 4, pp.96–104.

[62] Minsky, M., and Papert, S. (1969). “Preceptrons.” MIT Press.

[63] Rumelhart, D. E., Hinton, G. E., and Williams, R. J. (1986). Learning representations by backpropa-gation errors. Nature 323, pp. 533–536.

[64] Nauck, D., Klawonn, F., and Kruse, R. (1997). “Fundations of neuro-fuzzy systems.” John Wiley &Sons.

[65] Pedrycz, W., Card, H.C. (1992). Linguistic interpretation of self-organizing maps. Proc. of the 1st IEEEInt. Conf. on Fuzzy Systems, San Diego (CA), pp. 371–378.

[66] Nauck, D., Klawonn, F., and Kruse, R. (1996). “Designing neuro-fuzzy systems through backpropaga-tion”. In “Fuzzy modelling: Paradigms and practice” (W. Pedrycz, Ed.). Kluwer Academic Press.

[67] Brown, M., and Harris, C. (1994). “Neurofuzzy adaptive modeling and control.” Prentice-Hall.

[68] Harris, C. J., Moore, C. G., and Brown, M. (1993). “Intelligent control: Aspects of fuzzy logic andneural nets.” In “Robotics and automated systems”, Vol. 6. World Scientific.

[69] Holland, J. H. (1975). “Adaptation in natural and artificial systems.” Ann arbor: The University ofMichigan Press.

46

Page 47: Techniques for Learning and Tuning Fuzzy Rule-Based ...decsai.ugr.es/~casillas/downloads/papers/casillas-bc01-academic00… · II Fuzzy Rule-Based Systems In a very wide sense, an

[70] Goldberg, D. E. (1989). “Genetic algorithms in search, optimization and machine learning.” AddisonWesley.

[71] Michalewicz, Z. (1996) “Genetic algorithms + data structures = evolution programs.” Springer-Verlag.

[72] Grefensette, J. J. (Ed.) (1994). “Genetic algorithms for machine learning.” Kluwer Academic Press.

[73] Holland, J. H., and Reitman, S. (1978). Cognitive systems based on adaptive algorithms. In “Pattern-Directe Inference Systems” (D. A. Waterman and F. Hayes-Roth, Eds.). Academic Press.

[74] Smith, S. F. (1980). A learning system based on genetic adaptive algorithms. Ph. D. Thesis, Universityof Pittsburgh.

[75] Bonarini, A. (1996). Evolutionary learning of fuzzy rules: competition and cooperation. In “FuzzyModelling: Paradigms and Practice” (W. Pedrycz, Ed.), pp. 265–283. Kluwer Academic Press.

[76] Herrera, F., and Verdegay, J. L. (Eds.) (1997). “Genetic algorithms and soft computing.” Physica-Verlag.

[77] Herrera, F. (Ed.) (1997). Special issue on genetic fuzzy systems for control and robotics. Int. Journalof Approximate Reasoning 17(4).

[78] Herrera, F., Magdalena, L. (Eds.) (1998). Special issue on genetic fuzzy systems. Int. Journal of Intel-ligent Systems 13(10).

47