A fuzzy-filtered grey network technique for system state forecasting

Post on 25-Jan-2017

216 views 1 download

Transcript of A fuzzy-filtered grey network technique for system state forecasting

Soft ComputDOI 10.1007/s00500-014-1281-1

FOCUS

A fuzzy-filtered grey network technique for systemstate forecasting

Dezhi Li · Wilson Wang · Fathy Ismail

© Springer-Verlag Berlin Heidelberg 2014

Abstract A fuzzy-filtered grey network (FFGN) techniqueis proposed in this paper for time series forecasting and mate-rial fatigue prognosis. In the FFGN, the fuzzy-filtered rea-soning mechanism is proposed to formulate fuzzy rules cor-responding to different data characteristics; grey models areused to carry out short-term forecasting corresponding todifferent rules. A novel hybrid training method is proposedto adaptively update model parameters and improve train-ing efficiency. The effectiveness of the developed FFGN isdemonstrated by a series of simulation tests. It is also imple-mented for material fatigue prognosis. Test results show thatthe developed FFGN predictor can capture data characteris-tics effectively and forecast data trend accurately.

Keywords Fuzzy-filtered neural networks · Grey model ·Material fatigue prognosis

1 Introduction

System state forecasting is a process to predict future statesof a dynamic system based on its past observations. Reliableforecasting information is very useful in many real-worldapplications such as weather forecasting (Jiang et al. 2013;Liu et al. 2013) and stock market prediction (Ticknor 2013;

Communicated by E. Lughofer.

D. Li · F. IsmailDepartment of Mechanical and Mechatronics Engineering,University of Waterloo, Waterloo, ON N2L 3G1, Canadae-mail: d45li@uwaterloo.ca; fmismail@uwaterloo.ca

W. Wang (B)Department of Mechanical Engineering, Lakehead University,Thunder Bay, ON P7B 5E1, Canadae-mail: wwang3@lakeheadu.ca

Kazem et al. 2013). In industrial applications, predictor canbe used to foresee the future health state of a machine andforecast the remaining useful life of damaged equipment.Reliable prognostic information can be applied to machin-ery health condition monitoring, production quality control,and predictive maintenance scheduling rather than periodi-cally shutdown machines for manual inspection as commonlypracticed in a wide array of industries.

The classical time series forecast tools are based on sto-chastic models (Pourahmadi 2001). However, it is difficult toderive accurate analytical models for many complex dynamicsystems (Li and Lee 2005). The alternative approach is basedon the use of data-driven paradigms such as neural networks(NNs) (Wu and Liu 2012; Li et al. 2013b; Kulkarni et al.2013) and neuro-fuzzy (NF) systems (Svalina et al. 2013; Wuet al. 2014; Wang et al. 2012). Although these data-driven par-adigms do not require analytical system models, their reason-ing structure may become clumsy when the number of inputsbecome large. Bodyanskiy and Viktorov (2009a) presents acascade neo-fuzzy system with online learning algorithm;however, the singleton (a single weight) is used for each ruleoperation in the consequent, which may limit its capacity tomodel complex data features. The cascade growing NNs arepresented to learn data characteristics using quadratic neu-rons (Bodyanskiy et al. 2009b); however, its reasoning struc-ture is opaque to users. In addition, the online learning algo-rithms used in these two cascade systems may cause trainingprocess to be trapped in local optima, because both the gra-dient method and least square estimate are local searchingalgorithms and are sensitive to initial conditions.

The authors’ research team has proposed an enhancedfuzzy filtered neural network (EFFNN) (Li et al. 2013a) fortime series forecasting. The EFFNN aims to extract somerepresentative features from large input data sequence, soas to induce relatively simple reasoning architecture by the

123

D. Li et al.

use of fuzzy filters. The fuzzy filters in EFFNN can com-bine duplicated features and properly address sample vari-ations. Although EFFNN has several advantages over clas-sical NFs using fuzzy filters, it may not fully utilize datacharacteristics because of its limitation in modelling com-plex systems’ dynamics using simple linear functions in itsconsequent reasoning so that forecasting results may not besufficiently accurate in some applications.

The grey model method proposed by Deng (1982) can beused for time series prediction based on limited data informa-tion, which has been used in many applications such as airlinepassenger growth prediction (Benitez et al. 2013), financialindices prediction (Chen et al. 2010), and polluted gas emis-sion prediction (Lin et al. 2011). However, the grey modelsare mainly used for short-term prediction because of their rel-atively simple modelling structure. Their prediction accuracymay degrade when training data size becomes large (Wu etal. 2013). Furthermore, a grey model cannot fully utilize thewhole historical data, which may limit its prediction accu-racy.

To tackle the aforementioned problems in the EFFNN andthe grey model, a fuzzy filtered grey network (FFGN) tech-nique is proposed in this work for time series forecasting.The fuzzy filters are used in the antecedent part of the FFGNto partition the data space. The grey models are adopted toexplore the short-term data relationship corresponding to dif-ferent subspaces. The proposed FFGN technique is new in thefollowing aspects: (1) it uses a composite network architec-ture with fuzzy filters and grey models to improve predictionaccuracy. The fuzzy filters are used to generate rules, andthe grey models are applied to conduct short-term predictionfor each rule; (2) a new hybrid training method is proposedto adaptively optimize model parameters offline and improvetraining efficiency; (3) the proposed predictor is implementedfor material fatigue prognosis.

The remainder of this paper is organized as follows. Thedeveloped FFGN is described in Sect. 2. The proposed hybridtraining scheme is discussed in Sect. 3. In Sect. 4, the effec-tiveness of the proposed FFGN predictor is first examined bysimulations, and then it is implemented for material fatigueprognosis. Finally, some concluding remarks of this studyare summarized in Sect. 5.

2 Fuzzy filtered grey network system

The EFFNN can aggregate a large number of inputs into afew fuzzy channels; however, its capacity in capturing systemdynamics may be limited due to its simplified linear modelsused in the consequent part of the network. The proposedFFGN employs the grey models to enhance the modellingability of the EFFNN so as to further improve predictionaccuracy.

2.1 Discussion of the classical grey models

Grey models conduct forecast based on a small set of mostrecent data. The data to be used in the grey model must bepositive (Deng 1989) and the data are collected with a fixedsampling frequency (Kayacan et al. 2010). In grey modeltheory, GM(n, m) represents a grey model, where n is theorder of the difference equation and m denotes the numberof variables. GM(1, 1) is the most widely used grey modelbecause of its simple structure.

Given a time series data set, X0 = {x0(1), x0(2), . . . , x0

(q)}, where X0 is a non-negative sequence and q is the samplesize of the data. To fit a GM(1, 1) model to the given timeseries, the accumulating generation operator (AGO) is usedto formulate a monotonically increasing data sequence X1:

X1 = {x1(1), x1(2), . . . , x1(q)} (1)

where

x1(k) =k∑

i=1

x0(i), k = 1, 2, . . . , q. (2)

The generated mean sequence Z1 of X1 is defined as

Z1 = {(z1(1), z1(2), . . . , z1(q)} (3)

where z1(k) is the mean value of two adjacent data, and iscomputed as

z1(k) = 1

2x1(k)+ 1

2x1(k + 1), k = 1, 2, . . . , q − 1. (4)

The difference equation of GM(1, 1) will be

x0(k)+ αz1(k) = β (5)

The whitening equation is given as

dx1(t)

dt+ αx1(t) = β (6)

The parameter set {a, β} can be calculated from (5) using theleast square estimate (LSE) (Pratama et al. 2013; Pratamaet al. 2014; Rubio 2014; Rubio and Pérez-Cruz 2014;Buchachia 2012),

[α β]T = (AT A)−1 AT B (7)

where

A =

⎢⎢⎢⎣

−z1(2) 1−z1(3) 1...

...

−z1(q) 1

⎥⎥⎥⎦

B = [x0(2), x0(3), . . . , x0(q)]T

123

A fuzzy-filtered grey network technique

According to (6), x1(k + 1) can be computed as

x1(k + 1) =[

x0(1)− β

α

]e−αk + β

α(8)

The predicted value of the original data at time k + 1 can bedetermined using the inverse AGO,

x0(k + 1) =[

x0(1)− β

α

]e−αk(1 − eα) (9)

2.2 The proposed FFGN technique

The architecture of the proposed FFGN is illustrated in Fig. 1.It is a 4-layer feedforward network. Layer 1 is the inputlayer with each node representing an input data {Xi }, i =1, 2, . . . , n. The synaptic weights between layer 1 and layer2 are unity. The fuzzy filters are employed in layer 2.

Gaussian functions (Pratama et al. 2013; Rubio 2014;Rubio and Pérez-Cruz 2014; Buchachia 2012) have beenemployed as bandpass filters to preprocess input data, whichare given by

μB(xi ) = e− 1

2 (xi −ψiωi

)2

, (10)

where ψi and ωi represent the center and spread of theGaussian function, respectively. Sigmoid functions (Pérez-Cruz et al. 2013; Bordignon and Gomide 2014; Rubio 2012),as given in the following equation, are used in this case aslowpass and highpass filters to ensure that fuzzy rules cancover the whole feasible rule space:

μB(xi ) = 1

1 + e−εi (xi −γi ), (11)

where γi and εi are the center and spread of the Sigmoidfunctions, respectively. If εi < 0, μB(xi ) represents a low-pass filter; otherwise, if εi > 0, μB(xi ) corresponds to ahighpass filter. The initial conditions of the Gaussian andSigmoid functions are shown in Fig. 2.

T1

X1

X2

X3

Xn

YC

.

.

.

.

.

.

.

.

.

.

.

Layer 1 Layer 2 Layer 3 Layer 4

T2

T3

Tl

.

.

.

Fig. 1 The architecture of the proposed FFGN

0 0.2 0.4 0.6 0.8 10

0.2

0.4

0.6

0.8

1

Fig. 2 The initial conditions of the Gaussian functions and Sigmoidfunctions

The node output in layer 2 is formulated as

O j =∑

i μB(xi ) f (xi )∑i μB(xi )

, j = 1, 2, . . . , l; i = 1, 2, . . . , n

(12)

where l is the number of nodes in layer 2 and n is the numberof inputs. The output of each node O j denotes the firingstrength of the corresponding rule. f (xi ) is the distribution ofthe training data, which can be calculated using kernel densityestimation method (Bowman and Azzalini 1997). Each nodein layer 2 triggers only one rule/node in layer 3. The firingstrength O j represents the contribution of the correspondingrule/node in layer 3 to the final output.

Layer 3 represents the rule base. Each node in layer 3receives its inputs from the corresponding nodes in layer 2.The operation of each node/rule in layer 3 is a combinationof a grey model and a linear function represented by

Tj = G j + L j , (13)

where

G j =[

x(i − r + 1)− v1, j

v2, j

]e−αi (1 − eα) (14)

L j = w1, j x(i)+ w2, j x(i − 1)+ w3, j , (15)

where v1, j and v2, j are the parameters of grey models; wk, j

are the linear weights (k = 1, 2, 3) and j = 1, 2, . . . , l, and lis the number of nodes in layer 3. The data sequence input togrey models is {x(i − r + 1), x(i − r + 2), . . . , x(i)}, wherer is the number of the data sequence.

Defuzzification is performed in layer 4. If centroiddefuzzification method is used, the overall output will be

Y =∑

j O j Tj∑j O j

(16)

The parameters of fuzzy filters and weights of rules in layer 3are optimized by the proposed training technique discussedin the following section.

123

D. Li et al.

To improve prediction performance, the input data are par-titioned into subspaces by using fuzzy filters (10)–(11), andeach subspace will be addressed using a grey model in (14)and a linear function in (15). The output of each node in layer2 is the firing strength to trigger the corresponding grey modelin layer 3. When the firing strength is high, the correspond-ing grey model contributes more to the final prediction. Eachgrey model only processes a small size of the data sequence(e.g., r = 4 in this case) so as to conduct short-term prediction.

3 The proposed hybrid training algorithm

The proposed hybrid training algorithm is used offline tooptimize the system parameters of the developed FFGN.In system training, the genetic algorithm (GA) can be usedfor global system optimization. However, the classical GAusually suffers slow convergence. To overcome this prob-lem, a two-stage strategy is used in this work for optimiza-tion: first, the GA is used to locate the area containing theglobal optima. The second stage is to apply the conjugateLevenberg-Marquardt (CLM) technique to quickly find theoptima. The CLM has been proved to be an efficient super-vised local training algorithm, which can integrate the advan-tages of both conjugate gradient algorithm and Levenberg-Marquardt method. The details of the GA and CLM can befound from the authors’ prior work in (Li and Wang 2011,Li et al. 2013a).

Specifically, in the first stage, a real-coded GA is applied toprimarily optimize the fuzzy filters’ parameters [i.e., {ψi , ωi }in (10) and {γi , εi } in (11)], and linear weights wk, j in (15).A set of parameters {ψi , ωi , γi , εi , wk, j } corresponding toFFGN is considered as an individual at each generation ofGA optimization. The fitness of each set of parameters (orindividual) is tested by computing the root mean square valueof the prediction errors at all time steps using training dataset. At each time step, the firing strengths in (12) are cal-culated, and the grey model corresponding to the largest fir-ing strength is updated using least square estimate (LSE).The other grey models remain unchanged except the termx(i − r + 1) in (14).

In the second stage, the parameter optimization is carriedout using both CLM and LSE to accelerate training conver-gence. At each time step, the CLM is first used to optimizethe nonlinear system parameters {ψi , ωi } in (10) and {γi , εi }in (11). If the j th node in layer 2 generates the highest firingstrength, only the parameters of the j th grey model, {α β} in(14), are updated using the LSE in (7). Meanwhile the termx(i − r + 1) in (14) is updated for all grey models, wherei is a time index. The training data for the j th grey modelis the most recent r data in the time series. The predictionerror of the grey model G j is compensated by a linear func-tion Tj in (15). Only one grey model and the corresponding

Grey models (LSE)

Satisfying stop criterion ?

End

N

Fuzzy filters & linear weighs (GA)

Fuzzy filters (CLM)Linear weights

(LSE)

Y

Satisfying stop criterion ?

N

Y

Initialization

Grey models (LSE)

Fig. 3 The flowchart of the hybrid training process

linear function will be updated at each time step so that thetraining efficiency can be improved. The hybrid method canreduce the search space dimension compared to an individualtraining method, and reduce the possibility of trapping dueto local optima (Wang 2008).

The flowchart of the two-stage training process is illus-trated in Fig. 3. Specific implementation is discussed as fol-lows:

1. A population of m individuals is randomly generated.Each individual is a parameter vector of the fuzzy fil-ters’ parameters {ψi , ωi , γi , εi } in (10)–(11) and linearweights {wk, j } in (15).

2. The primary parameters are optimized by using GA toexplore the parameters space and search for the area con-taining the global optima. At each time step, the parame-ters of the grey model corresponding to the largest firingstrength are updated using the LSE.

3. Once the specified maximum number of generation isreached (stop criterion), the GA operation is terminated.Otherwise, step (2) is repeated until the stop criterion issatisfied. The individual with the least training error ischosen for further optimization.

4. After GA training, detailed search is conducted usinga hybrid method of CLM and LSE: the parameters offuzzy filters {ψi , ωi , γi , εi } in (10)–(11) are optimizedby CLM; the parameters of the grey model {v1, j , v2, j }in (14) and linear weights {wk, j } in (15) are updated byLSE at each training iteration.

5. The stop criterion of CLM and LSE training is eithera specified maximum number of iteration or a threshold

123

A fuzzy-filtered grey network technique

of training errors over two successive training processes.Otherwise, step (4) is repeated until the stop criterion isachieved.

4 Performance evaluation and applications

The effectiveness of the proposed FFGN predictor is firstexamined by simulation tests using three common bench-mark data sets, and then it is implemented for material fatigueprognosis.

To compare the performance of the proposed FFGN, thefollowing related techniques are applied:

1. EFFNN is used to compare the proposed reasoning struc-ture of the FFGN, which does not contain grey modelsin its architecture. The EFFNN has the same number ofnodes in layers 1-4 (i.e., 10-5-5-1) as in the FFGN inFig. 1, but it does not contain grey models. It is used tocompare the proposed reasoning structure of the FFGN.

2. DENFIS method (Kasabov and Song 2002) is used tocompare the prediction accuracy of the FFGN. It formsthe antecedent of rules using evolving clustering method,and trains consequent of the rules using LSE.

3. FFGN-F is used to compare the suggested hybrid trainingstrategy in the FFGN technique. FFGN-F has the samestructure as in the FFGN, but its grey models are updatedtogether with nonlinear parameters in fuzzy filters ratherthan optimized independently.

The prediction performance is evaluated using normalizedmean square error (NRMSE). The NRMSE is given by

NRMSE = RMSE

xmax − xmin

=√∑n

i=1 (xi −xdi )

2

n

xmax − xmin, (17)

where xi is the predicted value; xdi is the real value; n is the

data length; xmax is the maximum of xi ; xmin is the minimumof xi .

4.1 Mackey–Glass data forecast

The Mackey–Glass data set (Farmer 1982; Bodyanskiy andViktorov 2009a; Li et al. 2013b; Wang et al. 2012; Kasabovand Song 2002) is a commonly used benchmark data fortime series forecasting due to its chaotic, aperiodic and non-convergence properties, which is from the following equa-tion:

ds(t)

dt= 0.2s(t − τ)

1 + s10(t − τ)− 0.1s(t) (18)

0 0.5 10

0.2

0.4

0.6

0.8

1

Interval

Dis

trib

utio

n

Fig. 4 The distribution of Mackey–Glass training data

Table 1 The mean square error of Mackey–Glass data forecastingbased on different predictors

NRMSE Trainingerror

Predictionerror

Rules Executiontime (s)

EFFNN 0.0489 0.0507 5 58.507

DENFIS 0.0074 0.0079 24 0.594

FFGN-F 0.0072 0.0073 5 53.892

FFGN 0.0061 0.0066 5 50.174

In this simulation test, a data set is generated with the initialcondition of τ = 30, s(0) = 1.2, dt = 1 and s(t) = 0 fort < 0. 1,000 data are used for training, whose distribution isshown in Fig. 4. Another 1,000 data are used for testing. Themaximum number of generations in GA is set at 100.

The training error, prediction error, number of rules andexecution time of Mackey–Glass data forecasting are listed inTable 1. From Table 1, it can be seen that the proposed FFGN(as well as FFGN-F) outperforms EFFNN and DEFNISbecause of its more efficient modelling capability using greymodels. The grey models in FFGN and FFGN-F are strongerin capturing data features than linear models in DENFIS sothat FFGN and FFGN-F can achieve less NRMSE than DEN-FIS with fewer rules. In addition, DENFIS trains its nonlin-ear parameters only using clustering method, which may alsolead to inaccurate modelling. The DENFIS employs an onlinelearning algorithm; however, EFFNN, FFGN-F and FFGNutilize the offline training with GA, CLM and LSE. ThusDENFIS consumes the least execution time among all pre-dictors. The FFGN performs better than the FFGN-F becauseit upgrades grey models by extracting data characteristicsfrom the most recent r data samples rather than only the cur-rent data sample at each training epoch. So the modellingaccuracy can be improved.

4.2 Lorenz data forecast

Lorenz attractor (Chandra and Zhang 2012; Ardalani-Farsaand Zolfaghari 2010) is another commonly used chaotic, ape-

123

D. Li et al.

0 0.2 0.4 0.6 0.8 10

0.2

0.4

0.6

0.8

1

Interval

Dis

trib

utio

n

Fig. 5 The distribution of Lorenz training data

Table 2 The mean square error of Lorenz data forecasting based ondifferent predictors

NRMSE Trainingerror

Predictionerror

Rules Executiontime (s)

EFFNN 0.0711 0.0738 5 69.263

DENFIS 0.0381 0.0395 33 0.788

FFGN-F 0.0337 0.0358 5 67.194

FFGN 0.0244 0.0252 5 63.128

riodic data set, which can be described by the following equa-tions:

dx(t)

dt= σ(y(t)− x(t)) (19)

dy(t)

dt= x(t)(ρ − z(t))− y(t) (20)

dz(t)

dt= x(t)y(t)− βx(t)) (21)

The system parameters are selected as σ = 10, β = 8/3, andρ = 28. 1,000 data are used for training, with distributiondescribed in Fig. 5. Another 1,000 data samples are used fortesting. The maximum number of generations in GA is set at100.

Table 2 summarizes the test results of Lorenz data predic-tion. It can be seen from Table 2 that both FFGN-F and FFGNgenerate less NRMSE than EFFNN because of the effectivemodelling using grey models. DENFIS trains nonlinear para-meters using the clustering algorithm, which may not reachhigh accuracy. The FFGN-F and FFGN outperform DEN-FIS because their nonlinear parameters are updated using ahybrid training method with GA and CLM rather than a clus-tering algorithm. The FFGN generates the highest accuracy(with least NRMSE) because of its advanced grey modellingupdate mechanism.

0 0.5 10

0.2

0.4

0.6

0.8

1

Interval

Inte

nsity

Fig. 6 The distribution of Sunspot training data

Table 3 The mean square error of sunspot data prediction based ondifferent predictors

NRMSE Trainingerror

Predictionerror

Rules Executiontime (s)

EFFNN 0.1349 0.1654 5 10.673

DENFIS 0.2135 0.3013 32 0.190

FFGN-F 0.0826 0.1230 5 9.712

FFGN 0.0868 0.1070 5 8.798

Fig. 7 The experimental setup for material fatigue testing

4.3 Sunspot data forecast

Sunspot activity record (Chandra and Zhang 2012; Ardalani-Farsa and Zolfaghari 2010; Li et al. 2013a) is another com-monly used benchmark data set to evaluate the performanceof predictors. 262 data sets are used for test in this case. Thefirst 190 sunspot data (from years 1750 to 1939) are used for

123

A fuzzy-filtered grey network technique

200 400 600 800 1000 1200 1400 1600

0

10

20

30

40

Rel

ativ

e V

olta

ge

200 400 600 800 1000 1200 1400 1600

0

10

20

30

40

Rel

ativ

e V

olta

ge

200 400 600 800 1000 1200 1400 1600

0

10

20

30

40

Rel

ativ

e V

olta

ge

200 400 600 800 1000 1200 1400 1600

0

10

20

30

40R

elat

ive

Vol

tage

Time Duration in Cycles

(a)

(b)

(c)

(d)

Fig. 8 The performance of material fatigue prognosis in the first sce-nario. The blue solid line is the real data to estimate; the red dotted lineis the prediction performance of different schemes: a by EFFNN; b byDENFIS; c by FFGN-F; d by FFGN (color figure online)

training, and the remaining 72 data pairs (from years 1940to 2011) are used for testing. The distribution of the sunspottraining data is shown in Fig. 6.

The forecasting performance of the related predictors issummarized in Table 3. The DEFNSI has larger NRMSE thanEFFNN because its rules cannot be effectively extracted withthe clustering algorithm to improve prediction accuracy. TheFFGN-F predictor outperforms the EFFNN due to its effec-tive employment of the grey models in short-term prediction.But the developed FFGN generates the best forecasting per-formance and outperforms the other three counterparts dueto its enhanced modelling mechanism and its effective hybridtraining strategy.

4.4 Real-world application: material fatigue prognosis

As an application example, the proposed FFGN predictor isimplemented for material fatigue prognosis. Material fatigue

200 400 600 800 1000 1200 1400 1600

-20

0

20

40

Rel

ativ

e V

olta

ge200 400 600 800 1000 1200 1400 1600

-20

0

20

40

Rel

ativ

e V

olta

ge

200 400 600 800 1000 1200 1400 1600

-20

0

20

40

Rel

ativ

e V

olta

ge

200 400 600 800 1000 1200 1400 1600

-20

0

20

40

Rel

ativ

e V

olta

ge

Time Duration in Cycles

(a)

(b)

(c)

(d)

Fig. 9 The performance of material fatigue prognosis in the secondscenario. The blue solid line is the real data to estimate; the red dottedline is the prediction performance of different schemes: a by EFFNN;b by DENFIS; c by FFGN-F; d by FFGN (color figure online)

tests are usually very time-consuming. So an efficient pre-dictor is very helpful in industries to forecast the trend ofthe material’s dynamic property and estimate the material’sfatigue properties without using full-scale fatigue testing. Asa demonstration, a 3-mm thick aluminum specimen is testedin this case using the experimental setup shown in Fig. 7. Bothends of the specimen are clamped to the testing machine. Asmall hole is drilled in the middle of the specimen to ini-tiate crack propagation. The crack dimension is measuredeither directly (by calipers) or indirectly (by relative voltagemeasurement). The testing frequency is set at 20 Hz. Themeasurement is taken every 3,000 load cycles. The test isbased on an indirect measurement.

In material fatigue prognosis test, experiments are car-ried out corresponding to three different initial crack dimen-sions. In each testing scenario with a specific initial crackdimension, the tests are repeated over six times (i.e., six data

123

D. Li et al.

200 400 600 800 1000 1200 1400 1600 1800

-100

-50

0

50

100

Rel

ativ

e V

olta

ge

200 400 600 800 1000 1200 1400 1600 1800

-100

-50

0

50

100

Rel

ativ

e V

olta

ge

200 400 600 800 1000 1200 1400 1600 1800

-100

-50

0

50

100

Rel

ativ

e V

olta

ge

200 400 600 800 1000 1200 1400 1600 1800

-100

-50

0

50

100R

elat

ive

Vol

tage

Time Duration in Cycles

(a)

(b)

(c)

(d)

Fig. 10 The performance of material fatigue prognosis in the thirdscenario. The blue solid line is the real data to estimate; the red dottedline is the prediction performance of different schemes: a by EFFNN;b by DENFIS; c by FFGN-F; d by FFGN (color figure online)

sets). Among the six data sets in each scenario, two datasets are randomly selected for training and another data setis randomly chosen for testing. The maximum number ofgenerations in GA is set at 100. Figures 8, 9 and 10 showthe one-step-ahead forecasting results of three test scenarios,respectively, using these three predictors. The correspondingtest results are summarized in Tables 4, 5 and 6.

By comparing graph (a) and graph (c) in Figs. 8, 9 and 10,it can be seen that the FFGN-F predictor can track materialfatigue propagation trend more accurately than the EFFNNpredictor, because the grey models in FFGN-F can extractinformation from most recent data sequence to improve pre-diction accuracy. Comparing graph (b) and graph (c) in Figs.8, 9 and 10, it is seen that the FFGN-F predictor is superiorto the DENFIS predictor because of the more accurate mod-elling of the grey models, and the more effective nonlinearparameter training method. On the other hand, the FFGN

Table 4 The mean square error of material fatigue prognosis in the firstscenario based on different predictors

NRMSE Trainingerror

Predictionerror

Rules Executiontime (s)

EFFNN 0.0936 0.0120 5 51.523

DENFIS 0.0773 0.0076 87 0.405

FFGN-F 0.0736 0.0036 5 50.663

FFGN 0.0627 0.0011 5 48.833

Table 5 The mean square error of material fatigue prognosis in thesecond scenario based on different predictors

NRMSE Trainingerror

Predictionerror

Rules Executiontime (s)

EFFNN 0.0853 0.0174 5 58.380

DENFIS 0.0472 0.0067 27 0.114

FFGN-F 0.0447 0.0025 5 56.376

FFGN 0.0438 0.0010 5 54.743

Table 6 The mean square error of material fatigue prognosis in thethird scenario based on different predictors

NRMSE Trainingerror

Predictionerror

Rules Executiontime (s)

EFFNN 0.0695 0.0271 5 57.036

DENFIS 0.0729 0.0424 59 0.224

FFGN-F 0.0631 0.0037 5 53.164

FFGN 0.0584 0.0017 5 51.395

predictor (graph (d) in Figs. 8, 9 and 10) gives better predic-tion results than the FFGN-F predictor, because its adaptivegrey model training strategy can improve the dominant greymodel prediction accuracy as well as the overall forecastingperformance. It can catch the dynamic characteristics of thematerial crack propagation trend accurately.

5 Conclusion

A fuzzy-filtered grey network, FFGN, technique is developedin this work for time series forecasting and material fatigueprognosis. In the FFGN, the fuzzy filters extract features fromdata. The feature processing is enhanced using grey modelsto extract information from most recent data sequence. Theadaptive grey model training strategy can improve the train-ing efficiency of the FFGN. The effectiveness of the proposedFFGN predictor is verified using simulation of two bench-mark data sets. The proposed predictor is also implementedfor material fatigue prognosis. Test results have shown thatthe FFGN scheme is an efficient predictor. It is able to capture

123

A fuzzy-filtered grey network technique

the dynamic behaviour of the material properties quickly andtrack the crack propagation trend accurately.

Further research is under way to propose an analyticalmethod to properly select system orders. In addition, anonline training algorithm will be suggested to adaptively tuneFFGN model structures and optimize system parameters inreal-time.

Acknowledgments This work was partly funded by the Natural Sci-ences and Engineering Research Council of Canada (NSERC) andeMech Systems Inc.

References

Ardalani-Farsa M, Zolfaghari S (2010) Chaotic time series predictionwith residual analysis method using hybrid Elman-NARX neuralnetworks. Neurocomputing 73(13–15):2540–2553

Benitez RBC, Paredes RBC, Lodewijks G, Nabais JL (2013) Damptrend grey model forecasting method for airline industry. Expert SystAppl 40:4915–4921

Bowman AW, Azzalini A (1997) Applied smoothing techniques for dataanalysis. Oxford University Press Inc, New York

Bodyanskiy Y, Viktorov Y (2009a) The cascaded neofuzzy architectureand its on-line learning algorithm. Intell Process 9:110–116

Bodyanskiy Y, Viktorov Y, Pliss I (2009b) The cascade growing neuralnetwork using quadratic neurons and its learning algorithms for on-line information processing. Intell Inf Eng Syst 13:27–34

Bordignon F, Gomide F (2014) Uninorm based evolving neural net-works and approximation capabilities. Neurocomputing 127:13–20

Buchachia A (2012) Dynamic clustering. Evol Syst 3(3):133–134Chandra R, Zhang M (2012) Cooperative coevolution of Elman recur-

rent neural networks for chaotic time series prediction. Neurocom-puting 86(1):116–123

Chen C, Hsin P, Wu C (2010) Forecasting Taiwan’s major stock indicesby the Nash nonlinear grey Bernoulli model. Expert Syst Appl37:7557–7562

Deng J (1982) Control problems of grey systems. Syst Control Lett1:288–294

Deng J (1989) Introduction to grey system theory. J Grey Syst 1:1–24Farmer JD (1982) Chaotic attractors of an infinite-dimensional dynam-

ical system. Phys D: Nonlinear Phenomena 4:366–393Jiang Y, Song Z, Kusiak A (2013) Very short-term wind speed forecast-

ing with Bayesian structural break model. Renew Energy 50:637–647

Kasabov NK, Song Q (2002) DENFIS: dynamic evolving neural-fuzzyinference system and its application for time-series prediction. IEEETrans Fuzzy Syst 10(2):144–154

Kayacan E, Ulutas B, Kaynak O (2010) Grey system theory-based mod-els in time series prediction. Expert Syst Appl 37:1784–1789

Kazem A, Sharifi E, Hussain FK, Saberi M, Hussain OK (2013) Supportvector regression with Chaos-based firefly algorithm for stock marketprice forecasting. Appl Soft Comput 13(2):947–958

Kulkarni S, Simon SP, Sundareswaran K (2013) A spiking neural net-work (SNN) forecast engine for short-term electrical load forecast-ing. Appl Soft Comput 13(8):3628–3635

Li C, Lee H (2005) Gear fatigue crack prognosis using embeddedmodel, gear dynamic model and fracture mechanics. Mech Syst Sig-nal Process 19:836–846

Li DZ, Wang W (2011) An enhanced GA technique for system trainingand prognostics. Adv Eng Softw 42(7):452–462

Li D, Wang W, Ismail F (2013a) Enhanced fuzzy-filtered neural net-works for material fatigue prognosis. Appl Soft Comput 13:283–291

Li D, Wang W, Ismail F (2013b) Fuzzy neural network technique forsystem state forecasting. IEEE Trans Cybernet 43(5):1484–1494

Lin C, Liou F, Huang C (2011) Grey forecasting model for CO2 emis-sions: a Taiwan study. Appl Energy 88:3816–3820

Liu H, Tian H, Pan D, Li Y (2013) Forecasting models for wind speedusing wavelet, wavelet packet, time series and artificial neural net-works. Appl Energy 107:191–208

Pérez-Cruz JH, Rubio JJ, Pacheco J, Soriano E (2013) State estimationin MIMO nonlinear systems subject to unknown deadzones usingrecurrent neural networks. Neural Comput Appl. http://link.springer.com/article/10.1007%2Fs00521-013-1533-5

Pourahmadi M (2001) Foundations of time series analysis and predic-tion theory. Wiley, Hoboken

Pratama M, Anavatti SG, Lughofer E (2013) GENFIS: towards an effec-tive localist network. IEEE Trans Fuzzy Syst. http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=6521390&tag=1

Pratama M, Anavatti SG, Angelov PP, Lughofer E (2014) PANFIS: anovel incremental learning machine. IEEE Trans Neural Netw Learn-ing Syst 25(1):55–68

Rubio JJ (2014) Evolving intelligent algorithms for the modelling ofbrain and eye signals. Appl Soft Comput 14(Part B):259–268

Rubio JJ, Pérez-Cruz JH (2014) Evolving intelligent system for themodelling of nonlinear systems with dead-zone input. Appl SoftComput 14(Part B):289–304

Rubio JJ (2012) Modified optimal control with a backpropagation net-work for robotic arms. IET Control Theory Appl 6(14):2216–2225

Svalina I, Galzina V, Lujic R, Simunovic G (2013) An adaptive network-based fuzzy inference system (ANFIS) for the forecasting: the caseof close price indices. Expert Syst Appl 40(15):6055–6063

Ticknor JL (2013) A Bayesian regularized artificial neural network forstock market forecasting. Expert Syst Appl 40(14):5501–5506

Wang W (2008) An enhanced diagnostic system for gear system moni-toring. IEEE Trans Syst Man Cybernet Part B: Cybernet 38:102–112

Wang W, Li DZ, Vrbanek J (2012) An evolving neuro-fuzzy techniquefor system state forecasting. Neurocomputing 87:111–119

Wu J, Liu J (2012) A forecasting system for car fuel consumption usinga radial basis function neural network. Expert Syst Appl 39(2):1883–1888

Wu P, Hung Y, Lin Z (2014) Intelligent forecasting system based onintegration of electromagnetism-like mechanism and fuzzy neuralnetwork. Expert Syst Appl 41(6):2660–2677

Wu L, Liu S, Yao L, Yan S (2013) The effect of sample size on the greysystem model. Appl Math Model 37:6577–6583

123