Iterative Learning Control for Varying Tasks: Achieving ...In this paper discrete-time, linear,...

6
Iterative Learning Control for Varying Tasks: Achieving Optimality for Rational Basis Functions Jurgen van Zundert, Joost Bolder and Tom Oomen Abstract— Iterative Learning Control (ILC) can achieve su- perior tracking performance for systems that perform repeating tasks. However, the performance of standard ILC deteriorates dramatically when the task is varied. In this paper ILC is extended with rational basis functions to obtain excellent extrapolation properties. A new approach for rational basis functions is proposed where the iterative solution algorithm is of the form used in instrumental variable system identification algorithms. The optimal solution is expressed in terms of learning filters similar as in standard ILC. The proposed approach is shown to be superior over existing approaches in terms of performance by a simulation example. I. I NTRODUCTION For systems performing the same task over and over, the performance can be optimized by learning from previous executions. In Iterative Learning Control (ILC) [1]–[4], the repetitive behavior is exploited by updating the command signal using data from previous executions. ILC achieves optimal performance for a specific task only since the command signal is learned. For varying tasks the performance may severely deteriorate [5]. The main reason is that the command signal is not a function of the reference trajectory. In [6], the extrapolation properties are enhanced by constructing the task such that it comprises basis tasks that are learned in a training routine. Consequently, this approach is limited to tasks which can be constructed by a finite set of basis tasks. A more general approach is to parameterize the command signal in a set of basis functions [7], [8]. Examples of basis functions include polynomial basis functions [9]–[12], and more recently rational basis functions [13]. Rational basis functions are more general than polynomial basis functions, since the latter are recovered as a special case. The optimization associated with polynomial basis func- tions in ILC has an explicit analytic solution [3], whereas this is generally not the case for rational basis functions. In [13], an iterative solution based on Steiglitz-McBride is presented. The approach achieves fast convergence and is insensitive to local optima [14], however, the stationary point is not necessarily an optimum, as is well known in related system identification algorithms [15]. The authors are with the Eindhoven University of Technology, Department. of Mechanical Engineering, Control Systems Technology group, P.O. Box 513, 5600 MB Eindhoven, The Netherlands. E-mail: [email protected]. This research is supported by Oc´ e Tech- nologies, P.O. Box 101, 5900 MA Venlo, The Netherlands; and the Innova- tional Research Incentives Scheme under the VENI grant “Precision Motion: Beyond the Nanometer” (no. 13073) awarded by NWO (The Netherlands Organization for Scientific Research) and STW (Dutch Science Foundation). Although important contributions have been made by developing general rational basis functions in ILC, presently available optimization algorithms suffer from non-optimality or poor convergence properties. The aim of this paper is to show non-optimality of the pre-existing approach and develop a new solution algorithm for which the stationary point is always an optimum. The proposed approach has strong connections to instrumental variable system identification [16]. The contributions of this paper are threefold: I Non-optimality of the pre-existing approach in [13] for rational basis functions in ILC is illustrated. II A new iterative solution algorithm for rational basis functions in ILC is proposed and its optimality is shown which constitutes the main contribution of this paper. III It is shown that the proposed approach outperforms the pre-existing approach by a simulation example. The outline of this paper is as follows. In section II, the problem considered in this paper is introduced. The non- optimality of the pre-existing approach [13] is illustrated in section III. The proposed approach is presented in section IV. The iterative approaches are compared by use of a simula- tion example in section V, demonstrating that the proposed approach outperforms the pre-existing approach. Section VI contains conclusions. II. PROBLEM FORMULATION In this section the considered problem is defined. A. Notation In this paper discrete-time, linear, time-invariant (LTI), singe-input, single-output (SISO) systems are considered to facilitate the presentation. Let z C be the complex indeterminate and x(k) the signal x evaluated at time index k. Let h(l) be the impulse response of the system H(z). The output y(k) of the response of H(z) to input u is given by the [17] y(k)= l=-∞ h(l)u(k - l). Assuming u(k)=0 for k< 0 and k>N -1, then y[0] y[1] . . . y[N-1] | {z } y = h(0) h(-1) ... h(1-N) h(1) h(0) ... h(2-N) . . . . . . . . . . . . h(N-1) h(N-2) ... h(0) | {z } H u[0] u[1] . . . u[N-1] | {z } u , (1) with H the finite-time matrix representation of H(z), and u, y R N the input and output, respectively, with N Z + the trial length. Let j be the trial index and kx j k W := 2015 American Control Conference Palmer House Hilton July 1-3, 2015. Chicago, IL, USA 978-1-4799-8686-6/$31.00 ©2015 AACC 3570

Transcript of Iterative Learning Control for Varying Tasks: Achieving ...In this paper discrete-time, linear,...

Page 1: Iterative Learning Control for Varying Tasks: Achieving ...In this paper discrete-time, linear, time-invariant (LTI), singe-input, single-output (SISO) systems are considered to facilitate

Iterative Learning Control for Varying Tasks:Achieving Optimality for Rational Basis Functions

Jurgen van Zundert, Joost Bolder and Tom Oomen

Abstract— Iterative Learning Control (ILC) can achieve su-perior tracking performance for systems that perform repeatingtasks. However, the performance of standard ILC deterioratesdramatically when the task is varied. In this paper ILCis extended with rational basis functions to obtain excellentextrapolation properties. A new approach for rational basisfunctions is proposed where the iterative solution algorithm isof the form used in instrumental variable system identificationalgorithms. The optimal solution is expressed in terms oflearning filters similar as in standard ILC. The proposedapproach is shown to be superior over existing approaches interms of performance by a simulation example.

I. INTRODUCTION

For systems performing the same task over and over, theperformance can be optimized by learning from previousexecutions. In Iterative Learning Control (ILC) [1]–[4], therepetitive behavior is exploited by updating the commandsignal using data from previous executions.

ILC achieves optimal performance for a specific task onlysince the command signal is learned. For varying tasks theperformance may severely deteriorate [5]. The main reasonis that the command signal is not a function of the referencetrajectory. In [6], the extrapolation properties are enhancedby constructing the task such that it comprises basis tasksthat are learned in a training routine. Consequently, thisapproach is limited to tasks which can be constructed bya finite set of basis tasks. A more general approach is toparameterize the command signal in a set of basis functions[7], [8]. Examples of basis functions include polynomialbasis functions [9]–[12], and more recently rational basisfunctions [13]. Rational basis functions are more general thanpolynomial basis functions, since the latter are recovered asa special case.

The optimization associated with polynomial basis func-tions in ILC has an explicit analytic solution [3], whereasthis is generally not the case for rational basis functions.In [13], an iterative solution based on Steiglitz-McBride ispresented. The approach achieves fast convergence and isinsensitive to local optima [14], however, the stationary pointis not necessarily an optimum, as is well known in relatedsystem identification algorithms [15].

The authors are with the Eindhoven University of Technology,Department. of Mechanical Engineering, Control Systems Technologygroup, P.O. Box 513, 5600 MB Eindhoven, The Netherlands. E-mail:[email protected]. This research is supported by Oce Tech-nologies, P.O. Box 101, 5900 MA Venlo, The Netherlands; and the Innova-tional Research Incentives Scheme under the VENI grant “Precision Motion:Beyond the Nanometer” (no. 13073) awarded by NWO (The NetherlandsOrganization for Scientific Research) and STW (Dutch Science Foundation).

Although important contributions have been made bydeveloping general rational basis functions in ILC, presentlyavailable optimization algorithms suffer from non-optimalityor poor convergence properties.

The aim of this paper is to show non-optimality of thepre-existing approach and develop a new solution algorithmfor which the stationary point is always an optimum. Theproposed approach has strong connections to instrumentalvariable system identification [16]. The contributions of thispaper are threefold:

I Non-optimality of the pre-existing approach in [13] forrational basis functions in ILC is illustrated.

II A new iterative solution algorithm for rational basisfunctions in ILC is proposed and its optimality is shownwhich constitutes the main contribution of this paper.

III It is shown that the proposed approach outperforms thepre-existing approach by a simulation example.

The outline of this paper is as follows. In section II, theproblem considered in this paper is introduced. The non-optimality of the pre-existing approach [13] is illustrated insection III. The proposed approach is presented in section IV.The iterative approaches are compared by use of a simula-tion example in section V, demonstrating that the proposedapproach outperforms the pre-existing approach. Section VIcontains conclusions.

II. PROBLEM FORMULATION

In this section the considered problem is defined.

A. Notation

In this paper discrete-time, linear, time-invariant (LTI),singe-input, single-output (SISO) systems are considered tofacilitate the presentation.

Let z ∈ C be the complex indeterminate and x(k)the signal x evaluated at time index k. Let h(l) be theimpulse response of the system H(z). The output y(k)of the response of H(z) to input u is given by the [17]y(k) =

∑∞l=−∞ h(l)u(k− l). Assuming u(k) = 0 for k < 0

and k > N−1, then

y[0]y[1]

...y[N−1]

︸ ︷︷ ︸y

=

h(0) h(−1) ... h(1−N)h(1) h(0) ... h(2−N)

......

. . ....

h(N−1) h(N−2) ... h(0)

︸ ︷︷ ︸H

u[0]u[1]

...u[N−1]

︸ ︷︷ ︸u

, (1)

with H the finite-time matrix representation of H(z), andu, y ∈ RN the input and output, respectively, with N ∈Z+ the trial length. Let j be the trial index and ‖xj‖W :=

2015 American Control ConferencePalmer House HiltonJuly 1-3, 2015. Chicago, IL, USA

978-1-4799-8686-6/$31.00 ©2015 AACC 3570

Page 2: Iterative Learning Control for Varying Tasks: Achieving ...In this paper discrete-time, linear, time-invariant (LTI), singe-input, single-output (SISO) systems are considered to facilitate

fj

+

Pr

C+

−+ej yj

Fig. 1. Closed-loop system under consideration.

x>j Wxj , where xj ∈ RN and W ∈ RN×N . W is positivedefinite if x>Wx > 0, ∀x 6= 0 and positive semi-definite ifx>Wx ≥ 0, ∀x.

Let a system G(θ, z) be linearly parameterized as

G(θ, z) =

m−1∑

n=0

ξn(z)θ[n],

with parameters θ ∈ Rm and basis functions ξn(z), whichis equivalent to

y = ΨGuθ, (2)

with ΨGu =[ξ0u, ξ1u, . . . , ξm−1u

]∈ RN×m where

ξn ∈ RN×N are finite-time matrix representations of ξn(z)according to (1).

Example 1. Consider G(θ, z) with N = 3 and ξ0(z) = z,ξ1(z) = 1, ξ2(z) = z−2, i.e., G(z) = θ[0]z+θ[1]+θ[2]z−2.Using (1):[y[0]y[1]y[2]

]

︸ ︷︷ ︸y

=

[θ[1] θ[0] 00 θ[1] θ[0]θ[2] 0 θ[1]

]

︸ ︷︷ ︸G

[u[0]u[1]u[2]

]

︸ ︷︷ ︸u

=

[u[1] u[0] 0u[2] u[1] 0

0 u[2] u[0]

]

︸ ︷︷ ︸ΨGu

[θ[0]θ[1]θ[2]

]

︸ ︷︷ ︸θ

,

where the latter is in the form of (2).

B. System description

The considered ILC setup is shown in Fig. 1. Here is P ∈RN×N a single-input, single output system and C ∈ RN×Na stabilizing feedback controller. The aim is to determine thefeedforward fj ∈ RN such that the output yj ∈ RN followsthe reference r ∈ RN as accurately as possible, i.e., such thatthe tracking error ej = r − yj is minimized. The trackingerror for trial j is given by

ej = Sr − SPfj , (3)

with sensitivity S := (I + PC)−1. The tracking error

evaluated at trial j + 1 is given by

ej+1 = Sr − SPfj+1. (4)

Eliminating Sr from (4) by (3) yields the tracking errorpropagation from trial j to trial j + 1:

ej+1 = ej + SP (fj − fj+1) .

C. Norm-optimal ILC

The goal is to minimize the error signal ej+1 at the nexttrial using fj+1. ILC [1] enables the determination of fj+1

through an iterative procedure that only requires approximatemodel knowledge.

Norm-optimal ILC is an important class of ILC in whichthe feedforward signal fj+1 for the next trial is selected

as the signal that minimizes the performance criterion inDefinition 2.

Definition 2 (Performance criterion norm-optimal ILC). Theperformance criterion for norm-optimal ILC is given by

J (fj+1) := ‖ej+1‖We+ ‖fj+1‖Wf

+ ‖fj+1 − fj‖W∆f,(5)

with We ∈ RN×N a symmetric, positive definite weightingmatrix, and Wf ,W∆f ∈ RN×N symmetric, positive semi-definite weighting matrices.

Norm-optimal ILC is well-known and can be found in,for example, [1], [18]. In norm-optimal ILC the signal fjis learned over the trials. However, the optimal feedforwardsignal yielding ej = 0 in (3) is given by fj = P−1r, underthe assumption that P is invertible. This case correspondsto inverse model feedforward and shows that the optimalfeedforward signal is a function of the reference signal.Hence, the learned signal in norm-optimal ILC will only beoptimal for one specific reference signal and non-optimal fordifferent reference signals. In order to introduce extrapolationproperties, basis functions are exploited in the next section.

D. Introducing extrapolation properties with basis functions

Inspired by inverse model feedforward, extrapolation prop-erties are introduced in ILC by use of basis functions [9]:

fj = F (θj)r, θj ∈ Rm. (6)

Substitution of (6) in (3) yields

ej = S(I − PF (θj))r.

Hence, if F (θj) = P−1, ej = 0 ∀r.In this paper rational basis functions are chosen for the

feedforward filter F (θj) (see Definition 3 and Fig. 2) sinceit allows to fully describe the plant inverse P−1 and therebyobtain perfect tracking.

Definition 3 (Rational basis functions). Rational basis func-tions in the parameters θj ∈ Rm with reference r as basisare defined as in (6) with F (θj) the matrix representation ofF(θj , z) ∈ F ,

F =

{A(θj , z)

B(θj , z)

∣∣∣∣ θj ∈ Rma+mb

},

with

A(θj , z) =

ma−1∑

n=0

ξAn (z)θj [n],

B(θj , z) = 1 +

mb−1∑

n=0

ξBn (z)θj [ma + n],

where ξAn (z), n = 0, 1, 2, . . . ,ma−1, and ξBn (z), n =0, 1, 2, . . . ,mb−1 are basis functions.

The polynomial basis functions in [9]–[12] are recoveredby setting mb = 0. Interestingly, an analytic solution for thecase mb = 0 is presented in [9]. Such analytic solution doesnot exist for the general case mb > 0.

3571

Page 3: Iterative Learning Control for Varying Tasks: Achieving ...In this paper discrete-time, linear, time-invariant (LTI), singe-input, single-output (SISO) systems are considered to facilitate

fjF (θj)

+

Pr

C+

−+ej yj

Fig. 2. Implementation of rational basis functions (Definition 3) in Fig. 1.

Evaluating (5) for (6) yields the performance criterionstated in Definition 4, which is a function of the parametersθj+1. Instead of determining the optimal feedforward signal,the optimal parameters are to be determined. Since m =dim(θj+1)� dim(fj+1) = N , additional robustness againstmodel uncertainties is introduced.

Definition 4 (Performance criterion ILC with basis func-tions). The performance criterion for norm-optimal ILC withbasis functions is given by

J (θj+1) := ‖ej+1(θj+1)‖We+

‖fj+1(θj+1)‖Wf+

‖fj+1(θj+1)− fj(θj)‖W∆f,

(7)

with We ∈ RN×N a symmetric, positive definite weightingmatrix, and Wf ,W∆f ∈ RN×N symmetric, positive semi-definite weighting matrices.

E. Problem formulation

The goal in this paper is to solve Problem 5.

Problem 5 (Main problem). Given Definition 3 and θj ,determine

θ∗j+1 = arg minθj+1

J (θj+1),

with J (θj+1) in Definition 4.

III. NON-OPTIMALITY PRE-EXISTING APPROACH

In this section the non-optimality of the pre-existingapproach of [13] is demonstrated, forming contribution I.

A. Pre-existing approach

Similar to standard norm-optimal ILC [1], there is ananalytic solution to Problem 5 if both ej+1 and fj+1 arelinear in θj+1, which is not the case for mb > 0. In thepre-existing approach, which is closely related to Steiglitz-McBride system identification, the performance criterion inDefinition 6 is considered. Note that if θ〈q〉j+1 = θ

〈q−1〉j+1 =

θj+1, then J (θ〈q〉j+1) = J (θj+1).

Definition 6 (Weighted performance criterion).

J (θ〈q〉j+1) :=

∥∥∥B−1(θ〈q−1〉j+1 )B(θ

〈q〉j+1)e

〈q〉j+1

∥∥∥We

+∥∥∥B−1(θ

〈q−1〉j+1 )B(θ

〈q〉j+1)f

〈q〉j+1

∥∥∥Wf

+∥∥∥B−1(θ

〈q−1〉j+1 )B(θ

〈q〉j+1)f

〈q〉j+1 − fj

∥∥∥W∆f

.

Note that J (θ〈q〉j+1) is quadratic in θ〈q〉j+1. Hence, there is a

unique solution for the optimal parameters θ〈q〉∗j+1 , which canbe determined analytically from:

(dJ (θ

〈q〉j+1)

dθ〈q〉j+1

)>∣∣∣∣∣∣θ〈q〉j+1=θ

〈q〉∗j+1

= 0. (8)

The idea is to iteratively determine the optimal parametersθ〈q〉∗j+1 for J (θ

〈q〉j+1) in Definition 6, using (8). The idea is that

upon convergence of the parameters, i.e., θ〈q〉j+1 → θ〈q−1〉j+1 ,

θ〈q〉∗j+1 are also the optimal parameters for Problem 5, becauseJ (θj+1) is recovered from J (θ

〈q〉j+1) for θ〈q〉j+1 = θ

〈q−1〉j+1 =

θj+1. There is thus an analytic solution which after conver-gence provides the solution to Problem 5. The procedure ofthe iterative algorithm is formulated in Algorithm 7.

Algorithm 7 (Pre-existing algorithm). The pre-existing al-gorithm [13] for solving Problem 5 is given by the followingsequence of steps.

1) Let rj , fj , ej be given, set q = 1, and initializeθ〈q−1〉j+1 = θj .

2) Compute θ〈q〉∗j+1 from (8).3) Set q → q + 1 and go back to 2) until an appropriate

stopping criterion is satisfied.

The advantage of Algorithm 7 over Gauss-Newton itera-tion is that Algorithm 7 is insensitive for local optima andtypically converges in a few iterations [14].

B. Non-optimality pre-existing approach

In the pre-existing approach (8) is solved which yields theminimum of J (θ

〈q〉j+1). However,

dJ (θ〈q〉j+1)

dθ〈q〉j+1

= 0 ;dJ (θj+1)

dθj+1= 0.

Consequently, there is no guarantee that the found parametersare also the optimal parameters for J (θj+1) and therefore theperformance is non-optimal. A detailed proof is beyond thescope of the present paper and will be published elsewhere.The non-optimality of the pre-existing approach is confirmedby a simulation example in section V.

IV. PROPOSED APPROACH

In this section a new iterative solution algorithm is pro-posed and its optimality is shown which constitutes contri-bution II.

The starting point of the proposed approach is

dJ (θj+1)

dθj+1)= 0, (9)

with the gradient given by Lemma 9, exploiting the auxiliaryresult in Lemma 8. Note that the gradient is not linear inθj+1 for mb > 0. Hence, there will in general be no analyticsolution to (9).

3572

Page 4: Iterative Learning Control for Varying Tasks: Achieving ...In this paper discrete-time, linear, time-invariant (LTI), singe-input, single-output (SISO) systems are considered to facilitate

Lemma 8 (Gradient quadratic matrix function). Let x, b ∈RN , A ∈ RN×N , and W> = W ∈ RN×N , then

d (‖Ax+ b‖W )

dx= 2(Ax+ b)>WA.

Proof. Generalization of Proposition 10.7.1 in [19].

Lemma 9 (Gradient performance criterion). The gradientof the performance criterion J (θj+1) in Definition 4 withrespect to the parameters θj+1 is given by

(dJ (θj+1)

dθj+1

)>=

− 2

(dfj+1

dθj+1

)> ((SP )

>WeSP +W∆f

)fj

− 2

(dfj+1

dθj+1

)>(SP )

>Weej

+ 2

(dfj+1

dθj+1

)> ((SP )

>WeSP +Wf +W∆f

B−1(θj+1)A(θj+1)rj .

where dfj+1

dθj+1= Krj , with K the matrix representation of

K(z) =1

B(θj+1, z)

dA(θj+1, z)

dθj+1

− A(θj+1, z)

B2(θj+1, z)

dB(θj+1, z)

dθj+1.

Proof. Follows from applying Lemma 8 to (7).

The second step is to apply a weighting similar as inthe pre-existing approach, see Definition 10. Lemma 9 isrecovered from Definition 10 for θ〈q〉j+1 = θ

〈q−1〉j+1 = θj+1.

Definition 10 (Weighted gradient of performance criterion).LetdJ (θ

〈q〉j+1)

dθ〈q〉j+1

>

=

− 2ζ〈q〉(

(SP )>WeSP +W∆f

)B(θ

〈q〉j+1)fj

− 2ζ〈q〉 (SP )>WeB(θ

〈q〉j+1)ej

+ 2ζ〈q〉(

(SP )>WeSP +Wf +W∆f

)A(θ

〈q〉j+1)rk,

(10)

with

ζ〈q〉 =

df

〈q−1〉j+1

dθ〈q−1〉j+1

>

B−1(θ〈q−1〉j+1 ),

df〈q−1〉j+1

dθ〈q−1〉j+1

= Krj

with K the matrix representation of

K(z) =1

B(θ〈q−1〉j+1 , z)

dA(θ〈q−1〉j+1 , z)

dθj+1

−A(θ

〈q−1〉j+1 , z)

B2(θ〈q−1〉j+1 , z)

dB(θ〈q−1〉j+1 , z)

dθ〈q−1〉j+1

.

The term ζ〈q〉 in (10) is not a function of θ〈q〉j+1; (10) onlydepends on θ

〈q〉j+1 through linear dependencies on A(θ

〈q〉j+1)

and B(θ〈q〉j+1). Since both these filters are linear in θ〈q〉j+1, (10)

is also linear in θ〈q〉j+1. Therefore the solution of equating (10)to zero has an analytic solution, see Theorem 11.

Theorem 11 (Optimal parameters for weighted gradientperformance criterion). The optimal parameters θ〈q〉∗j+1 of (10)in Definition 10 are given by

θ〈q〉∗j+1 =

(ζ〈q〉Ψ〈q〉

)−1

ζ〈q〉(Q〈q〉fj + L〈q〉ej

), (11)

with

Ψ〈q〉 =[(

(SP )>WeSP +Wf +W∆f

)ΨArj , . . .

− (SP )>WeΨ

Bej −

((SP )

>WeSP +W∆f

)ΨBfj

],

Q〈q〉 = (SP )>WeSP +W∆f ,

L〈q〉 = (SP )>We.

Proof. Due to space restrictions, the proof will be publishedelsewhere.

The final step is the formulation of the algorithm. Inthe previous steps two key elements are derived. First, aweighted gradient is introduced in Definition 10, from whichthe gradient in Lemma 9 is recovered for θ〈q〉j+1 = θ

〈q−1〉j+1 =

θj+1. Second, an analytic solution for the parameters θ〈q〉∗j+1 =

θ〈q〉j+1 is obtained for which the weighted gradient in Defini-

tion 10 is zero. Hence, upon convergence, the actual gradientalso converges to zero and optimal performance is achieved.Combining these two elements, there is an analytic solutionto Problem 5 when there is convergence in the parameters,i.e., θ〈q〉j+1 → θ

〈q−1〉j+1 . The iterative algorithm to obtain this

solution is given by Algorithm 12. The proposed algorithm isclosely related to instrumental variable system identification[16].

Algorithm 12 (Proposed algorithm). The proposed algo-rithm for solving Problem 5 is given by the followingsequence of steps.

1) Let rj , fj , ej be given, set q = 1, and initializeθ〈q−1〉j+1 = θj .

2) Compute θ〈q〉∗j+1 from (11).3) Set q → q + 1 and go back to 2) until an appropriate

stopping criterion is satisfied.

V. SIMULATION EXAMPLE

In this section the pre-existing and proposed approach forsolving ILC with rational basis functions are analyzed by useof simulation. This section constitutes contribution III.

A. System description

The system under consideration is an Oce Arizona 550GT flatbed printer, see Fig. 3. In this simulation exampleonly the movement in y-direction is considered. Given arethe transfer from the input current of the carriage motor [A]

3573

Page 5: Iterative Learning Control for Varying Tasks: Achieving ...In this paper discrete-time, linear, time-invariant (LTI), singe-input, single-output (SISO) systems are considered to facilitate

gantry UV lights carriage printing surface

y

Fig. 3. Oce Arizona 550 GT at the CST Motion laboratory, EindhovenUniversity of Technology. The carriage moves in y-direction over the gantrywhich can be moved parallel to the y-direction over the printing surface.

Mag

nitu

de[d

B]

−150

−100

−50

100 101 102 103

Phas

e[d

eg]

−180

−90

0

90

180

Frequency [Hz]

Fig. 4. Bode plot of the FRF (—) andidentified 20th order model (- -) of thetransfer of the y-position of the carriage.

t [s]

0

r[m

]

0

0.02

0.04

0.06

0.08

0.1

2 4 6 8

Fig. 5. The reference signalr consists of a symmetric 4thorder polynomial.

to the output position of the carriage [m] as identified usinga frequency response function (FRF) measurement. A 20thorder model based on this FRF measurement is given andused as plant model. The Bode plots of the FRF measurementand the model are depicted in Fig. 4. The sampling time isTs = 0.001 s. The feedback controller consists of a weakintegrator, a lead-filter and a first-order low-pass filter, andyields a bandwidth of 25 Hz.

B. Simulation setup

Fig. 5 depicts the reference signal r which consists ofa fourth order polynomial, preceded and followed by zerosto allow for pre- and post-actuation, respectively and hasa trial length of N = 8000 samples. The weighting filtersin Definition 4 are set to We = 106I , Wf = W∆f = 0.Rational basis functions (Definition 3) are exploited with

A(z) = γ(z − 1)2, |A(z)P(z)|z=1 = 1,

B(θj , z) = 1 +

1∑

n=0

ξBn θj [n], θj ∈ R2,

with ξB0 (z) = z−1zTs

, and ξB1 (z) = ( z−1zTs

)2, respectively, i.e.,a first- and second-order differentiator.

C. Simulation results

The results are depicted in Fig. 6 and Fig. 7. From analysisof the performance criterion over a grid of parameters, itis concluded that there is only one minimum, namely forθj+1 ≈

[3.13, −0.18

]>.

Next, the convergence behavior of both approaches is ana-lyzed. The new parameters θ〈q〉j+1 only depend on the previousparameters θ〈q−1〉

j+1 . Therefore, given θ〈q−1〉j+1 , an approach will

always yield the same parameters θ〈q〉j+1. For both iterativeapproaches, the direction of θ〈0〉j+1 → θ

〈1〉j+1 is indicated in

Fig. 6a and Fig. 7a for a grid in θj+1. The result is a vectorfield indicating the direction the parameters θ〈1〉j+1 move togiven θ〈0〉j+1 = θj . Moreover, the evolution of the parametersθ〈q〉j+1 for q = 0, 1, 2, . . . , 7 of three trajectories is displayed:

starting from the upper-left (3) and lower-right (2) corner,and from the optimum (#). The corresponding performancecriterion J (θ

〈1〉j+1) is displayed in Fig. 6b and Fig. 7b. The

2-norm of the gradient with respect to the parameters isdepicted in Fig. 6c and Fig. 7c. The figures indicate thatboth approaches converge within a couple of iterations.

Pre-existing approach: Fig. 6a indicates that the pre-existing approach always converges to a non-optimal sta-tionary point. This can also be observed in Fig. 6c, whichshows that the gradient after convergence is still considerablylarge. Even if the approach is initialized in the optimum (#),it diverges to the non-optimal stationary point with poorerperformance.

Proposed approach: In contrast, for the proposed ap-proach the gradients for all three initial conditions convergeto a value close to zero, i.e., the approach converges to theminimum. Correspondingly, the value of J (θ

〈q→∞〉j+1 ) is also

significantly smaller than for the pre-existing approach.

D. Conclusion on simulation example

The simulation example shows that the pre-existing ap-proach is non-optimal. Even if initialized in the optimum,it diverges to a stationary point with worse performance. Incontrast, the proposed approach converges to a stationarypoint close to the optimum.

For cases with multiple optima the stationary point ofgradient-based algorithms, such as the Gauss-Newton algo-rithm, depends on the initial parameters. In contrast, theproposed algorithm in this case converges to the globaloptimum, independent of the initial parameters [14].

VI. CONCLUSION

In this paper, extrapolation properties of norm-optimalILC are enhanced through use of rational basis functions.The associated optimization problem is significantly morechallenging than for polynomial basis functions. The maincontribution of this paper is a new iterative solution algorithmfor rational basis functions in ILC. The proposed approachhas advantageous properties compared to the pre-existingapproach [13], as is well-known from a related algorithmused in system identification. In particular, upon convergence

3574

Page 6: Iterative Learning Control for Varying Tasks: Achieving ...In this paper discrete-time, linear, time-invariant (LTI), singe-input, single-output (SISO) systems are considered to facilitate

θj+1[0]2 3 4

θ j+1[ 1]

−0.4

−0.3

−0.2

−0.1

0

(a) Vector field and trajectories.

q0 1 2 3 4 5 6 7

0.016

0.018

0.022

0.024

0.020

J(θ

〈q〉

j+1)

(b) Performance criterion.

q0 1 2 3 4 5 6 7

0

0.02

0.04

0.06

0.08

0.10

∥ ∥ ∥ ∥dJ(θ〈q

〉j+

1)

θ〈q

〉j+

1

∥ ∥ ∥ ∥

(c) Norm of gradient

Fig. 6. The pre-existing approachconverges to a stationary point thatis different than the optimum, evenwhen starting in the optimum.

θj+1[0]2 3 4

θ j+1[ 1]

−0.4

−0.3

−0.2

−0.1

0

(a) Vector field and trajectories.

q0 1 2 3 4 5 6 7

0.016

0.018

0.020

0.022

0.024J(θ

〈q〉

j+1)

(b) Performance criterion.

q0 1 2 3 4 5 6 7

0

0.02

0.04

0.06

0.08

0.10

∥ ∥ ∥ ∥dJ(θ〈q

〉j+

1)

θ〈q

〉j+

1

∥ ∥ ∥ ∥

(c) Norm of gradient

Fig. 7. The proposed approachalways converges to the optimum,independent of the initial parame-ters θ〈0〉j+1.

the proposed approach is guaranteed to be in a minimum,whereas this is not the case for the pre-existing approach. Asa result, the proposed approach outperforms the pre-existingapproach.

The convergence behavior of both approaches is analyzedusing simulations of a complex industrial system. The sim-ulations confirm that the proposed approach is superior overthe pre-existing approach in terms of performance.

Currently, the simulation results are experimentally val-idated. Ongoing research focuses on: application to MIMOsystems, selection of basis functions, robustness analysis, andnumerical conditioning along the lines of [20].

ACKNOWLEDGMENT

The authors would like to acknowledge fruitful discussionswith and the contributions of Sjirk Koekebakker and MaartenSteinbuch that have significantly contributed to the resultspresented in this paper.

REFERENCES

[1] D. A. Bristow, M. Tharayil, and A. G. Alleyne, “A survey of iterativelearning control – A learning-based method for high-performancetracking control,” Control Systems, IEEE, vol. 26, no. 3, pp. 96–114,June 2006.

[2] K. L. Moore, Iterative learning control for deterministic systems.Springer-Verlag, 1993.

[3] S. Gunnarsson and M. Norrlof, “On the design of ILC algorithms usingoptimization,” Automatica, vol. 37, no. 12, pp. 2011–2016, December2001.

[4] E. Rogers, K. Galkowski, and D. H. Owens, Control Systems Theoryand Applications for Linear Repetitive Processes. Springer, 2007,vol. 349.

[5] S. Gunnarsson and M. Norrlof, “On the disturbance properties ofhigh order iterative learning control algorithms,” Automatica, vol. 42,no. 11, pp. 2031–2034, November 2006.

[6] D. J. Hoelzle, A. G. Alleyne, and A. J. W. Johnson, “Basis taskapproach to iterative learning control with applications to micro-robotic deposition,” Control Systems Technology, IEEE Transactionson, vol. 19, no. 5, pp. 1138–1148, September 2011.

[7] M. Q. Phan and J. A. Frueh, “Learning control for trajectory trackingusing basis functions,” in Decision and Control, 1996., Proceedingsof the 35th IEEE, vol. 3. IEEE, December 1996, pp. 2490–2492.

[8] S. J. Oh, M. Q. Phan, and R. W. Longman, “Use of decouplingbasis functions in learning control for local learning and improvedtransients,” Advances in the Astronautical Sciences, vol. 95, no. 2, pp.651–670, 1997.

[9] J. van de Wijdeven and O. Bosgra, “Using basis functions in iterativelearning control: analysis and design theory,” International Journal ofControl, vol. 83, no. 4, pp. 661–675, 2010.

[10] J. Bolder, T. Oomen, S. Koekebakker, and M. Steinbuch, “Usingiterative learning control with basis functions to compensate mediumdeformation in a wide-format inkjet printer,” Mechatronics, vol. 24,no. 8, pp. 944–953, December 2014.

[11] M. F. Heertjes and R. M. van de Molengraft, “Set-point variationin learning schemes with applications to wafer scanners,” ControlEngineering Practice, vol. 17, no. 3, pp. 345–356, 2009.

[12] X. Gao and S. Mishra, “An iterative learning control algorithm forportability between trajectories,” in American Control Conference(ACC), 2014, June 2014, pp. 3808–3813.

[13] J. Bolder and T. Oomen, “Rational basis functions in iterative learningcontrol - with experimental verification on a motion system,” IEEETransactions on Control Systems Technology, vol. 23, no. 2, pp. 722–729, June 2014.

[14] C. Bohn and H. Unbehauen, “Minmax and least squares multivariabletransfer function curve fitting: Error criteria, algorithms and compar-isons,” in American Control Conference, 1998., Proceedings of the1998, vol. 5. IEEE, June 1998, pp. 3189–3193.

[15] A. Whitfield, “Asymptotic behaviour of transfer function synthesismethods,” International Journal of Control, vol. 45, no. 3, pp. 1083–1092, March 1987.

[16] M. Gilson, H. Garnier, P. Young, and P. Van den Hof, “Optimal in-strumental variable method for closed-loop identification,” IET ControlTheory & Applications, vol. 5, no. 10, pp. 1147–1154, 2011.

[17] A. V. Oppenheim, A. S. Willsky, and S. H. Nawab, Signals andsystems, 2nd ed. Prentice-Hall, Inc., 1997.

[18] S. Gunnarsson and M. Norrlof, “Some aspects of an optimizationapproach to iterative learning control,” in Decision and Control, 1999.Proceedings of the 38th IEEE Conference on, vol. 2. IEEE, 1999,pp. 1581–1586.

[19] D. S. Bernstein, Matrix mathematics: theory, facts, and formulas.Princeton University Press, 2009.

[20] R. van Herpen, T. Oomen, and M. Steinbuch, “Optimally conditionedinstrumental variable approach for frequency-domain system identifi-cation,” Automatica, vol. 50, no. 9, pp. 2281–2293, September 2014.

3575