Robust Control
description
Transcript of Robust Control
Robust Control
References:
1. K. Zhou, and J. C. Doyle,Essentials of Robust Control,
Prentice-Hall, Inc., 1998
2. S. Skogestad, and I. Postlethwaite,Multivariable Feedback Control:
Analysis and Design, John Wiley & Sons Ltd., 1996
1
Introduction
1. Why Robust Control?Models are uncertain.
• Real environments of a control system may change and operatingconditions may vary from time to time.
• Even if the environment does not change, other factors are themodel uncertainties as well as noises.
• Any mathematical representation of a system often involvessimplifying assumptions. Nonlinearities are either unknown andhence unmodeled, or are modeled and later ignored in order tosimplify analysis. High frequency dynamics are often ignored atthe design stage as well.
In consequence, control systems designed based on simplifiedmodels may not work on real plants in real environments.
2
2. Robustness and Robust Control
• Robustness. The property that a control system must maintain
overall system performance despite changes in the plant is called
robustness. Any control system must possess this property for it
to operate properly in realistic situations.
Mathematically, this means that the controller must perform
satisfactorily not just for one plant, but for a family of plants. If a
controller can be designed such that the whole system to be
controlled remains stable when its parameters vary within certain
expected limits, the system is said to possess robust performance.
• Robust Control. The problem of designing controllers that satisfy
both robust stability and performance requirements is called
robust control.
3
3. Scope of Robust ControlSome of the key questions in robust control theory are:
1) Characterizing plant variation. In robust control theory, plant
variation plays a central role. What is a good way to describe
plant variations or uncertainty? Some descriptions attempt to
faithfully describe the variations that might be encountered (e.g.,
probability distributions on physical parameters). Other
descriptions are more convenient for the associated theory(e.g.,
bounds on singular values of transfer matrix errors).
2) Robustness analysis. Given a controller and plant, and some
description of the plant uncertainty, how can we determine such
things as “typical” or “worst-case” performance? How can we
predict the performance degradation caused by variation inthe
plant? How can we verify that some performance specifications
are met for some set of plants?
4
3) Robust controller synthesis. Given a plant and some descriptionof the plant uncertainty, how can we design a controller thatoptimizes “typical” or “worst-case” performance? How can wedesign a controller that meets some performance specificationsfor some set of plants (e.g., all or typical)?
It’s important to remember several things:
• First, these questions were asked, and partial answers obtained,before the term “robust control” was coined.
• Second, the qualifier “robust” shouldn’t be necessary sinceawell-designed controller must be able to tolerate the plantchanges or variations that can be expected. To put it anotherway,a controller that cannot tolerate variations in the plant that will beencountered in operation is simply a poorly-designed controller,not just a non-robust controller. (But the extra qualifier “robust”has helped re-focus attention on this important aspect of controlengineering.)
5
4. Development of Robust Control
• In classical control theory, gain and phase margins are usedtorepresent the robustness of a control system.
• During the development of state-space optimal control in 1960s,control system robustness received less attention, since the idea ofplant change, variation, or uncertainty played at best a secondaryrole.
– Some initial results showed that the state-feedbackimplementation of LQR was very tolerant of changes in theplant. This led to the hope that controllers designed to beoptimal for a fixed, known plant might automatically turn outto be robust, i.e., tolerant to changes in the plant. In a shortnote, Doyle pointed out that this is not the case.
• Robustness re-drew attention in 1980s. One of the most famouscontributions of robust control theory is the development of H∞
controller synthesis.
6
Contents
1. Norms of Signals and Systems
2. Uncertainty and Robustness
3. Performance Specifications
4. Robust Controller Synthesis
5. Model Reduction
6. Robustness Measure
7. Internal Model Control (IMC)
8. Wide-Range Robust Control
7
Chapter 1. Norms of Signals and Systems
A system can be regarded as a ‘gray box’ for manipulating signals. Todescribe the size of signals and systems, we need to work on functionspaces. There are two kinds of functions spaces that we are interested in:
• Linear spaces composed of real functions, especially, functions oftime f (t) with t ∈ R.
• Linear spaces composed of complex functionsF(s) with s∈ C.
Outline of this chapter:
• Vector Spaces
• Function spaces
• Norms of signals
• Norms of systems
8
1.1 Vector Spaces
Let F be a scalar field (it can be taken to be real numbersR, or thecomplex numbersC). SupposeV is a nonempty set, together with a set ofoperations: addition and scalar multiplication, then a linearvector spaceis the 4-tuple{V,F,+, ·} such that the following rules are satisfied for allx,y∈ V andα ,β ∈ F:
1) Commutativity:x+y= y+x
2) Associativity:(x+y)+z= x+(y+z)
α · (β ·x) = (αβ ) ·x
3) Identity: there exists an element 0∈ V and 1∈ F such that
x+0= x, 1·x= x
9
4) Inverse: x∈ V implies that there existsy∈ V such thatx+y= 0.
5) Distributivity:
α · (x+y) = α ·x+β ·y
(α +β ) ·x= α ·x+β ·x
Examples:
• Set of vectors (Rn) over the field of real numberR;
• Set of vectors (Cn) over the filed of complex numberC;
• Set of allm×n matrices (Cm×n) over the filed ofC.
• Set of all functions mapping fromR toRn ( f : R→ Rn) over the
fieldR.
10
1.1.1 Normed SpacesWe can define a norm on a vector space to denote the size of an element.
LetV be a vector space, a real-valued function‖ · ‖ : V→ R is said to be a
norm if it satisfies the following properties for anyx∈ V andy∈ V:
1) ‖x‖ ≥ 0 (positivity).
2) ‖x‖= 0 if and only ifx= 0 (positive definiteness).
3) ‖αx‖= |α |‖x‖ for any scalarα (homogeneity).
4) ‖x+y‖ ≤ ‖x‖+‖y‖ for anyx,y∈ V (triangular inequality).
A vector space together with a norm is called anormed spaceand is
denoted(V,‖ · ‖).A normed space iscompleteif every Cauchy sequence in it converges.
Such a space is referred to as aBanach space.
11
Vector Norms (Norms onCn)
For vector spaceCn, several norms can be defined:
• 1-norm.
‖x‖1 :=n
∑i=1
|xi |
• 2-norm.
‖x‖2 :=
√n
∑i=1
|xi |2
• ∞-norm.‖x‖∞ := max
1≤i≤n|xi |
They are special cases of thep-norm defined as
‖x‖p :=
(n
∑i=1
|xi |p)1/p
12
Matrix Norms ( Norms on Cm×n)
Let A∈ Cm×n, then theFrobenius norm of A is defined as
‖A‖F :=√
trace(A∗A) =
√m
∑i=1
n
∑j=1
|ai j |2
If we treatA as a linear transform fromCn toCm, then theinducedp-norm of A (induced by a vectorp-norm) is defined as
‖A‖p := supx 6=0
‖Ax‖p
‖x‖p
A most often used norm is the induced 2-norm (Euclidean norm).
‖A‖1 = maxj
m
∑i=1
|ai j |, ‖A‖∞ = maxi
n
∑j=1
|ai j |, ‖A‖2 = σmax(A)
From now on, if there is a norm without the subscript, then it refers to
(vector or matrix) 2-norm.
13
1.1.2 Inner Product SpacesAnother very important notion that we will frequently encounter is aninner product, and it is closely related to the idea of a norm.
Let V be a vector space, a function〈·, ·〉 : V×V→ C is said to be aninner product if it satisfies the following properties for anyx,y,z∈ V andα ,β ∈ C:
1) 〈x,x〉> 0 if x 6= 0
2) 〈x,αy+βz〉= α〈x,y〉+β 〈x,z〉3) 〈x,y〉= 〈y,x〉
A vector spaceV together with an inner product is called aninnerproduct spaceand is denoted(V,〈·, ·〉). A complete inner product spaceis called aHilbert space.
It is clear that the inner product defined above induced a norm
‖x‖ :=√
〈x,x〉
14
In particular, the distance between two vectorsx andy is d(x,y) = ‖x−y‖.
Two vectorsx andy in an inner product spaceV are said to beorthogonalif 〈x,y〉= 0, denotedx⊥ y. More generally, a vectorx is said to be
orthogonal to a setS⊂V, denoted byx⊥ S, if x⊥ y for all y∈ S.
Properties of inner product and the inner product induced norm .
Theorem LetV be an inner product space and let x,y∈ V. Then
1) |〈x,y〉| ≤ ‖x‖‖y‖ (Cauchy-Schwarz inequality). Moreover, the
equality holds if and only if x= αy for some constantα or y= 0.
2) ‖x+y‖2+‖x−y‖2 = 2‖x‖2+2‖y‖2 (Parallelogram law).
3) ‖x+y‖2 = ‖x‖2+‖y‖2 if x ⊥ y.
15
Examples of Hilbert spaces:
• Cn with the inner product defined below is a (finite dimensional)
Hilbert space:〈x,y〉 := x∗y for all x,y∈ Cn.
• Cm×n with the inner product defined below is a (finite dimensional)Hilbert space:〈A,B〉 := trace(A∗B),∀A,B∈ C
m×n.
• L2[a,b] (the space of all square integrable and Lebesgue measurablefunctions defined on an interval[a,b]) is an infinite dimensionalHilbert space with the inner product defined as
〈 f ,g〉 :=∫ b
af (t)∗g(t)dt
If the functions are vector or matrix-valued, the inner product isdefined accordingly as
〈 f ,g〉 :=∫ b
atrace[ f (t)∗g(t)]dt
16
1.2 Time-Domain Spaces
Consider the vector spaceL of all the Lebesgue measurable functions
mappingR toCn. We can define a norm‖ · ‖p as
‖u‖p :=
(∫ ∞
−∞‖u(t)‖p
pdt
)1/p
where‖u(t)‖p is thep-norm onCn. We define the spacesLp(−∞,∞) or
Lp as
Lp(−∞,∞) = {u∈ L such that‖u‖p < ∞}Lp spaces are Banach spaces.
We will useLp+ = Lp[0,∞) to denote the subspace ofLp(−∞,∞) with
functions zero fort < 0, andLp− = Lp(−∞,0] to denote the subspace of
Lp(−∞,∞) with functions zero fort > 0.
17
Some commonly used norms are:
• L1-Norm. (L1 space)
‖u‖1 =∫ ∞
−∞‖u(t)‖1dt =
∫ ∞
−∞
n
∑i=1
|ui(t)|dt =n
∑i=1
‖ui‖1
• L2-Norm. (L2 space)
‖u‖2=
(∫ ∞
−∞‖u(t)‖2
2dt
)1/2
=
(∫ ∞
−∞
n
∑i=1
ui(t)2dt
)1/2
=
(n
∑i=1
‖ui‖22
)1/2
• L∞-Norm. (L∞ space) The norm for the casep= ∞ is defined by
‖u‖∞ := max1≤i≤n
‖ui‖∞ = max1≤i≤n
supt∈R
|ui(t)|
Remark:L2 is an infinite dimensional Hilbert space with inner product
〈 f ,g〉 :=∫ ∞
−∞trace[ f (t)∗g(t)]dt
18
1.2.1 Norms of SignalsA signal can be regarded as a function inL. So with different norms usedwe can measure the ‘size’ of the signals. To illustrate, we consider the
scalar signals.
1. L2-norm of a signalu(t) is
‖u‖2 :=
(∫ ∞
−∞u(t)2dt
)1/2
It reflects thetotal energyof a signal. We often use theroot-mean-square(RMS)value to measure itsaverage power.
‖u‖rms :=
(
limT→0
12T
∫ T
−Tu(t)2dt
)1/2
This is a classical notion of the size of a signal, widely usedin manyareas of engineering.
Properties of RMS value:
19
• RMS is not a norm, since‖u‖rms can be zero whenu is nonzero.
Nevertheless, it is a useful, and often used, measure of a signal’s
size.
• It is known that even if the RMS value of a signal is small, the
signal may occasionally have large peaks, provided the peaks are
not too frequent and do not contain too much energy. In this
sense,‖u‖rms is less affected by large but infrequent values of the
signal.
• The RMS is asteady-statemeasure of a signal; the RMS value of
a signal is not affected by any transient. In particular, a signal with
small RMS value can be very large for some initial time period.
20
2. L∞-norm of a signal is the least upper bound of its absolute value:
‖u‖∞ := supt∈R
|u(t)|
Physical interpretation of this norm is the signal’s maximum or peakabsolute value, thus it is also calledpeak norm. A variation on thepeak norm is theeventual peakor steady-state peak:
‖u‖ss∞ := limT→∞
supt≥T
|u(t)|
• One simple but strict interpretation of “the signalu is small” isthat it is smallat all times, or equivalently, its maximum or peakabsolute value is small.
• The peak norm of a signal depends entirely on the extreme or
large values the signal takes on. If the signal occasionallyhaslarge values,‖u‖∞ will be large.
• The steady-state peak norm measures only persistent large
22
excursion of the signal. It is not affected by transients.
The peak norm can be used to describe a signal about which verylittle is known other than some bound on its peak or worst casevalue.Such a description is called anunknown-but-boundedmodel of asignal.
An example of a signalu and its peak‖u‖∞ is shown below.
23
3. L1-norm of a signalu(t) is the integral of its absolute value:
‖u‖1 :=∫ ∞
−∞|u(t)|dt
It reflects thetotal resource consumptionor total fuel. We often use
theaverage-absolute valueto measure asteady-state average
resource consumptionor average fuel.
‖u‖aa := limT→0
12T
∫ T
−T|u(t)|dt
This norm puts even less emphasis on large values of a signal.
24
An example of a signalu and its average-absolute norm is shownbelow.
We can think of the peak norm, the RMS norm and the average-absolutenorm as differing in the relative weighting of large versus small signalvalues:The peak norm is entirely dependent on the large values of a
signal; the RMS norm is less dependent on the large values, and the
average-absolute norm less still.
25
Comparing Norms of Signals
We have seen many norms for signals. A natural question is: howdifferent can they be? Intuition suggests that since these different normseach measure the ‘size’ of a signal, they should generally agree aboutwhether a signal is ‘small’ or ‘large’. However, this intuition is generallyfalse.
For scalar signals we have
‖u‖∞ ≥ ‖u‖rms≥ ‖u‖aa
For vector signals withn components we have the generalization
‖u‖∞ ≥ 1√n‖u‖rms≥
1n‖u‖aa
Another norm inequality, that gives a lower bound for‖u‖aa, is
‖u‖2rms≤ ‖u‖aa‖u‖∞
26
1.2.2 Norms of SystemsSeveral ways may be used to measure the ‘size’ of a system withinputw,
outputz, and transfer matrixG.
1. Norm of a Particular Response.The simplest general method for measuring the size of a system is to
measure the size of its response to aparticular input signal wpart,
e.g., a unit impulse, a unit step, or a stochastic signal witha particular
power spectral density.
‖G‖part := ‖Gwpart‖
2. Average Response Norm.A general method for measuring the size of a system, that directly
takes into account the response of the system to many input signals
(not just one particular input signal), is to measure theaveragesize
(expectation) of the response ofG to a specificprobability
27
distribution of input signals.
‖G‖avg := Ew‖Gw‖
3. Worst Case Response Norm.Another general method for measuring the size of a system, that takes
into account the response of the system to many input signals, is to
measure the worst case or largest norm of the response ofG to a
specificcollection of input signals.
‖G‖wc := supw∈W
‖Gw‖
whereW denotes the collection of input signals.
28
Gains of Systems
An important special case of a worst case norm is again, or inducednorm if a linear system is regarded as a linear operator (map) between
two normed spaces.
Suppose the size of the input is measured with norm‖ · ‖in and the size of
the output is measured with norm‖ · ‖out, then thegain of the system is
defined by:
‖G‖gn := sup‖w‖in 6=0
‖Gw‖out
‖w‖in= sup
‖w‖in≤1‖Gw‖out
So the gain is therefore the maximum factor by which the system can
scale the size of a signal flowing through it.
29
System Gains and Input-Output Relationships
• L2 (RMS) Gain.
‖G‖∞ := sup‖w‖2 6=0
‖Gw‖2
‖w‖2= sup
‖w‖rms6=0
‖Gw‖rms
‖w‖rms
• L∞ (Peak) Gain.
‖G‖1 := sup‖w‖∞ 6=0
‖Gw‖∞
‖w‖∞
It can be shown that the peak gain is equal to theL1 norm of its
impulse response:
‖G‖1 = max1≤i≤nz
∫ ∞
0
nw
∑j=1
|gi j (t)|dt
30
• L1 (Average-Absolute) Gain.
‖G‖∞ = sup‖w‖1 6=0
‖Gw‖1
‖w‖1= sup
‖w‖aa6=0
‖Gw‖aa
‖w‖aa
• H2-Norm. TheH2 norm of a system is the RMS value of its output
when the inputs are independentwhite noises, or unit impulses.
‖G‖2 =
(1
2π
∫ ∞
−∞trace[G( jω)∗G( jω)]dω
)1/2
Table: System Gains
‖u‖2 ‖u‖∞ ‖u‖rms
‖y‖2 ‖G‖∞ ∞ ∞
‖y‖∞ ‖G‖2 ‖G‖1 ∞
‖y‖rms 0 ≤ ‖G‖∞ ‖G‖∞
31
1.3 Frequency Domain Spaces
We have introduced time domain function spaces which are useful inrepresentations signals and systems. As we know sometimes it is moreconvenient to study them in the frequency domain. We now considercomplex function spaces (Hardy spaces).
1. Frequency domain spaces: Signals
• L2( jR) Space. L2( jR) or simplyL2 is a Hilbert space, whichconsists of functions mappingjR toC
n with the inner product
〈 f , g〉 :=1
2π
∫ ∞
−∞trace[ f ∗( jω)g( jω)]dω
and the inner product induced norm is given by
‖ f‖2 :=√
〈 f , f 〉
A function f : jR→ Cn is inL2( jR) if ‖ f‖2 < ∞.
32
• H2 Space. H2 space is a closed subspace ofL2( jR) with
functions f (s) analytic in Re(s)> 0 (open right-half plane).
The corresponding norm is defined as
‖ f‖22 := sup
σ>0
{1
2π
∫ ∞
−∞trace[ f ∗(σ + jω) f (σ + jω)]dω
}
=1
2π
∫ ∞
−∞trace[ f ∗( jω) f ( jω)]dω
So it can be computed just as inL2.
• H⊥2 Space. H⊥
2 is the orthogonal complement ofH2 in L2( jR);
that is the closed subspace of functions inL2 that are analytic in
Re(s)< 0 (open left-half plane).
33
2. Frequency domain spaces: SystemsL2( jR), H2, andH⊥
2 spaces can also be defined for systems. Otherfrequency domain spaces for systems are:
• L∞( jR) Space. L∞( jR) or simplyL∞ is a Banach space ofmatrix-valued functions that are essentially bounded onjR with
‖F‖∞ := ess supω∈R
σ [F( jω)]
• H∞ Space. H∞ is a closed subspace ofL∞( jR) with matrixfunctions that are analytic and bounded in Re(s)> 0 (openright-half plane). TheH∞ norm is defined as
‖F‖∞ := supRe(s)>0
σ [F(s)] = supω∈R
σ [F( jω)]
• H−∞ Space. H−
∞ is a subspace ofL∞ with functions that areanalytic and bounded in the open left-half plane. TheH−
∞ normcan be obtained similarly as theH∞ norm.
34
3. Characterization of real rational function spacesWe are interested in spaces of functions that arereal and rational.We denote real rational function spaces by prefixingR.
• RL2 consists of all real rationalstrictly propertransfer matriceswith no poles on the imaginary axis.
• RH2 consists of all real rationalstrictly propertransfer matriceswith no poles on the open right-half plane (stable).
• RH⊥2 consists of all real rationalstrictly propertransfer matrices
with no poles on the open left-half plane (anti-stable).
• RL∞ consists of all real rationalproper transfer matrices with nopoles on the imaginary axis.
• RH∞ consists of all real rationalproper transfer matrices with nopoles on the open right-half plane (stable).
• RH−∞ consists of all real rationalproper transfer matrices with no
poles on the left right-half plane (anti-stable).
35
Relationship Between Time- and Frequency-Domain Spaces
1. L2( jR) and L2(−∞,∞)
By Parseval’s theorem,L2 in time domain is isomorphic toL2 infrequency domain through a bilateral Laplace (Fourier) transform.
L2(−∞,∞) ∼= L2( jR)
L2[0,∞) ∼= H2
L2(−∞,0] ∼= H⊥2
As a result, ifG∈ L2(−∞,∞) and its bilateral Laplace transform isG(s) ∈ L2( jR), then
‖G‖2 = ‖G‖2
2. L∞( jR) and linear operator space onL2(−∞,∞)
An element in frequency domain spaceL∞( jR) is closely related tothe linear operator on the time domain spaceL2(−∞,∞).
36
(a) EveryG∈ L∞( jR) defines a linear operatorG on L2(−∞,∞),
wherez= Gu is defined by ˆz( jω) = G( jω)u( jω).
(b) For each linear operatorG onL2(−∞,∞), there exists a function
G∈ L∞( jR) such thatz= Gu satisfies ˆz( jω) = G( jω)u( jω) for
all u in L2(−∞,∞).
So for anyG∈ L∞( jR), we can define amultiplication operator :
G : L2 → L2, G f := G f
Theorem Let G∈ L∞( jR) be a transfer matrix, and G is the
corresponding multiplication operator. Then
‖G‖∞ = ‖G‖∞ = supu∈L2 6=0
‖Gu‖2
‖u‖2
So from now on if there is no confusion we will not distinguishasystem with gainG and a transfer matrix G.
37
1.4 ComputingL2L2L2 andH2H2H2 Norms
Several methods can be used to compute theL2-norm or theH2-norm.
1. LetG(s) ∈ L2, then
‖G‖22=
12π
∫ ∞
−∞trace[G∗( jω)G( jω)]dω =
12π j
∮
trace[G∼(s)G(s)]ds
2. Letei denote theith standard basis vector ofRm, wherem is the inputdimension of the system. Apply the impulsive inputδ (t)ei anddenote the output byzi(t) = g(t)ei . AssumeG(s) is strictly proper,thenzi ∈ L2 and
‖G‖22 =
m
∑i=1
‖zi‖22
3. Denote the impulse response matrix ofG(s) by g(t). Then
‖G‖22 = ‖g‖2
2 =∫ ∞
−∞trace[g∗(t)g(t)]dt
38
4. State space computation ofH2 norm.
Consider a transfer matrix
G(s) =
A B
C 0
with A stable. Then
‖G‖22 = trace(BTQB) = trace(CPCT)
whereQ andP are theobservability Gramian and the
controllability Gramian that can be obtained from the following
Lyapunov equations:
AP+PAT +BBT = 0
ATQ+QA+CTC = 0
Related MATLAB Commands:gram, h2norm.
39
1.5 ComputingL∞L∞L∞ andH∞H∞H∞ Norms
Let G(s) ∈ L∞, then‖G‖∞ = sup
ωσ [G( jω)]
A control engineering interpretation of the infinity norm ofa scalartransfer functionG is
• The distance in the complex plane from the origin to the farthestpoint on the Nyquist plot ofG.
• The peak value on the Bode magnitude plot of|G( jω)|.Hence theL∞ norm of a transfer function can be obtained graphically.
To get an estimate, set up a fine grid of frequency points:
{ω1, · · · ,ωN}
Then an estimate for‖G‖∞ is max1≤k≤N σ [G( jω)].
40
Computation in state space.Let γ > 0 and
G(s) =
A B
C D
∈ RL∞
Then‖G‖∞ < γ if and only if σ(D)< γ and the Hamiltonian matrixH
has no eigenvalues on the imaginary axis, where
H :=
A+BR−1DTC BR−1BT
−CT(I +DR−1DT)C −(A+BR−1DTC)T
andR= γ2I −DTD.
41
Bisection AlgorithmWe can use the following bisection algorithm to computeRL∞ norm:
(a) Select an upper boundγu and a lower boundγl such that
γl ≤ ‖G‖∞ ≤ γu.
(b) If (γu− γl )/γl ≤ specified level, stop;‖G‖∞ ≈ (γu+ γl )/2. Otherwise
go to the next step.
(c) Setγ = (γu+ γl )/2;
(d) Test if‖G‖∞ < γ by calculating the eigenvalues ofH for the givenγ.
(e) If H has an eigenvalue onjR, setγl = γ; otherwise setγu = γ; go
back to step (b).
The above algorithm applies toH∞ norm computation as well.
Related MATLAB Commands:sigma, hinfnorm .
42
Chapter 2. Uncertainty and Robustness
In this chapter we will describe various types of uncertainties that can
arise in physical systems, and obtain robust stability tests for systems
under various model uncertainty assumptions. Outline of this chapter:
• Model uncertainty description
• Small gain theorem and smallµ theorem
• Robust stability under unstructured and structured uncertainties
• Linear fractional transformation (LFT) and Main Loop Theorem
43
Model Uncertainty
Modeling of plant uncertainty can be done in two methods:
• Unstructured uncertainty. Unstructured uncertainty is the uncertainty
about which no information is available about its effects ona process,
except that an upper bound on its ‘size’ or magnitude as a function of
frequency can be estimated.
• Structured uncertainty. Structured uncertainty is the uncertainty
about which ‘structural’ information is available, which will typically
restrict to a section of a process model. It is also calledparameter
uncertainty.
44
2.1 Unstructured Uncertainty
Several methods can be used to model unstructured uncertainty.
1. Additive UncertaintySuppose that the nominal plant transfer matrix isP and consider
perturbed plant transfer matrices of the form
P∆ = P+∆W
HereW is a fixed stable transfer matrix, the weight, and∆ is a
variable stable transfer matrix satisfying
‖∆‖∞ ≤ 1
Furthermore, it is assumed that no unstable poles ofP are canceled in
forming P∆. Such a perturbation∆ is said to be allowable.
45
2. Multiplicative UncertaintySuppose that the nominal plant transfer matrix isP and considerperturbed plant transfer matrices of the form
P∆ = (I +∆W)P
HereW is a fixed stable transfer matrix, the weight, and∆ is avariable stable transfer matrix satisfying
‖∆‖∞ ≤ 1
We also assume that no unstable poles ofP are canceled in formingP∆.
K P
W
r
_
y
∆
+
+
47
3. Coprime Factor UncertaintyCoprime factor uncertainty may be described as:
P∆ = (M+∆M)−1(N+∆N)
whereP= M−1N is an lcf ofP, ∆M, ∆N ∈ RH∞. The uncertainty is
∆ = [ ∆M ∆N ]
Kr
_
y
∆M
+_
∆N
N~
1~ −M
48
Example of Uncertainties
Consider a process model
P(s) =ke−τs
Ts+1, 4≤ k≤ 9, 2≤ T ≤ 3, 1≤ τ ≤ 2
Take the nominal model as
P0(s) =6.5
(2.5s+1)(1.5s+1)
Then for each frequency, all possible frequency responses are in a box.
49
The additive perturbation is
∆a( jω) = P( jω)−P0( jω)
A weight such that|∆a( jω)| ≤ |Wa( jω)| can be chosen as
Wa(s) =0.0376(s+116.4808)(s+7.4514)(s+0.2674)
(s+1.2436)(s+0.5575)(s+4.9508)
50
The multiplicative perturbation is
∆m( jω) =P( jω)−P0( jω)
P0( jω)
A weight such that|∆m( jω)| ≤ |Wm( jω)| can be chosen as
Wm(s) =2.8169(s+0.212)(s2+2.6128s+1.732)
s2+2.2425s+2.6319
51
2.1.1 Small Gain TheoremThe basis for the robust stability criteria is the small-gain theorem.
Consider the system shown below:
Small Gain Theorem. Suppose M∈ RH∞ and letγ > 0. Then the
interconnected system is well-posed and internally stablefor all
∆(s) ∈ RH∞ with ‖∆‖∞ ≤ γ if and only if
‖M(s)‖∞ < γ−1
52
2.1.2 Robust Stability under Unstructured UncertaintiesFor the uncertainties discussed above, we have the following results:
1. Multiplicative UncertaintyLet P∆ = {(I +∆W)P : ∆ ∈ RH∞} and letK be a stabilizing
controller for the nominal plantP. Then the closed-loop system is
well-posed and internally stable for all‖∆‖∞ ≤ 1 if and only if
‖WPK(I +PK)−1‖∞ < 1
The condition‖WT‖∞ < 1 also has a nice graphical interpretation.
‖WT‖∞ < 1⇔∣∣∣∣
W( jω)L( jω)
1+L( jω)
∣∣∣∣< 1, ∀ω
⇔ |W( jω)L( jω)|< |1+L( jω)|, ∀ω.
The last inequality says that at every frequency, the critical point,−1,
lies outside the disk of centerL( jω), radius|W( jω)L( jω)|.
53
2. Additive UncertaintyLet P∆ = {P+∆W : ∆ ∈ RH∞} and letK be a stabilizing controller
for the nominal plantP. Then the closed-loop system is well-posed
and internally stable for all‖∆‖∞ ≤ 1 if and only if
‖WK(I +PK)−1‖∞ < 1
3. Coprime Factor UncertaintyLet P∆ = {(M+∆M)−1(N+∆N) : M, N,∆M,∆N ∈RH∞} with (M, N)
being an lcf of the nominal plantP. Let K be a stabilizing controller
for P. Then the closed-loop system is well-posed and internally
stable for all‖[ ∆M ∆N ]‖∞ ≤ 1 if and only if
∥∥∥∥∥∥
I
K
(I +PK)−1M−1
∥∥∥∥∥∥
∞
< 1
54
2.2 Structured Uncertainty
It is easy to see that the maximum singular value (H∞-norm) is useful in
analyzing the unstructured uncertainty. To analyze the structured
uncertainty, we need the concept of structured singular value.
The definition of the structured singular value, which is also calledµ,
depends on the underlying block structure of the uncertainty set∆∆∆.
Defining the structure involves specifying three things:
• The type of each block.repeated scalaror full block.
• The total number of blocks. The number ofrepeated scalarblocks is
denoted byS, and the number offull blocks is denoted byF .
• The dimension of each block. To bookkeep the dimensions, we
introduce positive integersc1, · · · ,cS andm1, · · · ,mF . Thei’th
repeated scalar block isci ×ci , and thek’th full block is mk×mk.
55
We define the block structure∆∆∆ ⊂ Cn×n as
∆∆∆ = {diag[δ1Ic1, · · · ,δSIcS,∆1, · · · ,∆F ] : δi ∈ C,∆k ∈ Cmk×mk}
For consistency among all dimensions, we must have
S
∑i=1
ci +F
∑k=1
mk = n
Structured Singular Value (SSV)
Given a matrixM ∈ Cn×n and the underlying structure∆∆∆, µ∆∆∆(M) is
defined as
µ∆∆∆(M) :=1
min{σ(∆) : ∆ ∈∆∆∆,det(I −M∆) = 0}
unless no∆ ∈∆∆∆ makesI −M∆ singular, in which caseµ∆∆∆(M) := 0.
56
For a transfer matrixM(s), define the set of all block diagonal and stablerational transfer functions that have block structures as∆∆∆:
M(∆∆∆) := {∆(·) ∈ RH∞ : ∆(s) ∈∆∆∆ for all s∈ C such that Res≥ 0}
then the structured singular value ofM(s) can be computed pointwise as
‖M(s)‖µ = µ∆∆∆(M(s)) = supω∈R∪{∞}
µ∆∆∆(M( jω))
The functionµ : Cn×n → R is not a norm, since it does not satisfy thetriangle inequality. However, for simplicity, we still use‖ · ‖µ .Properties ofµµµ :
• If ∆∆∆ = {δ I : δ ∈ C} (complex repeated scalar block,S= 1,F = 0),thenµ∆∆∆(M) = ρ(M).
• If ∆∆∆ = Cn×n (full block, S= 0,F = 1), thenµ∆∆∆(M) = σ(M).
• ρ(M)≤ µ∆∆∆(M)≤ σ(M), µ∆∆∆(M) = max∆∈∆∆∆,σ(∆)≤1 ρ(M∆).
Related MATLAB Command:mu(M,blk)
57
2.2.1 Upper and Lower Bounds ofµµµThe structured singular value lies between the spectral radius and the
maximum singular value ofM. However, the two bounds can be
arbitrarily far. To get a tight upper and lower bounds, we consider
transformations onM that do not affectµ∆∆∆(M), but do affectρ andσ .
To do this, define the following two subsets ofCn×n:
U := {U ∈∆∆∆ : UU∗ = In}D := {D ∈ Cn×n : detD 6= 0,D∆ = ∆D,∀∆ ∈∆∆∆}
= {D1, . . . ,DS,d1Im1, . . . ,dF ImF : Di ∈ Cci×ci ,d j ∈ C}
Note that for any∆ ∈∆∆∆, U ∈ U, andD ∈D, we have
U∗ ∈ U,U∆ ∈∆∆∆,∆U ∈∆∆∆,σ(U∆) = σ(∆U) = σ(∆)
D∆ = ∆D
58
Consequently, for allU ∈ U andD ∈D,
µ∆∆∆(MU) = µ∆∆∆(UM) = µ∆∆∆(M) = µ∆∆∆(DMD−1)
Therefore, the bounds can be tightened to
maxU∈U
ρ(UM)≤ µ∆∆∆(M)≤ infD∈D
σ(DMD−1)
• The lower bound is always an equality, i.e.,
maxU∈U
ρ(UM) = µ∆∆∆(M)
Unfortunately, the quantityρ(UM) can have multiple local maximathat are not global, so we can only obtain a lower bound.
• The upper bound can be formulated as a convex optimizationproblem, so the global minimum can be found. Unfortunately,theupper bound is not always equal toµ. We have
µ∆∆∆(M) = infD∈D
σ(DMD−1) if 2S+F ≤ 3
59
2.2.2 Smallµµµ Theorem and Robust StabilityThe basis for robust stability under structured uncertainties is the smallµtheorem. The system configuration is similar to that in smallgain
theorem, but∆ is allowed to be structured (block diagonal).
Small µ Theorem. Let γ > 0. Then the interconnected system is
well-posed and internally stable for all∆(·) ∈M(∆∆∆) with ‖∆‖∞ ≤ γ if and
only if
‖M(s)‖µ = supω∈R∪{∞}
µ∆∆∆(M( jω))< γ−1
Hence the peak value on theµ plot of the frequency response determines
the size of perturbations that the loop is robustly stable against.
60
2.3 LFT and LFT Uncertainty
Let M be a complex matrix partitioned as
M =
M11 M12
M21 M22
∈ C(p1+p2)×(q1+q2)
and let∆l ∈ Cq2×p2 and∆u ∈ Cq1×p1 be other two complex matrices.Then we can define alower LFT with respect to∆l as the map
Fl (M,•) : Cq2×p2 → Cq1×p1
withFl (M,∆l ) := M11+M12∆1(I −M22∆l )
−1M21
provided that the inverse(I −M22∆l )−1 exists. We can also define an
upper LFT with respect to∆u as
Fu(M,•) : Cq1×p1 → Cq2×p2
61
with
Fu(M,∆u) := M22+M21∆u(I −M11∆u)−1M12
provided that the inverse(I −M11∆u)−1 exists.
The motivation for the terminologies of lower and upper LFTsis clear
from the following diagram representations ofFl (M,∆l ) andFu(M,∆u).
we have
z1 = Fl (M,∆l )w1,z2 = Fu(M,∆u)w2
Related MATLAB Command:lft
62
Redheffer Star Products
An important property of LFTs is that any interconnection ofLFTs isagain an LFT. This property is by far the most often used and isthe heartof LFT machinery.
SupposeP andK are compatible partitioned matrices
P=
P11 P12
P21 P22
,K =
K11 K12
K21 K22
such that the matrix productP22K11 is well defined and square, andassume further thatI −P22K11 is invertible. Then thestar product of P
andK with respect to this partition is defined as
P⋆K :=
Fl (P,K11) P12(I −K11P22)
−1K12
K21(I −P22K11)−1P21 Fu(K,P22)
63
The interconnection is shown below.
So we have
Fl (P,Fl (K,∆)) = Fl (P⋆K,∆)
Most of the interconnection structures in control (e.g., feedback and
cascade) can be viewed as special cases of the star product.
Related MATLAB Command:starp
64
2.3.1 LFT UncertaintyLFTs are also useful in modeling uncertainty (unstructuredor structured):
P∆ = Fu(G,∆) = G22+G21∆(I −G11∆)−1G12
The uncertainty descriptions introduced earlier can be transformed to the
LFT form.
(1) Additive uncertainty
G=
0 W
I P
(2) Multiplicative uncertainty
G=
0 WP
I P
65
(3) Coprime factor uncertainty, with∆ = [ ∆M ∆N ]
G=
−M−1 P
0 −I
−M−1 P
since
Fu(G,∆) = P− M−1∆
I +
−M−1
0
∆
−1
P
−I
= P− (M+∆M)−1(−∆N +∆MP)
= (M+∆M)−1(N+∆N)
66
State Space Uncertainty
Suppose a state space realization ofP is
P=
A B
C D
The uncertain model in the state space can be expressed as
P∆ =
A+δA B+δB
C+δC D+δD
where the uncertainty is often assumed to meet the followingcondition:
δA δB
δC δD
=
E1
E2
∆[
F1 F2
]
with ∆ ∈ RH∞ and‖∆‖∞ ≤ 1.
67
The uncertainty can be expressed in LFT form:
P∆ = (D+δD)+(C+δC)(sI− (A+δA))−1(B+δB)
= Fu
A+δA B+δB
C+δC D+δD
, 1sI
= Fu
Fu
A B E1
C D E2
F1 F2 0
,∆
, 1
sI
= Fu
Fu
A E1 B
F1 0 F2
C E2 D
, 1
sI
,∆
= Fu
F1(sI−A)−1E1 F2+F1(sI−A)−1B
E2+C(sI−A)−1E1 D+C(sI−A)−1B
,∆
68
2.3.2 Pulling Out the∆∆∆’sThe basic procedure for getting the smallµ theorem configuration is
called “pulling out the∆’s”.
(a) Multiple sources of uncertainties (b) Pulling out the∆’s
69
2.3.3 Robust Stability with LFT UncertaintyThe general robust stability problem is shown below for the uncertainty
modelP∆ = Fu(G,∆).
G
∆
K
• For unstructured uncertainty, the
system is robustly stable for all
‖∆‖∞ ≤ 1 if and only if
‖Fl (G,K)‖∞ < 1
• For structured uncertainty, the
system is robustly stable for all
‖∆‖∞ ≤ 1 if and only if
‖Fl (G,K)‖µ < 1
70
Robust Stability with State Space Uncertainty
Consider the state space uncertainty described before, since
P∆ = Fu(
F1(sI−A)−1E1 F2+F1(sI−A)−1B
E2+C(sI−A)−1E1 D+C(sI−A)−1B
︸ ︷︷ ︸
G
,∆)
The condition for robust stability with state space uncertainty such that‖∆‖∞ ≤ 1 is
‖Fl (G,K)‖µ < 1
where
G=
A E1 B
F1 0 F2
C E2 D
71
2.4 Main Loop Theorem
Let M be a complex matrix partitioned as
M =
M11 M12
M21 M22
and suppose there are two defined block structures,∆∆∆1 and∆∆∆2, which arecompatible in size withM11 andM22. Define a third structure∆∆∆ as
∆∆∆ =
∆1 0
0 ∆2
: ∆1 ∈∆∆∆1,∆2 ∈∆∆∆2
then
µ∆∆∆(M)< 1⇐⇒
µ∆∆∆2(M22)< 1, and
max∆2∈∆∆∆2,‖∆2‖∞<1 µ∆∆∆1(Fl (M,∆2))< 1
72
2.5 Mixed µµµ Analysis
The parameter variations are more realistically modeled with real
uncertainties. Denote the number ofreal repeated scalarblocks byRand
the dimensions are represented by positive integersr1, · · · , rR, i.e., thei’th
real repeated scalar block isr i × r i . Together with the notation of complexstructured singular value, we define the block structure∆∆∆ ⊂ Cn×n as
∆∆∆ =
diag[δ R1 Ir1, · · · ,δ R
r Irr ,δC1 Ic1, · · · ,δC
c Icc,∆1, · · · ,∆F ] :
δ Ri ∈ R,δC
j ∈ C,∆k ∈ Cmk×mk
For consistency among all dimensions, we must have
r
∑i=1
r i +c
∑j=1
c j +F
∑k=1
mk = n
Then we can also define the structured singular value as before.
73
Chapter 3. Performance Specifications
In this chapter we consider further the feedback system properties.
1. Performance specifications: nominal and robust
2. Design tradeoff
3. Performance limitations
We will consider the following standard feedback configuration:
74
3.1 Performance Specifications
Consider the standard feedback configuration. Define
the input loop transfer matrix Li = KP
the output loop transfer matrix Lo = PK
the input sensitivity matrix Si = (I +Li)−1
the output sensitivity matrix So = (I +Lo)−1
the input complementary sensitivity matrix Ti = I −Si = Li(I +Li)−1
the output complementary sensitivity matrixTo = I −So = Lo(I +Lo)−1
the input return difference matrix I +Li
the output return difference matrix I +Lo
75
It is easy to see that the closed-loop system satisfies the following
equations:
y = To(r −n)+SoPdi +Sod
u = KSo(r −n)−Tidi −KSod
r −y = (Sor +Ton)−SoPdi −Sod
up = KSo(r −n)+Sidi −KSod
The equations show the fundamental benefits and design objectives
inherent in feedback loops. For example,
• The effects of disturbanced on the plant output can be made small by
making the output sensitivity functionSo ‘small’.
• The effects of disturbancedi on the plant input can be made small by
making the input sensitivity functionSi ‘small’.
76
With different assumption on the disturbance, we can formulate different
optimization problems for ‘optimal’ performance.
1. H∞H∞H∞ performance. The disturbance is assumed to be of bounded
power (RMS norm induced) or of bounded fuel (average-absolute
norm induced).
2. H2H2H2 performance. The disturbance is assumed to be white noise or
impulse, and the size ofy is taken as RMS value.
3. L1L1L1 performance. The disturbance is assumed to be bounded (L∞
induced).
77
H∞H∞H∞ Performance
As discussed in Chapter 4, theH∞-norm is the most used norm for asystem, thus it is a good measure for system performance. Note that theH∞-norm of a matrix equals the maximum singular value, so gooddisturbance rejection at the plant output (y) and the plant input (up)requires that
σ(So) = σ((I +PK)−1) =1
σ(I +PK)(for the effect ofd on y)
σ(SoP) = σ((I +PK)−1P) = σ(PSi) (for the effect ofdi ony)
σ(Si) = σ((I +KP)−1) =1
σ(I +KP)(for the effect ofdi onup)
σ(KSo) = σ(K(I +PK)−1) = σ(KSo) (for the effect ofd onup)
be small, particularly in the low-frequency range whered anddi areusually significant.
78
Note thatσ(PK)−1≤ σ(I +PK)≤ σ(PK)+1
σ(KP)−1≤ σ(I +KP)≤ σ(KP)+1
then1
σ(PK)+1≤ σ(So)≤
1σ(PK)−1
, if σ(PK)> 1
1σ(KP)+1
≤ σ(Si)≤1
σ(KP)−1, if σ(KP)> 1
These equations imply that
σ(So)≪ 1⇐⇒ σ(PK)≫ 1
σ(Si)≪ 1⇐⇒ σ(KP)≫ 1
σ(SoP)≈ σ(K−1) =1
σ(K), whenσ(PK)≫ 1 or σ(KP)≫ 1
σ(KSo)≈ σ(P−1) =1
σ(P), whenσ(PK)≫ 1 or σ(KP)≫ 1
79
Hence
• Good performance at plant output requires, in general, large output
loop gainσ(PK)≫ 1 in the frequency range whered is significant
and large enough controller gainσ(K)≫ 1 in the frequency range
wheredi is significant.
• Good performance at plant input requires, in general, largeinput loop
gainσ(KP)≫ 1 in the frequency range wheredi is significant and
large enough plant gainσ(P)≫ 1 in the frequency range whered is
significant.
So good multivariable feedback loop design boils down to achieving high
loop (and possibly controller) gains in the necessary frequency range.
80
3.2 Design Tradeoff
Despite the simplicity of the performance, feedback designis by no
means trivial. This is true because loop gains cannot be madearbitrarily
high over arbitrarily large frequency ranges. Rather, theymust satisfy
certain performance tradeoff and design limitations.
Typical tradeoffs are:
• Conflict between commands and disturbance error reduction (So
small) versus robust stability (To small for multiplicative uncertainty).
• Conflict between disturbance rejection (So small) and the sensor
noise reduction (To small).
• Conflict between disturbance rejection (So small) and the control
effort.
81
To summarize, good performance requires in some frequency range,typically some low-frequency range(0,ωl ),
σ(PK)≫ 1,σ(KP)≫ 1,σ(K)≫ 1
and good robustness and good sensor noise rejection in some frequencyrange, typically some high-frequency range(ωh,∞),
σ(PK)≪ 1,σ(KP)≪ 1,σ(K)≤ M
whereM is not too large. These design requirements are shown below.
82
3.3 Robust Performance
Stability is not the only property of a closed-loop system that must be
robust to perturbations. We also need ‘robust performance’, i.e., the
closed-loop performance will not degrade under perturbation.
We consider the following problem:
• Nominal Performance. Disturbance rejection on plant output
requires
‖We(I +PK)−1‖∞ < 1
• Robust Stability. Suppose the plant is under multiplicative
uncertainty, i.e.,P∆ = (I +∆SWy)P with ‖∆S‖∞ ≤ 1, then robust
stability requires
‖WyPK(I +PK)−1‖∞ < 1
83
• Robust Performance. Suppose the plant is under multiplicative
uncertainty, and we require the robust performance
‖We(I +P∆K)−1‖∞ < 1
for all P∆ = (I +∆SWy)P with ‖∆S‖∞ ≤ 1.
Define a full block∆S∆S∆S and a full block∆P∆P∆P as
∆S∆S∆S= {∆S∈ RH∞ with compatible dimension withWyT : ‖∆S‖∞ < 1}
∆P∆P∆P = {∆P ∈ RH∞ with compatible dimension withWeS: ‖∆P‖∞ < 1}Then the robust performance problem can be stated as:
µ∆S∆S∆S(WyT)< 1, andµ∆∆∆P(WeS(I +∆SWyT)
−1)< 1,∀∆S∈∆S∆S∆S
whereS= (I +PK)−1,T = PK(I +PK)−1.
84
Robust Performance and Robust Stability
The robust performance problem can be transformed to a robust stabilityproblem by the Main Loop Theorem. Define an augmented blockstructure
∆∆∆ :=
∆P 0
0 ∆S
: ∆P ∈∆P∆P∆P,∆S∈∆S∆S∆S
Since
We(I +P∆K)−1 =WeS(I +∆SWyT)−1 = Fl
WeS −WeS
WyT −WyT
,∆S
By the main loop theorem, the robust performance problem amounts to
µ∆∆∆
WeS −WeS
WyT −WyT
< 1
85
Condition of Robust Performance for SISO Systems
For a SISO system, a necessary and sufficient condition for robustperformance is
‖|WeS|+ |WyT|‖∞ < 1
that is
µ∆∆∆
WeS −WeS
WyT −WyT
= ‖|WeS|+ |WyT|‖∞
To prove this, for simplicity, assumeWe = 1,Wy = 1, the upper bound ofthe SSV at each frequency is equal to
maxω supd1
|1+PK|σ
1 0
0 d
1 −1
PK −PK
1 0
0 d−1
= maxω supd
√(1+P∗K∗d∗dPK)(1+d−1d−1∗)
|1+PK| = maxω1+|PK||1+PK| = maxω(|S|+ |T|)
86
3.4 Selection of Weighting Functions
The selection of weighting functions for a specific design problem ofteninvolves ad hoc fixing, many iterations, and fine tuning. It isvery hard togive a general formula for the weighting functions that willwork in everycase. For an SISO system,
1. Performance weight (forS). Bandwidth≥ ωb, peak≤ Ms.
We =s/Ms+ωb
s+ωbε
87
2. Control weight (forKS). Controller bandwidth≤ ωbc, peak≤ Mu.
Wu =s+ωbc/Mu
εs+ωbc
3. Robust stability weight (forT). Obtained from the information on theplant model. If no uncertainty is specified, a weight can be selectedas the control weight. Bandwidth≤ ωby, peak≤ Mp.
Wy =s+ωby/Mp
εs+ωby
88
3.5 General Design Framework
The general controller design framework is shown below.The uncertainty of the system is expressed in a general blockdiagonalblock ∆S. The performance of the system is reflected by the transfermatrix Tzw from w to z. Note here the feedback ispositive.
G
∆S
K
w z
∆P
G=
G11(s) G12(s) G13(s)
G21(s) G22(s) G23(s)
G31(s) G32(s) G33(s)
∆ :=
∆S 0
0 ∆P
M := Fl (G,K)
89
3.6 Summary of Controller Synthesis Methods
1. Nominal Performance (∆S= 0). Now the first row and first columnare zero, i.e.,G is partitioned by 2×2.
(a) H∞H∞H∞ performance. The disturbance is assumed to be of boundedpower (RMS norm induced) or of bounded fuel (average-absolutenorm induced).
minK
‖Fl (G,K)‖∞
(b) H2H2H2 performance. The disturbance is assumed to be white noiseor impulse, and the size ofy is taken as RMS value.
minK
‖Fl (G,K)‖2
(c) L1L1L1 performance. The disturbance is assumed to be bounded (L∞
induced).min
K‖Fl (G,K)‖1
90
2. Robust Stability (∆P = 0). Now the second row and second columnare zero.G is also partitioned by 2×2.
(a) H∞H∞H∞ analysis. For unstructured uncertainty,
minK
‖Fl (G,K)‖∞
(b) µµµ analysis. For structured uncertainty,
minK
‖Fl (G,K)‖µ
3. Robust Performance.
(a) RobustH∞H∞H∞ performance. Nominal performance expressed in
H∞ and robust stability expressed inH∞ or µ. By the Main Loop
Theorem, we have aµµµ synthesisproblem.
minK
‖Fl (G,K)‖µ
(b) RobustH2H2H2 performance.
91
4. Nominal Performance + Robust Stability.
Robust performance is not easy to obtained. Sometimes we areonly
concerned withnominal performance and robust stability, instead of
robust performance.
(a) Mixed H2/H∞H2/H∞H2/H∞ problem. Nominal performance inH2 and robust
stability inH∞.
minK
‖Fl (G1,K)‖2
under the constraint
‖Fl (G2,K)‖∞ < 1
(b) Multi-objective H∞H∞H∞ problem. Nominal performance inH∞ and
robust stability inH∞.
minK
‖Fl (G1,K)‖∞
92
under the constraint
‖Fl (G2,K)‖∞ < 1
(c) Mixed H2/µH2/µH2/µ problem. Nominal performance inH2 and robust
stability in µ.
minK
‖Fl (G1,K)‖2
under the constraint
‖Fl (G,K)‖µ < 1
(d) Mixed L1/H∞L1/H∞L1/H∞ problem. Nominal performance inL1 and robust
stability inH∞.
minK
‖Fl (G1,K)‖1
under the constraint
‖Fl (G2,K)‖∞ < 1
93
Chapter 4. Robust Controller Synthesis
We have discussed performance specifications and tradeoffs. One simple
method to achieve performance specifications and tradeoffsis loop
shaping. The procedure of designing controllers by changing the shapes
of the (closed- and/or open-) loops of a system is calledloop shaping.There are two kinds of loop shaping methods:
(1) Shape the closed loopsS, T, or KS;
(2) Shape the open loopL.
We will discuss these methods in this chapter.
94
4.1 Shaping the Closed Loops
We can directly change the shapes of the closed loops (sensitivity function
and the mixed sensitivity function) to achieve the desired performance.
1. S-T Approach.
Consider the robust performance problem:
• Nominal Performance:‖WeS‖∞ < 1
• Perturbed model:P∆ = (I +∆Wy)P with ‖∆‖∞ < 1.
A test for the robust performance problem of a SISO system is
‖|WeS|+ |WyT|‖∞ < 1
A compromise condition for the robust performance test is
‖(|WeS|2+ |WyT|2)1/2‖∞ < 1
95
since
1√2(|WeS|+ |WyT|)≤ (|WeS|2+ |WyT|2)1/2 ≤ |WeS|+ |WyT|
So we obtain the followingH∞ optimization problem∥∥∥∥∥∥
WeS
WyT
∥∥∥∥∥∥
∞
< 1
The problem can be formulated as in the following figure.
K P
Wy
r
_
y
We
e
z1 z2
It was first proposed by Safonov et. al., and is calledS-T approach.
96
2. S-KSApproach.
Consider the robust performance problem:
• Nominal Performance:‖WeS‖∞ < 1
• Perturbed model:P∆ = P+∆Wu with ‖∆‖∞ < 1.
A test for the robust performance problem of a SISO system is
‖|WeS|+ |WuKS|‖∞ < 1
A compromise condition for the test is
‖(|WeS|2+ |WuKS|2)1/2‖∞ < 1
or ∥∥∥∥∥∥
WeS
WuKS
∥∥∥∥∥∥
∞
< 1
97
This problem can be formulated as in the following figure.
K P
Wu
r
_
y
We
e
z1 z2
u
It was proposed by Postlewaith et al., and is calledS-KSapproach.
We see that by introducing the weightWu, we can put constraint on
the size of control input, which is required in practice.
98
3. Mixed-sensitivity Problem.Combination of theS-KSandS-T approach.
∥∥∥∥∥∥∥∥
WeS
WuKS
WyT
∥∥∥∥∥∥∥∥
∞
< 1
K P
Wy
r
_
y
We
e
z1 z3
Wu
u
z2
The significance of this approach is that not only the closed-loopshapes ofSandT can be pre-specified, but also the control input canbe constrained. So it is more realistic.
99
Remarks on Mixed-Sensitivity Approach
• When the plant is strictly proper, i.e.,Dp = 0, then the ‘D12’ block of
theS-T problem is singular, thus violates the regularity assumption
of the standardH∞ problem. To solve this problem, we can perturb
Dp to Dp = Dp+ ε I with a smallε .
• TheS-KS is regular as long asDu is nonzero, which can always be
guaranteed. So it is with the mixed sensitivity problem.
• When the plant contains integrators, i.e.,Ap has zero eigenvalues,
then the regularity assumption (A1) or (A3) might not be guaranteed.
In this case we can perturbAp to Ap = Ap+ ε I with a smallε .
• TheH∞ controller of the mixed sensitivity problems always cancels
the stable poles of the plant with its transmission zeros, which is
undesired if the plant contains slow-mode poles.
100
4.2 Shaping the Open-Loops
The mixed-sensitivity approach use the closed loop shapingmethod.Since open loop shape is closely related to closed loop shapes, we candesign by using open loop shaping method. The following figure showsthe desired shape of the open loop for good performance:
Due to the uncertainties in the plant model, loop shape should beguaranteed with a certain degree of robustness.
101
4.2.1 Loop ShapingH∞H∞H∞ DesignGiven a plantP, loop shapingH∞ design procedure goes as follows:
1. Loop Shaping. Pre-compensatorW1 and/or post-compensatorW2 are
used to shape the singular values ofP such that the shaped plant
P=W2PW1 has the desired open-loop shape.
2. Robust Stabilization. For the shaped plantP, find a controllerK
such that the robust stability marginε is maximized.
ε−1 = infK
∥∥∥∥∥∥
I
K
(I + PK)−1M−1
∥∥∥∥∥∥
∞
whereP= M−1N is an lcf ofP.
3. The final feedback controller is
K =W1KW2
102
It can be seen that in loop-shapingH∞ approach, the first step amounts todesign for performance, while the second step amounts to robustnessguaranty.
Some advantages of the loop shapingH∞ approach are as follows:
• The approach combines the classical loop shaping idea with therobust control idea.
• TheH∞ problem in step 2 is always regular and the infimum can becomputed explicitly without iteration. Further, the designedcontroller stabilizes all the plants in
P= {(M+∆M)−1(N+∆N) : ‖[ ∆M ∆N ]‖∞ < ε}
• There are no pole-zero cancellations between the plant and theH∞
controller.
• ε is a ‘design indicator’. A reasonable value indicates that the loopshapes can be well approximated together with good robust stability.
103
4.2.2 Robust Stabilization of Coprime FactorsThe second step of the loop shapingH∞ design is a robust stabilization
problem for coprime factor uncertainty for a plantP.
ε−1 = infK
∥∥∥∥∥∥
I
K
(I +PK)−1M−1
∥∥∥∥∥∥
∞
whereP= M−1N is an lcf ofP.
It is clear thatthe robust stabilization problem of coprime factors canbe regarded as a special case of theS-KSproblem, with M−1 as theweight. So it is always regular, and we can use Riccati equation-based
method or LMI to solve the optimization problem iteratively. However,
due to special structure of the problem, we will show that theproblem can
be solved without iteration.
104
To show that, without loss of generality, we can choose the lcf (M, N) of
P as anormalized one, i.e.,[
M N][
M N]∼
= I
In this case[ M N ] is co-inner, multiply it with a matrix will not
change theH∞ norm of the matrix.∥∥∥∥∥∥
S
KS
M−1
∥∥∥∥∥∥
∞
=
∥∥∥∥∥∥
S
KS
M−1[
M N]
∥∥∥∥∥∥
∞
=
∥∥∥∥∥∥
S
KS
[
I P]
∥∥∥∥∥∥
∞
Then the robust stabilization problem is equivalent to
ε−1 = infK
∥∥∥∥∥∥
(I +PK)−1 (I +PK)−1P
K(I +PK)−1 K(I +PK)−1P
∥∥∥∥∥∥
∞
105
SupposeP=
A B
C 0
is a stabilizable and detectable state-space
realization (Here for simplicity, we assume the plant P is strictly proper,
which is almost always the case in practical controller design).
The generalized plant for theH∞ problem is
I P −P
0 0 I
I P −P
=
A [ 0 B ] −B
C
0
I 0
0 0
0
I
C [ I 0 ] 0
Now the suboptimal (< γ) problem has a solution when the followingAREs have nonnegative definite stabilizing solutions andρ(X∞Y∞)< γ2.
ATX∞ +X∞A− (1− γ−2)X∞BBTX∞ +γ2
γ2−1CTC= 0
106
AY∞ +Y∞A−Y∞CTCY∞ +BBT = 0
Let
X := (1− γ−2)X∞, Y :=Y∞
then we have
ATX+XA−XBBTX+CTC= 0
AY+YAT −YCTCY+BBT = 0
The solutions are both independent ofγ, which suggests that the optimum
can be obtained without iteration.
Indeed, from the requirementρ(X∞Y∞)< γ2, we get
1+ρ(XY)< γ2
so the robust marginε equals
ε = γ−1min = [1+ρ(XY)]−1/2
107
In summary, theH∞ optimum for coprime factor uncertainty can be found
by solving two AREs without iteration:
ATX+XA−XBBTX+CTC= 0
AY+YAT −YCTCY+BBT = 0
and the maximum robustness margin can be computed explicitly
ε = [1+ρ(XY)]−1/2
Once the robust margin is obtained, a suboptimalH∞ controller can then
be constructed. Moreover, for this problem, an optimalH∞ controller can
be constructed using the following generalized state-space description:
Eq= (E(A+BBTX)+ ε−2YCTC)q+ ε−2YCTy
u= BTXq
whereE := (1− ε−2)I +YX.
108
4.2.3 Design Procedure ofH∞H∞H∞ Loop-Shaping
1. Scale the plant outputs and inputs. This is very importantfor mostdesign procedures and is sometimes forgotten. In general, scalingimproves the conditioning of the design problem, it enablesmeaningful analysis to be made of the robustness propertiesof thefeedback system in the frequency domain.
2. Order the inputs and outputs so that the plant is as diagonal aspossible. The relative gain array (RGA) can be useful here.
3. Select the pre- and/or postcompensators to obtain the shaped plantP=W2PW1. The desired shapes normally mean high gain at lowfrequencies, roll-off rates of approximately 20 dB/decadeat thedesired bandwidth(s), and higher rates at high frequencies.
(a) W2 is usually chosen as a constant, reflecting the relativeimportance of the outputs to be controlled.
109
(b) In generalW1 has the formW1 =WpWaWg.
• Wp contains dynamic shaping. Integral action for lowfrequency performance; phase-advance for reducing the roll-offrates at crossover; and phase-lag to increase the roll-off rates athigh frequencies should all be placed inWp if desired.
• Wa is a constant that aligns the singular values at a desiredbandwidth (optional). This is effectively a constant decouplerand should not be used if the plant is ill-conditioned in terms oflarge RGA elements.
• Wg is an additional gain matrix to provide control over actuatorusage (optional). It is diagonal and adjusted so that the actuatorrate limits are not exceeded for reference demands and typicaldisturbances on the scaled plant outputs.
4. Robustly stabilize the shaped plantP. If the robust margin is toosmall (ε < 0.25), then go back to the previous step and modify theweights.
110
5. Analyze the design and if all the specifications are not metmake
further modifications to the weights.
6. Implement the controller. The configuration shown below has been
found useful when compared with the conventional feedback.This is
because the references do not directly excite the dynamics of K,
which can result in large amounts of overshoot (classical derivative
kick).
P(s)r y
_
+
)(~sK
)(1 sW
)(2 sW
)0()0(~
2WK
Related MATLAB Commands:ncfsyn, cf2sys, emargin
111
4.2.4 Guidelines for Loop Shaping DesignSome guidelines for the loop-shaping design:
• The loop transfer function should be shaped in such a way thatit has
low gain around the frequency of the modulus of any right-half plane
zeroz. Typically, it requires that the crossover frequency be much
smaller than the modulus of the right-half plane zero; say,ωc <|z|2
for any real zero andωc < |z| for any complex zero with a much
larger imaginary part than the real part.
• The loop transfer function should have a large gain around the
frequency of the modulus of any right-half plane pole.
• The loop transfer function should not have a large slope nearthe
crossover frequencies.
These guidelines are consistent with the rules used in classical control
theory.
112
Chapter 5. Model Reduction
Simple linear models/controllers are preferred over complex ones incontrol system design for some obvious reasons:
(1) Simple models are much easier to do analysis and synthesis with.
(2) Simple controllers are easier to implement and are more reliable.
(3) In the case of infinite dimensional system, the mode/controllerapproximation becomes essential.
A model order-reduction problem can, in general, be stated as follows:Given a full-order model G(s), find a low-order model (say, an rth ordermodel Gr ), such that G and Gr are close in some sense. There are manyways to define the closeness of an approximation. The most used norm istheL∞-norm. So the model reduction problem can be formulated as
infdeg(Gr )<r
‖G−Gr‖∞
113
5.1 Truncation Methods
Truncation methods of model reduction seek to remove or truncateunimportant states from state-space models. There are two options:
1. State-space truncation.Consider a linear, time-invariant system with the realization
G=
x(t) = Ax(t)+Bu(t), x(0) = x0
y(t) = Cx(t)+Du(t)
Divide the state vectorx into two components:
x(t) =
x1(t)
x2(t)
where ther-vectorx1(t) contains the components to be retained, andthe(n− r)-vectorx2(t) contains the components to be discarded.
114
Now partition the matricesA, B andC conformably withx to obtain
A=
A11 A12
A21 A22
,B=
B1
B2
,C=[
C1 C2
]
By omitting the states and dynamics associated withx2(t), we obtain
the low-order system
Gr =
p(t) = A11p(t)+B1u(t), p(0) = p0
q(t) = C1p(t)+Du(t)
Therth-order truncation of the realization(A,B,C,D) is given by
Tr(A,B,C,D) = (A11,B1,C1,D)
115
Properties of model truncation:
• The truncated system may be unstable even if the full-order
system is stable.
• The truncated system realization may be nonminimal even if the
full-order system realization is minimal.
• All reduced-order models obtained by truncation match the
full-order model at∞, i.e.,Gr(∞) = G(∞) = D.
• The steady-state error associated with state-space truncation is
G(0)−Gr(0) =CA−1B−C1A−111 B1
• The truncation error is
G(s)−Gr(s) = C(s)(sI− A(s))−1B(s)
whereA(s) = A22+A21(sI−A11)−1A12,
B(s) = B2+A21(sI−A11)−1B1, C(s) =C2+C1(sI−A11)
−1A12.
116
2. Singular perturbation approximation (SPA) .The state-space truncation does not retain the steady-state error, so itis sometimes not acceptable in applications. We can use singularperturbation approximation to improve the low-frequencycharacteristics.
If x2(t) represents the fast dynamics of the system, we mayapproximate the low-frequency behavior by setting ˙x(t) = 0. Thisgives
0= A21x1(t)+A22x2(t)+B2u(t)
SupposeA22 is nonsingular. Eliminatingx2(t), we have
Gr =
p(t) = (A11−A12A−122 A21)p(t)+(B1−A12A
−122 B2)u(t)
q(t) = (C1−C2A−122 A21)p(t)+(D−C2A−1
22 B2)u(t)
Therth-order singular perturbation approximation is given by
Sr(A,B,C,D) = (A11, B1,C1, D)
117
where
A11 = A11−A12A−122 A21, B1 = B1−A12A
−122 B2
C1 =C1−C2A−122 A21, D = D−C2A−1
22 B2
Singular perturbation approximation is equivalent to truncation for
the plantG(1s). That is, if we setH(s) = G(1
s), and performing a
state-space truncation ofH(s) to obtainHr(s), then the SPA is
Sr(G) = Hr(1s).
Since the singular perturbation and truncation operationsare related in a
straight-forward way, it suffices to develop all the theoretical results for
state-space truncation. When the low-frequency is important, the singular
perturbation approximation is the method of choice. Conversely, direct
truncation should be preferred when good high-frequency modelling is
the central concern.
Related MATLAB Commands:strunc, sresid
118
5.2 Balanced Realization
State-space truncation can be used to reduce a model. However, we
cannot do it with any state-space realization, because truncation errorcannot be guaranteed. We need to transform a realization to somestandard form. For example, we can put theA-matrix in Jordan canonicalform, then we have the classical modal truncation (dominated pole)method. The most used method is to put the realization to abalanced
form.
1. Motivation for balanced realization.
Given a modelG∈ RH∞ with a state-space realization
x = Ax+Bu
y = Cx+Du
Suppose a reduced-order model isGr ∈H∞. A natural criterion with
119
which to measure the absolute error is
‖G−Gr‖∞ = supu∈L2
‖y−yr‖2
‖u‖2
It is aL2-induced norm, so for the error to be small, we should deletethose components of the state-vectorx that are at least involved in theenergy transfer from the inputu to the outputy. This observationleads us to consider two closely related questions:
(1) What is the output energy resulting from a given initial statex(0) = x0?
(2) What is the minimum input energy required to bring the statefrom zero to the given initial statex(0) = x0?
The solutions are well known:
(1) Supposex(0) = x0 is given and thatu(t) = 0. TheL2[0,∞) normof y is given by‖y‖2
2 = xT0 Qx0, in whichQ is the observability
gramian.
120
(2) Consider the LQ problem
minu∈L2(−∞,0]
∫ 0
−∞uT(t)u(t)dt
subject to ˙x= Ax+Buwith x(0) = x0. Defineτ =−t, p(τ) = x(t)
andv(τ) = u(t), an equivalent problem is
minv∈L2[0,∞)
∫ ∞
0vT(τ)v(τ)dτ
subject to ˙p(τ) =−Ap(τ)−Bv(τ) with p(0) = x0. By LQ theory,
the optimal control isu(t) = BTP−1x(t) whereP is the
controllability gramian, and
minu∈L2(−∞,0],x(0)=x0
∫ 0
−∞uT(t)u(t)dt = xT
0 P−1x0
121
Combining the two solutions we get
maxu∈L2(−∞,0],x(0)=x0
∫ ∞0 yT(t)y(t)dt∫ 0−∞ uT(t)u(t)dt
=xT
0 Qx0
xT0 P−1x0
=αTP
12 QP
12 α
αTα
wherex0 = P12 α.
These calculation suggests that in order to keep‖G−Gr‖∞ small, the
state-space for the truncated system should be the space spanned by
the eigenvectors corresponding to the larger eigenvalues of P12 QP
12 .
That is, we should truncate a realization in whichP12 QP
12 is diagonal,
with the eigenvalues ordered in descending order.
122
2. Balanced realization.The realization withP
12 QP
12 diagonal is known as a balanced
realization. It always exists for a stable minimal realization.
Definition. A realization(A,B,C) is balancedif A is asymptotically
stable and the controllability and observability gramiansare equal
and diagonal. That is
AΣ+ΣAT +BBT = 0
ATΣ+ΣA+CTC= 0
in which
Σ =
σ1Ir1 0 · · · 0
0 σ2Ir2 · · · 0...
......
...
0 0 · · · σmIrm
,σ1 > σ2 > · · ·> σm > 0
123
The valueσi ’s are called theHankel singular valuesof (A,B,C).
Remarks:
• For a balanced realization, the basis for the state space is equallycontrollable and observable, with its “degree” of controllabilityand observability given by the corresponding diagonal entry of Σ.
• Suppose(A,B,C) is a balanced realization and the initialx0 ispartitioned accordingly asΣ, then
maxu∈L2(−∞,0],x(0)=x0
∫ ∞0 yT(t)y(t)dt∫ 0−∞ uT(t)u(t)dt
=m
∑i=1
σ2i xT
i xi
This shows thatσ2i is a measure of the extent to which the
correspondingr i dimensional subspace of the state space isinvolved in the transfer of energy from past inputs to futureoutputs.
• A given realization(A,B,C) can be transformed to a balancedrealization if and only if it is stable and minimal.
124
The state transformation matrixT can be obtained by:
(1) Find the controllability and observability gramiansP andQ.
(2) Perform a Cholesky factorization ofP:
P= RRT
(3) Perform a singular value factorization ofRTQR:
RTQR=UΣ2UT
(4) The transformation matrix is
T = Σ12UTR−1
The balanced realization is(TAT−1,TB,CT−1), with theobservability and controllability gramians
TPT∗ = (T−1)∗QT−1 = Σ
Related MATLAB Command:sysbal
125
5.3 Model Reduction by Balanced Truncation
Suppose(A,B,C) is a balanced realization ofG. PartitionΣ as
Σ =
Σ1 0
0 Σ2
with
Σ1 = diag{σ1Ir1, · · · ,σl Ir l },Σ2 = diag{σl+1Ir1+1, · · · ,σmIrm}
If (A,B,C) is partitioned conformably withΣ, we can obtain areduced-order modelGr = (A11,B1,C1,D), with r = r1+ . . .+ r l bystate-space truncation. We have
(1) (A11,B1,C1) is also a balanced realization, with the controllabilityand observability gramians equal toΣ1. SoA11 is stable and(A11,B1,C1) is minimal.
126
(2) The error bound is
‖G−Gr‖∞ ≤ 2(σl+1+ . . .+σm)
If we want to retain the steady-state performance, then we need to use
SPA to truncate(A,B,C). The error bound is the same as that using
state-space truncation.
For example, consider a 5th-order system
G(s) =1
(s+1)5
The Hankel singular values ofG(s) is
0.7292,0.2826,0.0601,0.0069,0.0003
The last three are small compared with the first two. So we can reduce the
model to order 2, with an error less than 0.1347.
127
The reduced-order model by state-space truncation is
Gr1 =−0.0738s+0.1107
s2+0.3979s+0.1239
and by the SPA truncation is
Gr2 =0.1070s2−0.2328s+0.1994
s2+0.7536+0.1994
Step Response
Time (sec)
Am
plitu
de
0 5 10 15 20 25 30−0.2
0
0.2
0.4
0.6
0.8
1
1.2
Singular Values
Frequency (rad/sec)
Sin
gula
r V
alue
s (d
B)
10−3
10−2
10−1
100
101
102
−90
−80
−70
−60
−50
−40
−30
−20
−10
Related MATLAB Commands:sysbal, strunc, sresid, hksv
128
5.4 Optimal Model Reduction
The motivation for the balanced truncation method of model reduction
comes from energy transmission arguments. If a high-order modelG
mapsu to y via y= Gu, then the idea is to select a low-order modelGr ,
which mapsu to yr , such that
e= supu∈L2(−∞,0]
(∫ ∞0 (y−yr)
T(y−yr)dt∫ 0−∞ uTudt
)
is small whenu(t) = 0 for t > 0. The quantity can be thought of as the
energy gain from past inputs to future outputs.
The balanced realization method guarantees that the gain issmall. It is
preferred if we can make it minimum. Thus we have the optimal Hankel
norm approximation.
129
Hankel Operators
TheHankel operator of a linear system is the prediction operator thatmaps past inputs to future outputs, assuming the future input is zero.
Suppose the systemG∈ RH∞ is defined by the minimal state-spacerealization
x = Ax+By, x(−∞) = 0
y = Cx+Du
If u∈ L2(−∞,0], then future outputs are determined by
y(t) =∫ 0
−∞CeA(t−τ)Bu(τ)dτ , t > 0
If we setv(t) = u(−t), theny(t) = (ΓGv)(t) for t > 0, whereΓG : L2[0,∞) 7→ L2[0,∞) is the Hankel operator:
(ΓGv)(t) :=∫ ∞
0CeA(t+τ)Bv(τ)dτ
130
Hankel Norm
TheHankel norm of a system is theL2[0,∞) induced norm of its
associated Hankel operator.
‖G‖H = ‖ΓG‖=(
supu∈L2(−∞,0]
( ∫ ∞0 yTydt∫ 0−∞ uTudt
)) 12
For a givenx(0) = x0, we have (c.f. section 13.2)
supu∈L2(−∞,0],x(0)=x0
( ∫ ∞0 yTydt∫ 0−∞ uTudt
)
=xT
0 Qx0
xT0 P−1x0
Thus
‖G‖2H = sup
x0
xT0 Qx0
xT0 P−1x0
= λmax(PQ) = σ21 (largest Hankel singular value)
131
Remarks:
(1) The Hankel norm of a system is a measure of the effect of itspast
input on its future output, or the amount of energy that can bestored
in and then retrieved from the system.
(2) The Hankel singular values of a system are, in fact, the singular
values of its Hankel operator. They can be obtained by
σi(ΓG) = λ12
i (PQ)
The Hankel norm equals the largest Hankel singular value.
(3) ‖G‖H ≤ ‖G‖∞, since for an arbitrary unit energy input inL2(−∞,0],
‖G‖2H is the least upper bound on the energy of thefuture output,
while ‖G‖2∞ is the least upper bound on the energy of thetotal output.
132
(4) For any anticausal systemF, if u∈ L2(−∞,0], then(Fu)(t) is zero
for t > 0. Thus the future output is unaffected by an addition of any
anticausalF and it is immediate that
‖G‖H ≤ ‖G−F‖∞
is satisfied for any anticausalF. In fact, sinceF ∈H−∞ impliesF is
anticausal, we have
‖G‖H = minF∈H−
∞‖G−F‖∞
which is referred to as Nehari’s theorem.
133
Optimal Hankel Norm Model Reduction
The optimal Hankel norm model reduction problem is:Given a full-order
model G(s), find a low-order model Gr such that
infdeg(Gr )<r
‖G−Gr‖H
The error is now in the sense of Hankel norm rather thanL∞ norm.
The optimum of the problem is
infdeg(Gr )<r
‖G−Gr‖H = σr+1
For the optimal Hankel norm approximationGr , the error in theL∞ norm
is
‖G−Gr‖∞ ≤ (σr+1+ . . .+σm)
so it is just half of the bound for balanced truncation method.
134
Since‖ · ‖H is bounded by‖ · ‖∞, so for anyrth-order model we have
‖G−Gr‖∞ ≥ σr+1
Example. Consider the 5th-order system discussed above. The 2rd-ordermodel obtained by the optimal Hankel norm reduction method is:
Gr(s) =0.0578s2−0.1409s+0.1465
s2+0.5248s+0.1554Step Response
Time (sec)
Am
plitu
de
0 5 10 15 20 25 30−0.2
0
0.2
0.4
0.6
0.8
1
1.2
Singular Values
Frequency (rad/sec)
Sin
gula
r V
alue
s (d
B)
10−2
10−1
100
101
102
−70
−60
−50
−40
−30
−20
−10
Related MATLAB Command:hankmr
135
5.5 Coprime Factorization Reduction
Another method for model/controller reduction is to use thecoprime
factorization of a model. Suppose an nlcf of a plantG is G= M−1N.
If we reduce the ‘numerator’ and the ‘denominator’ of the model
simultaneously tor-th order,Mr , Nr , thenGr := M−1r Nr appears to be a
‘close’ approximation toG.
To show that, let
D(s) :=
M
N
, Dr(s) :=
Mr
Nr
Consider the approximation ofD(s) by Dr(s). Suppose there exists a
Q(s) ∈ RH∞ such that
Dr(s) = D(s)Q(s)
136
then the approximation error‖D(s)−Dr(s)‖∞ < γ means∥∥∥∥∥∥
M
N
−
M
N
Q
∥∥∥∥∥∥
∞
< γ
soGr belongs to
{(M+∆M)−1(N+∆N) : ‖[ ∆M ∆N ]‖∞ ≤ γ}
Thus if γ is small, thenGr is close toG in the coprime factorization sense
(gap metric). Thus we can use normalized coprime factorization to
perform a model reduction, where the approximation ofD(s) by Dr(s)
can be done either using balanced truncation or optimal Hankel reduction
method.
137
Example. Consider the 5th-order system discussed above. The 2rd-order
model obtained by the coprime factorization reduction method is:
Gr(s) =0.0968s2−0.2155s+0.1899
s2+0.7165s+0.1899
Step Response
Time (sec)
Am
plitu
de
0 5 10 15 20 25 30−0.2
0
0.2
0.4
0.6
0.8
1
1.2
Singular Values
Frequency (rad/sec)
Sin
gula
r V
alue
s (d
B)
10−2
10−1
100
101
102
−70
−60
−50
−40
−30
−20
−10
Related MATLAB Commands:sncfbal, sresid, cf2sys
138
Summary
• The balanced truncation method and the optimal Hankel method are
not applicable to unstable models. If a model contains unstable poles,
we need first decompose it into a stable part and an unstable part. The
order can only be reduced for the stable part. The final reduced-order
model is then obtained by adding the unstable part and the
reduced-order stable part.
• The coprime factorization reduction method can be applied to both
stable and unstable models, so there is no need to have a
decomposition first.
• The coprime factorization reduction method makes sense in the gap
metric. It does not guarantee a bound for the error in the infinity
norm.
139
5.6 PID Approximation
About 90% of industrial controllers are of PID-type. So it isuseful if the
high-order controllers can be reduced to PID control structure. Two
methods can be used to make the approximation:
1. In the state-space domain.
2. In the frequency-domain.
140
PID Approximation in Time-Domain
Consider a controllerK(s), given by a state-space realization of the form
x= Akx+Bky
u=Ckx+Dky
Find a similarity transformationT such that
TAkT−1 =
0 0
0 A2
This transformation can be computed using the eigenvalue decompositionof Ak. With thisT, the new state-space realization is given by
˙x= Akx+ Bky,
u= Ckx+ Dky,
141
with Ak = TAkT−1, Dk = Dk, and
Ck =CkT−1 =
[
C1 C2
]
, Bk = TBk =
B1
B2
A PID approximation of the form
KPID(s) = Kp+Ki/s+Kds
can now be obtained by truncating the Maclaurin expansion ofthecontroller with respect to the variables:
K(s) =[
C1 C2
]
sI 0
0 sI−A2
−1
B1
B2
+Dk
= C1B1s +(Dk−C2A−1
2 B2)−C2A−22 B2s+ · · · ,
So we have
Kp = Dk−C2A−12 B2, Ki =C1B1, Kd =−C2A−2
2 B2.
142
It is then clear that the resulting PID controller achieves good
approximation at low frequencies, especially the integralaction, so we
can expect that the resulting PID controller will retain thedisturbance
rejection performance of the high-order controller.
The following steps need to be considered after the reduction procedure:
1. Due to the minimum-phase requirement of a PID controller,the signs
of the proportional gain, the integral gain and the differentiator
should be the same. So if the corresponding elements ofKp, Ki and
Kd do not have the same sign, then we need to discard the term that
has the opposite sign with the element inKi , which is the most
important term.
143
2. The derivative action should be taken with care. We find that a PID
approximation with ideal differentiators sometimes destabilizes a
process even though the original high-order controller works well.
This seldom happens for single-loop processes, but it is quite
common for multivariable processes. We believe the reason is that
the phase information of the original controller is lost when using an
ideal differentiator. Thus, a filter of the form1αs+1 should be used for
the differentiator.
144
PID Approximation in Frequency Domain
The standard procedure for obtaining PID parameters from IMCcontrollers is to expand the final controller into MacLaurian series and getthe coefficients of the first three terms. However, the procedure is notconvenient and sometimes not sufficient.
One-point Approximation
A new method to approximate any high-order controller with PID in thefrequency domain with one point approximation goes as follows:
1. Given any controllerK(s), get a frequency range of interest, computethe frequency response ofK(s).
2. Find the frequencyωz such that the magnitude ofK( jω) achieves itsminimum value.
145
3. Then the approximated PID is
KPID(s) = Kp+Ki
s+Kds
with
Ki = |K( jωs)|ωs
Kp = Re(Kwz)
Kd = Im(Kwz)/ωz
whereωs is any frequency that is small (sayωs = 0.001), and
Kwz= K( jωz)−Ki/( jωz)
146
10−2
10−1
100
101
102
103
−10
0
10
20
30
40
50
Frequency(rad/sec)
Mag
nitu
de(d
B)
The procedure amounts to approximatingK(s) by a PID controller withthe same integral action and the lowest turning point, and the resulted PIDcontroller retains the magnitude of the IMC controller at low to mediumfrequency range, thus the load-rejecting performance willbe guaranteedclose to that of the IMC controller.
147
From the design practice, it is noted that the above approximation
procedure works well for stable processes, however, for lightly-damped
and unstable processes, an additional lead compensator maybe needed to
cascade with the PID controller to retain the performance ofthe
high-order controller, i.e., the final controller should bein the form
KPIDm(s) = (Kp+Ki
s+Kds)
αs+1βs+1
The parameters of the lead compensatorαs+1βs+1 are determined by
approximating the phase of theKPIDm(s) with that of the original
controller at a certain frequency.
148
The procedure to determineα andβ goes as follows (continue from the
above procedure):
4) Find the frequencyωp such that the phase ofK( jω) achieves its
maximum value.
5) Let
φm = ∠K( jωp)−∠KPID( jωp)
Soφm is the phase advance between the original controller and the
approximated PID controller atωp.
6) If φm < 0, then it is not necessary to add the lead compensator (i.e.,
α = 0,β = 0); Otherwise, let
a=1+sin(φm)
1−sin(φm)
149
and set
β =1
ωp√
a
α = aβ
The procedure is just a modification of the design of a lead compensator
in frequency domain in classical control theory, which can be found in
any undergraduate textbook.
The Bode plot of a TDF-IMC controller for a lightly-damped, unstable
plant and its PID approximations by the above procedure are shown
below.
150
10−3
10−2
10−1
100
101
102
103
0
20
40
60
80
100
120
140
Mag
nitu
de(d
B)
10−3
10−2
10−1
100
101
102
103
−100
−50
0
50
100
150
200
Frequency(rad/sec.)
Pha
se(° C
)
It is observed that the magnitude and phase of a pure PID approximation
are indeed close to the IMC controller at low frequency. However, the
frequency range is too small (0.3 rad/s) compared with a modified PID
approximation, which extends the range to 5 rad/s. So betterperformance
can be retained.
151
Multiple-point Approximation
Given a range of frequencyωi (i = 1, · · · ,m), suppose we want to use the
following nth-order transfer function to approximateK( jωi),
K(s) =ansn+an−1sn−1+ · · ·+a1s+a0
bnsn+bn−1sn−1+ · · ·+b1s+1
then ideally the following equations should be satisfied:
an( jωi)n+an−1( jωi)
n−1+ · · ·+a1( jωi)+a0
bn( jωi)n+bn−1( jωi)n−1+ · · ·+b1( jωi)+1= K( jωi), i = 1, · · · ,m
Define
x=[
an · · · a1 a0 bn · · · b1
]T
b=[
K( jω1) · · · K( jωm)]T
152
Θ=
( jω1)n · · · ( jω1) 1 −K( jω1)( jω1)
n · · · −K( jω1)( jω1)...
......
......
......
( jωm)n · · · ( jωm) 1 −K( jωm)( jωm)
n · · · −K( jωm)( jωm)
So the equations become
Θx= b
To make sure the solutionx is real, we define
A=[
Real(Θ) Imag(Θ)]
,z=[
Real(b) Imag(b)]
where Real (Imag) denotes the real (imaginary) part of a matrix. So we
have
Ax= z
A least-squared solution can be found.
153
It is observed that the least-squared solution may not be positive, so the
positiveness constraint is added and the problem becomes
minx>0
(Ax−z)T(Ax−z)
A standard quadratic programming algorithm can be used to solve it
efficiently.
154
Chapter 6. Robustness Measure
In this chapter a simple method will be proposed to measure system
robustness and performance. It is shown that the method can be applied to
various industrial processes, no matter they are stable, integrating, or
unstable; single-loop, cascade, or multivariable.
155
6.1 Classical Robustness Measures
6.1.1 Robustness Measure in terms ofMs
Gain and phase margins are generally used in the classical control theory
to measure system robustness for a single-input-single-output (SISO)
process. If the open-loop transfer function of a linear system is
L(s) := G(s)K(s), then the gain margin (GM) is
GM =1
|L( jωp)|
whereωp is the phase crossover frequency, and the phase margin (PM) is
PM= ∠L( jωg)−180◦
whereωg is the gain crossover frequency.
156
A Nyquist plot interpretation of the gain and phase margins is shown
below
-1
Phase Margin
gω
Pω
|)(| PjL ω
)( ωjL
|)(1|min ωω
jL+
Re
jIm
157
Clearly, GM and PM measure the distance from the Nyquist plotL( jω) to
the critical point(−1, j0) in two different viewpoints: GM indicates the
closeness of the intersection of the negative real axis by the Nyquist plot
of L( jω) to the(−1, j0) point; while PM indicates the distance with the
variation in phase. A more compact indicator would be
Ms := ‖S‖∞ = maxω
∣∣∣∣
11+L( jω)
∣∣∣∣=
1minω |1+L( jω)|
whereS= 11+GK is the sensitivity function. Thus 1/Ms is the minimal
distance fromL( jω) to the critical point.
Note that the peak ofSusually occurs at the low and mid-frequencies, so
Ms is a measure of system robustness against low and mid-frequency
uncertainty (e.g, gain variation).
158
6.1.2 Robustness Measure in terms ofMp
Consider the standard unity feedback control system
K Gr y
_
e
d1
+
d2
+
+
+
+
If the open-loop systemG is subject to multiplicative uncertainty(G∆ = (1+∆W)G), then the closed-loop system is robustly stable if andonly if
Mp := ‖T‖∞ = maxω
∣∣∣∣
L( jω)
1+L( jω)
∣∣∣∣<
1|W( jω)|
whereT = GK1+GK is the complementary sensitivity function. Thus 1/Mp is
the maximum bound of the allowable perturbations that will destabilizethe control system, soMp is a measure of system robustness against mid-and high frequency uncertainty (e.g, unmodelled dynamics).
159
6.1.3 Robustness Measure in terms of Coprime FactorClearly, as measures of system robustness,Ms andMp alone are not
sufficient. A combination ofMs andMp seems more appropriate.
Furthermore, since uncertainty in a multivariable system may exhibit
directions, the case for robustness measure for a multivariable system is
more complex. To solve this problem, we can adopt the coprimefactor
uncertainty. Now the uncertain model is represented as:
G∆ = (M+∆M)−1(N+∆N)
whereG= M−1N is a left coprime factorization of the nominal plant
model, and the uncertainty structure is
∆ = [ ∆M ∆N ]
160
the closed-loop system is robustly stable if and only if
ε :=
∥∥∥∥∥∥
I
K
(I +GK)−1M−1
∥∥∥∥∥∥
∞
≤ 1
‖∆‖∞
So 1/ε measures the maximum bound of the allowable uncertainties thatwill destabilize the control system.
If the coprime factor is normalized, i.e.,
[ M N ][ M N ]∼ = I
then
ε =
∥∥∥∥∥∥
I
K
(I +GK)−1M−1[ M N ]
∥∥∥∥∥∥
∞
=
∥∥∥∥∥∥
I
K
(I +GK)−1[ I G ]
∥∥∥∥∥∥
∞
161
For a single loop control system, it is easy to show that
ε = ‖M‖∞ = ‖MT‖∞
=
∥∥∥∥∥∥
0 1
1 0
MT
0 1
1 0
∥∥∥∥∥∥
∞
=
∥∥∥∥∥∥
G
1
1
1+GK
[
−K 1]
∥∥∥∥∥∥
∞
whereM is given by
M :=
11+GK
G1+GK
− K1+GK − KG
1+GK
.
andMT denotes the transpose ofM.
162
The following example shows that sometimesε is not suited for
robustness measure.
Consider a high-order plant having a large roll-off rate at high frequencies
G=1
(s+1)5 ,
and a practical PID controller tuned by the Ziegler-Nicholsmethod
K = 1.7313(1+1
4.3240s+
1.0810s1.0810/10s+1
).
163
10−2
10−1
100
101
102
0
2
4
6
8
10
12
14
16
18
20
Frequency(rad/s)
Mag
nitu
de
σ(M(jω))
It can be read from the figure thatε = 19.0. The peak ofσ(M( jω)) at the
high frequencies gives a conservative indication of the robustness of the
closed-loop system, since it can be easily reduced with little performance
degradation by incorporating a low-pass filter in the controller. The peak
at the mid-frequency makes more sense.
164
6.2 Robustness Measure for SISO Systems
We note that the coprime factor uncertainty clearly ignoresthe structure
of ∆M and∆N. So more detailed structures will probably overcome the
conservatism given byε .
Consider a plantG having the following uncertainty structure :
G∆ =1+∆2
1−∆1G, with ∆1,∆2 ∈ H∞.
It represents simultaneous input multiplicative and inverse output
multiplicative uncertainty.
∆∆∆∆2
G
∆∆∆∆1
+ +
d2 d1
u y+ +
165
Suppose a normalized left coprime factorization ofG is G= M−1N, then
G∆ =N+ N∆2
M− M∆1=:
N+∆N
M+∆M.
So compared with a coprime factor uncertainty, additional structure
information (∆M =−M∆1 and∆N = N∆2) is contained in the uncertainty
expression, thus the conservatism of the robust analysis can be reduced.
A unity feedback control system with the uncertain plantG∆ is shown
below.
166
To analyze the robust stability of the closed-loop system, we transfer the
uncertain system to anM-∆ structure.HereM is the transfer matrix from
signalsd1, d2 to signalsy, u and given by
M :=
11+GK
G1+GK
− K1+GK − KG
1+GK
.
M
∆∆
2
1
2
1
d
d
u
y
167
Define
∆ :=
∆1 0
0 ∆2
.
By the smallµ theorem, the closed-loop system is robustly stable if and
only if
εm := µ∆(M)<1
‖∆‖∞
Thusµ∆(M) is a measure of system robustness.
168
µ∆(M( jω)) of the unity feedback control system for above example.
10−2
10−1
100
101
102
1
1.5
2
2.5
3
3.5
4
Frequency(rad/s)
Mag
nitu
de
µ(M(jω))
It can be read from the figure thatε = 3.7, suggesting that 1/3.7≈ 27%uncertainty in both the input and output can be allowed. Moreover, itoccurs at the mid-frequency. So it is more reasonable than the peak ofσ(M( jω)) that occurs at the high frequencies.
169
We see that the additional structure information (∆M =−M∆1 and
∆N = N∆2) used in the proposed method leads the peak in the mid-range
frequencies, which is very important since simple methods can be used to
deal with the uncertainties at low and high frequencies. Furthermore, for
a single-loop system, we have
εm = µ∆
1
1+GK
1
−K
[ 1 G ]
= supω
1+ |GK||1+GK|
= supω(|S( jω)|+ |T( jω)|)
Compared withMs or Mp, the proposed measure is more appropriate
since it bounds bothMs andMp simultaneously.
170
Application: Comparison of PID tuning formulas
Evaluation of various PID design or tuning methods found in the
literature can be done by using some performance criteria. However,
since a specific method might be effective for a specific plantmodel, it is
hazardous to draw general conclusions on which method is thebest (in
fact, no best at all). What we can conclude is that some methods show
better performance in disturbance rejection and/or robustness than other
methods.
In this section, we will apply the criteria proposed in the previous section
to analyze several PID tuning techniques found in the literature.
171
The process model is first-order with deadtime (FOPDT)
G(s) =k
Ts+1e−τs.
The following PID tuning formulas are considered:
1. Ziegler-Nichols (Z-N) method
2. Internal model control (IMC) method.
3. Gain-phase margin (G-P) method.
4. Optimum integral error for load disturbance (IAE-load, ITAE-load,
ISE-load, ISTE-load), and for setpoint change (IAE-setpoint,
ITAE-setpoint, ISE-setpoint, ISTE-setpoint) methods.
172
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 12.5
3
3.5
4
4.5
5
5.5
6
Normalized delay (τ/T)
Rob
ustn
ess
mea
sure
men
t
IMCG−PIAE−setpointITAE−setpointISE−setpointISTE−setpoint
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 12
4
6
8
10
12
14
16
18
Normalized delay (τ/T)
Rob
ustn
ess
mea
sure
men
t
Z−NC−CIAE−loadITAE−loadISE−loadISTE−load
(a) Setpoint-based methods (b) Load-based methodsRobustness measures
173
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10
1
2
3
4
5
6
7
8
9
10
Normalized delay (τ/T)
Nor
mal
ized
inte
gral
gai
n (K
ikτ)
IMCG−PIAE−setpointITAE−setpointISE−setpointISTE−setpoint
0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10
10
20
30
40
50
60
70
80
90
Normalized delay (τ/T)
Nor
mal
ized
inte
gral
gai
n (K
ikτ)
Z−NC−CIAE−loadITAE−loadISE−loadISTE−load
(a) Setpoint-based methods (b) Load-based methodsIntegral gains
174
The following can be observed:
1. IMC and G-P methods can be regarded as setpoint-based tuning
methods, since the integral gains of the PID controllers tuned by the
two methods are close to those of the setpoint-based optimum
integral error methods. Similarly Z-N and C-C methods can be
regarded as load-based tuning methods.
2. The robustness measures of the setpoint-based methods are less than
4 except the ITAE-setpoint method for processes with large delay;
while those for the load-based methods are greater than 4, except the
Z-N method for processes with small delay and the ITAE-load
method for processes with large delay. So the PID controllers tuned
by the setpoint-based methods are usually more robust than those
tuned by the load-based methods.
175
3. The PID controllers tuned by the setpoint-based methods for
processes with small delay have very small integral gains and thus
have sluggish load responses. On the other hand, the PID controllers
tuned by the load-based methods have very large integral gains but
are generally not robust.
4. All of these methods give similar integral gains for processes with
large delay, but the setpoint-based methods (except ITAE-setpoint)
have smaller robustness measures, so they are better than the
load-based methods for processes with large delay.
5. Among the load-based methods, only Z-N and ITAE-load methods
are acceptable. ISE-load, ISTE-load, IAE-load and C-C methods
should not be used since the PID controllers tuned by them aretoo
aggressive. The robustness of the PID controllers tuned by the Z-N
method become worse as the delay becomes larger, so it shouldonly
be used for processes with small delay.
176
6. Among the setpoint-based methods, IMC and G-P methods have
almost constant robustness measures, but the G-P method hassmaller
robustness measure and larger integral gain, so it is slightly better
than the IMC method. The IAE-setpoint method has the smallest
integral gains and the best robustness measures thus it is too
conservative. The ISE-setpoint method has larger integralthan IMC
and G-P methods with a sacrifice on robustness, and the
ISTE-setpoint method has smaller integral than IMC and G-P
methods with slightly larger robustness. Like the Z-N method, the
robustness of the PID controllers tuned by the ITAE-setpoint method
become worse as the delay becomes larger, so it should only beused
for processes with small delay, too.
177
7. Extensive simulations show that when the robustness measureεm is
larger than 5, then the closed-loop system will not have sufficiently
large robust margin. On the other hand, if it is less than 3, the integral
gain will not be sufficiently large. The best compromise for the
robustness measure is between 3 and 5, which amounts to an internal
or external disturbance of 20% to 33%.
178
6.3 Robustness Measure for Cascade Systems
Cascade control structure is quite common in industrial processes.
K2 G2r y2=u1
_
Gd2
y1
d2
+
G1K1
Gd1
d1
u2+ +
+ +
+
_
179
The advantages of a cascade control compared with a conventional
single-loop feedback control can be summarized as follows:
1) Disturbances arising at the inner loop are corrected before they
influence the controlled variable, thus the performance of acascade
structure is better than a conventional feedback structurewith regard
to disturbance attenuation.
2) Parameter variations at the inner process can be corrected for within
its own loop. This property can be used to reduce the static
nonlinearity of the actuator.
3) If the inner loop has a faster response than the original inner process,
then the speed of the overall system response can be improved.
180
Design and tuning of a cascade controller are usually done insequence:
The inner loop is tuned first and then the outer loop is tuned with the inner
loop closed. Methods used in a conventional feedback structure have been
extended to the cascade structure.
Several problems exist in cascade controller tuning:
1) Interaction. In a cascade structure, the inner controller affects the
outer loop, and the outer controller also affects the inner loop. This
interaction makes the tuning of a cascade controller more challenging
than that of a conventional feedback controller. The sequential tuning
method clearly takes into consideration the effect of the inner
controller on the outer loop, but it neglects the outer controller’s
effect on the inner loop. This is undesirable since the innerloop is
very important in a cascade control.
181
2) Inner controller mode. The inner controller can be chosen as a P, a
PI, a PD or a PID controller. There always exists a ‘What if ...’
dilemma for the inner controller mode selection.
3) Robustness. Robustness should always be considered in controller
tuning since there exists loop interaction in a cascade structure. In the
sequential tuning, we need to consider how aggressive the inner and
the outer controller can be tuned without affecting the robustness of
the other loop.
182
For the cascade control structure suppose that the inner andouter models
have the following uncertainty structures:
G1∆ =
1+∆2i
1−∆1iG1 : ∆i :=
∆1i 0
0 ∆2i
G2∆ =
1+∆2o
1−∆1oG2 : ∆o :=
∆1o 0
0 ∆2o
The cascade uncertain system is then
K2
∆∆∆∆ 2i
G2
∆∆∆∆ 1i
+ +
_
d2i d1i
y1
∆∆∆∆ 2o
G1
∆∆∆∆ 1o
+ +
d2o d1o
K1_ y2=u1u2
+
+ ++ +
183
To analyze the robustness, we transfer the system to anM-∆ structure,
M
∆∆
∆∆
o
o
i
i
2
1
2
1
o
o
i
i
d
d
d
d
2
1
2
1
1
1
2
2
u
y
u
y
184
whereM is the transfer matrix from signalsd1i , d2i ,d1o andd2o to signals
y2, u2, y1 andu1.
M =
S SG2 −SG2K2K1 −SG2K2K1G1
−K2(1+K1G1)S −K2(1+K1G1)SG2 −K2SK1 −K2SK1G1
G1S G1SG21
1+G1T2K1
G11+G1T2K1
S SG2 −SG2K2K1 −SG2K2K1G1
where
S:=1
1+G2K2+G2K2K1G1
T2 :=G2K2
1+G2K2
185
Suppose the uncertainties for the inner and outer models areindependent,
define the overall system uncertainty as
∆ :=
∆i 0
0 ∆o
=
∆1i 0 0 0
0 ∆2i 0 0
0 0 ∆1o 0
0 0 0 ∆2o
then by the smallµ theorem, the configuration is robustly stable if and
only if
µ∆(M)<1
‖∆‖∞
Thusµ∆(M) is a robustness measure of the cascade system.
186
We can also analyze the system robustness against the inner (outer) model
uncertainty individually. If there is only inner model uncertainty, then we
just need to consider the upper 2×2 block ofM; and if there is only outer
model uncertainty, then we need to consider the lower 2×2 block ofM.
We denote the corresponding blocks by
Mi :=
11+G2K2(1+K1G1)
G21+G2K2(1+K1G1)
− K2(1+K1G1)1+G2K2(1+K1G1)
− K2(1+K1G1)G21+G2K2(1+K1G1)
Mo :=
11+G1T2K1
G11+G1T2K1
− T2K11+G1T2K1
− T2K1G11+G1T2K1
Compared with a conventional feedback structure it is clearthat the ‘real’
inner loop controller becomesK2(1+K1G1), while the ‘real’ outer
controller becomesT2K1. The interaction between the inner and the outer
loop is clear.
187
A typical plot for µ∆(M) is shown below with the solid line. Also shown
are robustness measures for inner model only uncertainty (dashed line)
and outer model only uncertainty (dotted line).
10−2
10−1
100
101
102
1
1.5
2
2.5
3
3.5
Frequency(rad/s)
Rob
ustn
ess
Mea
sure
188
It can be observed from this figure:
i) The robustness measure for a cascade structure has two peaks at two
different frequencies, which clearly correspond to the effect of the
inner loop (high frequency peak) and the outer loop (low frequency
peak). This property makes it easy to point out which loop is not
robustly tuned.
ii) The plot also shows the relative interaction between theinner loop
and the outer loop:
• If the inner loop is fast compared with the outer loop, then the
effect of one loop on the other loop is small. In this case the high
peak frequency is away from the low peak frequency.
• If the inner loop is slow, i.e., has almost the same speed as the
outer loop, the effect of one loop on the other loop is large. In this
case the high peak frequency is close to the low peak frequency.
189
Example
Consider a cascade system with the following inner and outermodels:
G1 =1
(s+1)2 e−s;G2 =1
αs+1e−αs
Two cases will be considered:
• Case 1.α = 0.2. A fast inner process.
• Case 2.α = 0.8. A slow inner process.
190
Fast Inner Loop (α = 0.2)
The following methods are considered:
1) Parallel Compensation (PI/LL)
2) IMC (PID/PID)
3) Independent tuning (PID/PID)
4) Independent tuning (PID/P)
K1 K2
1) 0.432(1+ 12.448s) 2.8291.382s+1
2.448s+1
2) 0.850(1+ 12.239s+0.502s) 0.889(1+ 1
0.267s+0.05s)
3) 0.832(1+ 12.332s+0.519s) 1.2(1+ 1
0.3s+0.0667s)
4) 1.624(1+ 12.405s+0.601s) 1.131
191
10−2
10−1
100
101
102
1
2
3
4
5
6
7
Frequency(rad/s)
µ ∆(M)
Robustness measure (solid: (3); dashed: (2); dashdotted: (4); dotted: (1))
It can be observed that the two peaks are reasonably far away so theinteraction of the the two loops is small. It is clear that therobustnessmeasure of the PI/LL setting is too large for the inner process uncertaintyand too small for the outer process uncertainty.
192
Performance Analysis
The overall inner loop integral gains of the PI/LL and the independent
tuning (PID/P) settings are 0.499 and 0.764, and those of thePID/PID by
the IMC method and the independent tuning method are 3.547 and 4.363.
It is clear that since PI/LL and PID/P do not have integral action in their
inner loops they have very weak integral actions. Best performance is
achieved by the proposed PID/PID controller due to its largest integral
action.
The step responses for the setpoint and the inner loop disturbance of the
closed-loop systems are shown below. It clearly verifies theclaim above.
193
0 5 10 15 20 25 30 35 40 45 50−0.1
−0.05
0
0.05
0.1
0.15
0.2
0.25
0.3
Time
y 1
0 5 10 15 20 25 30 35 40 45 50−0.2
0
0.2
0.4
0.6
0.8
1
1.2
Time
y 1(a) Disturbance response (b) Setpoint response
194
Slow Inner Loop (α = 0.8)
The following methods are considered:
1) Parallel Compensation (PI/LL)
2) IMC (PID/PID)
3) IMC (PID/P)
4) Independent tuning (PID/P)
K1 K2
1) 0.874(1+ 12.023s) 2.6681.171s+1
2.023s+1
2) 0.775(1+ 12.739s+0.787s) 0.889(1+ 1
1.067s+0.2s)
3) 0.775(1+ 12.739s+0.787s) 0.889
4) 1.198(1+ 12.332s+0.519s) 1.131
195
10−3
10−2
10−1
100
101
102
1
2
3
4
5
6
7
8
9
Frequency(rad/s)
µ ∆(M)
Robustness measure (solid: (4); dashed: (2); dashdotted: (3); dotted: (1))
Now the two peaks are close which means that the interaction of the thetwo loops is large. The integral action at the inner loop willmake theoverall integral gain too large to be useful. Thus it is better to just use aproportional gain.
196
Performance Analysis
Now the overall inner loop integral gains of the PI/LL and proposed
PID/P settings are 1.153 and 0.581, and those of the PID/PID and PID/P
by the IMC method are 1.730 and 0.252. The IMC PID/PID controller
has too large integral action at the inner loop so the disturbance response
is oscillatory, while the IMC PID/P has too small integral action so the
response is sluggish.
197
0 5 10 15 20 25 30 35 40 45 50−0.3
−0.2
−0.1
0
0.1
0.2
0.3
0.4
0.5
Time
y 1
0 5 10 15 20 25 30 35 40 45 500
0.2
0.4
0.6
0.8
1
1.2
1.4
Time
y 1(a) Disturbance response (b) Setpoint response
198
6.4 Robustness Measure for Multivariable Systems
Almost all practical industrial processes are multivariable. The methods
in the previous sections are not only applicable to single-variable system,
but also to multivariable systems.
For a multivariable system, a plant with uncertainty is described as
G∆ = (I −∆1)−1G(I +∆2), where∆1,∆2 ∈ H∞.
G+
+
+
+
+ +
+
+
+++
+
...
......
... ... ...
... ...2∆ 1∆
199
Similarly, to analyze the robustness of the closed-loop system, it is
transfered to theM-∆ structure, where
M :=
(I +GK)−1 (I +GK)−1G
−K(I +GK)−1 −K(I +GK)−1G
.
Let
∆ :=
∆1 0
0 ∆2
.
µ∆(M) can measure the robustness of the closed-loop system.
200
6.5 Robust PID Tuning
As discussed before, a robust controller should maximize the integral gain
and simultaneously maximizes the robustness measure. However, a large
integral gain would destabilize the robustness requirements. So there must
be a compromise. Generally the optimization problem goes asfollows
maxσ(Ki)
under the constraint
µ∆(M)< γm
whereγm is the desired degree of robustness, andM is given by
M :=
(I +GK)−1 (I +GK)−1G
−K(I +GK)−1 −K(I +GK)−1G
.
201
The problem is nonconvex, thus a global optimal solution is hard to
obtain. However, a sub-optimal solution can be found via loop-shaping
H∞ method.
Suppose the pre- and post-compensators areW1 andW2, and the left
coprime factorizationM, N of the shaped plantG=W2GW1 is
normalized, then[ M N ] is a unit, so
∥∥∥∥∥∥
I
K
(I + GK)−1M−1
∥∥∥∥∥∥
∞
=
∥∥∥∥∥∥
I
K
(I + GK)−1M−1[ M N ]
∥∥∥∥∥∥
∞
=
∥∥∥∥∥∥
I
K
(I + GK)−1[ I G ]
∥∥∥∥∥∥
∞
=
∥∥∥∥∥∥
W2 0
0 W−11
M
W−1
2 0
0 W1
∥∥∥∥∥∥
∞
202
Thus ifW2 andW1 are commutive with the uncertainty blocks∆1 and∆2,
then it is the upper bound of the structured singular value. In this case
µ∆(M)≤
∥∥∥∥∥∥
W2 0
0 W−11
M
W−1
2 0
0 W1
∥∥∥∥∥∥
∞
thus a loop shapingH∞ controller will be a sub-optimal solution to the
optimization problem.
203
PID Tuning via Loop Shaping
In the usual loop-shaping design procedure, the pre-compensator has two
parts
W1 =WaWi
whereWa is usually chosen as a static decoupler andWi is diagonal. For
controller tuning purpose, we can simplify the choice by setting the
post-compensatorW2 as an identity matrix and the pre-compensatorW1 as
a diagonal PI one.
204
The reasons for just using a PI instead of a more sophisticated high-order
pre-compensator are as follows:
1) The final controller will be approximated by a PID one. Comparing
with a simple PI pre-compensator, a high-order pre-compensator
might result in a betterH∞ controller but not a better PID controller
due to the approximation. It may not be worth to find more
sophisticated open-loop shapes.
2) The time-domain performance of the closed-loop system isaffected
by the PI parameters just as in a single-loop process, which makes it
possible to tune the final PID controller by tuning the diagonal PI
pre-compensator.
3) As pointed out in (Skogestad-Postlethwaite:1996) for single-loop
processes, a PI pre-compensator is the ‘best’ for input disturbance
rejection.
205
However,Wa is usually not diagonal, thusW1 does not commute with∆2.
So the uncertainty which theH∞ problem tries to stabilize is only for the
shaped plant, not the actual plant, which is the main source of criticism
for the method. We note that the closed-loop system will showgood
disturbance rejection as long as the integral action is reasonable; the
coupling of the process will only affect the setpoint tracking. So in
loop-shaping design we can first ignore the coupling effect of a
multivariable process, and after the design we can choose a setpoint filter
to modify the setpoint tracking properties.
For example, in a 2×2 system, the final setpoint filter has the following
form:
F =
1 − f1
− f2 1
To completely decouple the setpoint responses,f1 and f2 should be
206
chosen as
f1 = T−111 T12, f2 = T−1
22 T21
whereT is the complementary sensitivity matrix and the subscriptsrefer
to its denoted elements. To realize, a first-order approximation can be
used. Note thatT12 andT21 have derivative action; so the filter will have
the form of asbs+1. In general, we can choose the setpoint filter as
F(s) =
1λ11s+1 − λ12s
λ12s+1
− λ21sλ21s+1
1λ22s+1
where the parameters are tuned to reduce the overshoots and the coupling
in the setpoint responses.
207
In summary, PID tuning procedure based on loop shapingH∞ approach
goes as follows:
1) Select a diagonal PI compensatorW1 such that the valueεmax is
reasonable (between 0.2 and 0.5, which makes the robustness
indicator lie between 2 and 5). In this step we also try to makethe
integral action ofW1 as large as possible.
2) Approximate the resultingH∞ controller with a PID.
3) If necessary, tune a setpoint filter to decouple the setpoint responses.
4) If the time-domain performance is unsatisfactory, follow the idea of
tuning single-loop PI controllers to find another pre-compensator,
and repeat the above steps.
208
Chapter 7. Internal Model Control (IMC)
Internal Model Control (IMC) is a popular control structurefound in
process control.
Pr y
+
_
_Q
u
+
P~
whereP is the plant,P is the plant model, andQ is the IMC controller.
209
The equivalent conventional feedback structure is:
P r y +
+ _
Q u
+
P ~
K
For the feedback controller,K, we have:
K = Q(I − PQ)−1
Q = K(I + PK)−1
210
The closed-loop transfer matrices are given by:
(I +PK)−1 = (I − PQ)(I +(P− P)Q)−1
K(I +PK)−1 = Q(I +(P− P)Q)−1
(I +PK)−1P = (I − PQ)(I +(P− P)Q)−1P
(I +KP)−1 = (I +Q(P− P))−1(I −QP)
PK(I +PK)−1 = PQ(I +(P− P)Q)−1
K(I +PK)−1P = (I +Q(P− P))−1QP
It is clear they are all related to the model errorP− P.
211
When the model is perfect, i.e.,P= P, we have
S:= (I +PK)−1 = I − PQ
T := PK(I +PK)−1 = PQ
K(I +PK)−1 = Q
(I +KP)−1 = I −QP
(I +PK)−1P = (I − PQ)P
K(I +PK)−1P = QP
So when we have a perfect model, the closed-loop transfer functions are
all affine functions of the controllerQ.
212
7.1 Internal Stability of IMC
In order to test for internal stability we exam the transfer matrices between
all possible system inputs and outputs. From the discussionabove, when
we have a perfect model, all the transfer matrices involveQ, PQ or QP,
and(I −PQ)P, so the closed-loop system is internally stable if and only if
(i) Q stable.
(ii) PQ stable.
(iii) (1−PQ)P stable.
213
So we have:
Theorem: Assume that the model is perfect (P= P). Then the IMC
structure is internally stable if and only if both the plantP and the
controllerQ are stable.
Thus the structure cannot be used to control plants which areopen-loop
unstable.
Nevertheless, even for unstable plants we can exploit the features of the
IMC structure for control system design and then implement the
controllerK in the classic manner.
214
7.2 IMC Design
To design an IMC controller, consider the following problem:
1. Nominal Performance. H2 or H∞ performance.
minQ
‖WeS‖α = minQ
‖We(I − PQ)‖α , α = 2,∞
2. Robust Stability. Multiplicative uncertainty.
‖WyT‖∞ = ‖WyPQ‖∞ < 1
3. Robust Performance.
minQ
‖WeS‖α , α = 2,∞, ∀P∈ {(I +Wy∆)P,∀‖∆‖∞ < 1}
215
Clearly what we need in practice is robust performance. However, for
IMC design, we consider a mixedH2/H∞-type problem for the following
reasons:
• It is hard to solve the optimal robust performance problem (3).
• H2-type performance is an integral-square-error (ISE) objective.
• Since the plant changes and the model quality degrades with time, it
is desirable to provide for convenient on-line robustness adjustment.
To solve the problem, a two-step design method is used. The method has
no inherent optimality characteristics but should providea good
engineering approximation to the optimal solution. It guarantees
robustness but the performance is generally not optimal in any sense.
216
The design procedure goes as follows:
Step 1: Nominal Performance
The controllerQ is selected to yield a “good” system response for the
input(s) of interest, without regard for constraints and model
uncertainty. Generally we will chooseQ such that it is
integral-square-error (ISE) orH2-optimal for a specific set of
reference inputv.
minQ
‖(I − PQ)v‖2
Step 2: Robust Stability and Performance
The aggressive controllerQ obtained in Step 1 is detuned to satisfy
the robustness requirements. For that purposeQ is augmented by a
filter F of fixed structure
Q= QF
217
The filter parameters are adjusted to meet the robustness requirement.
Sometimes a more complicated form is required.
In general, it might not be possible to meet the robust performance
requirement. The reason could be that the design procedure fails to
produce an acceptableQ.
On the other hand, there might not exist anyQ such that the
requirement is satisfied. Then the performance requirements have to
be relaxed and/or the model uncertainty has to be reduced.
218
Nominal Performance
Nominal performance can be guaranteed by solving theH2-optimization
problem. Generally the plantP can be factored into a stable all-pass
portionPA and an minimum-phase (MP) portionPM such that
P= PAPM
HerePA andPM are stable andPHA ( jω)PA( jω) = I .
Similarly, the set of inputsv can be factored as
v= vMvA
wherevM is minimum-phase andvA is all-pass.
219
SincevA andPA are all-pass, we have
‖(I − PQ)v‖2 = ‖(I −PAPMQ)vMvA‖2
= ‖P−1A vM︸ ︷︷ ︸
M
−PMQvM︸ ︷︷ ︸
N
‖2
GenerallyM andN are inL2. Since we require that the system beinternally stable, we must haveN ∈ H2. DecomposeM as the sum of thestable and the unstable part:
M = M++M−, with M+ ∈H2,M− ∈H⊥2
then
minN∈H2
‖M−N‖22 = min
N∈H2‖M++M−−N‖2
2
= minN∈H2
(‖M+−N‖22+‖M−‖2
2) = ‖M−‖22
The minimum is achieved whenN = M+.
220
So we have
Theorem: TheH2-optimal controllerQ is given by
Q= P−1M {P−1
A vM}+v−1M
The operator(P−1A vM)+ denotes the stable part of the transfer matrix
P−1A vM. It can be obtained by omitting all terms involving the polesof
P−1A after a partial fraction expansion ofP−1
A vM.
Generally the IMC controller is the inverse of the minimum-phase part of
the plantP. Moreover, if the plant is minimum-phase, i.e.,P= PM, then
the IMC controller is just the inverse of the plant.
221
Robust Stability and Performance
The controllerQ is to be detuned through a lowpass filterF such that for
the detuned controllerQ both the robust stability and the robust
performance conditions are satisfied.
In principle the structure ofF can be as complex as the designer wishes.
However, in order to keep the number of variables in the optimization
problem small, a simple structure like a diagonalF with first- or
second-order terms is recommended.
In most cases this is not restrictive because the controllerQ designed in
the first step of the IMC procedure is a full matrix with high order
elements. Some restrictions must be imposed on the filter in the case of an
open-loop unstable plant. Also a more complex filter structure may be
necessary in cases of highly ill-conditioned systems.
222
Generally the filterF is chosen to be a diagonal rational function
F(s) = diag{ f1(s), . . . , fn(s)}
and it must satisfy the following requirements:
(a) Pole-zero excess. The controllerQ= QF must be proper.
(b) Internal stability. Q, PQand(I −PQ)P must be stable.
(c) Asymptotic setpoint tracking and/or disturbance rejection. (I −PQ)v
must be stable.
In most cases this is not restrictive because the controllerQ designed in
the first step of the IMC procedure is a full matrix with high order
elements. Some restrictions must be imposed on the filter in the case of an
open-loop unstable plant.
223
Open-loop Stable Plants
Since the open-loop is stable, (b) and (c) are satisfied whenever F isstable. A reasonable choice of a filter elementfl (s) would be:
fl (s) =1
(λs+1)m
wherem is chosen to makeQ proper.
Open-loop Unstable Plants
Let πi(i = 1, . . . ,k) be the open RHP poles ofP. Let π0 = 0 andm0l be thelargest multiplicity of such a pole in any element of thel -th row ofv. Itfollows that thel -th element,fl of the filterF must satisfy:
Condition (b): fl (πi) = 1, i = 0,1, . . . ,k
Condition (c): d j
dsj fl (s)|s=π0= 0, j = 1, . . . ,m0l −1
224
The requirements clearly show the limitation that RHP polesplace on the
robustness properties of a control system designed for an open-loop
unstable plant.
Since one cannot reduce the closed-loop bandwidth of the nominal system
at frequencies corresponding to the RHP poles of the plant, only a
relatively small model error can be tolerated at those frequencies.
Experience has shown that the following structure for a filter element
fl (s) is reasonable:
fl (s) =avl−1,l svl−1+ · · ·+a1,l s+a0,l
(λs+1)r+vl−1
For a specific tuning parameterλ the numerator coefficients can be
computed to satisfy (b) and (c).
225
7.3 Two-degree-of-freedom (TDF) IMC
It is shown that IMC control can achieve very good tracking performance.
However, the load disturbance rejection performance sometimes is not
satisfactory. So a second controller can be added to improvethe
disturbance-rejection performance.
The TDF-IMC structure is shown below
Pr y
+
_
_Q
u
+
P~
Qd
226
The design ofQd goes as follows:
Design a disturbance-rejecting IMC controller of the form
Qd(s) =αmsm+ · · ·+α1s+1
(λds+1)m
whereλd is a tuning parameter for disturbance rejection,m is the
number of poles ofP(s) such that theQd(s) needs to cancel. Suppose
p1, · · · , pm are the poles to be canceled, thenα1, · · · ,αm should
satisfy
(1− P(s)Q(s)Qd(s))∣∣s=p1,··· ,pm
= 0
227
It can be shown that the TDF-IMC structure is equivalent to the
conventional TDF feedback structure, where the feedback controller K
equals
K =QQd
1− PQQd
Pr y+
+_
u
+
P~
K
Q¡1
d QQd
228
7.4 Modified IMC Structure
A modified IMC structure for unstable processes with time delays is
P
P*
r y
+_
_
1d 2d
++
K2
_K1
K0
_
_se θ−
u
229
The advantages of the modified IMC structure are:
• IMC structure can be retained for unstable processes. Controllers do
not have to be converted to conventional ones for implementation.
• Setpoint tracking and disturbance rejection can be designed
separately. The setpoint tracking design follows the standard IMC
design for a stable plant.
• Robustness and disturbance rejection mainly rely on a controller in a
feedback loop. Robustness of the whole structure can be considered
by tuning this controller.
230
Properties of Modified IMC
It is easy to verify that the modified IMC structure is equivalent to
P
P* se θ−
r y
+
_
_
1d 2d
++
K2
_K1
u
0*1
1
KP+
*G
1u
2u
0*1
1
KP+
+ + ++
231
We have
y =PK1(1+P∗e−θsK2)
(1+P∗K0)(1+PK2)+(P−P∗e−θs)K1r
+P(1+P∗K0−P∗e−θsK1)
(1+P∗K0)(1+PK2)+(P−P∗e−θs)K1d1
+1+P∗K0−P∗e−θsK1
(1+P∗K0)(1+PK2)+(P−P∗e−θs)K1d2
If the plant model is perfect, i.e.,P= P∗e−θs, then
y=PK1
1+P∗K0r+
P1+PK2
1+P∗K0−PK1
1+P∗K0d1+
11+PK2
1+P∗K0−PK1
1+P∗K0d2
232
Let
G∗ :=P∗
1+P∗K0
G :=P
1+P∗K0= G∗e−θs
then
y= GK1r︸ ︷︷ ︸
yr
+(1−GK1)P
1+PK2d1
︸ ︷︷ ︸
yd1
+(1−GK1)1
1+PK2d2
︸ ︷︷ ︸
yd2
It follows thatK2 has no effect on the setpoint responseyr .
Similarly, when model is perfect,
u= K1r − (K2+K1
1+P∗K0)
P1+PK2
d1− (K2+K1
1+P∗K0)
11+PK2
d2
If K2 = 0, thenK1 can be shown to be an IMC controller forG.
233
The structure thus proposed has three compensators, namely, K0, K1, and
K2, each having a distinctive use and influence on the overall closed loop
response:
• K0 is used to stabilizeP∗, the original (unstable) plant, ignoring the
time-delay.
• K1 is an IMC controller for the new modelG.
• K2 is used to stabilize the original unstable systemP, with the delay
θ . It is crucial for the internal stability of the structure.
234
Robustness Analysis
Suppose the model has the following uncertainty structure
P∆ = (1−∆1)−1P(1+∆2), with ∆1,∆2 ∈ H∞
and uncertainty block
∆ :=
∆1 0
0 ∆2
thenµ∆(M) is a robustness measure for modified IMC, with
M =
−(K2+
K11+P∗K0
) P1+PK2
−(K2+K1
1+P∗K0) 1
1+PK2
(1− PK11+P∗K0
) P1+PK2
(1− PK11+P∗K0
) 11+PK2
235
7.5 Double TDF Scheme
A double TDF control scheme for processes with time delays is
P
P*
r y
+_
_
1d 2d
++
K2
K1
K0
_
+ se θ−
u
K3
236
It is equivalent to
P
P*
r y
+_
_
1d 2d
++
K2
_K1
K0
_
_se θ−
u
K3
WhenK3 = K1, it is reduced to modified IMC.
237
Chapter 8. Wide-Range Robust Control
Robust controller synthesis aims to achieve acceptable performance under
model uncertainties. A common source of model uncertainties is the
change of operating points due to system nonlinearity. If a process has
severe nonlinearity, then a single robust controller may not achieve global
performance, hence wide-range robust controllers are required.
• Gain scheduling control is a possible method to achieve wide-range
performance, however gain scheduling needs to have a detailed
nonlinear model or complete knowledge of the operating points,
which is not practical. Moreover, the cost of implementing again
scheduling controller is high.
238
• Another possible method to achieve wide-range performanceis to
use multi-model control. The method divides the operating range into
several ‘linear’ range, and designs linear controllers at each local
operating point, and combines them into a multi-model controller.
The idea is simple and the simulation results shown that the method
is effective.
• A still simpler method is to ‘avoid’ the nonlinearity of a unit by
carefully choosing the operating points. The idea is that some of the
operating points are seldom met in practice so even if the linear
controller may not work well under such operating points it can still
get good global performance as long as the controller does not enter
such an operating range.
239
8.1 Gap Metric
The robust control design techniques assume that we have some
description of the model uncertainties (i.e., we have a measure of the
distance from the nominal plant to the set of uncertainty systems). This
measure is usually chosen to be a metric or a norm (e.g.,H∞-norm).
However, theH∞ norm can be a poor measure of the distance between
systems with respect to feedback control system design.
240
For example, consider
G1 =1s, G2 =
1s+0.1
The closed-loop complementary sensitivity functions corresponding toG1
andG2 with unity feedback are relatively close and their difference is
‖ G1
1+G1− G2
1+G2‖∞ = 0.0909
but the difference betweenG1 andG2 is
‖G1−G2‖∞ = ∞
This shows that the closed-loop behavior of two systems can be very
close even though the norm of the difference between the two open-loop
systems can be arbitrarily large.
241
To deal with such problems, the gap metric was introduced into the
control literature by Zames and El-Sakkary as being appropriate for the
study of uncertainty in feedback systems.
Definition and Computation
Let G(s) be ap×m rational transfer matrix and letG have the following
normalized right coprime factorization
G= NM−1, with M∼M+N∼N = I
The graph ofG is a closed subspace ofH2 given by
G(G) =
M
N
H2
242
The gap between two linear systemsG1 andG2 is defined by
δg(G1,G2) :=∥∥ΠG(G1)−ΠG(G2)
∥∥
whereΠK denotes the orthogonal projection ontoK.
It was shown that the gap metric can be computed as follows:
Theorem: Let G1 = N1M−11 andG2 = N2M−1
2 be normalized right
coprime factorizations. Then
δg(G1,G2) = max{~δ (G1,G2),~δ (G2,G1)}
where~δg(G1,G2) is the directed gap and can be computed by
~δg(G1,G2) = infQ∈H∞
∥∥∥∥∥∥
M1
N1
−
M2
N2
Q
∥∥∥∥∥∥
∞
243
Properties of Gap metric
1. If δg(G1,G2)< 1, then
δg(G1,G2) = ~δg(G1,G2) = ~δg(G2,G1)
2. The gap metric can be thought of as a measure of the ‘distance’
between two linear systems. It is an extension of the common
measure – the magnitude (the∞-norm) of the difference between two
systems. The gap metric is not only applicable to stable systems, but
also to integrating and unstable systems. For example, the distance in
the gap metric sense for the two systems considered above is
G1 =1s, G2 =
1s+0.1
,δ (G1,G2) = 0.0995.
244
3. The reason that the gap metric applies to integrating and unstable
systems is that it measures the ‘distance’ in the closed-loop sense
instead of the open-loop sense. Even though the open-loop systems
may look different, their distance can be close. For example, consider
G1 =100
2s+1, G2 =
1002s−1
.
The gap metric betweenG1 andG2 is
δ (G1,G2) = 0.0205,
which shows that they are indeed very close. In fact, we can show
that the closed loops corresponding toG1 andG2 with unity feedback
are close.
G1(1+G1)−1 =
50s+50.5
,G2(1+G2)−1 =
50s+49.5
‖G1(1+G1)−1−G2(1+G2)
−1‖∞ = 0.02
245
So a small distance between two systems in the gap metric sense
means that there exists at least one feedback controller that stabilizes
both systems and the distance between the closed loops is small in
the∞-norm sense. Here the closed loops contain the ‘gang of four’
transfer functions.
I
K
(I +GK)−1[
I G]
246
Robust Control and Gap Metric
Gap metric can be used to describe the uncertainties. The connection
between the uncertainties in the gap metric and the uncertainties
characterized by the normalized coprime factors is as follows.
Theorem: Let G0 have a normalized coprime factorizationG0 = N0M−10 .
Then for all 0< r ≤ 1,
{G :~δg(G,G0)< r}=
G : G= (N0+∆N)(M0+∆M)−1,
∥∥∥∥∥∥
∆N
∆M
∥∥∥∥∥∥
∞
< r
247
Let G be a linear system, andK be a stabilizing controller ofG. Let
bG, K :=
∥∥∥∥∥∥
I
K
(I +GK)−1[ I G ]
∥∥∥∥∥∥
−1
∞
.
Then we have:
Theorem: Suppose the feedback system with the pair(G,K) is stable. Let
G := {G∆ : δ (G,G∆)< γ}.
then the feedback system with the pair(G∆,K) is stable for allG∆ ∈ G if
and only if
γ ≤ bG, K .
248
8.2 Nonlinearity Measure
It is generally accepted that most industrial processes arenonlinear.
However, there is no definite quantification of the nonlinearity of a
process, specifically,
• How nonlinear is it?
• Can alinear controller be used to cover the whole operating range?
These are fundamental issues in the control system design. Without a
thorough understanding of the nonlinearity, the operatingrange and
performance of alinear controller cannot be guaranteed.
249
One way to approach this problem is to study the nonlinearityof a
process. The nonlinearity measure attracted much attention in the past
years, and several definitions and computation methods wereproposed.
Roughly speaking, a nonlinearity measure can be regarded asthe
‘distance’ between a nonlinear system and a class of feasible linear
systems.The first nonlinearity measure is defined as
v := infL∈Λ
‖N−L‖,
whereN is the nonlinear system considered, and the infimum is taken
over all the linear operatorsL in the feasible setΛ. The norm here can be
any appropriate norm, such asL2 or L∞. A largerv means that the system
is ‘more’ nonlinear; and in this case, a linear control may not achieve
good global performance.
250
The available nonlinearity measures have the following limitations:
1) Almost all the examples found in the literature are for SISO systems.
The computation is rather difficult for MIMO systems.
2) The available nonlinearity measures are not applicable for integrating
and unstable systems, since distances between unstable systems
cannot be measured in terms of standard norms.
A method to compute the distance between a nonlinear system and a
(fixed) linear system based on the gap metric concept was proposed. The
distance measure is closely related to nonlinearity measures in that a
nonlinearity measure computes the distance between a specified nonlinear
system and any of the feasible linear systems. So the distance measure is
the basis for computing nonlinearity measures.
251
With the gap, we can define a nonlinearity measure as
vd := infL∈Λ
δd(N,L) = infL∈Λ
supr
δ (LrN,L),
whereLrN is the linearization ofN along trajectoryr.
A more convenient measure is
vg := supr0
δ (Lr0N,L),
whereLr0N is the linearization ofN at the operating pointr0. The
difference betweenLrN andLr0N is that the linearization along a
trajectory is usually time varying, while at an operating point it is
time-invariant.
252
While vd is more appropriate for quantifying the nonlinearity of a system
andvg only reflects the nonlinear dynamics near an operating point, there
are several advantages in usingvg:
1) The computation ofvg is simple.vd involves the gap between a linear
time-varying systemLrN and a linear time-invariant systemL. There
is no efficient method to compute it. On the contrary, the gap
between two linear systems can be easily computed [?].
2) In practical industrial processes, normal operation is confined to the
neighborhood of the equilibrium points; sovg is a reflection of the
nonlinear dynamics due to the operating point change, as long as the
change is slow.
253
3) Linear controller design is usually based on a nominal linear model,
either from linearization of a nonlinear model or identification from
real data. The distance between a nonlinear system and a nominal
linear system can help determine the operating range of a linear
controller designed using the nominal model.
Thus a distance measure between a nonlinear system and a nominal linear
system can determine the operating range of a linear controller designed
based on the nominal model. In contrast, a nonlinearity measure such as
vd can help determine whether a linear controller is enough forthe whole
operating range, and if not, where the operating points should be selected.
254
Example: A Nonlinear Boiler-turbine Unit
Consider a nonlinear boiler-turbine unit. The dynamics of the unit isgiven by
x1 =−0.0018u2x9/81 +0.9u1−0.15u3,
x2 = (0.073u2−0.016)x9/81 −0.1x2,
x3 = (141u3− (1.1u2−0.19)x1)/85,
y1 = x1,
y2 = x2,
y3 = 0.05(0.13073x3+100acs+qe/9−67.975),
The model is based on the basic conservation laws, and the parameterswere estimated from the data measured from the Synvendska Kraft ABPlant in Malmo, Sweden. The plant is oil-fired and the rated power is160MW.
255
In the modelx1, x2, andx3 denote drum pressure (kg/cm2), electric output
(MW), and fluid density (kg/m3), respectively. The inputs,u1, u2, andu3
are the valve positions for fuel flow, steam control, and feedwater flow,
respectively. The outputy3 is the drum water level (m) andacs andqe are
steam quality and evaporation rate (kg/s), respectively and are given by
acs=(1−0.001538x3)(0.8x1−25.6)
x3(1.0394−0.0012304x1),
qe = (0.854u2−0.147)x1+45.59u1−2.514u3−2.096.
Due to actuator limitations, the control inputs are subjectto the following
constraints:
0≤ ui ≤ 1(i = 1,2,3), −0.007≤ u1 ≤ 0.007,
−2≤ u2 ≤ 0.02, −0.05≤ u3 ≤ 0.05.
256
Some typical operating points of the model are shown below.
#1 #2 #3 #4 #5 #6 #7
xo1 75.60 86.40 97.20 108 118.8 129.6 140.4
xo2 15.27 36.65 50.52 66.65 85.06 105.8 128.9
xo3 299.6 342.4 385.2 428 470.8 513.6 556.4
uo1 0.156 0.209 0.271 0.34 0.418 0.505 0.6
uo2 0.483 0.552 0.621 0.69 0.759 0.828 0.897
uo3 0.183 0.256 0.340 0.433 0.543 0.663 0.793
yo3 -0.97 -0.65 -0.32 0 0.32 0.64 0.98
257
The linear control design for the unit found in the literature usually takes
the linearized model at operating point #4 as the nominal model. The
linearized model is
G0 :=
δ x = Aδx+Bδu
δy = Cδx+Dδu
with
A=
−0.0025 0 0
0.0694 −0.1 0
−0.0067 0 0
,B=
0.9 −0.349 −0.15
0 14.155 0
0 −1.398 1.659
,
C=
1 0 0
0 1 0
0.0063 0 0.0047
,D =
0 0 0
0 0 0
0.253 0.512 −0.014
.
(1)
258
To analyze the nonlinearity of the unit, we will compute the measurevg to
the linearized model at operating point #4.
1. Drum pressurey1 = 108, electric outputy2 varies from 20 to 160, and
drum levely3 varies from−0.5 to 0.5.
−0.5
0
0.5 2040
60 80100
120 140160
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
Electric Output (MW)
Drume Level (m)
v g
For fixed pressure operation, there is a maximum electric output. The
linearized models at the region with large drum levels have asmall
distance to the nominal model.
259
2. Drum pressurey1 varies from 60 to 160, electric outputy2 = 66.65,
and drum levely3 varies from−0.5 to 0.5.
−0.5
0
0.5 6080
100120
140160
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
Steam Pressure (kg/cm2)
Drume Level (m)
v g
Different combination of drum pressure and drum level can result in
the same electric outputs. The linearized models at the region with
small drum pressures and large drum levels have a small distance to
the nominal model.
260
3. Drum pressurey1 varies from 60 to 160, electric outputy2 varies
from 20 to 160, and drum levely3 = 0.
20
40
60
80
100
120
60
80
100
120
140
1600
0.2
0.4
0.6
0.8
Electric Output (MW)Steam Pressure (kg/cm2)
v g
For fixed drum level operation, only the linearized models atthe region
with small electric outputs have a small distance to the nominal model.
261
It is also observed
• Larger values ofvg occur at the region with large drum pressures,
large electric outputs, and small drum levels; The maximum is about
0.9.
To evaluate the nonlinearity of the unit, we need to computevg by taking
L as linearized models at other operating points. It is found that the
maximum ofvg is always larger than 0.7, so we conclude that the Bell and
Astrom model shows severe nonlinearity.
262
8.3 Linear Control of Nonlinear Processes
If a process shows severe nonlinearity, does it mean that only nonlinearcontrollers can be used to achieve wide-range performance?The answeris NO.
In this section we will show that by careful choice of operation points thecontrol system can avoid the nonlinearity and a linear controller is enoughin the operating range.
Example: (continue)
The plots ofvg show us how to ‘avoid’ the nonlinear dynamics due to theoperating point change. Since a boiler-turbine unit must follow theelectricity demand from the grid, and the same amount of electric outputcan be obtained with different combination of drum pressureand drumlevel, the operating range of the unit should be carefully chosen.
263
For instant, if a linear controller is designed at operatingpoint #4, then we
should avoid driving the unit to operating points that have large drum
pressures, large electric outputs, and small drum levels, since at these
operating points the dynamics are quite different from the nominal one.
A good operation is to increase the drum pressure and drum level as the
electric output is increased. In this case the dynamics at the operating
region will not be far away from the nominal, thus stability of the system
can be guaranteed.
To verify the argument, we compute the gaps between the linearized
models at the typical operating points. It is clear that the gaps between
them are small, the largest being between models at operating point #1
and #7. If the nominal model is taken at #4, then as long as the designed
linear controllerK satisfiesbG,K > 0.195, then the controller can
guarantee the closed-loop stability at other operating points.
264
Gaps between linearized models at typical operating points
#1 #2 #3 #4 #5 #6 #7
#1 0 0.074 0.139 0.195 0.245 0.289 0.329
#2 0.074 0 0.066 0.123 0.174 0.219 0.261
#3 0.139 0.066 0 0.058 0.110 0.157 0.199
#4 0.195 0.123 0.058 0 0.053 0.10 0.142
#5 0.245 0.174 0.110 0.053 0 0.048 0.091
#6 0.289 0.219 0.157 0.10 0.048 0 0.043
#7 0.329 0.261 0.199 0.142 0.091 0.043 0
265
To show it clearly, we design a linear controller at operating point #4 via
the loop shapingH∞ approach.
The pre-compensator we select isW1 =WaWi andW2 = I3,
W1 =Wa
5+ 1s 0 0
0 1+ 1s 0
0 0 5+ 5s
whereWa is a constant that aligns the singular values of the model at
0.001 rad/s. It is effectively a constant decoupler, and given by
Wa =
0.001095 0.00373 0.02136
−0.004273 0.007065 0
−0.000374 0.00595 0.1281
.
With this pre-compensator we get anH∞ controller of order 8.
266
Since we are using weights, we need to computevgw for the model to
proceed. The procedure and the plots are similar to those in the previous
section, hence omitted here for brevity.
Considering the practical implementation issue, we simplify the final
loop-shapingH∞ controller using the PID reduction procedure. After
eliminating some small terms, we finally get the following multivariable
PI controller:
K(s) =
0.0485+ 0.0012s 0 1.2091+ 0.0486
s
0 0.0197+ 0.0045s 0
0 0 7.2548+ 0.2914s
.
267
The singular value plots of the originalH∞ controller and the reduced PIcontroller are:
10−4
10−3
10−2
10−1
100
101
102
−80
−60
−40
−20
0
20
40
60
80
Singular Values
Frequency (rad/sec)
Sin
gula
r V
alue
s (d
B)
They are close at the low frequencies but different at the high frequencies,so the PI controller will have similar performance as theH∞ controller,but the robustness will degrade. In fact,bW2G0W1, K for the PI controller is0.27, smaller than that for theH∞ controller. However, the PI controllercan still guarantee the stability of the system at other operating points.
268
Simulation Results
1. From the nominal point to another operating point that is ‘close’. The
weighted gap between the linearized model at the final operating
point and the nominal model is 0.154.
0 100 200 300 400 500 600105
110
115
120
125
y 1 (kg
/cm
2 )
0 100 200 300 400 500 60060
80
100
120
140
y 2 (M
W)
0 100 200 300 400 500 600−0.2
0
0.2
0.4
0.6
y 3 (m
)
Time(sec.)
269
2. From the nominal point to another operating point that is ‘far’. The
weighted gap between the linearized model at the final operating
point and the nominal model is 0.727, larger thanbW2G0W1,K .
0 100 200 300 400 500 600100
120
140
160
y 1 (kg
/cm
2 )
0 100 200 300 400 500 60060
70
80
90
100
y 2 (M
W)
0 100 200 300 400 500 600−100
0
100
200
y 3 (m
)
Time(sec.)
270
3. If one wants to increase the drum pressure from 108 to 150 withoutcausing instability, one method is to increase the drum level at thesame time. For instant, att = 100 increasingy3 from 0 to 0.5, thenthe weighted gap between the linearized models at the final operatingpoint and the nominal model is only 0.183, so the system can enterthe final operating point in a stable fashion.
0 100 200 300 400 500 600100
120
140
160
y 1 (kg
/cm
2 )
0 100 200 300 400 500 60060
70
80
90
100
y 2 (M
W)
0 100 200 300 400 500 600−0.5
0
0.5
1
y 3 (m
)
Time(sec.)
271
4. A large operating point change. To show that the linear controller can
operate well in the careful chosen operating range, we consider the
operating point change from #1 to #7 att = 100.
0 100 200 300 400 500 600 700 800 900 100050
100
150
200
y 1 (kg
/cm
2 )
0 100 200 300 400 500 600 700 800 900 10000
50
100
150
y 2 (M
W)
0 100 200 300 400 500 600 700 800 900 1000−1
−0.5
0
0.5
1
y 3 (m
)
Time(sec.)
272
8.4 Multimodel Control
If a single linear controller cannot achieve the performance in the desiredrange for a nonlinear system, then nonlinear strategies should beconsidered. The simplest method is the so called ’multimodel control’.The method represents the nonlinear system as a combinationof linearsystems.
Several problems are critical for multimodel control:
1) How many models are sufficient in design? Where the models shouldbe selected?
2) How to combine local controllers? Switching or weighting?
3) How to prove that the performance can be guaranteed for themultimodel controller?
These problems are not yet completely solved.
273
Selection of Operating Points
Gap metric is suggested as a guideline for selecting local models. The
idea is that the ‘distance’ of two selected models should notbe larger than
a prescribed level. Since local controller can be designed to robustly
stabilize all the models within the prescribed level, models selected in this
way can guarantee the global stability of the closed-loop systems as long
as the change of models is ‘slow’.
Theoretical background in applying the gap metric in selecting operating
points for multimodel controller design is Theorem ?. It shows that a
linear controllerK can stabilize all the linear systems that have a distance
to G less thanbG, K .
274
Suppose now a certain operating point has been selected. Thelocal model
is G and the local controller isK. Then the next operating point should be
selected at a distance (in the gap metric sense) no larger than bG, K , since
all models with a gap metric less thanbG, K to the given model can be
stabilized by the local controllerK. To have a minimal set of the
operating points, the next operating point should be selected at a distance
exactly equal tobG, K , or just a little smaller than it.
In practice, the local controller is not available before selecting the
operating points, so we can first prescribe a distance level,and then
starting from an initial operating point, compute the next operating point
till the whole range of the operating points are covered.
275
There are two drawbacks when applying the gap metric in selecting the
operating points for multimodel controller design:
1) The gap metric is only related to robust stability, that is, the local
controllerK can only guarantee that it can stabilize the models at
operating points close to the given operating point. However, stability
is not the only issue in control system design. Other performance
should also be guaranteed in selecting operating points.
2) In multimodel controller design,bG, K should be checked to make
sure it is larger than the prescribed distance level, otherwise robust
stability cannot be guaranteed.
276
Motivated by the loop shapingH∞ approach, we can include performance
weights in the gap metric computation.
Since
bopt = maxK
bG, K
Sobopt is the maximum of the robustness margin for the shaped plantG,
and we can directly obtainbopt with loop shapingH∞ design.
The gap metric between the shaped plants thus has potential applications
in selecting operating points for multimodel controller design. We
compute the distance between the shaped models instead of the original
models. Then a shaped model with a distance less thanbopt to a given
(shaped) model at one operating point can be stabilized withthe local
optimal controller and the performance can be guaranteed.
277
Multimodel Controller Design
The procedure of multimodel controller design goes as follows:
1. Compensator selection for performance. We have discussed some
guidelines in choosing the pre- and/or post-compensators to reflect
the performance requirements. However, for multimodel controller
design, we have two options in choosing the compensators:
(a) Choose a fixed set of pre- and/or post-compensators for all
operating points.
(b) Choose different sets of pre- and/or post-compensatorsat
different operating points.
278
Obviously, the first method is simple. However, due to different gains
of the models at different operating points, the resulting open-loop
shapes will certainly be different, that means at differentoperating
points the performance specifications are different, whichis
undesired.
The most desired is that at all the operating points, we specify the
same open-loop shapes. However, it is not always possible due to
system nonlinearity. Note that in process control most of the
disturbance arises at the low frequency, so we will choose sets of
compensators such that the desired open-loop shapes at all operating
points are almost the same at the low frequency, which means that the
closed-loop systems will have similar disturbance rejection ability at
all operating points.
279
2. Local controller design. It is a standard practice to design such loop
shaping controllers. One thing to remember is to check whether the
designedbG,K is actually less thanbopt, otherwise the operating
points selected are not appropriate and new operating points needed
to be re-selected.
3. Constructing the multimodel controller. Once local controllers are
designed, we can form the global controller by switching, orusing
weights (fuzzy logic). We will use membership functions to create a
transition region according to the operating pointz,
u(t) =k
∑i=1
ui(t)ρi(z)
wherek is the number of operating points,ui(t) is the output of the
ith local controller, andρi(z) is the membership function of theith
local controller.
280
Example
Consider the following two-input-two-output (TITO) process
x1 = x2−u1
x2 = (4.75−4.5x1)u2−0.7(x2−u1)−0.25x1
y1 = x1
y2 = x2
.
In a previous literature, five operating points were used to design a
multimodel controller.
281
With the pre-compensator
W1 = G(0)−1
3+1/s 0
0 2+4/s
and the post-compensatorW2 = I , we find that only two operating points
are needed. HereG(0) denotes the dc gain of the local model, which
serves as a static decoupler. The two operating points correspond to
#1. x10 = 0.679,x20 = 0,u10 = 0,u20 = 0.1;
#2. x10 = 0.950,x20 = 0,u10 = 0,u20 = 0.5.
282
The sigma plots of the shaped open-loop systems at the two operating
points andG0 (corresponding tox10 = 0) andG3 (corresponding to
x10 = 1) are shown below.
10−2
10−1
100
101
102
−60
−40
−20
0
20
40
60
Frequency(rad/s)
Mag
nitu
de(d
B)
283
The gap metrics between the shaped models are shown below.
δ G0 G1 G2 G3 bopt
G0 0 0.474 0.836 0.919 –
G1 0.474 0 0.589 0.768 0.619
G2 0.836 0.589 0 0.320 0.597
G3 0.919 0.768 0.320 0 –
The robustness margins at the two operating points are shownin the last
column of the table. It is clear that we just need two operating points for
the specified performance.
284
The responses of the closed-loop systems for the global controller and
two local controllers are shown below.
0 20 40 60 80 100 120 140 160 180 200−0.2
0
0.2
0.4
0.6
0.8
1
1.2
Time
x 1
0 20 40 60 80 100 120 140 160 180 200−0.2
0
0.2
0.4
0.6
0.8
1
1.2
Time
x 2
(a) x1 (b) x2
It is clear that the local controllers only performs well at their own
operating regimes. The global controller achieves the desired
performance at the whole operating range.
285