Optimal Estimation in Approximation Theory

302

Transcript of Optimal Estimation in Approximation Theory

Page 1: Optimal Estimation in Approximation Theory
Page 2: Optimal Estimation in Approximation Theory

Optimal Estimation in Approximation Theory

Page 3: Optimal Estimation in Approximation Theory

THE IBM RESEARCH SYMPOSIA SERIES

Computational Methods in Band Theory Editors: P.M. Marcus, J .F. Janak, and A.R. Williams

Computational Solid State Physics Editors: F. Herman, N.W. Dalton, and T.R. Koehler

Sparse Matrices and Their Applications Editors: D.j. Rose and R.A. Willoughby

Complexity of Computer Computations Editors: R.E. Miller and J .W. Thatcher Associate Editor: J.D. Bohlinger

Computational Methods for Large Molecules and Localized States in Solids

Editors: F. Herman, A.D. Mclean, and R.K. Nesbet

lon Implantation in Semiconductors and Other Materials

Editor: Billy L. Crowder

Stiff Differential Systems Editor: Ralph A. Willoughby

Optimal Estimation in Approximation Theory Editors: Charles A. Micchelli and Theodore J. Rivlin

A Continuation Order Plan is available for this series. A continualion order will bring delivery of each new volume immediately upon publication. Volumes are billed only upon actual shipment. For further information please contacl the publisher.

Page 4: Optimal Estimation in Approximation Theory

Optimal Estimation in Approximation Theory

Edited by

Charles A. Micchelli and Theodore j. Rivlin IBM

Yorktown Heights, New York

SPRINGER SCIENCE+BUSINESS MEDIA, LLC

Page 5: Optimal Estimation in Approximation Theory

Library of Congress Cataloging in Publication Data

International Symposium on Optimal Estimation in Approximation Theory, Freud­enstadt, Ger., 1976. Optimal estimation in approximation theory.

(The IBM research symposia series) lncludes index. 1. Approximation theory-Congresses. 2. Mathematical optimization-Congresses.

3. Computational complexity-Congresses.l. Micchelli, Charles A. 11. Rivlin, Theodore J., 1926- III. Title. IV. Series: International Business Machines Corporation. IBM research symposia series. QA297 .5.157 1976 511 '.4 774329 ISBN 978-1-4684-2390-7 ISBN 978-1-4684-2388-4 (eBook) DOI 10.1007/978-1-4684-2388-4

Proceedings of an International Symposium on Optimal Estimation in Approximation Theory held in Freudenstadt, Federal Republic of Germany,

September 27-29, 1976

© 1977 Springer Science+Business Media New York Originally published by Plenum Press, New York in 1977 Softcover reprint of the hardcover 1 st edition 1977

Ali rights reserved

No part of this book may be reproduced, stored in a retrieval system,or transmitted, in any form or by any means, electronic, mechanical, photocopying, microfilm ing,

recording, or otherwise, without written permission from the Publisher

Page 6: Optimal Estimation in Approximation Theory

PREFACE

The papers in this volume were presented at an International Symposium on Optimal Estimation in Approximation Theory which was held in Freudenstadt, Federal Republic of Germany, September 27-29, 1976.

The symposium was sponsored by the IBM World Trade Europe/Middle East/Africa Corporation, Paris, and IBM Germany. On behalf of all the participants we wish to express our appreciation to the spon­sors for their generous support.

In the past few years the quantification of the notion of com­plexity for various important computational procedures (e.g. multi­plication of numbers or matrices) has been widely studied. Some such concepts are necessary ingredients in the quest for optimal, or nearly optimal, algorithms. The purpose of this symposium was to present recent results of similar character in the field or ap­proximation theory, as well as to describe the algorithms currently being used in important areas of application of approximation theory such as: crystallography, data transmission systems, cartography, reconstruction from x-rays, planning of radiation treatment, optical perception, analysis of decay processes and inertial navigation system control. It was the hope of the organizers that this con­frontation of theory and practice would be of benefit to both groups.

Whatever success th•~ symposium had is due, in no small part, to the generous and wise scientific counsel of Professor Helmut Werner, to whom the organizers are most grateful.

Dr. T.J. Rivlin IBM T.J. Watson Research Center Yorktown Heights, N. Y. Symposium Chairman

Dr. C.A. Micchelli IBM T.J. Watson Research Center Yorktown Heights, N. Y.

Dr. P. Schweitzer IBM Germany Scientific and Education Programs Symposium Manager

v

Page 7: Optimal Estimation in Approximation Theory

CONTENTS

A Survey of Optimal Recovery • • . • • • • • • • • • C. A. Micchelli and T. J. Rivlin

The Setting, Basic Bounds and Relationship to Previous Work

Optimal Estimation of Linear Functionals Examples of Optimal Recovery of Linear

Functionals by Linear Methods Example of Optimal Recovery of a Function Optimal Recovery by Restricted Algorithms

n-Widths and Optimal Interpolation of Time- and Band-Limited Functions • • • • •

Avraham A. Melkman Introduction n-Widths Optimal Interpolation

Computational Aspects of Optimal Recovery Carl de Boor

Introduction The Optimal Recovery Scheme of

Micchelli, Rivlin, and Winograd The Envelope Construction The Construction of Norm Preserving

Extensions to all of IL1 Construction of the Knots for the

Optimal Recovery Scheme Construction of the Optimal Interpolant

Interpolation Operators as Optimal Recovery Schemes for Classes of Analytic Functions • •

Michael Golomb Introduction Interpolating ~ Splines Optimal Recovery Schemes for a Ball in bV.

vii

1

55

69

93

Page 8: Optimal Estimation in Approximation Theory

viii

Spaces of Analytic Functions bf(DR)-Splines from Interpolation Data bf(DR)-Splines Satisfying Initial Conditions.

Hyperoptimal J#'(AR) -Splines bf(AR) ,- bfd~b)- and bf(ER)-Splines

Comparison with n-Widths Stabilization of the Recovery Schemes

Optimal Degree of Approximation by Splines . . . Karl Scherer

The Problem Inverse Theorems

Minimal Projections Carlo Franchetti

Estimation Problems in Crystallography . . . . Robert Schaback

Introduction and Problem Formulation Determination of Unit Cell Dimensions Estimation of the Relative Positions of

Atoms Within the Unit Cell

Estimation Problems in Data-Transmission Systems G. Ungerboeck

Introduction General Solution by Means of a Single

Likelihood Function Estimation of Timing Phase and Carrier Phase Adaptive Equalization Signal Detection Conclusion

CONTENTS

139

151

159

181

Optimal Approximation in Automated Cartography . . . . . . . 201 Wigand Weber

Introduction Determination of Information Content A Model of Cartographic Generalization Evaluation of the Generalization Model Some Available Partial Solutions of

Automated Cartographic Generalization Optimal Approximation in Other Domains

of Automated Cartography

Reconstruction from X-Rays •••... K. T. Smith, S. L. Wagner, and R.

Mathematical Generalities Computerized Axial Tomography

B. Guenther 215

Page 9: Optimal Estimation in Approximation Theory

CONTENTS

Discrimination Between Cander and Fibrocystic Disease in the Breast

Noninvasive Angiography

Planning of Radiation Treatment Udo Ebert

Introduction Foundations of Radiotherapy Model for Spatial Distribtuion of Fields

Some Aspects of the Mathematics of Limulus • • • • K. P. Hadeler

Facts from Biology The Model The Equilibrium States Oscillating Solutions The Vector System Related Problems from Ecology

Analysis of Decay Processes and Approximation by Exponentials • • • • • • • •

Dietrich Braess

Optimal State Estimation and Its Application to Inertial Navigation System Control

W. Hofmann

Index

Introduction Modelling of Physical and Random

Disturbing States The Linear, Gaussian Estimation Problem State Estimation and Platform Error

Angle Control of a Doppler-Aided Inertial Navigation System

SUDDnary

ix

229

241

257

267

297

Page 10: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY

C. A. Micchelli and T.J. Rivlin

Research Staff Members of IBM

IBM- Yorktown Heights, New York 10598

ABSTRACT: The problem of optimal recovery is that of approximating as effectively as possible a given map of any function known to be­long to a certain class from limited, and possibly error-contamin­ated, information about it. In this selective survey we describe some general results and give many examples of optimal recovery.

TABLE OF CONTENTS

1. The Setting, Basic Bounds and Relationship to Previous Work

2. Optimal Estimation of Linear Functionals

3. Examples of Optimal Recovery of Linear Functionals by Linear Methods

4. Example of Optimal Recovery of a Function

5. Optimal Recovery by Restricted Algorithms

Page 11: Optimal Estimation in Approximation Theory

2 C. A. MICCHELLI AND T. J. RIVLIN

1. THE SETTING, BASIC BOUNDS AND RELATIONSHIP TO PREVIOUS WORK

Let X be a linear space andY, Z normed linear spaces. Sup­pose K to be a subset of X and U to be a given linear operator from X into Z. Our object is to approximate Ux for xEK using limited in­formation about x. A linear operator, I, the information operator, maps X into Y, and we assume that Ix for xEK is known possibly only with some error. Thus in our attempt to recover Ux we actually know yEY satisfying I jix-yj I~E for some given E~O. An algorithm, A, is any transformation - not necessarily a linear one - from IK+ES, where S = {yEY: I jyj 1~1}, into Z. The process is schematized in Figure 1.

y Figure 1

The algorithm A produces an error (1) EA (K,E) sup jjux - Ayjj.

We call (2) E(K,E)

{ xEK II Ix-yjj ~E

inf EA(K,E) A

the intrinsic error in the recovery problem, and any algorithm A satisfying (3) EA(K,E) = E(K,E)

is called an optimal algorithm.

Our interest is in determining E(K,E) for various choices of X,Y,Z,U,I,K and E and finding optimal or nearly optimal algorithms.

A lower bound for E(K,E) which is most useful is given by the following simple result. Recall that K is balanced if xEK implies -xEK.

Theorem 1. If K is a balanced convex subset of X then

(4) e(K,E) sup II Ux II SE (K, E) • (

XEK

IIIxii~E

Proof. Let 0 denote the zero vector in Y. For every xEK satisfying I jixj I~E we have, for any algorithm A,

Page 12: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY

and JJux-AOjjs;EA(K,e:)

Jju(-x)-Aoll = Jlux+AOII s; EA (K,e:) in view of (1). Therefore,

Jlzux II = II Ux-AO+Ux+AO II s; 2EA (K, e:) and (4) follows, since A was arbitrary.

3

Remark. This result is still valid if U and I are possibly nonlinear transformations with U(-x)=-Ux and I IIxl I = I II(-x)l I for all xe:X.

An upper bound for E(K,e:) in terms of e(K,e:) can also be ob­tained.

Theorem 2. If K is a balanced convex subset of X then E(K,e:) s; 2e(K,e:).

Proof. First we prove the theorem when e:=O and then reduce the general result to this special case.

Suppose e:=O, so that an algorithm is a transformation from IK into z. Consider the algorithm, A, defined on IK as follows: For each ye:Y put

B(y) = inf{c:c>O, {ze:X:Iz=y} n cK~0}.

Note that when K is a unit ball in X, B(y)=inf{l lzl I :Iz=y}, the distance of the hyperplane Iz=y from the origin in X. Also observe that B(Ix)s;l, if xe:K. Hence given any fixed o>O and any xe:K there is a z(Ix)e:X such that I(z(Ix)) = Ix and z(Ix)e:cK for some c satisfying O<c<l+o. Define the algorithm A by

A: Ix -+ Uz (Ix).

This algorithm produces an error EA(K,O) sup llux-Uz(Ix)ll =sup llu(x-z(Ix))ll.

xe:K xe:K

But since if xe:K also (-1/c)z(Ix)e:K, and hence x-z(Ix)e:(l+c)K. Thus, because I(x-z(Ix)) = 0 we have

EA(K,O) s; (l+c) sup I luxl I s; (2+o)e(K,O), { xe:K

Ix=O and since o>O was arbitrary the theorem holds when e:=O.

Next we show that optimal recovery with error is, at least formally, equivalent to recovery with exact information. This fact is just the observation that knowing both y and the actual error in computing Ix is the same as having exact information. Thus we

h h

introduce the space X= XxY, and in X the convex balanced subset

K = {(x,y) : xe:K, IIYII:;; 1}.

Page 13: Optimal Estimation in Approximation Theory

We extend U to X by defining U(x,y) information operator as

I(x,y) = Ix + Ey.

C. A. MICCHELLI AND T. J. RIVLIN

Ux, and finally define the

In the context we are now considering Figure 1 is replaced by

" XxY -----~

u z

Figure 2

Now we observe that E("K,O) inf

A sup A I IU(x,y)-A(I(x,y))l I

(x,y)EK inf

A

inf A

sup II Ux-A(Ix+Ey) II { XEK

IIY II ~1 sup llux-Ayll E(K,e:).

( XEK IIIx-yll ~E

Hence, applying the part of the theorem we have already proved, we have

E(K,E) ~ 2e(K,O). But

e(K,O) sup llu<x,y) II sup !lux !I { I(x,y)=O

(x,y)EK

= sup lluxll { IIIxii~E

XEK and the theorem is proved.

{ Ix+Ey=O IIYII~l. XEK

e(K,e:),

It is possible, in some fairly general circumstances, to close the gap between Theorems 1 and 2 and solve the optimal recovery problem. For example, suppose K is convex and balanced, and

We have

K = {xEK : Ix = 0}. 0

Page 14: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY

Theorem 3. Suppose there exists a transformation G : IX_... X

5

such that x-Glx€K for all X€K, then e(K,O) = E(K,O) and D 0

UG is

an optimal algorithm.

Proof. sup II Ux-UGix II X€K sup II U (x-Glx) II X€K

s sup lluxll = e(K,O). X€K

0

Remark. See Morozov and Grebennikov [36] for related material.

Let us look at some examples of Theorem 3. Example 1.1. Suppose X to be a Hilbert space, K the unit ball in X and K1 = {x€X:Ix=O} to be closed. For x€X let Q be the projection

of X on the subspace K1 , that is,

llx-Qxll min Z€Kl

Note that since IIQxllsllxll,

(5) G : Ix _... x - Qx

II x-z II.

Qx€K for x€K. 0

The transformation

is now seen to be a well-defined linear operator of IX into X which satisfies the hypothesis of Theorem 3. Thus UG, where G is defined in {5) is seen to be an optimal algorithm and e(K,O) = E(K,O).

For example, if we assume that I is a bounded linear operator from X onto a Hilbert space, Y, so that, in particular, K1 is closed, then if-r;-denotes the adjoint of I we define

(6), G = I*<n*>-1

and it is easy to verify that (5) is satisfied.

In particular, suppose Y to beRn, let z1 , •.• ,zn be linearly

independent elements of X and define I by Ix=((x,z1), ... ,(x,zn)). * Then II is the Gram matrix g(z1 , ..• ,zn) and (6) reduces to a

familiar formula,

G(Ix)

where a -1 g (Ix).

n I:

j=l a.z.

J J

Remark. In this example we do not require that Z be a Hilbert space.

Page 15: Optimal Estimation in Approximation Theory

6 C. A. MICCHELLI AND T. J. RIVLIN

Example 1.2. Let X be iP(t), p~l, the space of sequences of complex numbers, {cj}, -~<j<~, satisfying

rlcjlp < ~.

Let a:{a.} be any given doubly infinite sequence of complex numbers, J

and choose

Fix integers m,n, m<n and put Ic = (c ,c +1 , ••• ,c) so that m m n Y = tn-m+l. Consider

G: (c , ••• ,c)+ ( ••• o, ... ,O,c , ••• ,c ,o, ... ,o, ... ). m n m n

Since x-Gix = ( ••• ,cm_1 ,o, ... ,O,cn+l'""") is in K0 if xEK, we have

another illustration of Theorem 3.

Example 1.3. Suppose n~l and X is the space of functions having n-1

continuous derivatives on [0,1] having an nth derivative in L2[0,l] and equipped with the L2 norm on [0,1]. Z=X and U is the identity operator. Suppose O~x1~x2~ ••• ~xn+r~l, with xi<xi+n' i=l, ••• ,r. For fEX define I by

If= (f(x1), ••• ,f(xn+r))

with the convention that if x <x -x - -x < then i-1 i- i+l- ••• - i+s xi+s+l

f(x.+.) = f{j)(x.), j=l, ••• ,s. In other words, our information ~ J ~

is obtained by sampling the functions at a given set of points. Put

K = {fEX: llf(n)ll 2 ~ 1}.

We now construct the required operator G. Let S be the set of X

natural splines of degree 2n-l with respect to the knots x:(x1 , ••• ,xn+r). Given fEX let Sf be the unique element of sx satisfying

(Sf)(xi) = f(xi), i = l, ••• ,n+r. Note that SfEX and, as is well known,

(7)

(8)

Thus if we define G by

G: If+ Sf,

Page 16: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 7

I(f-G(If)) = If-I(Sf) =If-If= 0 and in view of {7), if fEK also f-GifEK. G defined by (8) satisfies our requirements, and Theorem 3 tells us that natural spline interpolation is an optimal algorithm.

For another optimal method of recovery for this example see the end of Section 5.

Whenever our information is obtained by sampling at a given set of points, as in the present case, it is natural to seek the best set of points. That is, the intrinsic error, E(K,O), depends on the points x. For what choice of x is it minimal? Melkman and Micchelli [23] show that in the present case one considers the differential equation

y(2n)(t) = lJY(t),

y{j) (0) = y(j) (1) = 0, j = n, ••• ,2n-1. Its eigenvalues satisfy

0 = " = " - - " < " < ~1 ~2 - ••• - ~n ~n+l ••• •

and if yi(t) is the ith orthonormal eigenfunction of this system,

then yn+r+l(t) has exactly n+r simple zeros satisfying

0 < n1 < n2< ••• < nn+r < 1,

and they are the optimal sampling points.

Generalizations of this problem are considered by Melkman and Micchelli [23] and its relationship to the n-widths of the appro­priate spaces is revealed by them. In this connection see also Melkman [22].

One may be tempted to conjecture on the basis of these ex­amples (as well as other examples to be discussed) that the lower bound of Theorem 1 is always sharp (as was asserted by S. Winograd in [49], see also the article of M. Schultz [43], in the same pro­ceedings for material related to Theorems 1 and 2 and Example 1.1.) We will provide a counterexample to this conjecture and also point out that the upper bound in Theorem 2 may be reduced to 1:2e(K,E) when Z is a Hilbert space.

First, we discuss the relationship of the problem of optimal recovery to the notion of Chebyshev center of a set and to the hypercircle inequality (Cf. [12], [14], [21]).

We have already seen in the proof of Theorem 2 that the prob­lem of optimal recovery with error may be reduced to the case when E=O. Thus we confine our attention to this case.

Recall that if M is a bounded subset of Z then an element z EZ is called a (Chebyshev) center for M if

0

Page 17: Optimal Estimation in Approximation Theory

8 C. A. MICCHELLI AND T. J. RIVLIN

(9) sup llz0 -all = inf sup liz-all· aEM ZEZ aEM

The number on the right-hand side in {9) is called the Chebyshev radius of M and is denoted by r(M), see [14, p. 177]. We define the "hypercircle" determined by the intersection of the "ball" UK and the hyperplane Ix = y as

H(y) = {Ux:xEK, Ix=y}.

Clearly, H(y) ~ 0 if, and only if, yEIK.

Let us assume that for every yEIK, H(y) has a center, c(y)EZ. Then we have

* Theorem 4. A : y ~ c(y) is an optimal algorithm and

{10) E(K,O) = sup r(H(y)). yEIK

Proof. Suppose that A is any algorithm. Then for all xEK

I lux- A(Ix)l I ~ EA(K,O).

In particular

sup llw-z II ~ EA (K,O) ZEH(y)

where w = A{y). Hence r(H(y)) ~ EA(K,O) and therefore

(11) sup yEIK

r(H(y)) ~ inf A

EA(K,O) = E(K,O).

* To establish the reverse inequality and the optimality of A we observe that for any yEIK and xEK such that Ix=y

I lux - c(y) II ~ r(H(y)) ~ sup r(H(y)) • yEIK

Hence E(K,O) ~ sup

XEK and the theorem is proved.

* I lux- A (Ix)l I ~sup r(H{y)) , yEIK

Remark. The validity of {10) can be proved without assuming that H(y) has a center for every yEIK. Indeed our proof of (11) did not use c(y). But if

sup r(H(y)) < oo, yEIK

the only case of interest, is a c0 (y)EZ such that

llux-c6 (y) II ~

for any xEK with Ix = y.

choose o>O and then for every yEIK there

o + r(H(y)) ~ o + sup r(H(y)), + yEIK

Letting o ~ 0 gives the desired equality.

Page 18: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 9

Remark. Theorem 1, in the case E = 0, is a consequence of Theorem 4 since, as is easily verified

r(H(O)) sup xe:K Ix=O

I lux! I.

Remark. In the above discussion, when E>O, H(y) is replaced by H (y) = {Ux:xe:K, I jix-yj l~e:} and therefore r(H (0)) = e(K,e:). e: €

Our next example shows that the lower bound in Theorem 1 is not sharp. We owe our appreciation to Dr's. A. Melkman and I. Meilijson for their help in providing this example.

Example 1.4. Let ~ be the n-simplex defined as n

n -~n = {(x1 , ••• ,xn): E xi= 1, xi~ 0}, n~2,

i=l and suppose K is the smallest convex balanced subset of Rn which n

n contains ~ •

n Hence K is the i 1

n unit ball {x: E lxil~l} and i=l

is the convex hull of the points ±e1 , ••• ,±en, - ( 0 , i;'j

(ei)j -1 , i=j

For our example we choose X= Rn, K = K , Y • R1 z n

n-space) and Ux = x, n

Ix = E i•l

En(Euclidean

Thus we wish to recover a vector x which is known to be in K n from the sum of its components. The hypercircle n

H(d) • {x: E xi = d, x&Kn} is the convex hull of tA;1points

d+l d-1 C = {~ ei + *2: ej : i;'j}

1 and its center of gravity- (d, ••• ,d), being equally spaced from n all the points of Cis also the Chebyshev center of H(d). Hence

r(H(d)) and

E(K ,0) = max r(H(d))= r(H(l)) n jdj~l

However e(K ,0) n

than E(K ,O). n

(l-l/n)112

r(H(o)) = 2-l/2 hi h . f >3 . 1 1 w c 1s, or n- , str1ct y ess

Page 19: Optimal Estimation in Approximation Theory

10 C. A. MICCHELLI AND T. J. RIVLIN

The motivation for this example is a classical result of Jung (Cf. [14]) that states that for any bounded subsetM of fO

where r(M) ~ <2:+2) 1 / 2 diam(M)

diam(M) sup llx-y II x,yEM

and equality occurs if and only if M is a regular simplex. For any (infinite) Hilbert space we have

r(M) ~ 2-l/2 diam(M)

(Routledge, Cf. [14]). Note that the face of K which determines I n

is a regular simplex.

Clearly, for any yEIK

diam(H(y)) ~ 2e(K,O).

Hence, if Z is a Hilbert space we obtain the estimate

E(K,O) ~ /Z e(K,O)

(an improvement over Theorem 2). If, in addition, we have diam(H(y))~ n, all yEIK, e.g. Y = Rs, Z =Ens, then Jung's theorem tells us that

2n 1/2 E(K,O) ~ (;;y) e(K,O).

Now, the previous example show that this inequality is sharp

for all n. Since, in this case, dim (H(y)) and therefore in general we have

E(K,O) ~ l:f (l-l/n) 1 / 2 e(K,O).

1 n n-l,Y=R,Z=E,

However, r(O) = e(K,O) and thus equality is achieved for K = K . n

Let us next interpret the solution to the optimal recovery problem that is provided in Example 1.1 by (6) in the context of Theorem 4. Thus in the setting of the first sub-example of Ex­ample 1.1 we are required to find the center of the hypercircle

H(y) = {Ux: Ix = y, I lxl I ~ 1}, yEIK.

The result is a minor variation on [14, p. 197].

Theorem 5. Let X be a Hilbert space, K the unit ball in X and I a bounded linear operator of X onto the Hilbert spaceY.

* * -1 Then UI (II ) y is the center of H(y) and its radius is

Page 20: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 11

Proof.

ever Ix = 0.

* * -1 x = I (II ) y satisfies Ix = y and (x,x ) = 0 when-a 2 2 °2 2 ° Thus llx0 -xll =llxll -llx0 ll Sl-llx0 ll if Ix = Y and

llxii~L Consequently

sup llux-Ux II 0

sup llu(x-x ) II ( Ix=y 0

cn:r1~1 llxll~l ~ (1-llx 11 2/ 12

0

and we see that r(H(y)) ~ (1-l lx I 12>112 e(K,O). 0

To establish the reverse inequality put s =

Then if Ix=O and II x II ~1 II x ±sx 11 2 ~ II x 11 2 +s 2 = 1

0 0

and also I (x ±sx) = y. Thus for any zE z 0

llu<x +sx)- zll ~sup llux-zll 0

{fj:11~1 and

II U(x -sx)-z II 0

~ sup llux-zll·

<li:ll~l Hence by the triangle inequality,

slluxll ~ sup llux-zll

{fi:ll~l and thus we conclude that

and the proof is complete.

*

sup

{ Ix=O llxll ~1

lluxll

Remark. When Z = R, the real numbers, and UEX the set H(y) is an interval and the specification of this interval given by the theorem is sometimes called the hypercircle inequality.

Page 21: Optimal Estimation in Approximation Theory

12 C. A. MICCHELLI AND T. J. RIVLIN

If Z = R, while X is not necessarily a Hilbert space, the center of our hypercircle is

1 c(y) = 2 (sup Ux + inf Ux)

{ Ix=y {Ix=y xe:K xe:K

and c(Ix) provides an optimal algorithm by Theorem 4 (Cf. Bojanov [3]). In general c(y) is a nonlinear function of y. In the next section we show that when Z = R we can, in fairly general circum­stances, find optimal algorithms which are linear. Before turning to this let us observe that when Z = R it is easy to see that we have equality in (4) of Theorem 1. Since in this case the hyper­circle H{y) is an interval and thus its diameter is twice its radius. However, we already pointed out that the diameter of H(y) is ~ 2r{H(O)) and therefore r(H{y)) ~ r(H(O)).

Let us now show that sometimes an optimal algorithm may be obtained by variational methods.

We assume that E(K,O) < ~. Then there is an algorithm, A, such that

sup lux-A(Ix)l < M < ~. xe:K

Hence if we put K(y) {xe:K:Ix = y} then for all ye:IK

and we define

sup Ux < M + A{y) xe:K(y)

'(y;K) = '(y) = sup Ux. xe:K{y)

Then r(H{y)) = ('(y) + '(-y))/2 and c{y) = ('{y)-,(-y))/2.

Let us observe that '(y) is a concave function of y. For, if yie:IK, i = 1,2 and xie:K(yi), i = 1,2 then AYl + (l-A)y2e:IK and

Ax1+(1-A)x2e:K(Ay1+(1-A)y2) for 0 ~ A ~ 1. Thus U(Ax1+(1-A)x2) ~

'(Ay1+(1-A)y2) for any xie:K(yi), i 1,2. Hence A'(y1)+{1-A),{y2)~

HAY1+(1-A)y2).

Choosing A = 1/2 and y1 = -y2 = y we obtain 1

r(H(y)) = 2(,{y) + '(-y)) ~ '(0) = sup Ux, { Ix=O

xe:K

and in view of Theorem 4 we have again shown that we have equality in {4).

Next we note that the inequalities -'(-y) ~ Ux ~ '(y)

Page 22: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 13

are sharp for x€K(y). Furthermore, the concavity of$ implies that, for u>l, $(uy1+(1-u)y2) ~ u$(y1) + (l-u)$(y2). Hence we have for x€K(y) and 0 < A ~ 1

Therefore, for any A, 0 < A ~ 1

t(Ay)+<P(-Ay) ~ 2

(I ) = p(Aix)-p(-AixJ CA X 2A

<P(O) e(K,O).

is also an optimal algorithm.

verges to a limit as A + 0+. at zero then by definition

But, since$ is concave cA(Ix) con­If, in addition, $ is differentiable

lim cA(Ix) = F(Ix) A+O

where F is a linear functional which is necessarily bounded on IK. Thus the hypothesis of differentiability of $ at zero insures the existence of an optimal algorithm which is linear.

As it turns out optimal algorithms which are linear exist even when this condition fails. The existence of linear optimal algo­rithms is treated in the next section. We will also demonstrate in this section that the differentiability of $(y) at zero is equiva­lent to the uniqueness of the optimal linear algorithm.

Remark. For €>0, ~(y) is replaced by

$ (y) sup Ux € X€K (y)

K (y) {X€K IIIx-yjj~d. €

2. OPTIMAL ESTIMATION OF LINEAR FUNCTIONALS

In this section we shall suppose that X is a linear space over R (the real numbers), that Z =Rand that K is a balanced convex

subset of X. At the end of our discussion we shall indicate how our methods can be adjusted in the case that R is replaced by ¢ (the complex numbers). Our major aim is to show under what circum­stances there is an optimal algorithm which is continuous and linear, that is, an element of Y*.

* For any L€Y , X€K andy satisfying I jix-yj I~E we have jux-Lyj = jux-Lix+L(Ix-y)j

Page 23: Optimal Estimation in Approximation Theory

14 C. A. MICCHELLI AND T. J. RIVLIN

~ sup lux-Lix+ELYI

{ xe:K IIYII~l

Thus, recalling (4) we have

~ sup xe:K

lux-Lixl + £ IlLII·

(12) e(K,£) ~ E(K,£) ~ inf Le:Y*

sup lux-Lix+£Lyl ~ d(K,£)

{ xe:K IIYII~l

where we put (13) d(K,£) lux-Lixl + £IlL I P · The inequality we are dealing

inf {sup Le:Y* xe:K

between E(K,£) and with linear spaces

d(K,£) in (12) will be useful when over (,.

A basic question for us is: when is e(K,£) = d(K,£)? this is true if d(K,£) = 0. Suppose, then, that d(K,£) > 0.

C = {(Ux,Ix) : xe:K}

Clearly Put

and note that C is a convex subset of RxY. Then e(K,£) d(K,£) holds if, and only if, given any positive d < d(K,£) there is an xe:K satisfying I fixl I ~ £ such that

d ~ Ux ~ d(K,£), that is, for every positive d < d(K,£) the convex subset of RxY

cd = {(t,y) : d ~ t ~ d(K,£), I IYI I ~ £} has a point in common with C.

We are thus led to consider linear separation of convex sets.

Definition. Two convex sets A and B of R x Y are separated if there exists a non-trivial linear functional on R x Y and a real constant b such that

Hz ~ b, ze:A, Hz ~ b, ze:B.

A and B are strictly separated if Hz ~ b-o for ze:A and Hz ~ b+o for ze:B, for some o>O.

Lemma 1. H is a linear functional on RxY if, and only if, H(t,y) at+Ly, ae:R, La linear functional on Y.

Lemma 2. i) If £>0 and O<d<d(K,£) then Cd and C cannot be sepa­rated. ii) If e:~O and d=d(K,£), Cd and C cannot be strictly sepa­rated.

Proof. i) Suppose Cd and C can be separated. Then there exist b,ce:R and a linear functional L on Y such that (14) cd + Ly ~ b II Y II ~ £ and (15) Setting y b ~ 0 and

cUx + Lix ~ b , xe:K. 0 in (14) gives cd ~ b and putting x = 0 in (15) yields

therefore c ~ 0. Moreover, if c = 0 then b = 0 and (14)

Page 24: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 15

implies that L = 0. This is impossible because H(d,y) = cd+Ly is a non-trivial linear functional. Thus c < 0 and we have

sup (Ux+c-1Lix) ~ ~ XEK c

and £ II L II ~ b-ed

so that, noting that lb/cf = b/c,

d(K,£) ~sup lux+c-1Lrxl + £1 lc-1LI I XEK

~ b ! {b-ed) = d < d(K,£), c c

a contradiction which establishes i).

ii) Suppose Cd and C are strictly separated. (14) and (15) are now replaced by (16) cd(K,£) + Ly ~ b-o, IIYII ~ £ and (17) cUx + Lix ~ b+o , XEK for some o>O. Setting y=O cd(K,£) ~ b-o and b+o ~ 0.

in (16) and x=O in (17) yields Hence cd(K,£) < b+o ~ 0 and so c < 0.

Thus, as in i) d(K,£) ~ sup lux+c-1Lixl + £llc-1LII

X€K b+O 1 2o

~ -c- - c(b-o-cd(K,£)) = d(K,£) + c-< d(K,£) ,

and the contradiction proves the Lemma.

The value of Lemma 2 is that if we can show that Cd meets C by an argument based on a standard separation theorem, then we will have established that e(K,£) = d(K,£). Before we carry out this program let us give a condition under which i) of Lemma 2 remains valid for £=0.

A convex set, K, is absorbing if for every xEX there is a A>O such that AXEK.

Lemma 3. If IK when £ = 0.

{Ix:xEK} is absorbing then Lemma 2 i) is true

Proof. We need only prove that c+O in (14) and (15). But if c=O then b=O and hence Llx ~ 0 for XEK, and since IK is balanced and absorbing L = 0, which is impossible.

We recall the following: Definition. If K is a convex subset of the linear space X,xEK is an internal point of K if given any yEX there is a o>O such that x+AyEK for all IAI<o.

Page 25: Optimal Estimation in Approximation Theory

16 C. A. MICCHELLI AND T. J. RIVLIN

Note that if K is a balanced absorbing convex set then 0 is an internal point of K. We now quote (Cf. Royden [40]).

Theorem A. Let K1 and K2 be two disjoint convex subsets of the

linear space X, and suppose that one of them has an internal point. Then K1 and K2 can be separated.

We now have

Theorem 6. If g>O then e(K,E) = d(K,E). If IK is absorbing then e(K,O) = d(K,O).

Proof. Suppose d{K,E)>O and there exists a positive d<d(K,E) such that Cd and C are disjoint. If g>O, Cd has an internal point, while if IK is absorbing then C is a balanced absorbing set. The theorem now follows from Theorem A, Lemma 2 i) and Lemma 3.

Our next result will be used to compute the directional deriva­tive of the function ~ (y) which was introduced at the end of the e: previous section.

Theorem 7. Let e:~O. If e:>O or IK is absorbing andy is an in­ternal point of IK then

Proof.

~ {y) sup Ux e: { II Ix-y II ~e:

xe:K

= inf* {sup lux-Lixl + e: IlLII + Ly} Le:Y xe:K

Let I x = Ix-y then the y

sup Ux { II Iyx II ~e:

xe:K

above equation reads

{sup (Ux-LI x) + e:l ILl I} xe:K Y

This equation is similar to the duality relation we proved in Theorem 6. However, in the present case I is not linear but

y rather affine, i.e. preserves convex combinations. Nevertheless the previous proof applies in this case as well provided we make the following additional observations. The set C previously defined is still convex when I is affine. Hence for e:>O there is no change in the proof. When e:=O we must show that if L(I x) ~ 0, xe:K, and y

y an internal point of IK then L = 0. This is easily done. Given ze:Y there is a ~>0 such that y+Aze:IK for IAI<~. Therefore it fol­lows that Lz = 0.

Note that Theorem sup lux!

{ xe:K II Ixii~E

6 concludes that inf* {supiUx-Lixl + El ILl I} Le:Y xe:K

Page 26: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY

inf* sup lux-Lyl LEY (XEK

II Ix-yii~E

17

E(K,E),

but that nothing is said about whether the various limits involved are attained. We turn next to an examination of this question in which topological considerations are important.

* We require a feasibility condition:

sup lux-Lixl < ®' XEK

There exists LEY such that

and also need Theorem B (Banach-Alaoglu (Cf. [41])). Let X be a linear topo­logical space. If N is a neighborhood of 0 and if

* V = {gEX : lgxl ~ 1, all XEN}

* then V is weak compact.

Theorem 8. If E>O an optimal linear algorithm exists. If E=O and IK is a neighborhood of the origin then an optimal linear al­gorithm exists.

*Proof. The feasibility condition implies that there exists L EY such that

0 sup XEK

I Ux-L Ix I + E II L II ~ B < ®. 0 0

Thus a linear optimal algorithm must lie in the set

* {LEY : I (L-L )Ixl ~ 2B; xEK}. 0

* This set is weak compact when IK is a neighborhood of the origin. The function

*

G(L) = sup lux-Lixl XEK

is a weak lower semi-continuous function of L, and so the minimum exists when E=O. When E>O, a linear optimal algorithm must lie in

* the weak compact set

{LEY*: IlLII ~ ~} and the proof is complete.

The first part of Theorems 6 and 8 were proved using a version of the min-max theorem by Micchelli [27]. In the case that Y is a finite dimensional space see also Marchuk and Osipenko [20], and when E=O, Bakhvalov [1] and Smolyak [44].

The "off center" version of Theorem 8 is

Page 27: Optimal Estimation in Approximation Theory

18 C. A. MICCHELLI AND T. J. RIVLIN

Theorem 9. Let E~O. If E>O or IK is an absorbing neighborhood of

* * y then there exists a weak compact set M~Y such that

cpE(y) =min {suplux-Lixl + EIILII + Ly}. LEM XEK

Proof. Let L and B be as defined in Theorem 8. For E>O we choose * 0 1 M ={LEY : I ILl I ~ E- (B+L y)}. When E=O, the set

* 0 * {LEY : I (L-L )I xi~2B} is a weak compact set in which the minimum lies. 0 y

Next we examine the question of the existence of a worst function, that is, an xEK such that llrxll~e: and Ux=e(K,e:)=d(K,e:). Existence of a worst function is equivalent to having

cd(K,e:) n c ~ 0 .

According to Lemma 2 Cd(K,e:) and C cannot be strictly separated, but we recall

Theorem C. (Cf. [41]). If A and Bare disjoint non-empty convex sets in a locally convex topological linear space X, and A is com­pact and B is closed, then A and B can be strictly separated.

Thus, for example, we have

Theorem 10. i) If C is norm compact a worst function exists. ii) If C is weakly closed (equivalently, since C is

convex, C is norm closed) and Y is reflexive then a worst function exists.

Proof. i)

ii)

C is closed in the norm topology on Y. d (K, e:)

C is weakly compact. d(K,e:)

We turn next to the question of characterizing an optimal linear algorithm.

Theorem 11. Suppose * x is a worst function, then L EY 0 0

optimal algorithm if, and only if i) e: II L0 II

ii) max lux -XEK

L Ix 0 0

L Ixl Ux - L Ix . 0 0 0 0

Proof. If i) and ii) hold then

Conversely,

(18)

max I Ux - L Ix I + e: II L II = Ux XEK 0 0 0

if L is an optimal algorithm 0

max I Ux-L Ix I + e: II L II = Ux , XEK 0 0 0

d (K,.e:).

is an

Page 28: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 19

hence Ux - L Ix + £ II L II :s; Ux , or £ II L II ::;; L Ix • 000 0 0 0 00

But since L Ix :s; el IL I I, i) holds and ii) follows from (18). 0 0 0

Remark. If L is

0

~and ii) may also be used to characterize worst functions. an optimal estimator then x is a worst function if and

0

only if i) and ii) hold.

Before turning to the case that X is a linear space over the complex numbers we settle the question of the differentiability of $ (y) at zero. Recall, that the right directional derivative of

E

$ (y) at y in the direction w is defined as E 0

$ (y +Aw)-$ (y ) 1 . E 0 E 0

1m+ A A-+0

The left derivative is defined similarily and will be denoted by ($')(y ,w). We will say that$ is differentiable at y0 , if its

E- 0 E

right and left derivative exist for any direction and are equal.

Since $ (y) is a concave function its right and left deriva-e: tive exist. We will compute these quantities by using Theorem 9 which expresses $ (y) as a min function. The study of the differ­entiability of mid (max) functions apparently originated with Danskin [8]. The result we will use is proved in Micchelli [24].

Theorem D. Let T be a compact space and E a linear topological space. Suppose G(t,x) is a real-valued function defined on TxE such that for each xe:E, G(t,x) is continuous on T and for each te:T, G(t,x) is a convex function on E. Suppose g(x) =max G(t,x).

te:T Then

g+(x,y) = max G+(t,x,y) te:S(x)

where S(x) = {t:te:T,g(x) = G(t,x)} and G+(t,x,y) is the right de­rivative of G with respect to its second argument in the direction y.

Proof. If te:S(x) then for A>O

g(x+Ay)-g(x) ::;; A

G(t,x+Ay)-G(t,x) A

Letting A-+0+ we obtain g+(x;y)~G+(t,x,y) for all te:S(x) which implies

g+(x,y) ~max G+(t,x,y). te:S(x)

To prove the reverse inequality we use the fact that f(A)/A is a nondecreasing function for A>O whenever f(O) = 0 and f is a convex function on [O,oo).

Page 29: Optimal Estimation in Approximation Theory

20 C. A. MICCHELLI AND T. J. RIVLIN

Set g+(x,y) = o then from our previous remark it follows that

g(x+Ay) ~ AO + g(x), A>O. Define

h(A,t) = G(t,~Ay)-g(x) , A>O,

then for each t€T, h(A,t) is a nondecreasing function of A because

h(A t) = G(t,x+Ay)-G(t,x) _ g(x)-G(t,x) ' A A

Therefore the sets AA={t:t€T, h(A,t)~o} are nonempty, closed and nested for A>O. Thus the compactness of T implies that there exists a t 0 €AQOAA, i.e.,

G(t ,x+Ay) ~ AO+g(x), A>O. 0

Thus t 0 €S(x), G+(t0 ,x,y) ~ o, and therefore

g+(x,y) ~ G+(t0 ,x,y).

Remark. If max is replaced by min and convex by concave Theorem D is obviously still valid.

Theorem 12. Let e~O. If €>0 or IK is a neighborhood of the origin then the function ~ :IK+R is differentiable at the origin if and only if there is a uniqu~ optimal linear algorithm L • In this case, ~ '(0) L • 0

€ 0

Proof. Given any y€Y for IAI sufficiently small, w = AY is an in­terior point of IK. Hence by Theorem 9

~ (w) =min (suplux-Lixi+Lw+el ILl I>· €

LeM xeK * The function H(L,w) = supiUx-Lixi+Lw+el ILl I is weak continuous in L

X€K and linear in w. Thus by Theorem D if we setB =the set of all optimal linear algorithms then

and

~ (Ay)-~ (0) € E

lim+ A A+O min Ly LeB

(~E)~(O,y) =max Ly L€B

The theorem easily follows from these formulas.

For a related result see Osipenko [39].

Finally, as promised we examine the complex case. First note

Page 30: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 21

* that for any L€Y sup{Ux-Llx+&Ly)

{ X€K IIYII ~1

~ sup Ux { X€K

II Ixll ~e: Then if the hypothesis of Theorem 3 is assumed, (12) and Theorem 6 imply

d(K,e:) = sup Ux ~ inf* sup (Ux-Llx + e:Ly) ~ d(K,e:). { X€K L€Y (X€K

II Ix II ~e: II Y II ~1 Thus,

sup lux! sup (Ux-Llx + e:Ly). { X€K

II Ixll ~e: { X€K

II Y II ~1

Suppose now that X is a linear space over t, the complex num­bers, that K is balanced relative to ¢, i.e. AX€K for X€K and

* IAI = 1, A€¢. Also let Z = t and let Y (¢) be the complex valued

* continuous linear functionals on Y, that is, L€Y (¢) if, and only * if, L = L1+iL2 , L1 , L2€Y •

Clearly, sup lux! sup Re Ux

( X€K II Ixll ~e:

{ X€K II Ixll ~e:

But by applying our results in the real case with U replaced by ReU we obtain

sup ReUx = inf* ( x€K L1 €Y

II Ixll ~e: inf* sup Re(Ux-Llx + e:Ly), L€Y (¢) K

{ X€ IIY II ~1

inf* sup lux-Llx + e:Lyl, L€Y (¢) K

( X€ IIYII ~1

in view of the fact that K and I IYI 1~1 are balanced.

It remains now to show that for any Z€¢,

Page 31: Optimal Estimation in Approximation Theory

22 C. A. MICCHELLI AND T. J. RIVLIN

sup I z + e:Ly I = I z I + e: II L II , IIYII~l

a fact which is easily established. Thus we conclude that Theorem 6 continues to hold in the complex case.

In summary, then, in this section we have shown that under rather weak hypotheses a linear functional can be recovered opti­mally by means of a linear functional of the information. The next section is devoted to examples illustrating this fact.

3. EXAMPLES OF OPTIMAL RECOVERY OF LINEAR FUNCTIONALS BY LINEAR METHODS

Unless otherwise specified we suppose K to be a balanced con­vex subset of X, a linear space over R, and that Z = R. Also we suppose that K is such that the additional hypotheses in Theorems 6 and 8 are satisfied.

Example 3.1. Let Y = Rn normed with the sup norm, and suppose that for xeX

Ix = (I1x, ••• ,Inx)

where I., j = l, ••• ,n are linear functionals on X. Suppose also J

that the feasibility condition holds, then Theorem 6 implies that

(19) sup Ux

{ xeK IIjxl~e:

n n {sup lux-E cii.xl+e: E lcil}.

XEK i=l 1 i=l

3.1.1. For example. Take X= K = P 1 , polynomials of degree at n-most n-1. Take Iix = x(ti), i = l, ••• ,n where ti are distinct

n-1 points of the real line. If x{t) = a +a1t+ ••• +a 1t suppose Ux = a u + ••• + an-l un-l' Then ° n-

oo ~1 n Ux-Lix E aiui - E cjx(tj)

i=O j=l n-1 n i

E ai(ui- E cjtj), i=O j=l

from which it is clear that the feasibility condition is satisfied if and only if

u = i

n i E cjtj , i = O, ••• ,n-1.

j=l These equations determine c1 , ••• ,cn uniquely, and if we let

Page 32: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 23

Ut .• Thus (19) becomes, ~

for e:>O,

max Ux = lx(t.) l::;e:

~

n d i=l

lut .1. ~

A linear optimal algorithm is seen to consist of Lagrange inter­polation followed by U, as is trivially the case when e: = 0.

3.1.2. Suppose W is a normed linear space and w1 , ... ,Wn€W are

* linearly independent. Choose X = W ,

K= {T€X: IITII::; 1},

I.T = Tw., i=l, ••• ,n and UT = Tw for some W€W. Now the feasibility ~ ~

condition is clearly satisfied, therefore (19) yields

(20)

when e: o, (20)

(21)

n n min c.

J

{sup I Tw - E c. Twj I +e: E I c. I }. IITII;S;l j=l J i=l J

n n min { II w- E c . w. II +e: E I c j I }. cj j=l J J j=l

reduces to the familiar duality formula

= dist(w,M) = min cj

n

llw - E cJ.wJ.II , j=l

where M is the linear space spanned by w1 , ... ,wn. Note that when

e:>O, (20) depends on a particular choice of basis in M, which is not the case in (21).

Example 3.2 Suppose X= (X,I·I> and Y=(Y,I 1·1 I> are Hilbert

spaces and K = {x€X:Ixl::;l}, Let [x,x] = lxl 2 , (y,y) = IIYII 2 , Ux = [u,x], U€X, uiO and suppose I is a bounded linear operator from X onto Y. The feasibility condition being obviously satisfied we have, according to our theory

max[u,x] =min (max ([u,x] - (y,Ix) + e:l IYI I>·

C lxi;S;l y€Y lxl ::;1

IIIxll::;€

*

(22)

The adjoint I Y + X of I is defined by

* (y,Ix) = [I y,x] , X€X, y€Y and (22) becomes (23) max[u,x] min

( lxl ;S;l yd

IIIxll::;€

* ( I u-I Y I + e: II Y II ) •

Page 33: Optimal Estimation in Approximation Theory

24 C. A. MICCHELLI AND T. J. RIVLIN

Let x(E) be a worst function and y(E) an optimal linear algorithm (see Theorems 8 and 10 of Section 2). Clearly, y(E) is unique since (23) shows it to be the value at which a strictly convex function attains its minimum. According to Theorem 11 we have

E IIY(E) II = (y(E), Ix(E))

* * (24) and (25) max [u-I y(E),x] = [u-I y(E), x(E)].

lxlsl

Using the criteria for equality in Schwarz's inequality we see that (24) and (25) are equivalent to

(26) y(E) = 0 or Ix(e) = Ey(E) and IIY(E) II

(27) * v = u-I y(E) 0 or x(E) v -I vi

It is convenient to analyze these equations by considering several cases. Case 1. y(E) = 0.

In this case v=u ~ O, hence x(E) = v/lvl by definition, I IIx(e)l Is€ we conclude that

(28) 2( ) 2 _ (Iu,Iu) < 2 El u = El - [u,u] - E

u/lul, and since,

Conversely, if {28) holds, then x(E) = u/lul satisfies lx(E)Isl, I IIx(E)I Is€ and is a worst function. Hence e(K,E) = lu and therefore, since the minimum problem in (23) has a unique solution it must be y(E) = 0. Note that we have also proved that if E~El then x(E) = u/lul is the unique worst function.

Case 2. y(E) ~ 0 and v ~ 0. We may solve {26) and (27) for x{E) and y(e). the linear operators

(II* + AJ) -li and

To this end we define

where J denotes the identity map. Since I maps X onto Y the mappings PA and QA are well-defined for A~O. A simple calculation shows that

y(E) p U ll

Qllu

I Qllu I

and

x(e)

ll is the uniquely determined solution of the equation, g{A) = E

Page 34: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY

where IIP).ull

g().) = A I Q ul • ).

Again in this case x(e) is unique.

* * Case 3. U€R(I ) = range of I •

* Suppose u=I t for t€Y. We claim that v=o if, and only if,

(29) 2 < 2 ( ) = .,.2 (t. t) €-Et"' *l

0 * 0 ((II ) - t, t)

25

Clearly, and x(e) Let s be

if v=O then I y(e) = u, and since I maps X onto Y y(e)=t is any vector in X with lx(e)l~l and Ix(e) = etllftl I. any element of Y then, if e>O,

(s,t) = (s,Ix(e)) ll!ll = [I*s,x(e)] l1!ll € €

so that, for e~O

(30) * < II sllltll € -

l<s,t)l * -1 If we choose s = (II ) t we obtain (29).

Conversely, suppose (29) holds. Since, for any s€Y

so that

* * * -1 II* I > [I s,I (II ) t]

s - * * 1 II (II )- tl

* mi ~ Yn T<S:tn

S€

€ 0 nm

then, (29) implies that (30) holds. Now

II t II - II t-s II ~ lfitHl and we conclude that for s€Y

ell til ~ II*sl + d It-s II,

(s,t)

or, equivalently, * min ( I u-I Y I + E II Y II) E II t II • yd

hence y(e) = t and v = 0.

Note that we also have the inequality e (t) ~ e (u).

For if e1 (u) ~ E < e0 lt) then v = y(e) = 0 which implies u = 0,

contrary to our assumption.

Page 35: Optimal Estimation in Approximation Theory

26 C. A. MICCHELLI AND T. J. RIVLIN

* In summary, we see that if U€R(I ) we have the following classification according to the size of E.

Case 3.1. E :s; E 0

y(E) t and x(E) is any vector in X such that

t Ix(E) = € TTtTT and lx(E) I :S;l.

Case 3.2. E < E < El. 0

x(E) and y(E) are as described in case 2.

Case 3.3. El :s; E.

x(E) and y(E) are as described in case 1.

Finally, let us note that lim g(>..) El >..-+oo

* while if u = I t lim g(>..) =

ceo (t)'

* >..->0 0 ' if u I. R(I ).

Since y(E) is unique we also conclude that g(>..) is strictly in­creasing for >.. ~ 0, a fact that we used in our discussion of Case 2.

* Observe that in Case 3, when u=I t, t€Y, then Ux=[u,x]=(t,Ix). Thus Ux is a function of the exact information. However, the choice y(E) = t, continues to be optimal only for E :s; E •

0

3.2.1. As an example consider the problem of predicting the value of a function f at time T > 0 from its values in the past, {f(t), t<O}, which we can compute with error at most E>O. The case E=O is implicit in the discussion of an example given by Wiener [48, p. 65].

Let X= {f:f,f'€L 2 (R)} and suppose

[f,h] 1 f(t)h(t)dt + / 00 f'(t)h'(t)dt. -oo -oo

Take Y = L2 (-oo,O) and suppose (f,h) = ! 0 f(t)h(t)dt. -00

Let I be the identity mapping of X onto Y given by (If)(t)=f(t), t<O.

The space X has adjoint of I is seen

* (I h)(t)

a reproducing kernel K(x,t) to be

1 o -lx-tl = - f e h(x)dx. 2 -oo

1 -lx-tl -e so the 2

The linear functional Uf f(T) can be represented as [u,f] where

Page 36: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY

1 -IT-tl * * u(t) = 2e , teR. Since f(T)iY we conclude that uiR(I ). Thus, according to cases 1 and 2 of Example 3.2, an optimal al­

-T gorithm is obtained by estimating f(t) by~ if e~e1=(e )/2, while if e<e1 we must examine the integral equation

1 ° -I t-x I 1 -I T-t I (31) 2 f e y(x)dx + Ay(t) = 2e , t<O, -co

2 where yeL (-co,O) and A>O. The solution of (31) is . -1 1/2

YA(t) = ({l+A-1)1/2_1)e-Te(l+A ) t

27

and 1/2 A II y A II 1 (l+A -1) t

g(A)= =----;-~,.----.....:;;~~~-~~-~~"7'::' e lu-r*yAI ((l+A-l)l/2+1)(((1+A-l)l/2_1)(e2T_l)+e2T)l/2

-T Thus when e<(e )/2 the optimal recovery of f satisfying lfl~l, from the information, {f(t), t<O}, is given by

0 -T ax (a-1) r e e f(x)dx,

where a is uniquely determined by the equation

(a+l)2 [(a-1) (e2T_l) + e2T] = €-2 •

Example 3.3 We now consider optimal recovery of the derivative of a smooth func~~on1 Let (Zm) co wzm(R) = {f:f m- ) abs. cont. on every finite interval, f eL (R)}, for m ~ 1. Given real numbers 0 ~ t 1 ~ ••• ~tm' define

m d2 2 (Lf)(x) = IT (dxz- tj)f(x).

j=l

Let, X = w2m(R) nLco(R) equipped with the ess sup norm,

K = {feX : IILfllco~ 1},

Uf= f'(O), Y = Lco(R) and I be the identity operator. The intrin­sic error in this case is

suplf'(O)I

II f llco~e IILfllco~l

Micchelli [25] gives a detailed discussion of this extremal problem, as well as determining a best linear algorithm, which, in general, requires sampling the information at a denumerably infinite number of points. In the special case that m-1 and t 1=0 his results, in

agreement with those of Newman [37] and Schoenberg [45] for this

Page 37: Optimal Estimation in Approximation Theory

28 C. A. MICCHELLI AND T. J. RIVLIN

simple situation, yield

Ag = g{t)-g(-t) 2t

where t = (2€) 1 / 2 and g{t) is the available data.

An optimal algorithm for the differentiation operator is easily obtained from Micchelli's result by replacing f(x) by f(x+T).

Example 3.4 Next we consider optimal recovery of the value of a function at a given point from a finite sampling of the function, which is assumed to be "smooth" on an interval. This leads to optimal recovery of the function globally. To be precise, suppose Osx1sx2s ••• s xn+rsl with xi<xi+n' n~l and i = l, ••• ,r. Let

. .n n-1 (n-1) (n) oo X= W [0,1) = {fEC [O,l):f abs. cont.; f EL (0,1)},

equipped with the sup norm, and take K = {fEX: I lf(n)l lsl}.

Choose Y to be Rn+r and let I be the mapping that assigns to fEX the vector (f(x1), ••• ,f(xn+r)) with the convention that if

xi_1<xi = xi+l = ••• = xi+s<xi+s+l then f(xi+j) = f(j)(xi), j l, •.• ,s. Finally, suppose Uf=f(T) for some fixed Tin [0,1] and € = 0.

The general theory of Sections 1 and 2 provides us with sever­al approaches to this problem. The first that we discuss is based on finding the midpoint of the interval

H(y) = {f(T): llf(n) llsl, f(xi) = yi, i = 1, ... ,n+r}.

This problem may be solved by appealing to a result of Micchelli and Miranker [29] on envelopes of smooth functions.

Let m, M, m<M, be two given constants. A function, P, is called an (m,M) perfect spline of degree n with knots ~1 ••••• ~k

satisfying -oo = ~0<~1 ••• <~k<~k+l = oo if: i) Pis a polynomial of

degree at most n in each interval (~i'~i+l), i = O, ••• ,k. ii) PECn-l(-oo,oo), and iii) P(n)(t) = m, ~0<ts~1 , P(n)(t) M,

~1<ts~2 and so on alternately. Thus the n-th derivative of an (m,M)

perfect spline begins with m. If P is either an (m,M) or an (M,m} perfect spline we say it is an {m,M} perfect spline. An {m,-m} perfect spline is called a perfect spline.

The following result was proved in [29] where applications to finding the zero of real valued functions on an interval were given.

Page 38: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 29

Theorem E. Given O~x1~ ••• ~xn+r~l, y:(y1 , ••• ,yn+r) and m<M.

If there exists an f€wn[o,l] with

(32) f(xi)=yi,i=l, ••• ,n+r; ~f(n)(t)~, O~t~l, then there exist two {m,M} perfect splines of degree n, Pm and PM' with ~r knots in (x1 ,xn+r) such that

Pm(xi) = PM(x1) = yi' i=l,2, ••• ,n+r,

and for every f satisfying (32)

(-l)n+r+i(P (x)-f(x)) M

(-l)n+r+i(P (x)-f(x)) ~ 0 m

(xo = 0, xn+r+l = 1).

Before we comment on the use of this theorem we will digress from Example 3.4 to describe the proof given in [29].

tion The first observation to be made is the fact that the interpola­conditions f(x.) = y., i = 1,2, ••• ,n+r are easily seen to be

~ ~

equivalent to the x + = b, then

following moment conditions on f(n). Let x1=a,

n r b

f M(xi, ••• ,xi+n;t)f(n)(t)dt = [yi, ••• ,yi+n]' i=l, ••• ,r, a

where zi = [yi, ••• ,yi+n] is the divided difference of yi, ••• ,yi+n at

xi, ••• ,xi+n and1ui(t) = M(:!i ••• ,xi+n;t) is similarly the divided

difference of (n-l)! (x-t)+ at x = xi, ••• ,xi+n" This reduction

to moment conditions on f(n) is also used by H. Burchard [6] who examined the analog Theorem E corresponding to mFO and M=® (of

course, by this we mean that df(n-l)(x) is to be interpreted as a nonnegative measure). For further references concerning this problem see [6].

Now the main fact we require about the functions u1 (t), ••• ,

ur(t) is that they form a weak Chebyshev system. This means that u1 , ••• ,ur are continuous (we may assume that xi<xi+n-l since other-

wise the problem splits into independent subproblems on each of which u.(x) is continuous) linearly independent functions on [0,1] and ~

Page 39: Optimal Estimation in Approximation Theory

30

u (t ) r r

C. A. MICCHELLI AND T. J. RIVLIN

~ 0

for all O<t1< ••• < tr<l (the results in [6] also depend on this

property). A Chebyshev system has the additional property that strict inequality always holds above.

It is the case that every weak Chebyshev system may be "smoothed" into a Chebyshev system, Cf. Karlin and Studden [16]. Thus for 6>0 there exists a sequence u1 (x;6), ••• ,ur(x;6) of functions

which is a Chebyshev system and lim+ui(x;6) = ui(x), X€(0,1), 6-+0

boundedly in 6 and uniformly on every compact subinterval of (0,1). Now, the proof of the existence of Pm and PM follows easily from

the following theorem of Krein, Cf. [18], [19] or [16]

The (restricted) moment space generated by u1 , ••• ,u is defined as b b r

F (m,M)={( f u1 (t)h(t)dt, ••• , f u (t)h(t)dt):m5h(t)~, a.e., u r a a tda,b]} •

We will say h€L 00, mSh(t)SM, a.e. t€[a,b] represents Z€F (m,M), if

u

. . . '

b J ui(t)h(t)dt. Equivalently h represents z=(z1 , ••• ,zr) a [yi, ••• ,yi+n] iff there is an f€wn[o,l] with f(xi)=yi,i=l,2,

n+r and f(n) = h on [a,b] •

F is a convex compact subset of Rr with nonempty interior.

Theorem F. Let v1 , ••• ,vr be a Chebyshev system on [0,1]. Then

given any w = (w1 , ••• ,w) in the interior ofF (m,M) there exists r v two distinct step functions hm'~' which represent w, and have ex-

actly r jumps; each function satisfies (h(x)-m)(h(x)-M) = 0, a.e., X€[0,1], and hm(x) = m, ~(x) =Min some interval beginning at x=O.

The procedure to prove Theorem E is clear; we apply Theorem F to the smoothed system {u1 (x;6), ••• ,u (x;6)} and then pass to the

+ r limit as 6-+0 •

Page 40: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 31

It is interesting to note here that+the case mFO, M=~ treated by Burchard encounters at the stage "o-+0 " some difficulties which do not enter into our analysis. In particular, Burchard is dealing

+ with measures and for general data (z1 , ••• ,zr)' as o-+0 , infinite mass at either x = 0 or x = 1 may occur. However, this difficulty does not always occur, see Micchelli and Pinkus, [31].

Proof of Theorem E: Let z = (z1 , ••. ,zr)' zi = [yi, ••• ,yi+n]' i = l, ••• ,r. Then our hypothesis states that ZEF (m,M). Clearly u then for any m'<m<M<M', z is in the interior ofF (m',M') and for u fixed m',M' z is likewise in the interior ofF (m' ,M') (the moment

uo space determined by the smoothed system) for cS sufficiently small. Now, according to Theorem F there are two step functions h ,(t;o),

m h ,(t;o) which represents z. We define h ,(t)=lim+h ,(t;o) (through -"M (n) m o-+0 m a convergent subsequence), and P ,(t)=h ,(t) and similarly for m m PM 1 (t). Then (by adding on the appropriate polynomials to Pm' and

PM,) P ,(xi)= PM,(xi) = y., i=l,2, ••• ,n+r, and P ,, PM' haves r kiiots. m 1 m

Let f be any function in wP[O,l] with f(xi)=yi, i=l,2, ••• ,n+r ~f(n)(t)SM, a.e.tE[O,l], and define g=PM,-f. Then g vanishes at

(n-1) (n-1) each xi and hence g has at least r+l zeros. However, g is clearly strictly monotonic between the knots of PM'" Thus PM' must have exactly r knots and strictly cross f only at each xi. Since g(n-l)(x) is increasing beyond its last zero we have (-l)n+r+i (PM 1 (x)-f(x))>O, for xi<x<xi+l" A similar analysis shows that

n+r+i (-1) (Pm1 (x)-f(x))<O, for xi<x<xi+l" Letting m'-+m, M'-+M proves

the theorem.

Remark. The extremal properties of Pm and PM are not surprising.

It is well known that the principal representation described in Theorem F extremizes the "next moment". This fact is one form of the Markoff-Krein inequality (Cf., [14], Chapter 4, §8) and, of course, in our context, the next moment is f(x). Our proof that P ,

m PM are envelopes uses Rolle's theorem which is simple and direct.

This type of argument is, of course, the basis for the general form of the Markoff-Krein inequalities discussed in [16], [19].

Let us note some further important properties of Pm and PM.

Page 41: Optimal Estimation in Approximation Theory

32 C. A. MICCHELLI AND T. J. RIVLIN

First we suppose that z = (z1 , ••• ,zr), zi = [yi, ••• ,yi+n]'

i l, ••• ,r is an interior point ofF (m,M). Then clearly there is a function f such that u

0

- (n) [ ] f (xi)- y., i = 1,2, ••• ,n+r, m<f (t)<M,a.e.,t€ 0,1. 0 l. 0

Since f(n)(t) a.e.t€[0,1] lies strictly in the closed interval 0 [m,M] we may use Rolle's theorem as before, comparing f to P ,PM o m

and conclude that Pm and PM have exactly r knots, intersect only at

xi and are respectively (m,M), (M,m) perfect splines. Now, this same argument also shows the following additional fact.

Let ~1< ••• <~r be the knots of PM (a similar argument holds for

p ) and suppose s(x) is any spline function of degree n-1 with knots m

at ~1' • • • '~r' n-1

a.tj r n-1

s (t) I: + I: cj (t-~j)+ j=o J j=l

which vanishes at x=x1 , ••• ,xn+r" Then for any A, f 0 +As is in

wP(~i'~i+l), i = O,., ••• ,r and hence we conclude by the previous

Rolle's theorem argument that the function f 0 +As-PM has a constant

sign in each of the intervals, [O,x1 ], [x1 ,x2], ••• ,[x +r'l]. Thus letting A-+±co we conclude that s::O. n

We have thus shown that the determinant

K(O,l, ••• ,n-1,~1 , •••• ~r) xl ,x2, • • • ' xn+r

1 1

• n-1 • n-1 (xl-~r) + (xn+r -~r) +

is non zero. It is known that for any choice of ~i' xi, ~1< ••• <~r'

~< ••• <x the above determinant is nonnegative. Using these facts, .L n+r

we will now show that PM (as well as Pm) is unique. This will be

Page 42: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 33

shown by demonstrating that any (M,m) perfect spline interpolating the data and having r knots is alternately the upper and lower en­velope for all functions satisfying (32).

Let xi<x<xi+l and define

u(t)

K (O,l, .•. ,n-1,~1 , ... ,~r't)

xl,x2, • • • ' xn+r'x

where ~1 , ...• ~r are the knots for an (M,m) perfect spline with n+i+i

PM(xi) = yi, i=l, ••• ,n+r. Then (-1) u(t)~O, ~i<t<~i+l and

u(t) = O, tf(x1 ,xn+r). Thus by Taylor's theorem with remainder (or

integration by parts)

f xn+r PM(x)-f(x) = u(t) (P~n)(t)-f(n) (t))dt

xl

(-l)n+r+i

n+r+i Hence (-1) (PM(x)-f(x))~O, X€(xi,xi+l). This proves the unique-

ness of PM and a similar argument holds for Pm.

We end this discussion by making some observations about repre­sentations of boundary points ofF (m,M). First we recall some

u general facts, due to Krein, about restricted moment spaces.

A point z=(z1 , ... ,zr) is in Fu(m,M) if and only if for every al, ••• ,ar

r b r b r (33) E aizi$M f ( E aiui(t))+dt+m f ( E aiui(t))_dt

i=l a i=l a i=l

((f(t))+ = max{f(t),O}, (f(t))_ = min{f(t),O}). z is a boundary point

of F (m,M) u r

u0 (t) = E i=l

if and only if there is some nontrivial function

a~u.(t) which yields equality above. Moreover, if u0 (t) 1 1

is any such function and h, ~h(t)SM, a.e. t€[a,b] is any representer of z then

Page 43: Optimal Estimation in Approximation Theory

34 C. A. MICCHELLI AND T. J. RIVLIN

(34) h(t) = { M, m,

if u0 (t) > 0

if u0 (t) < 0

It is these facts that prove see [16 p. 235] that a boundary point of a restricted moment space corresponding to a Chebyshev system is uniquely represented by a step function h, (m-h(x)) (M-h(x)) = 0, a.e. xE[O,l] with sr-1 discontinuities. Conversely, any z represented by such a step function is necessarily a boundary point. For weak Chebyshev systems, in particular for the functions u1 , ••• ,u , the following facts are valid. If

b r z. = J u.(t)h(t)dt where (m-h{x)){M-h(x)) = 0, a.e. XE[a,b] and h

1 1 a

has sr-1 discontinuities then z is a boundary point.

Proof. Since for every o>O, u1 (x;o), ••• ,u (x;o) is a Chebyshev system, there is a function, r

a.(o)u.(x;o) such that 1 1

h(t) = {M,

m, otherwise

Cf. [16]. r

We normalize the coefficients of u0 so that E a~(o)=l i=l

and letting o-+0+ through a convergent subsequence we produce a non-r

trivial function u0 (t) = E 0 aiui(t) with

i=l

C' if u0 {t) > 0 h(t)

u0 (t) m, if < 0

0 Hence, u gives equality in (33) and so z is a boundary point of F (m,M).

u

Now, let us demonstrate that every boundary point may be so represented (but not uniquely).

The proof of this fact uses exists points ~0 = a<~1< ••• <~k<b

k ~j+l

Lemma 6 of Micchelli [26]. ~k+l' ksr-1 such that

E (-l)j f j=o

ui(t)dt 1,2, ..• ,r

There

Page 44: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY

b min 1

r a I: aiwi=l

i=l

r I I: a.ui(t)ldt i=l l.

Since z is a boundary point, (33) implies that w=(w1 , ••• ,wr) ~ 0 2 and A M=i. Hence

h(t) = M;m(-l)j + ~. ~i<t<~i+l is the desired representer of z.

35

Finally, let us observe that, since every nontrivial function r

u0 (t) = I: a~u.(t) must be nonzero on an interval of the form i=l l. l.

(xi,xm)' ~i~, lSi<mSn+r we conclude from (34) that every repre-

senter of z agrees on some "core" subinterval (xi,xm).

We summarize these facts in

Theorem G. a) If z = (z1 , ••• ,zr), zi = [yi, ••• ,yi+n] is a boundary point

of F (m,M) then there is an {m,M} perfect spline P of degree n, u with sr-1 such that P(xi) = yi' i=l,2, ••• ,r.

b) If z is a boundary point then there exists a "core" inter­val (xi,xm), ~i~ such that if f(xi) = yi' i=l,2, ••• ,n+r and

(n) mSf (x)SM, a.e. xE(x1 ,xn+r) then f(x) = P(x), xE(xi,xm).

c) If there is an {m,M} perfect spline P with sr-1 knots such that P(xi) = yi, i=l,2, ••• ,n+r then z is on the boundary of Fu(m,M).

This theorem is a statement, in the terminology of moment theory of some results of Fisher and Jerome [10] and Karlin [15] (see also de Boor [4]). To explain this further let M = -m =Land suppose z is a boundary ofF (m,M). Then

u 1 r -1

L

Now, according to

= 1 I I: aiui(t)ldt. r o i=l

min

I: aizi=l i=l

Lemma 6 of[26]

Page 45: Optimal Estimation in Approximation Theory

36 C. A. MICCHELLI AND T. J. RIVLIN

1 L min{ I lhl I: f h(t)ui(t)dt zi, i=l, ••• ,r}

0

(35)

min{ II f<n> II: f(xi>

<I 1·1 I= ess sup on (x1 ,xn+r)).

yi' i=l,2, ••• ,n+r},

Thus part a) is Karlin's result which states that the extremal problem (35) has a perfect splin.e solution with sr-1 knots. c) is also Karlin's observation that any perfect spline with sr-1 knots which interpolates the data solves the extremal problem. Part b) of Theorem G is Fisher and Jerome's result which shows that any two solutions of (35) agree on some core subinterval of (x1 ,xn+r). The Envelope Theorem gives sharp pointwise bounds on all solutions to the extremal problem. The envelope is necessarily squeezed to­gether on the core subinterval and in general neither the upper or lower envelope is a Karlin solution to (35).

Returning to Example 3.4, we use Theorem E as follows:

Specializing Theorem E to M = -m = 1 we see that the midpoint of the interval H(y) is c(y) = (P_1 (<) + P1 (T)/2, and its length is

IP_1 (T)- P1 (<)l/2. Thus c(y) is an optimal algorithm for recover­

ing f(T).

Gaffney and Powell [11] discovered Theorem E independently and used it as follows: As pointed out in Section 1, for any A,O<ASl,

P_1 (t;Ay)+P1 (t;Ay) lim+ 2 A~

s(t;y)

is also an optimal algorithm. Gaffney and Powell proved that s(t;y) is a spline interpolant of degree n-1 to the data. The knots of s are determined by the perfect spline of degree n,q(t), defined by

P1 (t;Ay)-P_1 (t;Ay) lim+ 2 A~

q(t).

q has r knots and is zero at each x., i=l, ••• ,n+r. This determines q uniquely (see Micchelli, Winograd1 and Rivlin [35]).

The approach in Micchelli, Winograd and Rivlin [35] begins with q(t), defines a linear recovery which is the spline interpolant s(t;y) and proves directly that it is optimal. Neither approach uses the duality formula proved in Theorem 6, which reads in this case,

Page 46: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY

min c1, ••• ,cn+r

n+r sup lf(T) - L cif(xi)l.

i=l

37

The minimum problem can be written, using Taylor's theorem with re-mainder, as

1 min ! ls(t)ldt,

0 where the minimum is taken over all (spline) functions

n+r 1 n-L cJ. (xJ. -t) + }

j=l S(t) 1 n-1

(n-1) ! { ( T-t) +

subject to the boundary conditions S(i)(O) 0, i=O, •.. ,n-1. The remarks in [26] imply that this minimum problem has a unique solu­tion. The perfect spline, q(t), can then be obtained directly from the minimum problem. Also, since in this case the minimum problem has a unique solution

is differentiable at zero and its derivative in the direction w=(w1 , ••• ,wn+r) is s(t;w).

Remarks. 1. Suppose that U is the identity operator in W0[0,l]. Since the perfect spline q mentioned above is independent of T it is clear that

max II f II = II q II . ~~~:~~)~~:~i•l, ... ,n+r

Moreover, the above-mentioned inter~olating spline approximates f globally with error not exceeding I lql I· It fails to be an optimal

algorithm because it may lack one derivative of being in W0[0,l]. However, by smoothing it appropriately we can produce a bona-fide algorithm as close to optimal as we desire. This proves that for the identity operator, U,we still have e(K,O) = E(K,O).

These facts also hold if we measure the error in recovery in

any monotone norm. Thus if U:X+Lp[O,l], l~p~oo the recovery error

Page 47: Optimal Estimation in Approximation Theory

38 C. A. MICCHELLI AND T. J. RIVLIN

is I lql I , the Lp norm of q. p

2. Other linear functionals such as f' (T) where T = xi or TE[O,l]-(x1 ,xn+r) are served by the same optimal algorithm, as is

clear from the proof of Micchelli, Rivlin, Winograd [35].

* 3. There is a set of nodes, x , with respect to which E(K,O) is minimal. These nodes are the zeros of the Chebyshev perfect spline of degree n with r knots in [0,1] when the error is mea­sured in L®, see Tihomirov [46]. For the Lp case, p~l, see Micchelli and Pinkus [30], [34].

4. There are also results known where X is

W~n)[O,l] = {fECn-l[O,l]: f(n-l)abs. cont.;f(n)ELP[O,l]},

l~p. For instance, Bojanov [2] studies this problem for some re­stricted classes of nodes, x.

5. If X is taken to be the set1of n-fold integrals of functions of bounded variation on [0,1] the L norm, and K is the subset of functions stemming from those of total variation not exceeding 1, 1 then Micchelli and Pinkus [33] obtain what may be considered the L analog of the result mentioned in Remark 1. They exhibit an opti­mal algorithm, which is spline interpolation of the information and show that e(K,O) = E(K,O). They also determine the set of points at which to sample in order to minimize the intrinsic error. ([32])

6. A periodic version of this problem is studied by Bojanov, [3].

7. The optimal recovery of f(T) from f(x1), ••• ,f(xn+r) when

E>O is discussed in [28]. The approach is based on Theorem 6.

We now turn to some problems of optimal recovery of analytic functions in the plane.

®

Example 3.5 Recall that H is th~ set of bounded analytic function in the unit disc D:lzl<l. If fEH, I lfl I= suplf(z)l, ZED. Put X= H® and let K = {fEX:I lfl 1~1}. z1 , ••• ,zn are given points of D,

Y = tn and for each fEK, If= (f(z1), ••• ,f(zn)), with the convention

that if some of the points z1 , ••• ,zn coincide the corresponding

functions values are replaced by consecutive derivatives in the ob­vious way. Suppose that ~ED is given, Uf = f(~) and E = 0.

Let

Page 48: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY

B (z) = n

; =.t. j=l 1-z.z

J

39

then the elegant theory of extremal problems in H~ for bounded linear functionals which stem from rational kernels (Cf. Duren [9, Ch. 8]) tells us that the coefficients, a1 , .•• ,a of the opti-mal recovery are determined uniquely from n

(36) n

Lf = f(~)+ L a.f(z.) j=l J J

and Bn(z) is the unique (up to rotation) worst function and IBn(')i

is the intrinsic error. In making the proper choice of the kernel in (36) we use Theorem 6 of Section 2 in this problem. The calcu­lation of the parameters a1 , ••• ,an of an optimal algorithm

from (36) is easy. For example, if n=l and z1=0 we obtain

a1 = -(1-1~1 2 > and E(K,O) = 1'1· But note that (Cf. Golusin [13,

p. 287])

if(~)-f(O) 1-1,1 2 <

12l 121 - I~ I 1- f(O) '

l-lf<o>l 2 ~ l-lfCo>l 2 1,1 2

which informs us that in addition to the linear optimal algorithm

(l-l~i 2 )f(O), there is a non-linear optimal algorithm, namely,

f(O),

and indeed there are uncountably many non-linear optimal algorithms.

Remark. Osipenko [38] solves the optimal recovery problem when in our notation K is the set of functions analytic and bounded by a fixed constant in a simply connected domain symmetric with respect to the real axis and containing an interval J, of the real axis. He as­sumes that ' as well as the sampling points are situated in J. He also studies the question of the optimal choice of sampling points so as to minimize the maximum intrinsic error for all '€J.

Example 3.6. Assumptions are all the same as in Example 3.5 except that Uf = f'(,). Our main result and the general theory lead to the conclusion that if a1 , ... ,an are the parameters of an optimal al­gorithm then

(37) Lf

where either

n f'(') + L a.f(z.)

j=l J J

1 2iTi f K(z)f(z)dz

lzl=l

Page 49: Optimal Estimation in Approximation Theory

40

(38) K(z)

or

(39) K(z) -iy Ce

C. A. MICCHELLI AND T. J. RIVLIN

(z-a) (1-az) 1 2 - 2B:('Z')

(z-~) (1-~ z) n

for approiately chosen C(>O), a satisfying lal~l and a real y, all depending on ~. In other words for each ~ in D there is a C, y and ~ determined. The corresponding (unique) worst functions are

(40) F(z) = iy z-a B ( ) e ------· z - n

and 1-az

(41) F(z) eiy B (z) n

respectively.

In order for (37) to hold we must have

(42) 0

which enables us to determine a in terms of ~. The complex number

Ceiy is then determined by the fact that the coefficient off'(~) in (37) is 1, so that

[K(z)(z-~) 2 ] = 1. z=~

Once K is thus specified the calculation of the parameters a1 , ••. ,a of the optimal algorithm is straightforward. n

Let us now carry out the computation indicated by (42) when K is given by (38) and then (39).

a) [<l-az) 2 1 J ' = 0 - 2 ITz)

(1-~z) n z=~ implies that

(1-I~1 2>B'(~)-2CB <~> a = ------~~n~----~n __ ___

~(l-I~1 2>B'(~)-2B (~) n n

(43)

The condition that a as given by (43) is satisfactory is lal~l, or

equivalently, lal 2~~. which is satisfied, if, and only if

(44) I B~ (~) I 2 Bn (~) ~ -1--~-~-~2~

Thus if ~ED satisfies (44) then a is given by (43) and with this a

Page 50: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 41

the problem is solved as indicated above. The set of ~€D for which ~4) is violated must then lead to the kernel given in (39). Let us see next how this works out explicitly.

b) r~z-~) (~-az) J ' = 0

Ul-~z) B (z z=~ n

implies that

1 a-~ 1-~a (45) 2 {- + ~} 1-~a:

Now

w = .!:.L 1-~a:

B' (~) 1-1~12 n Bm' 2 n

gives a 1-1 map of lils;l onto lwls;l and

gives a 1-1 map of

(46) I B~ (~) l rn> n

lwls;l onto

> 2

- 1-1~1 2

Hence

if, and only if, (45) holds for a satisfying lals;l, and if ~€D satisfies (46) the a given by (45) solves the problem, as we have seen.

Thus D is partitioned into two subsets Da and Db which meet along the curve

I B~ (~) I B (~) n

2

These subsets can pose z1=z2=

sometimes be simply described. For example, sup­z = 0, so that If= (f(O), f'(O), ••• , n

f(n-1) (O)). In this case B (z) n

zn and so the curve becomes

1~1 2 + ~1~1 - 1 = 0 n

or

I~ I r n

Th "f I I in case b), eiy zn is a worst function and the us 1 ~ s;rn we are

intrinsic error is n. If l~l~r we are in case a). We obtain n

Page 51: Optimal Estimation in Approximation Theory

42

a = (l-lr,;l 2)n-2lr,;l 2

r,;[n(l-lr,;l 2>-2l

C. A. MICCHELLI AND T. J. RIVLIN

and then use (38), {40) and (37) to find all the required parameters.

When n = 1 we recover a theorem of Dieudonne (Cf. Caratheodory [7, p. 19].).

Example 3.6 The following example was communicated to us by Donald J. Newman. Again X= H00

, but now K = {fEX: I If' I 1~1}, z1=z2= ••• = zn= 0 and If= (f(O), ••• ,f(n-l)(O)), while Uf = f(r;),

lr,;l<l and E = 0. Put r,; =rei~. and

00 j f(z) = I: aJ.z •

j=O Clearly, g{z) = zn/n is inK, Ig = 0 and lg(r,;) I = lr,;l n/n. Thus e(K,O)~Ir,;ln/n. But consider the linear functional

~ n+k n-l n-k a r,;n + I: an+k r,; + k--I:l ii+k n k=l

2k n-k r an-kr,;

n-1 f(r) + I: (_j__ r 2 (n-j) - 1) r;ja.

"' 2n-J" J j=O

f(r,;) - A(If)

we have

(47) 21T 00 k

ILfl ~ lr,;ln !_ r I!+ 2 L r cos k(~-e>l de. 21T 0 n k=l ii+k k The sequence {r I (n+k)} is convex, hence the quantity whose ab-

solute value is the integrand in (47) is non-negative, as summation by parts twice reveals. The absolute value signs can be erased and we obtain ILfl~lr,;ln/n; which shows that A is optimal.

4. EXAMPLE OF OPTIMAL RECOVERY OF A FUNCTION

This section is devoted to a specific example of the optimal recovery of a function (i.e. U is the identity operator) which we

Page 52: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 43

believe has some interesting features. It is a generalization of Example 1.2.

Let X= tp(t) for p~l, that is f {fj}, fj€ t , j=O,±l,±2, ••• is contained in X if, and only if,

1

o: I Llp)p < 00 •

J

(Throughout this indicated.) Let Put

section all sums are from -co to co unless a= {a.}, a.€¢, j=O,±l,±2, ••• be a given

J J 1

K= K(a,p) {f€X:(Eia.f.lp)p ~ 1}. J J

otherwise sequence.

K is a convex (in view of Minkowski's inequality) balanced subset of X. Let J = {k1 , .•• ,km} be a given set of m distinct integers,

and let J' denote its complement in the set of all integers. Take

Y = ~. equipped with the p-norm, define I by

If (fk , ••• ,fk) , 1 m

and suppose Uf Put

f, all f€X, so that Z

and if A>O

A = inf I a. I ; B j€J' J

B A

c .

Recall that, in the present setting

e(K,£)

Then we have

X.

(48)

Lemma 4. i) If A > 0, C ~ 1 and EB ~ 1 then

e(K,£) (A-p + (1-Cp) Ep) l/p

(49)

ii) If A > 0, C ~ 1 and EB ~1 then 1

e(K,E) B •

Page 53: Optimal Estimation in Approximation Theory

C. A. MICCHELLI AND T. J. RIVLIN

(50) iii) If A > 0 and C > 1 then

1 e(K,E) = A . iv) If A = 0 than e(K,e) = ~

Proof. Suppose A> 0, then

Therefore,

and

Note that

since

Thus:

(51)

(52)

(53)

Ap E If. I p ~ E I a. f I p ~ jEJ' J jEJ' J j

a Elf.lp ~ A-p + L (1-I:)IP) lf.lp

J jEJ A J

~ A-p + (1-cP) E If. IP . jEJ J

L If .IP ~ B-p jEJ J

i) if c ~ 1 and EB ~ 1 then 1

e(K,£) ~ (A -p + (1-Cp) Ep)p ii) if C ~ 1 and EB ~ 1 then

e(K,E) < ! - B iii) If C > 1 then

e(K, E) < ! . - A

We show next that equality holds in (51)-(53) by exhibiting extremal functions. Choose any 8 > 0. Consider f defined by fj=O. j ~ k,q, where kEJ, qEJ', B = 1~1, laql <A+ 8 and

with

f = (l-A.PBP) 1/p fk = A ; q Ia lp

q

Page 54: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY

For this choice of f we have

and 1 p p

Elf. lp = Ap + -A B > J Ia IP

q -1 Now choose A = min(E,B ) and then A = 0, and since 6 was arbitrary

as well. (48)-(50) hold, and (iv) is proved

We turn now to optimal algorithms. If yEY define y'Eip by Yj = yj for jEJ, Yj = 0 for jEJ'. We now define the following three

algorithms.

(54) \ : y + y'

(55) 1:; y + 0

(56) s y + (1-cP)y'

Since there is no optimal algorithm if A = 0 we suppose henceforth that A > 0. Then we have

Theorem 12 • i) If C = 0 or £ = 0, i is an optimal algorithm

with E(K,£) = (A-p + £-p)l/p. ii) If C ~ 1 or C < 1 and £B ~ 1, 1:; is an optimal

algorithm with E(K,£) =max (A-1 , B-1). iii) If 0 < C < 1 and £B ~ 1, S is an optimal algorithm

with E(K,£) = (A-p+Ep(l-Cp))l/p.

- - -Proof. Let f (f f ) denote an element of Y. kl' • •• ' km

Since

i)

i)

ii)

Ep (K,£) 1

follows

E~ (K,£)

from (48) and (50).

sup Elf lp

{ Elf.-f ip~Ep j€J J j

Page 55: Optimal Estimation in Approximation Theory

46 C. A. MICCHELLI AND T. J. RIVLIN

Since we always have

ii)

I: Jf.JP $ B-p jEJ J

follows from Lemma 4.

iii) E~(K,e) sup

As usual we have

(57) E I f . I P $ L - cP E I f . I P . jEJ' J Ap jEJ J

Also, an application of Minkowski's inequality yields

I: if - (1-Cp) f JP = I: J(l-Cp)(f.-f.) + cPf.JP j EJ j j j EJ J J J

$ [< E I (1-cP)(f.-f.) JP/IP + (E JcPf.JP>l/p~P j EJ J J j EJ J j

$[(1-Cp) (I: Jf.-f.JP)l/p +Cp( I: if JP)l/p]p jEJ J J jEJ j

$ [(1-Cp)u + cPv]P ,

where we put

u = (I: Jf.-f. JP)l/p ; v = (I: Jf.Jp)l/p jEJ J J jEJ J

Since the function tP, p>l is convex for O<t we have for p~l

so that,

I: Jf.-(1-Cp)f. JP $ (1-Cp) I: jEJ J J jEJ

Page 56: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 47

~ (1-Cp)Ep + cP L lfjlp j€J

which taken with (57) and (48) completes the proof of (iii) and the theorem. Remark 1. Theorem 12 is summarized in Figure 3.

E

,&,...----------~c (,

Figure 3

Remark 2. Suppose X= L2 [0,l],

(58) K = {fEX: ! 1 (f'(t)) 2dt ~ 1}, 0

2k+l 2 Y = t with the L -norm, If = (f_k fk) where 1 •...•

fj = f f(t)e-2~ijtdt, 0 and E~O. Thus we consider the problem of recovering a function

satisfying (58) from, possibly noisy, values of its Fourier co­efficients of order at most k.

Page 57: Optimal Estimation in Approximation Theory

48 C. A. MICCHELLI AND T. J. RIVLIN

If f~E c.e2nijt then (58) is equivalent to J

4n2 E j 2 lcjl 2 ~ 1,

and by identifying X with i 2 (¢) and 2nj with a., Theorem 12 can be J

invoked. Since C = 0, Figure 3 tells us that an optimal algorithm is obtained by simply forming the trigonometric polynomial of degree k whose coefficients are the observed Fourier coefficients

f -k, ... 'fk .

5. OPTIMAL RECOVERY BY RESTRICTED ALGORITHMS

The material in this section is suggested by some recent re­sults of H. F. Weinberger [47].

Referring back to Example 3.4 we see that the dimension of the optimal recovery spline is equal to the number of data values. If this number is excessively large it may be desirable to have an al­gorithm which requires fewer parameters. For instance, we may de­sire a method of recovering f(T) which is a polynomial of low degree, e.g., a linear function of '•

A(f(x1), ••• ,f(xn+r)) = a0 (f(x1), ••• ,f(xn+r))+Ta1 (f(x1), ••• ,f(xn+r)).

In [47] Weinberger treats such questions in Hilbert spaces.

Following Weinberger we c.onsider the following problem of re­stricted recovery from exact data. Let U,X,Y,Z,K and I be as before, except that X is a Hilbert space, K is its unit ball (actually K may be the unit ball determined by a Hilbert space semi-norm), Y = Rn,

and e: = 0. However, instead of allowing

any transformation of Rn into Z as an admissible algorithm we require the transformation to have the form MA(Ix) where M is some fixed m n ----- m mapping from R into Z and A is any transformation from R into R . Figure 1 now becomes

Figure 4

Page 58: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY

We suppose that the range of M is a subspace of Z of dimension m, which we also denote by M. Thus algorithms are restricted to map into McZ. As before we assume that I maps X onto Rn, and the in­trinsic error is defined to be

E(M) = inf SUJ? I lUx - MA Ixll .. A llxll$1

49

The analog of Theorem 1 in the !?resent context is easily seen to be e (M) = max ( sup II Ux II , sup inf II Ux-z II) $E (M) .

{ Ix=O llxll$1 ZEM

Weinberger MAI, where space.

I fxll$1 proved that the intrinsic error in approximating U by A is an nxm matrix is precisely e(M), when Z is a Hilbert

Theorem H (Weinberger) If X and Z are Hilbert spaces and A(n,m) is the set of nxm real matrices then

min SUJ?

AEA(n,m) I lxl 1$1 I lux.:. MAixll = e(M).

Hence there is always a linear optimal recovery for U of the form MA, AEA(m,n).

This theorem of Weinberger has some interesting consequences for unrestricted recovery of U using the information I. We showed in Example 1.1 that the (unrestricted) intrinsic error satisfied

E(K,O) = sup lluxll. Ix=O llxiJ$1

Hence, if M is any finite dimensional subspace of Z such that sup lluxll ;;:: sup inf I lux - z II

{ Ix=O llxll$1 ZEM llxll $1

then there is an (unrestricted) optimal recovery of the form MAix where A is an nxm matrix. In many cases this inequality is easily demonstrated.

For example, if U is a compact operator and A1;;::A 2;;:: ••• are the

* eigenvalues of U U then by the min-max characterization of the

* eigenvalues of U U we have

max II Ux 112 ;;:: An+l·

(Ix=O

llxll$1

Hence if M is any n dimensional subspace of Z such that

sup inf llux-zii 2 =An+l llxll$1 ZEM

then we obtain an optimal algorithm of the form MA. Since the

Page 59: Optimal Estimation in Approximation Theory

50 C. A. MICCHELLI AND T. J. RIVLIN

n-width of the set U = {Ux:l lxllsl}

d inf sup inf llux-z II n dim MSn llxllsl ze:M

equals An+l this requirement on M means that it is an optimal sub­space for approximating the set U. There are many subspaces which have this desired property (Cf. Karlovitz [17], Melkman and Micchelli [23]). If the associated eigenfunctions are given by

* U U'i = Ai,i'('i''j) = oij' then a classical choice for such a sub-

space is the one spanned by u,l, ••• ,u,n.

For the special Example 1.3 these remarks imply that in addi­tion to natural spline interpolation to the data we may also choose an optimal recovery which is an "approximation" by natural splines with knots at n1 , •.. ,nn+r" Thus in the notation of Example 1.3 we have

d + (K) n r 1

lln+r+l

sup inf II f-g 11 2 II f(n) 112sl ge:Sn

where S = the space of natural splines of degree 2n-l with knots at n nl' ••• , nn+r and K = { f: II f (n) 11 2sl}.

Thus we have a trade-off between two desirable properties. The first method of recovery, natural spline interpolation Sf, has the advantage of being interpolatory. However the knots of the splines, being at x1 , ... ,xn+r change if we alter even one sampling point. On

the other hand, the second method uses the same set of knots regard­less of the sampling points. However, the method of approximation is more complicated. We must find an n+r x n+r matrix, B, such that

n+r min max II f - I: giAi/(xj) 112

Aij II f (n) ll 2sl i=l

n+r max I If - I: giBijf(xj) 112'

II f (n) ll 2sl i=l

where (g1 , ... ,gn+r) is a basis for the natural splines of degree

2n-l with knots at n1 , ... ,nn+r"

Page 60: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 51

Acknowledgement: We are grateful to Dr. A. Melkman for reading an earlier version of this manuscript and suggesting many improvements.

REFERENCES

1. Bakhvalov, N. S., On the optimality of linear methods for operator approximation in convex classes of functions, USSR Computational Mathematics and Mathematical Physics 11 (1971), 244-249.

2. Bojanov, B.D., Optimal methods of interpolation in

W(r)L (M·a b) Comptes Rendus de l'Acadamie Bulgare des q ' ' '

Sciences 27 (1974), 885-888.

3. Bojanov, B.D., Best methods of interpolation for certain classes of differentiable functions, Matematicheskie Zametki 17 (1975), 511-524.

4. Bojanov, B.D., Favard's interpolation and best approximation of periodic functions, preprint.

5. deBoor, C., A remark concerning perfect splines, Bull. Amer. Math. Soc. 80(1974), 724-727.

6. Burchard, H., Interpolation and approximation by generalized convex functions, Ph.D., Dissertation, Purdue University, Lafayette, Indiana, 1968.

7. Caratheodory, C., Theory of Functions of a Complex Variable, Volume Two, 2nd English Edition, Chelsea, New York, 1960.

8. Danskin, John M., The theory of max-min, with applications. SIAM Journal, 14 (1966), 641-664.

9. Duren, Peter L., Theory of Hp spaces, Academic Press, New York 1970.

10. Fisher, S.D. and J.W. Jerome, The existenc!, characterization and essential uniqueness of solutions of L extremal problems Trans. Amer. Math. Soc. 187(1974), 391-404.

11. Gaffney, P.W. and M.J.D. Powell, Optimal Interpolation, C.S.S. 16 Computer Science and Systems Division, A.E.R.E., Harwell, Oxfordshire, England 1975.

12. Golomb, M., and H.F. Weinberger, Optimal approximation and error bounds, in On Numerical Approximation, R.E. Langer ed. The University of Wisconsin Press, Madison (1959), 117-190.

Page 61: Optimal Estimation in Approximation Theory

52

13.

14.

15.

16.

17.

18.

19.

20.

21.

22.

23.

24.

25.

26.

C. A. MICCHELLI AND T. J. RIVLIN

Golusin, G.M. Geometrische Funktionentheorie, VEB, Berlin 1957.

Holmes, R.B., A Course on Optimization and Best Approximation Lecture Notes Series 257, Springer-Verlag, Berlin 1972.

Karlin, S., Interpolation properties of generalized perfect splines and the solution of certain extremal problems I, Trans. Amer. Math. Soc. 206 (1975), 25-66.

Karlin, S. and W.J. Studden, Tchebycheff Systems with Appli­cations in Analysis and Statistics, Interscience Publishers, New York 1966.

Karlovitz, L.A., Remarks on variational characterization of Eigenvalues and n-widths problems, J. Math. Anal. Appl. 53 (1976), 99-110.

Krein, M.G., The L-problem in abstract linear normed space in some questions in the theory of moments (N.I. Ahiezer, M.G. Krein, Eds.), Translations of Mathematical Monographs, Vol. 2 Amer. Math. Soc. Providence, R.I., 1962.

Krein, M.G., The ideas of P.L. Chebyshev and A.A. Markov in the theory of limiting values of integrals and their further devel­opments, Amer. Math. Soc. Transl. 12 (1951), 1-122.

Marchuk, A.G., K. Yu Osipenko, Best approximation of functions specified with an error at a finite number of points, Matemati­cheskie Zametki 17 (1975), 359-368.

Meinguet, J., Optimal approximation and error bounds in semi­normed spaces, Numer. Math. 10 (1967) 370-388.

Melkman, A.A., n-widths and optimal interpolation of time- and band-limited functions , these proceedings.

Melkman, A.A. and C.A. Micchelli, On nonuniqueness of optimal

subspaces for L2 n-width, IBM Research Report 6113 (1976).

Micchelli, C.A., Saturation classes and iterates of operators, Ph.D. Dissertation, Stanford University, 1969.

Micchelli, C.A., On an optimal method for the numerical differ­entiation of smooth functions, J. Approx. Theory 18(1976)189-204.

1 Micchelli, C.A., Best L -approximation by weak Chebyshev systems and the uniqueness of interpolating perfect splines, to appear in J. Approx. Theory.

Page 62: Optimal Estimation in Approximation Theory

A SURVEY OF OPTIMAL RECOVERY 53

27. Micchelli, C.A., Optimal estimation of linear functionals, IBM Research Report 5729 (1975).

28. Micchelli, C.A., Optimal estimation of smooth functions from inaccurate data, in preparation.

29. Micchelli, C.A. and W. Miranker, High order search methods for finding roots, J. of Assoc. of Comp. Mach. 22(1975), 51-60.

co 30. Micchelli, C.A. and A. Pinkus, On n-widths in L , to appear

Trans. Amer. Math. Soc.

31. Micchelli, C.A. and A. Pinkus, Moment theory for weak Chebyshev systems with applications to monosplines, quadrature formulae

1 and best one-sided L -approximation by spline functions with fixed knots, to appear in SIAM J. of Math. Anal.

32. Micchelli, C.A. and A. Pinkus, Total positivity and the exact n-width of certain sets in Ll, to appear in Pacific Journal of Mathematics.

33. Micchelli, C.A. and A. Pinkus, On a best estimator for the class r M using only function values, Math. Research Center, Univ. of Wisconsin, Report 1621 (1976).

34. Micchelli, C.A., and A. Pinkus, Some problems in the approxi­mation of functions of two variables and the n-widths of inte­gral operators, to appear as Math. Research Center Report, University of Wisconsin.

35. Micchelli, C.A., T.J. Rivlin, S. Winograd, Optimal recovery of smooth functions, Numer. Math. 260 (1976), 191-200.

36. Morozov, V.A. and A.L. Grebennikov, On optimal approximation of operators, Soviet Math. Dokl. 16(1975), 1084-1088.

37. Newman, D., Numerical differentiation of smooth data, preprint.

38. Osipenko, K. Yu, Optimal interpolation of analytic functions, Mathematicheskie Zametki 12(1972), 465-476.

39. Osipenko, K. Yu, Best approximation of analytic functions from information about their values at a finite number of points, Matematischeski Zametki 19(1976), 29-40.

Page 63: Optimal Estimation in Approximation Theory

54 C. A. MICCHELLI AND T. J. RIVLIN

40. Royden, H.L., Real Analysis, MacMillan Company, New York 1963.

41. Rudin, Walter, Functional Analysis, McGraw-Hill, New York 1973.

42. Sard, A., Optimal approximation, J. Funct. Anal. 1 (1967), 222-244; Addendum 2(1968), 368-369.

43. Schultz, M.H., Complexity and differential equations, in Analytic Computational Complexity, Ed. J.F. Traub, Academic Press 1976.

44. Smolyak, S.A., On an optimal restoration of functions and functionals of them, Candidate Dissertation, Moscow State University 1965.

45. Schoenberg, I.J., The elementary cases of Landau's problems of inequalities between derivatives, Amer. Math. Monthly 80(1973), 121-158.

46. Tihomirov, V.M., Best methods of approximation and interpola­tion of differentiable functions in the space C[-1,+1], Math. USSR Sbornik, 9(1969), 275-289.

47. Weinberger, H.F., On optimal numerical solution of partial differential equations, SIAM J. Numer. Anal. 9(1972), 182-198.

48. Wiener, N., Extrapolation, Interpolation, and Smoothing of Stationary Time Series, the Technology Press of MIT and J. Wiley New York, 1950.

49. Winograd, S., Some remarks on proof techniques in analytic complexity, in Analytic Computational Complexity, Ed., J.F. Traub, Academic Press, 1976.

Page 64: Optimal Estimation in Approximation Theory

n-WIDTHS AND OPTIMAL INTERPOLATION OF TIME- AND BAND-LIMITED

FUNCTIONS

Avraham A. Mellaaan

IBM Thomas J. Watson Research Center

Yorktown Heights, New York 10598

1. INTRODUCTION

Let ~2 (R) denote the space of complex valued square inte­grable functions on the real line. Define

§•{fE~2 (R)I 1 lf(t)l 2 .dt!>E~, 1 lt<w>l 2dw!>n!}(l) It I >T lwl >a

with f the Fourier transform of f. Following Slepian [7] § may be regarded as the class of functions time-limited to (-T,T), at level ET• and bandlimited to (-a,a), at level na (only

f = 0 is strictly time- and band-limited). For this class we con­sider the n-widths and the optimal recovery of a function from its sampled values as well as the interrelation between them.

Section 2 is devoted to n-widths. It is based on the classi­cal papers of Landau, Pollak and Slepian [3,4,8] and Slepian [7]. We have attempted to make the proofs more geometrical in nature and to provide a unified framework. The main result for the n-widths is 2 2

d2( G;~2(R)) • .! (ET+na) +.! (ET-na) (2) n 2 l-IA'" 2 l+IA'"

n n

where A are decreasing numbers dependent only on aT. This n

result can be put to two uses of particular interest. By the definition of the zero-width aa the radius of § it follows that

for all f E §, llfll 2 s (E~+n! + 2~ ET na)/(1-A0). This

55

Page 65: Optimal Estimation in Approximation Theory

56 AVRAHAM A. MELKMAN

inequality may then be turned around to provide, for any f, a relationship between the fraction of its norm outside (-T,T) and outside (-a,a), the uncertainty relationship of Landau and Pollak [3]. Another use of formula (2) rests upon an examination of A

2 2 2 n for aT large. This shows that for any E > ET + n0 and o > 0.

that for n = (l+o)4aT d2 < E2 n

aT can be chosen large enough so 2 2 while for n = (l-o)4aT d > E • Slepian [7], in rephrasing n

Landau and Pollak's [4] dimension theorem, interpreted this fact to mean that the approximate dimension of the set of time- and band-limited signals is asymptotically 4aT as a or T becomes large.

In section 3 we turn to the optimal recovery of a function n from its values at a fixed sampling set {si}l. To make the

problem meaningful n =0 is required which turns ~ into an a ellipsoidal set. Consequently it follows from general arguments, e.g., Micchelli and Rivlin in these Proceedings, that the optimal

n procedure is to interpolate with the set of functions {K(t,si)}1 , 2 K(t,t') a suitable kernel (if the ET bound is replaced by

£:1f(t)l 2 dt $ 1 the resulting kernel is simply

sin2~a(t-t')/~(t-t')). Since the procedure is a particular kind of linear approximation the worst error that can be encountered in recovery of the class ~cannot be less than then-width d •

n A natural question is therefore whether this bound can be ac-hieved. This is indeed the case, the corresponding points being the zeros of the "worst" function for the n-width problem. Thus as a fringe benefit this approach identifies an (additional) op­timal approximating subspace, which moreover achieves the n-width by the simple device of interpolation. Similar results have been described by Melkman and Micchelli [6] for totally positive ker­nels. The present result shows that the condition of total posi­tivity is not necessary.

2. n-WIDTHS

We briefly review some basics of n-widths; for more details consult Lorentz [5]. The n-width of a subset A of a normed linear space X is obtained by considering its distance from an arbitrary n-dimensional subspace X c X, and then looking for the

* n optimal approximating subspace X which achieves the least dis­n

tance, the n-width with respect to X,

Page 66: Optimal Estimation in Approximation Theory

TIME- AND BAND-LIMITED FUNCTIONS 57

d (A;X) = inf sup infll f-sll n X eX f€A .9'€X

n n

(3)

When X is a Hilbert space A is for which n-widths results are known the present set ~ is not ellipsoidal sider a simple example of this kind:

typically an ellipsoidal set in great generality. Although it is illuminating to con-

3 I 2 X= R and A= {(x0,x1 ,x2) L xi/Ai i=o

!> 1}

with AO ~ Al ~ A2 > O. Then the zero width is the radius of the

smallest ball containing J(, d0 (A;R3) = ~· For the 1-width one

has to look for the optimal line through the origin. The major

axis is clearly such a line, d1 (A;R3) =~,but any other line

in the x0-x2 plane with slope not exceeding I(A1-A2)/(A0-A1) is

optimal too.

The same line of reasoning shows that non-uniqueness of the optimal subspace prevails for general ellipsoidal sets, e.g., Kar­lovitz [1], and even in the case at hand which is not quite ellip­soidal. But more on that in the next section.

The above example intimates that a convenient representation for ~, defined in (1), will greatly facilitate the calculation of its n-widths. To that end let us introduce the following notation.

Denote by ~the space of functions f €~2 (R) vanishing out­side ( -T, T) (strictly time-limited), by fill those with Fourier transform vanishing outside (-cr,cr) (strictly band-limited). Let D and B be the projections on these spaces, i.e.,

{f(t) ltl !> T

Df(t) - 0 It! > T

Bf(t) • 1~ f(t') sin 2ncr(t-t') dt' -~ n(t-t')

(4)

(5)

corresponding to time-limiting and band-limiting. Landau and Pollak [3] made the crucial observation that the minimum angle formed between ~ and {ill is non-zero, meaning

Re(d b) sup lldll f lb II < 1

d dlJ b€{i//

with the usual inner product

Page 67: Optimal Estimation in Approximation Theory

58 AVRAHAM A. MELKMAN

co (d,b) =£co d(t) b(t)dt, (f,f) • llfll 2•

Using this property they prove

Lemma 1.

The space qj + (jlJ is closed. Consequently any f E .¥'2 (R) may be decomposed as

f = d+b+g with d € qj, b E (jlJ , Dg • Bg • 0.

There is no convenient representation for the multitude of functions g with Dg • Bg • 0, but fortunately they play no essential role here. The spaces qj and (jlJ on the other hand possess bases with remarkable properties as shown by Slepian and Pollak [8], Landau [2].

Let Wi and Ai , i • 0,1,... be the eigenfunctions and

eigenvalues of the integral equation

IT sin 2~cr(t-t') w(t') • Aw(t) -T ~(t-t')

(6)

Noting that it has a completely continuous, positive definite, symmetric kernel the Ai may be assumed nonnegative decreasing

to zero and the wi real and normalized (wi,wj) = oij" Denoting

~i<t> = _!_ nwi<t> (7)

~ the integral equation implies

B~ i = ~ W i ' ( ~ i '~ j) • 0 ij •

Clearly wi € (jlJ' ~i E qj. Out of the many other properties of these functions to be found in the cited references we will find the following particularly useful.

1. The set {wi}; is complete in (jlJ, the set {~i}: in qj •

2. Wi has exactly i simple zeros in (-T,T).

lim Aj • 0. j-+<»

4. The eigenvalues depend only on the product aT,

A[4crT]+l(aT) ~ •5 • A[4crT]-l (aT)~ •5 '

Page 68: Optimal Estimation in Approximation Theory

TIME- AND BAND-LIMITED FUNCTIONS

lim An (crT) "' {~ crT -+co

n • [ (l+n)4aT] n • [(l-n)4crT]

59

5. Of all functions f e ~ orthogonal to ~0 , ••• ,~n the one

most concentrated in (-T,T) is ~n+l' i.e., this function achieves

llnf 11 2 sup • A

f E {jJ II f 11 2 n+l (f,~i)=O i=o, ••• ,n

The last property provides the rationale behind the ~i

it leads directly to the integral equation BDf = Af, i.e., (6). It may be worth mentioning that the ~i are also (regular) solu-

tions of a singular Sturm-Liouville type differential equation, the prolate spheroidal wave equation, so that much is known about them.

Property 1 and Lemma 1 immediately provide the sought after representation for § , which is then used to calculate its n-widths.

Lemma 2.

Any function f e ~2 (R) has the representation co

f(t) = E (di'i(t) + bi~i(t)) + g(t) i•o

where Dg ..

II (1-D)fll 2

Bg .. 2

:$; ET

0. If in addition

, I 1<1-B)fl 12:$; n2 0'

f e §,i.e.,

then co

E lbil2(1->.i) i=o

Theorem 1. 2 2

1 (eT+ncr) 1 (eT-ncr) - - + - _;;;;..__...;....._

2 1-IA 2 1+/A d (§;~2 (R))

n n n

-an_d_an--=-y_se_t_o_f_th_e_f_o_rm_ { o ' (t) + ~ ~ {t) }n-1 ii ii 0,

Re oi~i ~ 0, spans an optimal subspace.

(8)

(9)

(10)

(11)

Page 69: Optimal Estimation in Approximation Theory

60 AVRAHAM A. MELKMAN

Proof.

We show first that the right hand side is an upper bound by calculating the distance between .!/ and the particular space spanned by

gi(t). oi~i(t) + ai ~i(t).

Let P f denote the projection of n

A short calculation using (~i'~j) • shows

(f,gi) 2 I ldi~i + bi~i - (gi,gi> gil I

n-1 f given by (8) on {gi}o •

(~it~ j) - cS ij t (~it~ j) .. ~ cS ij

2 2 16idi+iibil 2 + 2~ (jbil 2+jdij 2)Reoiii .. [ jbij +jdij - 2 2 - ](1-Ai)

!oil +!ail + 2~Re oiai

since Re oiii ~ 0. Hence

jj£-P fjj 2 n

n-1 m

~ E ( jbij 2+jdi j2)(1-Ai) + E [ jbi j2+jdi 12 + 2~ Re di hi ]+jj gjj 2 i=o i=n ~ ~ ~

E jbij 2(1-Ai)+jjgjj 2+ E jdij 2 (1-Ai)+jjgjj 2+2~ E jbidij (1-Ai) i=o i•o i=o

~ 1-A n

ET2 + n2 + 2/A ETn ~ a n a

1-A n

making use of (~i'g) • (~i,g) • 0, 0 < ).i < ).n < 1 i ~ n and

Cauchy-Schwarz. Examination of the inequalities reveals that

equality is attained only for h • (n ~ + ET ~ )/11-~ • n an n n

In order to show that no n-dtmensional subspace can improve this bound we use the standard technique of finding an n+l-dtmen­sional ball of this radius contained in .!/ , i.e., a set {hi}~ such that

Page 70: Optimal Estimation in Approximation Theory

TIME- AND BAND-LIMITED FUNCTIONS 61

n 2 2 2 I I E aihil I ~ (ET + n + 2~ ETn )/(1-A ) implies i:=o cr n cr n

n E aihi E ~. Indeed, the distance of this ball from an arbitrary

i:=o n-dimensional subspace equals its radius, for there always is a function on the ball which is orthogonal to the subspace and hence its best approximation is zero. Taking

hi ,.. (ncr~i + ET ~i)/11-Ai

one has II (1-D)hi II .. ET' II (1-B)hi II = n0 and hence

n n 2 • E aihi E ~if E I ai I ::;; 1. That the ball is indeed contained in 1'"'0 i=o

~ follows therefore

Remark. From the proof it follows that h = (n ~ + ET ~ )/11-A n cr n n n is the unique function with the largest norm of all those in ~

n-1 orthogonal to {hi}o • In this sense the system {hi} is the

most natural one to approximate with, though from the n-width point of view one may choose the approximating subspace independently from ET and n0 •

In this theorem tris approximated on the whole real line. If one is interested in obtaining an approximation only on (-T,T)

then ~ should be considered as a subset of ~2 (-T,T) instead of

~2 (R). This is done by Slepian [7] who, using variational argu­ments derived the bounds

2 2 r:;- 2 A E~/(1-A ) ::;; d (~~ (-T,T)) S (YA ET+n ) /(1-A ). nT' n n n cr n

The next theorem shows that the upper bound is sharp.

Theorem 2.

1-A n

d {"' }n-l imal b an ~i 0 spans an opt su space.

(12)

Page 71: Optimal Estimation in Approximation Theory

62 AVRAHAM A. MELKMAN

Proof.

The proof of the previous theorem showed that, with P the n-1 n

projection on {~i}o

2 II (l-D)fll 2+11 (l-B)fii 2+21A" II (1-D)fll·ll (1-B)fll II f-P n f II :;; 1-). n

n

Since D~i • ~i , f-Pnf • (1-D)f+D(f-Pnf) and so

2 A II (l-D)fl 2+11 (l-B)fii 2+21AII (1-D)fll·ll (1-B)fll JID(f-Pnf) II :;; n 1-). n

n

establishing the right hand of (12) as an upper bound. The ball argument again shows that equality holds.

Two consequences of these results have particularly interesting interpretations. The first one addresses the question to what extent a signal can be concentrated in the time interval (-T,T) and simultaneously in the frequency interval (-cr,cr). This problem was solved by Landau and Pollak [3]. We give here an equivalent version which instead shows how small the fractional concentrations outside the time and band intervals, I 1<1-D)fl Ill lfl I and I I (1-B)fl Ill lfl 1. can be.

Corollary 1. For f E:$ 2 (R) let II f II = 1. Then the possible values of~T = I j(l-D)flr--and ncr= J 1<1-B)fl I fill up

the unit sQuare [O,l]x[O,l] except for the points (0,1), (1,0) and the region inside the ellipse

1.

Proof.

The zero-width of theorem 1 establishes the ellipse as a

boundary since, by definition, I If I I :;; d (~;~(R)) for any f € Jr. The boundary is attained by the ~unctions

(13)

h = (n ~ +E:T$ )/~ • The point (1,0) is not permitted since o cr o o o a strictly band-limited function cannot vanish identically in an interval. However the points (l,).i) come arbitrarily close and

are attained by (l-D)$i/ll-).i • The point (1,1) corresponds to

all functions in -t'2 (R)-(~+~). Finally all other points may be reached by the device of shifting frequency via a factor

21Tict e This leaves E:T intact and increases ncr to 1 as c increases.

Page 72: Optimal Estimation in Approximation Theory

TIME- AND BAND-LIMITED FUNCTIONS 63

The other application we want to bring concerns the notion that there are 4crT signals of duration 2T and bandwidth cr. It is based upon an analysis of the eigenvalues as functions of crT, as summarized in property 5. The idea is to show that 4crT func-tions suffice to approximate ~ on the real line to within the

2 2 order of magnitude of ET + ncr •

Corollary 2. Let N(ET,ncr;E) be the least integer such that

~(~;~2 (R)) $ E • Then

a.

b.

2 221": Landau and Pollak [4]: forE a2(ET+ncr+Y2ETncr)

[4crT]-l $ N(ET,ncr;E) $ [4crT]+l

Slepian [7]: chosen large

(l-o)[4crT]

for anv enough so

2 2 2 E > ET + n and o > 0 that cr

$ N(ET,ncr; E) $ (l+o)[4crT]

3. OPTIMAL INTERPOLATION

and all crT

crT can be

Consider the problem of recovery of f E~, on the real line or (-T,T), from the knowledge of its values at a fixed set of

sampling points {si}~ • Since point evaluations have no meaning

in the context of ~ we confine attention to the subset ~ for which ncr • 0

(14)

Thus ~ is a subset of {i8 the Paley Wiener class of entire

functions of exponential type $ 2~cr • For comparison we shall A 2

also consider the recovery problem for f E {i8 • { f E {/81 II f II $ 1}.

Taking the latter problem first, recall that {i8 has the repro­ducing kernel

sin 2~cr(t-t') K(t,t') = ~(t-t') (15)

i.e., f(t) = (K(t, •) ,f) for all f E {i8 (this may also be deduced from (5) and Bf • f). For this situation an optimal scheme for recovery on (-T,T), or even pointwise, is well known, see, e.g., the survey of Micchelli and Rivlin, and consists of

projecting f on {K(t,si)}~. Indeed, projection obviously de­

creases the norm and moreover it uses only the available informa­tion because the orthogonality conditions imply it is equivalent to interpolation

Page 73: Optimal Estimation in Approximation Theory

AVRAHAM A. MELKMAN

Hence the worst error that can be encountered as f ranges over ~ occurs for zero data,

A

E( {i/;S) • max A llnfll fE{i/ f(si)co i•l, ••• ,n

The procedure of projection-interpolation with the functions sin2~cr(t-si)/~(t-si) is of course widely used in practice. How-

ever this procedure is not optimal for ~,for in that case we 0

should use the kernel K (t, t') which reproduces {iJ with respect 0 to the inner product

<f,g> = (f,(l-D)g) (16)

This kernel can be defined implicitly by the integral equation

~(t,t') = K(t,t') + £i K(t,s) ~(s,t')ds (17)

or explicitly by its expansion

(18)

Proceeding as before one finds that an optimal recovery pro­

cedure is projection with respect to <•,•> on {~(t,si)}~

which again is equivalent to interpolation by this set. The error in recovery of f E ~ on (-T,T) is therefore

0

E( /7_ ;S) = 0

max llnfll • ETE(fi;S)/(1-E(~;S)) f E §

0

f(si)•O,i=l, ••• ,n

(19)

Since the optimal recovery scheme involves linear approximation from a particular n-dimensional subspace, the inequality

E( ~;S) ~ d ( /7_; -P2 (-T,T)) o n o ~

is immediate. of the points

A

We want to show next that with a propitious choice equality can be achieved. A similar result will hold

for ~ • The following fact will be needed.

Page 74: Optimal Estimation in Approximation Theory

TIME- AND BAND-LIMITED FUNCTIONS

Lemma 3. A

Let f*E~ achieve E(~;S) or E( ~;S). Then the only 0

zeros of f* in [-T,T] are those si lying in the interval.

Proof.

Suppose f* does have an additional zero, s 0 , in [-T,T]. Consider the function

2 T -ts 0

fl(t) = f*(t) T(t-s) • 0

65

2 f 1 is an entire function of exponential type 2Ticr and f 1 E ~ (R)

because the same is true for f* and f*(s ) = 0. Hence by the 0

Paley-Wiener theorem £1 E ~. However 1£1 (t) I ~ lf*(t) I for

ltl ~ T, lf1 (t)l ~ lf*(t)l for ltl ~ T and therefore

IIDflll IIDf*ll I I (1-D)fll I > I 1<1-D)f*l I

Thus f* cannot possibly attain E( ~;S).

Theorem 3.

E( § ;S) 0

nor, from (19),

Let the (simple) zeros of ~n(t) in (-T,T) be = = {~i}~ •

Then the recovery error of the optimal scheme based on these points equals the n-width

E( ~; => = fA n

E( d ·=) • E fA /(1-). ) o' T n n

Thus is an o_ptimal sampling set.

Proof.

(20)

(21)

We prove (20) since (21) follows from it via (19). What is

needed is an explicit evaluation of max I Inti 1211 If I 12 , the maximum being taken over the space f E~ , f(~i) = 0 i=l, ••• ,n.

This is most easily done by considering the reproducing kernel for this space.

Page 75: Optimal Estimation in Approximation Theory

66 AVRAHAM A. MELKMAN

where the numerator and denoadnator are appropriatt Fredholm determinants of K(t,t'). With this notation E(~;E) is the

largest eigenvalue of the integral operator H • KDnKD with kernel

H(s,t) = £i Kn(s,t')Kn(t',t)dt'.

The eigenfunction associated with the top eigenvalue is the "worst" function for recovery. Clearly 111 (t) is an eigen­n function of this operator, with eigenvalue A , since

n (Kn(t,•),w ) • 111 (t), (Kn(t,•),Dw ) = A 111 (t), by the choice of E. n n n n n The question is whether it is the largest eigenvalue. Any eigen­function associated with a different eigenvalue is orthogonal to 111 (t) on (-T,T). Moreover, all eigenfunctions must vanish at the n ~i" Since these are the only zeros of 111n in (-T,T) the ortho-

gonality implies that all other eigenfunctions (which may be taken real) must vanish in addition somewhere else in (-T,T). Lemma 3 now finishes the proof of (20).

,. 2 That d (~;~ (-T,T)) • ~ follows by a standard argument, n n ,.

cf. Lorentz [5], once it is observed that ~ is an ellipsoidal

set with respect to~(-T,T). Indeed, while any f E~(-T,T) may be expanded as

... f(t) ~ L ci~i(t)

i•o

the condition ,. ... 2

f E ~ is equivalent to L lcil /Ai s 1. i•o

Corollary 3. The subspace spanned by n {DK0 (t,~i)}1 is an

optimal n-dimensional approximating subspace for X> with ~· 2 2 respect to both ~ (-T,T) and~ (_..., ... ), in addition to the sub-

spaces given in theorems 1, 2.

Page 76: Optimal Estimation in Approximation Theory

TIME- AND BAND-LIMITED FUNCTIONS

Proof.

Optimality for ~2 (-T,T) is inherent in theorem 3. From it the optimality on the whole real line follows since, denoting by

Pnf the interpolant of f by {DK0 (t 1 ~i)}~

which equals d ( ~:~2 (-~,~)), by theorem 1. n o

Remark. We have been informed that Logan (unpublished) obtained a similar theorem on the basis of a formula, valid for all f € {ZJ, expressing

{lf<~>l 2 l where

II Df 11 2 - A II f 11 2 as a weighted sum of n {~} are the zeros of ~n(t) in (~,~) and

the weights are positive for 1~1 < T and negative otherwise.

67

The above proof has been patterned after Melkman and Micchelli [6]. There we considered sets of the form f • Kh, I lhl I s 1 with K a totally positive kernel, and used the total positivity to arrive at similar results. Though total positivity.does not hold here, e.g., sin2~a(t-t')/~(t-t 1 ) is not TP for all t,t', an important consequence of it, the zero properties of the eigen­functions,continues to hold. This, together with lemma 3, was the crux of the proof. An open problem is what the most general con­ditions are under which a conclusion such as theorem will be valid. The following example demonstrates that the zero properties in themselves are neither sufficient nor necessary.

2 Let {fi}O be real orthonormal functions, A0 ~ Al ~ A2 > 0, and consider the set of functions

2 f(t) • r aifi(t)

i•o

An optimal recovery scheme when given f(~) consists of inter­polation by

2 K(t,~) • r Aifi(t)f1 (~)

i•o

and the largest error E encountered occurs for zero data. Thus, taking ~ to be a zero of f 1 , one has to find the maximum of

2 2 r a1 under the conditions r a1/A1 • 1, a0f0 (~) + a2f2 (~) • 0.

Page 77: Optimal Estimation in Approximation Theory

68 AVRAHAM A. MELKMAN

This condition can be made to hold, by choosing Ai' when ~ is

not the only zero of f 1 , or made to fail when it is.

References

1. L. A. Karlovitz, Remarks on variational characterization of eigenvalues and n-width problems, J. Math. Anal. Appl. 53 (1976), 99-110.

2. H. J. Landau, The eigenvalue behavior of certain convolution equations, Trans. Am. Math. Soc. 115 (1965), 242-256.

3. H. J. Landau and H. 0. Pollak, Prolate spherical wave functions, Fourier analysis and uncertainty. II, Bell System Tech. J. 40 (1961), 65-84.

4. , Prolate spheroidal wave func-tions. Fourier analysis and uncertainty. III, Bell System Tech. J. 41 (1962), 1295-1336.

5. G. G. Lorentz, Approximation of functions. Holt, Rinehart and Winston, 1966.

6. A. A. Melkman and C. A. Micchelli, On non-uniqueness of opti­

mal sUbspaces for L2 n-width, IBM Research Report, RC 6113 (1976).

7. D. Slepian, On bandwidth, Proc. IEEE, 64 (1976), 292-300.

8. D. Slepian and H. o. Pollak, Prolate spheroidal wavefunctions, Fourier analysis and uncertainty, I, Bell System Tech. J. 40 (1961), 43-64.

Page 78: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY

Carl de Boor

Mathematics Research Center, U. Wisconsin-Madison

610 Walnut St., Madison, WI 53706 USA

1. INTRODUCTION

This paper offers a Fortran program for the calculation of the optimal recovery scheme of Micchelli, Rivlin and Winograd [15]. A short derivation of that recovery scheme is given first, in order to make the paper selfcontained and also to provide an alternative to the original derivation in [15]. For the latter reason, a derivation of the related envelope construction of Gaffney and Powell [8] is also given. From a computational point of view, these schemes are special cases of the following computa­tional problem: to construct an extension with prescribed norm of a linear functional on some finite dimensional linear subspace to all of 1L 1 [a,b].

2. THE OPTIMAL RECOVERY SCHEME OF MICCHELLI, RIVLIN AND WINOGRAD

This scheme concerns the recovery of functions from partial n

information about them. Let n > k and lett T := (<i)l be non-

decreasing, in some interval [a,b], with 'i < 'i+k' all i. For f E IL(k) := {g E C(k-l)[a,b] : g(k-l) abs. cont., g(k) E lL 00}, de~ote by f,, the restriction of f to the data point

sequence T, i.e., f,, := (fi)~ with

69

'· 1-m

Page 79: Optimal Estimation in Approximation Theory

70 CARL DE BOOR

We call a map S : 1L (k) + 1L a recovery scheme (on L (k)) ® 00 00

with respect toT provided Sf depends only on f'T' i.e.,

provided fjT = g'T implies that Sf = Sg. The (possibly infinite) constant

consts := sup{ II f - Sf II I II f(k) II : f E :n. (k)}

measures the extent to which such a recovery scheme S may fail to recover some f since it provides the sharp error bound

(1) II f - Sf II ~ consts II f(k) II , all f E E. (k) • 00

Here and below,

II &II := ess.sup{ lg(t) I a < t < b}

A recovery scheme S is optimal if its consts is as small as possible, i.e., if the worst possible error is as small as possible, as measured by (1). We write

const := inf consts T S

for the best possible constant. Here is ~ quick lower bound for that constant. We have

consts = sup f

>

>

sup II g- Sf II I II g(k) II g1T=f1T

sup II g - so II 1 II g Ck> II g1T=O

sup max{ II g - so II, II -g - so II}/ II g (k) II giT=O

sup II g lit II g <k> II giT=O

for every recovery scheme S, hence

(2) const > c(T) T-

Page 80: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY

But, actually, equality holds here. For the proof, we need the notion of perfect splines.

A perfect spline s ~l < < ~r in [a,b]

r s(x) P(x) + c E

i=O

of degree k with (simple) knots is any function s of the form

k-1 - y)+ dy

71

with P E 1Pk := polynomials of degree < k and c some constant.

In other words, such a function s is any k-th anti-derivative of an absolutely constant function with r sign chan~es or jumps in [a,b]. Their connection with the present topic stems from

Proposition 1 (Karlin [10]). If s is a perfect spline of degree k with n - k knots then

b,

and

r =

II s (k) II = inf{ II f (k) II : f E lL !k) ' fIT = s IT} .

This proposition follows directly from the following lemma.

Lemma 1. If hE ~!k) is such that. for a= ~O < ••• < ~r+l

cr.h(k) 1

hiT = n - k,

> 0 a. e. on (~i' ~i+l) for some cri E {-1,1}, i=O, ••• ,r ,

0.

then

then r > n - k. Further. if r < n - k, hence

(3) Ti < ~i < 'i+k' all i •

Assuming this lemma for the moment, we see that if, in

Proposition 1, we had II s (k) II > II f (k) II for some f E ~ !k) with

fiT = siT• then h := s - f would satisfy the hypotheses of the

lemma for some r < n - k, which would contradict the conclusion of the lemma.

Proof of Lemma 1: By Rolle's theorem, there must be points

tl < < tn+l-k

at which h(k-l) vanishes, and for which

Page 81: Optimal Estimation in Approximation Theory

72 CARL DE BOOR

(4) all i 'i 2 ti 2 'i+k-1,

Then, each interval [~ .• ~.+1 ] contains at most one of the tJ. 1 1 (k-1)

since, on each such interval, h is strictly monotone, by assumption. Therefore, if n. denotes the number of such

J intervals to which tj belongs, then

n + 1 - k < ~ nj 2 r + 1 ,

or n - k < r. This also shows that, if n - k r, then n. 1, all j, and so

~i-l < ti < ~i' all i (except that, possibly, ~O t 1 ,

tn+l-k = ~r+l)

which, combined with (4) then implies (3). QED.

We also need

J

Theorem 1 (Karlin [10]). If f E lL (k) 00 , then there exists a

perfect spline of degree k with < n - k knots which agrees with f at '·

A simple proof can be found in [2].

We are now fully equipped for the attack on the optimal recovery scheme. By Karlin's theorem, the set

Q := {q : q 1"

is a perfect spline of degree k with < n - k knots, q,, 0}

is not empty. By Lemma 1, each q E Q \{0} T

has, in fact, exactly

n - k knots and does not vanish off have n + 1 zeros yet only n - k

derivative, contrary to the lemma).

T (since otherwise it would sign changes in its k-th

Therefore, if f E lL (k) and 00

f,, = 0 and x' ,,then (f(x)/q(x))q is a well defined perfect

spline of degree k with n - k knots which agrees with f at the n points of 1" and the additional point x, hence, Proposition 1 implies that

cs> llf<k> II~ lf<x>lq<x> Ill q<k> II

It follows that

Page 82: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY 73

(6) sup{ If (x) I I llf (k) II : f E 11 (k), f I = 0} = I q (x) I I II q (k) II • oo T

Since the left side is independent of q, this shows that Q, is A

the span of just one function, say of q, normalized to have

It also shows that

(7) c(T) = II ~~~

This gives equality const

T

a means of computing c(T), but not quite yet the = c(T) nor the optimal recovery scheme. For this,

let ~l'"""'~n-k be the n - k knots of q and consider

$ :=splines of order k with simple knots ~1 , ... ,~n-k

1P C(k-2) := k,~ n

(k-2) I := {f E C : f (~i'~i+l) E Fk, all i}

Theorem 2 (Micchelli, Rivlin & Winograd [15]). The rule

<ff) fIT and .6.

$, f E 11 (k) Sf E all T 00

A (read "ess crown(ed)") on 11 (k) defines a map s which is

00

linear and an optimal recovery scheme with respect to T.

Proof. First, we prove that f is well defined. Since, by = the lemma, T. < ~. < '·+k' we ask for matching f(k-l) at some 1 1 1

. t 1 h h .I 1:" • h (k-l) ( ) k po1n Ti on y wen sue T. ~ ~. 1.e., wen s T. ma es 1 1

sense for each s E $. Secondly, dim $ = k + number of polynomial

pieces - 1 = n, therefore f is well defined if and only if

0 and f E $ implies f = 0 .

For this, we would like to use inequality (5), but, since $ is

not contained in 11 (k) but only in the larger space 00

11 (k) 00' ~

n-k := 11 (k-1) n

00

Page 83: Optimal Estimation in Approximation Theory

74 CARL DE BOOR

we must first extend (5) to such spaces, with the convention that

for f E U. (k) co,E; •

This requires the following slight strengthening of Lemma 1.

Lemma 1'. If h e n.. (k) for some a = z;; 0 < ••• < z;;r+l b and,for=80me a e-{-1,1}, "",z;;

a.e. on i o, ... ,r'

then hi = 0 implies n - k < r. -- 1:

Here, we mean by h(k-l)(x) = 0 that h(k-l)(x-)h(k-l)(x+) ~ 0, in case the number x occurs k-fold in 1:.

Proof. Rolle's theorem gives again n + 1 - k distinct -- (k-1)

zeros of h which again consists of r + 1 strictly monotone pieces, but may fail to be continuous across the points z;;1 , ••• ,z;; •

But, since h(k) alternates in sign, this latter fault can easil; be remedied by appropriate local linear interpolation across

a small neighborhood of each discontinuity z;;i without disturbing

the other two properties and now n - k < r follows as before; Q.E.D.

This gives

Proposition 1'. If s is a perfect spline with knots, say with knots z;;1 < ••• < z;;r where r < n- k,

< n - k and of

Page 84: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY 75

degree k, then even

II s (k) II = inf { II f (k) II : f E n. ~~~, fIT = sIT} •

II qA{k) II-- 1, From this we conclude, with that

<5' > II f <k> II :: I f <x> 1 c1 <x> I for X { T, f E 1L (k)

oo,E;;' fl, = 0 •

But this implies (8) since Ji

II f<k> II = o for f E $, and so shows

that S is well defined. Finally, (5') also implies that, for f E 1L (k)

00 '

II f <k> II = II <f - ~f> <k> II > If <x> - ef <x> I 1 I c! <x> I

or

(9) lf(x) - tf(x) I ~ lq(x) Ill f(k) II showing, with (2) and (7), that t is an optimal recovery scheme. This proves Theorem 2.

MRW actually insist that a recovery scheme S WSP into lL~k), hence they are not quite done at this point, since ~ only maps

into n.~k-l). But, since $ can be viewed as spline functions

of degree k with double knots at the E;;i' we can produce an

element of n. (k) arbitrarily close to ~f merely by pulling all 00

these double knots apart ever so slightly. This shows that inf const8 = c(T) even if the inf is restricted to S mapping into n.~k) but now the inf is not attained apparently for k > 1.

3. THE ENVELOPE CONSTRUCTION

The preceding discussion allows a simple derivation of sharp

estimates for the value of a function f E n. (k) at some point x, 00

given the vector fj, and a bound L on its k-th derivative on [a,b], as follows.

with

We are to construct the set

I := {f{x) : f E F} X

Page 85: Optimal Estimation in Approximation Theory

76 CARL DE BOOR

for some given a. and L. closed interval,

If F is not empty, then I is a X

I = • [a , b ] X X X

say, since F is closed, convex and bounded and [x] : f ~ f(x)

is a continuous linear functional on n. (k). Assume that 00

F0 : = { f E F : II f (k) II < L} "#. t/J

Then

(10) (a ,b ) = [x]F0 • X X

Karlin's theorem then implies that, for x i T,

Q := {q : q is perfect spline of degree k with < n - k x knots, qiT = a., q(x) = ax}

is not empty. Further, Qx ~ F, since, by definition of ax'

there exists f E F with f(x) = ax while each q E Qx agrees

with such an f at the n + k + 1 points T and x, therefore

II q (k) II ~ II f (k) II ~ L' by the proposition. On the other hand, 0 0

Qx n F = t/J, since, if q E Qx n F , then a = q(x) E (a ,b ), X X X

by (10), a contradiction.

It follows that

for all y, a < q(y) < b y- - y

0 But, for any g E F , h := g - q has < n - k sign changes in its k-th derivative and vanishes at T, therefore q has exactly n - k knots and h does not vanish off T, by Lemma 1. Hence, if Ti+1 , •.• ,Ti+j are all the points of T between x and y iT,

then

(-)j(g(y)- q(y)) > 0 .

This shows that

q(y) ~{_:~} if the interval between x andy contains au{_:::~

Page 86: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY 77

number of points of T. It follows that Qx contains exactly one

function and that this function supplies half of the entire boundary of

(11) {(x,f(x)) : x E [a,b], f E F} ,

the other half being supplied by the perfect spline p of degree k with n-k knots for which piT = a and p(x) = bx.

We note the curious corollary that the perfect spline s of Karlin's theorem is unique if there exists g which agrees with

fat all points of T but one and for which II g(k) II< II s(k) II, i.e., if at least one of the interpolation constraints is active. Put differently, it says that if there are two different perfect srlines s, s of degree k with < n - k knots for which S = Sl =a and which agree at some X~ T, then

T T

lls(k)ll= lls(k)ll= La:= min{llf(k)ll: f E lL!k), fiT a}.

Let now q be the half of the envelope of (11) with

q(k)(O+) = L, and let p be the one with p(k)(O+) = -L. Gaffney and Powell [8] choose

as a good interpolant, its graph being clearly the center of (11). Since p and q are uniquely defined for L > L by the requirement a that they are perfect splines of degree k with n - k knots, equal to a at T and to a , resp. b at x, they are necessarily continuous

X X functions of L and a in that range. In particular, with q =:qL , ,a we have qL,a Lql,a/L and ql,a/L + ql,O = q as L + oo, and,

similarly, PL,a/L + -q as L + 00 • This shows that, with

u1 < < un-k the knots of q and v1 < ... < vn-k the knots of p,

lim u. = lim vi = ~i' i L+oo ~ L+oo

l, ... ,n- k .

In particular, for large enough L, the sequence

with

Page 87: Optimal Estimation in Approximation Theory

78 CARL DE BOOR

is nondecreasing and

(S ct) (k) c 0 L = ±2L

on

on

hence

II (SLct) (k) Ill ~ 2LII u- vlll----+0 . L-+<»

This shows that SLCL converges to an element of $, while

SLCLIT = CL for all L

therefore, SLCL converges to the optimal interpolant for the data.

It was in this way that Gaffney and Powell [8] constructed, quite independently from Micchelli, Rivlin and~Winograd [15], the same optimal recovery or interpolation scheme ~.

The problem of constructing the set I was posed originally X

in the basic paper by Golomb and Weinberger [9], although they gave detailed attention to such problems oniy when the (semi)norm involved comes from an inner product. Micchelli and Miranker [14] solved the problem posed at the beginning of this section in the sense that they correctly described the boundary of (11) as being given by just two perfect splines of degree k, each with n - k knots, and with their k-th derivative equal to L in absolute value. In fact, Micchelli and Miranker consider the slightly more general

situation where f(k) is only known to map [a,b] into some interval [m,M]. They state that these matters could be proved along the lines used by Burchard [6] to solve a related restricted moment problem and refer specifically to Karlin and Studden [11, VIII, §3] for requisite facts concerning principal representations of interior points of moment spaces. Of course, these facts go back to Krein [12]. Quite independently, Gaffney and Powell [8] also solved this problem, with the proofs in Gaffney's thesis based on Chebyshev type inequalities as found in Karlin and Studden [11, VIII, §8] and adapted by him to weak Chebyshev systems.

4. THE CONSTRUCTION OF NORM PRESERVING EXTENSIONS TO ALL OF IL 1 4

Both the optimal recovery scheme S of Section 2 and the envelope of Section 3 require the construction of an absolutely constant function h with no more than a specified number of jumps

Page 88: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY 79

which provides an integral representation of a linear functional given on a linear space of spline functions.

In the optimal recovery, we are to construct a perfect spline h

s of degree k with n - k knots and with II 9<k>11 = 1 which vanishes

at the given n-point sequence T. n Let (Mi)l be the sequence of

B-splines of order k with knot sequence T, each normalized to have unit integral, i.e.,

(12) M. (x) := M. k (x) 1 1, , T

with [Ti, .•• ,Ti+k]f the k-th divided difference of the function

fat the points Ti, ... ,Ti+k" Then, from Taylor's expansion with

integral remainder,

_ Ti+k (k) I (k) [T., ... ,T.+k]f- J -11. k (x)f (x)dx/k. for all fElL .

1 1 1, ,T oo

Ti

The points ~ = (~i)~-k are therefore characterized by the require­

ment that the function

n-k h~(x) := sign r-f (x - ~i)

i=l ±s(k)(x)

be orthogonal to each of then- k functions M1 , ... ,Mn-k"

Before considering the computational details of determining ~ from this orthogonality condition, I want to comment on the fact that this is a problem of representing or extending a linear functional on some subspace of IL 1 and is therefore closely related

to the problem of computing the norm of a linear functional on some subspace of IL 1 . This is also explored in a forthcoming paper by Micchelli [ 13] •

Indeed, if Tis a linear subspace of ~l of dimension n + 1 - k, and A is a linear functional on T, we might ask for ~

and a so that

Ag for all g E T •

But then, in particular, h~ is orthogonal to ker A, a subspace of

dimension n - k. Conversely, if we have already found h~ ortho­gonal to ker A then there will be exactly one a so that ali~

Page 89: Optimal Estimation in Approximation Theory

80 CARL DE BOOR

represents A on Tin the sense of (13), unless h~ is even ortho­

gonal to all of T. But this latter event cannot happen in case T is weak Chebyshev since h~ has only n - k jumps.

It is clear that any such representation h~ for A produces

an upper bound for the norm of A. In fact, II A II = I al in case T is weak Chebyshev. This is actually how I became interested two years ago in the numerical construction of representers of linear functionals [3], [4]. I was interested in computing, or at least closely estimating, the norm of certain linear functionals on certain spline subspaces in IL 1 . In a way, this is a trivial

problem, viz. the maximization of a linear function over a finite dimensional compact convex set, and there was the feeling that there ought to be special methods available. Perhaps some reader can steer me towards such methods. I found, for the particular cases of concern to me in which T was always weak Chebyshev, nothing more effective for calculating II All than to construct such a representation (13).

Finally, the envelope construction corresponds to the slightly different situation where dim T = n - k, A E T' and a with lal > II All is prescribed and one seeks ~ so that again (13) holds.

We have now one less condition to satisfy but also one less para­meter to do it with.

5. CONSTRUCTION OF THE KNOTS FOR THE OPTIMAL RECOVERY SCHEME

n-k As we saw in the preceding section, the knots ~ = (~i)l

for the optimal recovery scheme are the solution to the following

problem. We are given (Ti)~ nondecreasing, with Ti < Ti+k' all i,

and n > k, and wish to construct ~O < ••• < ~r with

r: n-k+l

and with ~0 = a := Tl' ~r b := T so that n

r ~-(14) I s. J ] M. k 0, i 1, ... , n - k

j=l ] c 1

1,

]-

while also

0, j l, ... ,n-k.

Page 90: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY 81

Extend T by

T =: 'n+l =: 'n+2 =: ... =: 'n+k . n

Then one verifies easily that

X

J M. k = I Nm,k+l(x) for x > a a l., m>i

with

'm+k+l - 'm Nm,k+l := k + 1 Mm,k+l' all m .

Therefore, (14) is equivalent to

r

L S. L (N k+l(~.) - N k+l(~._1 )) = 0, i = l, ..• ,n- k, j=l J m~i m, J m, J

or, on subtracting equation i from equation i- 1 fori= 2, •.. ,n-k,

r L S.(N. k+l(~.)- Ni,k+l (~j-l)) = 0, i l, ..• ,n- k- 1

j =1 J l., J

r

L SJ. L (Nm,k+l(~j)- Nm,k+l(~J.-1)) 0 • j=l m>n-k

Using the fact that Nm,k+l(~0 ) written

n-k L (SJ.- SJ.+l)Ni,k+l(~j)

j=l

n-k

0 for m ~ 1, this can also be

l, .•. ,n-k-1

L (SJ.- SJ.+l) L Nm k+l(~.) = -Sr L Nm k+l(~r) j""l m>n-k ' J m>n-k '

Since Ni,k+l(~r) = 0 fori< n- k, while L k Nm,k+l(~r) = 1,

the right side becomes simply (0, ••• ,0,-S 1~n-r

Choose now S = -1 to make things definite. Then S. = (-)r-j-l n-k-j r J

(-) by (15), and (19) and (15) are seen to be equivalent to

(16) F(~) = 0

n-k n-k with F : lR + :R given by

Page 91: Optimal Estimation in Approximation Theory

82

i < n - k

(17) F(~)i := n

where

L am, i m=n-k

n - k

·= I {n-k

• j=l

n-k-j (-) Ni,k+l(~j),

-1/2

CARL DE BOOR

i l, ... ,n-1

i n •

We solve (16) by Newton's method. From the current guess ~. we compute a new guess ~* = ~ + o~, with o~ the solution to the linear system

(19) F' {~)o~ -F(~)

Since Ni,k+l = Mi,k - Mi+l,k' addition of equation i to equation

i-1, i = n- k, •.• ,2, in (19) produces the equivalent linear system

n-k n-1 j~l m~i (Mm,k-Mm+l,k)(~j) (-)n-k-jo~j

n L am' i=l, ••• ,n-k.

m=i

n-1 But, since L (M k

i m,

for t ~ b, this shows that (19) is equivalent to the linear system

(20) Cx = d

with n

x., J

di=( L-a )(T.+k- T.)/k, m=i m 1 1

c .. l.J

i,j = l, ••• ,n- k.

The matrix C is totally positive and (2k - 1)-banded, hence can be stored in 2k - 1 bands of length n - k each, and can be factored cheaply and reliably within these bands by Gauss elimination with­out pivoting (see [5]).

The iteration is carried out in the program SPLOPT below, starting with the initial guess

(22 ) ~i = (Ti+l + ••• + 'i+k-1)/(k 1), i l, ... ,n- k .

Page 92: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY 83

A first version of the program was equipped to carry out Modified Newton iteration: ~* is computed as the first vector in the sequence

~ + 2-ho~, h = 0,1,2, •••

for which II F(~ + 2-hoO II 2 < II F(O 11 2 • But, in all examples

tried, the initial guess (22) was apparently sufficiently close to the solution to have always h = 0, i.e.' II F(~ + oO II 2 < IIF(~) II z• In fact, the termination criterion

was usually reached in three or four clearly quadratically converg­ing iterations. For this reason, the program SPLOPT below carries out simple Newton iteration. It would be nice to prove that Newton iteration, starting from (22), necessarily converges. But, such a proof would necessarily have to be in control of the norm of the inverse of the matrix F'(~), hence in control of the norm of the inverse of C = (Ni(~j)) as a function of~. Good estimates for

II (Ni (~j)) -lll have been searched for in the past by some who were

interested in bounding (the error in) spline interpolation, but without much success. E.g., the simple conjecture that, for the

initial guess (22), II (Ni(~j))-1 11 1 .::_ constk independent of~ has

been proved so far only for k .::_ 4. SUBROUTINE SPLOPT C TAUr Nr Kr SCRTCHr Tr !FLAG >

COMPUTES THE KNOTS T FOR THE OPTIMAL RECOVERY SCHEME OF ORDER K C FOR DATA AT TAU CI>ri=l••••• N • TAU MUST BE STRICTLY INCREASING. C SEE TEXT FOR DEFINITION OF VARIABLES AND METHOD USED. C !FLAG 1 OR 2 DEPENDING ON WHETHER OR NOT T WAS CONSTRUCTED. C DIMENSION SCRTCHCCN-K>*C2*K+3>+5*K+3>r T<N+K>

DIMENSION TAUCN>rSCRTCHC1)rTC1> DATA NEWTMX,TOLRTE I 10r.000001/ NMJ< = N-K IF CNMK>

1 PRINT 601rNrK 1r56r2

601 FORMATC13H ARGUMENT N =ri4r29H IN SPLOPT IS LESS THAN K =ri3>

2 IF CK .GT. 2) PRINT 602rK

602 rORMATC13H ARGUMENT K =ri3r27H IN

3 NMKH1 = NMK - 1 FLOATK = K KPK = K+K KP1 = K+l KPKP1 = K+i<P1 KM1 = K-1 KPKM1 = K+KM1 KPN = K+N SIGNST = -1.

GO TO 999 GO TO 3

SPLOPT IS LESS THAN 3> GO TO 999

IF <NMK .GT. CNMK/2>*2> SIGNST 1.

Page 93: Optimal Estimation in Approximation Theory

84

C SCRTCH<I> = TAU-EXTENDEDCI>r I=l•••••N+K+K NX = N+KPKP1

C SCRTCHCI+NX> = XICI>ri=O•••••N-K+1 NA = NX + NHK + 1

C SCRTCH<I+NA> = -A<I>r I=l•••••N ND = NA + N

C SCRTCHCI+ND> = XCI> OR DCI)r I=l•••••N-K NV = ND + NHK

C SCRTCHCI+NV> = VNIKXCI)ri=l•••••K+l NC = NV + KP1

CARL DE BOOR

C SCRTCHCCJ-1>*<N-K)fl + NC> = CHATCirJ)ri=lr ••• ,N-KrJ=1•••••2*K-1 LENGCH = NMK*KPKM1

C EXTEND TAU TO A KNOT SEQUENCE AND STORE IN SCRTCH. DO 5 J=1rK

SCRTCHCJ> = TAUC1> 5 SCRTCHCKPN+J> = TAUCN>

DO 6 J=1rN 6 SCRTCHCK+J> = TAUCJ)

C FIRST GUESS FOR SCRTCH C.+NX> XI , SCRTCH<NX> = TAUC1) SCRTCHCNMK+1+NX> = TAU<N> DO 10 J=lrNMK

SUM = O. DO 9 L=1rKM1

9 SUM = SUM + TAUCJ+L> 10 SCRTCHCJ+NX) = SUM/KM1

C LAST ENTRY OF SCRTCH C.+NA> -A IS ALWAYS ••• SCRTCHCN+NA> = ,5

C START NEWTON ITERATION. NEWTON = 1 TOL = TOLRTE*CTAUCN> - TAUC1))/NMK

C START NEWTON STEP COMPUTE THE 2K-1 BANDS OF THE MATRIX C AND STORE IN SCRTCHC .i·NC>, C AND COMPUTE THE VECTOR SCRTCHC.+NA> ~-A,

20 DO 21 I=1rLENGCH 21 SCRTCHCI+NC> = o.

DO 22 I=2rN 22 SCRTCHCI-1+NA> = O.

SIGN = SIGNST ILEFT = KPl DO 28 J=1rNMK

XIJ = SCRTCHCJ+NX> 23 IF CXIJ .LT. SCRTCHCILEFT+1>> GO TO 25

!LEFT = ILEFT + 1 IF CILEFT .LT. KPN> GO TO 23 ILEFT = ILEFT - 1

25 CALL BSPLVNCSCRTCHrKr1rXIJriLEFTrSCRTCHC1+NV>> ID = HAXOCOriLEFT-KPK> INDEX = NC+CJ-ID+KM1>*NMK+ID LLHAX = MINOCKrNMK-ID> LLMIN = 1 - MINOCOriLEFT-KPK> DO 26 LL=LLMINrLLMAX

INDEX = INDEX - NMKM1 26 SCRTCH<INDEX> = SCRTCHCLL+N''>

CALL BSPLVNCSCRTCHrKP1r2rXIJtiLEFTrSCRTCH<1+NV>> ID = HAXOCOriLEFT-KPKP1) LLMIN = 1 - MINOCOriLEFT-KPKP1) DO 27 LL=LLMINrKP1

ID = ID + 1 27 SCRTCHCID+NA> = SCRTCHCID+NA> - SIGN*SCRTCHCLL+NV> 28 SIGN = -SIGN

Page 94: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY

CALL BM.fFAC C SCRTCH CJ+NC > rNMK, NMt\ ri\PKM1, K·r I FLAG>

44 PRINT 644 644 FORMATC32H C IN

SCRTCH c.+N£t> 46 I=Nr2r-1

SPLOPT

D

GO TO C45r44>riFLAG

IS NOT INVERTIBLE> RETURN

FROM SCRTCH C.+NA> =-A • COMPUrE 45 DO 46 SCRTCHCI-l+NA> = SCRTCHCI-1+NA> + SCRTCHCI+NA>

DO 49 I=lrNMK 49 SCRTCHCJ+ND> = SCRTCH<I+NA>*CTAUCI+K>-TAU<I>>/FLOATK

COMPUTE SCRTCH (.+ND> = X • CALL BANSUBCSCRTCHC1+NC>rNMKrNMKrKPKM1rKrSCRTCHC1+ND>>

COMPUTE SCRTCH C.+ND> =CHANGE IN XI • MODIFYr IF NECESSARY, TO C PREVENT NEW XI FROM MOVING MORE THAN 1/3 OF THE WAY TO ITS C NEIGHBORS. THEN ADD TO XI TO OBTAIN NEW XI IN SCRTCHC.+NX>.

DELMAX = O. SIGN = SIGNST DO 53 I=1rNMK

DEL = SIGN*SCRTCHCI+ND> DELMAX = AMAX1CDELMAXrABSCDEL>> IF CDEL .GT. 0.) GO TO 51 DEL = AMAX1CDELrCSCRTCHCI-1+NX>-SCRTCH<I+NX))/3.>

GO TO 52 51 DEL= AMIN1CDELrCSCRTCHCI+1+NX>-SCRTCHCI+NX))/3.> 52 SIGN = -SIGN 53 SCRTCH<I+NX> = SCRTCH<I+NX> + DEL

CALL IT A DAY IN CASE CHANGE IN XI WAS SMALL ENOUGH OR TOO MANY C STEPS WERE TAKEN.

IF <DELHAX .LT. TOL) NEWTON = NEWTON + 1 IF <NEWTON .LE. NEWTMX> PRINT 653rNEWTMX

GO TO 54

GO TO 20

85

653 FORMATC33H NO CONVERGENCE IN SPLOPT AFTERri3r14H NEWTON STEPS.) 54 DO 55 I==1rNMK 55 TCK+I> = SCRTCHCI+NX> 56 DC 57 I=lrK

57 TCI > = TAUCl) T<N-i-I> = TAU<N>

999 IFLAG = 2

END

RETURN

RETURN

Page 95: Optimal Estimation in Approximation Theory

CARL DE BOOR

The subroutine SPLOPT has input TAU(i) = Ti' i = l, ••• ,n,

assumed to be nondecreasing and to satisfy Ti < Ti+k' all i, the

integer N = n and the desired order k in K. The routine needs a work array SCRTCH, of size > (n - k) (2k + 3) + 5k + 3; n(2k + 3) is more than enough. The routine has output T(i) = t.,

1

i = l, ••• ,n + k, the knot sequence for the optimal recovery scheme, in case !FLAG = 1. For !FLAG = 2, something went wrong.

The routine uses the subroutine BSPLVN for the evaluation of all B-splines of a given order on a given knot sequence which do not vanish at a given point. This routine, and others for dealing computationally with splines and B-splines, can be found in [1]. For completeness, we also list here the subroutines BANFAC and BANSUB, used in SPLOPT for the solution of the banded system (20).

SUBROUTINE BANFAC C A, NROW• N, NDIAG• MIDDLE, IFLAG > DIMENSION A<NRnW,NDIAG> !FLAG = 1 ILO = MIDDLE - 1 IF CILO) 999r10r19

10 DO 11 I=1rN IF CACir1)) 11r999r11

11 CONTINUE RETURN

19 IHI = NDIAG - MIDDLE IF CIHI> 999,20r29

20 DO 25 I=l•N IF CACI,MIDDLE>> 21r999r21

21 JMAX = MINOCILO,N-I> IF CJHAX> 25r25r22

22 DO 23 J=1•JHAX 23 ACI+JrMIDDLE-J> ACI+JrMIDDLE-J)/ACirHIDDLE> 25 CONTINUE

RETURN 29 DO 50 I=1rN

DIAG = ACirMIDDLE> IF CDIAG> 31r999r31

31 JMAX = MINOCILO,N-I> IF CJHAX> 50r50,32

32 KMAX = MINOCIHirN-I> DO 33 J=1,JHAX

IPJ = I+J HMJ = MIDDLE-J ACIPJ,MHJ> = ACIPJrMMJ)/DIAG DO 33 K=1,KMAX

33 A<IPJ,MMJtK> = A<IPJ,MMJtK> - ACIPJrMMJ>*A<I•MIDDLE+K> 50 CONTINUE

RETURN 999 IFLAG = 2

RETURN END

Page 96: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY 87

BANFAC factors an N x N band matrix C whose NDIAG bands are con­tained in the columns of the NROW x NDIAG array A, with the MIDDLE column containing the main diagonal of C. It uses Gauss elimina­tion without pivoting and stores the factors in A.

c

SUBROUTINE BANSUB < Ar NROWr Nr NDIAGr MIDDLEr B ) DIMENSION ACNROWrNDIAG>rB<N> IF <N .EQ. 1> GO TO 21 ILO = MIDDLE - 1 IF <ILO) 21r21r11

11 DO 19 I=2rN JMAX = MINO<I-1riLO> DO 19 J=1rJMAX

19 B<I> = B<I> - B<I-J>*A<IrMIDDLE-J>

21 I = N IH~ = NDIAG-MIDDLE DO 29 II=1rN

JMAX = MINO<N-IriHI> IF <JMAX> 28r28r22

22 DO 25 J=lrJMAX 25 B<I> = B<I> - B<I+J>*A<IrMIDDLE+J> 28 B<I> = B<I>IA<IrMIDDLE> 29 I = I - 1

END

BANSUB then uses the factorization of C into the product of a lower and an upper triangular matrix computed in BANFAC to solve the equation Cx = b for given b (input in B) by forward and back substitution. The solution x is contained in B, on output.

6. CONSTRUCTION OF THE OPTIMAL INTERPOLANT

With the break points ~l < ••• < ~n-k for the optimal inter­

polant ~f determined from T in SPLOPT, it remains to compute ~f. This we propose to do by determining its B-spline coefficients.

n+k SPLOPT has produced the knot sequence ! = (ti)l , with

l, ... ,n-k,

n Let (Ni)l be the corresponding sequence of normalized B-splines

of order k, i.e.,

Ni(x) := N. k t(x) 1, ,_

Page 97: Optimal Estimation in Approximation Theory

88 CARL DE BOOR

Then, according to Curry & Schoenberg [7], every piecewise poly­nomial function of order k on [a,b] := [T1 ,Tn], with k- 2

continuous derivatives and break points ~1 , ... ,~n-k' i.e., every

spline of order k on [a,b] with knot sequence !• has a unique representation as a linear combination of the n functions N1 , ... ,Nn. Therefore

with a =

n (23) I

j=l

n I aiN.

i=l 1

n (ai)l the solution of the linear system

N.(T.)a. = f(T.), i = l, ••• ,n. J 1 J 1

In case T is strictly increasing, -the only case considered here,­Lemma 1 implies that t. < T. < t.+k' all i, which, together with

1 1 1

the fact that N. vanishes outside the interval [t.,t.+k], all j, J J J

shows that the coefficient matrix of (23) is 2k - 1 banded. Since the coefficient matrix is also totally positive, we can therefore (see [5]) solve (23) by Gauss elimination without pivoting and within the 2k - 1 bands required for the storage of the nonzero entries of the matrix. The following subroutine SPLINT generates the linear system (23), given on input the arrays TAU(i) = T., FTAU(i) = f(Ti), i = l, ••• ,N, T(i) = ti' i = l, ••• ,N + K and1 the

numbers N and K. The system is then solved, using BANFAC and BANSUB given in Section 5, and using a working array Q, of size N(2K- 1). The output consists of the B-coefficients a.= BCOEF(i), i = l, ••• ,N, in case !FLAG= 1. If !FLAG= 2, then

1

the linear system (23) was not invertible.

Page 98: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY 89

K~ Q, BCOEFr !FLAG ) c c c c c c

SUBROUTINE SPLINT ( TAU, FTAU• T• Nr SPLINT PRODUCES THE B-SPLINE COEFF.S K WITH KNOTS T <I>• !=1••••• N + K r FTAU <I> AT TAU <I>• !=1••••• N •

BCOEF OF THE SPLINE OF ORDER WHICH TAKES ON THE VALUE

c

TAU IS ASSUMED TO BE STRICTLY INCREASING. SEE TEXT FOR DESC~IPTION OF VARIABLES AND METHOD. DIMENSION T<N+K>•G<N•2*K-1>

DIMENSION TAU<N>r FTAU<N>• T(1)r Q(N,1)r BCOEF(N)

13

15

NP1 = N + 1 KPKM1 = 2*K - 1 ILEFT = K DO 30 I=1rN

TAUI = TAU<I> ILP1MX = MINO<I+KrNP1> DO 13 J=l.KPKM1

Q(J,J) = o. !LEFT = MAXO<ILEFT•I> IF <TAUI .LT. T<ILEFT>>

IF <TAUI .LT. T<ILEFT+1>> !LEFT = !LEFT + 1 IF (!LEFT .LT. ILP1MX>

ILEFT = !LEFT - 1

GO TO 998 GO TO 16

GO TO 15

IF <TAUI .G~. T(ILEFT+1>> GO TO 998 16 CALL BSPLVN ( Tr 1(, 1• TAUir ILEFTr BCOEF >

NOTE THAT BCOEF IS USED HERE FOR TEMP.STORAGE. L = !LEFT - I DO 30 J=1rK

L = L+1 30 Q(IrL> = BCOEF<J>

NP2MK = N+2-K CALL BANFAC < Q, Nr Nr KPKM1• Kr !FLAG )

40 DO 41 I=1•N 41 BCOEF<I >

CALL BANSUB

998 IFLAG = 2 999 PRINT 699

GO TO <40,999), !FLAG

FTAU<I> Q, Nr N• KPKMlr K, BCOEF >

RETURN

699 FORMAT<41H LINEAR SYSTEM IN SPLINT NOT INVERTIBLE> RETURN

END

Note that [1] contains programs which might facilitate sub­sequent use of the optimal interpolant determined in this way via SPLOPT and SPLINT.

Finally, the linear system (23) can be generated in O(nk2) operations and, because of the band structure, can be solved in O(nk) operations. The linear system (20), to be generated and solved at each Newton step for finding ~. is of similar nature (with n - k rather than n equations and a coefficient matrix which is the transposed of the kind of matrix appearing in (23)) hence requires a similar effort for its construction and solution. Therefore, if it takes indeed only three to four Newton iterations

Page 99: Optimal Estimation in Approximation Theory

90 CARL DE BOOR

to find ~ to sufficient accuracy, then it takes only four to five times as much work to construct the optimal interpolant rather than any spline interpolant to the same data. Also, the total effort is only O(nk2) which, for large n, compares very favorably with such interpolation schemes as polynomial interpolation which takes O(n2) operations, or more general schemes which take as much as O(n3) operations.

ACKNOWLEDGMENT

This work was sponsored by the United States Army under Con­tract No. DAAG29-75-C-0024.

REFERENCES

1. C. de Boor, Package for calculating with B-splines, MRC TSR 1333 (1973); SIAM J. Numer. Anal. (to appear).

2. C. de Boor, A remark concerning perfect splines, Bull. Amer. Math. Soc. 80 (1974) 724-727.

3. C. de Boor, A smooth and local interpolant with 'small' k-th derivative, in "Numerical solutions of boundary value problems for ordinary differential equations", A. K. Aziz ed., Academic Press, New York, 1975, 177-197.

4. C. de Boor, On local linear functionals which vanish at all B-splines but one, in "Theory of Approximation with Applica­tions", A. G. Law & B. N. Sahney eds., Academic Press, New York, 1976, 120-145.

5. C. de Boor & A. Pinkus, Backward error analysis for totally positive linear systems, MRC TSR 1620, Jan. 1976; Numer. Math. (to appear).

6. H. G. Burchard, Interpolation and approximation by generalized convex functions, Ph.D. Thesis, Purdue U., Lafayette, IN., 1968.

7. H. B. Curry and I. J. Schoenberg, On Polya frequency func­tions. IV: The fundamental spline functions and their limits, J. d'Analyse }futh. 17 (1966) 71-107.

8. P. W. Gaffney & M. J. D. Powell, Optimal interpolation, in "Numerical Analysis", G. A. Watson ed., Lecture Notes in Math. Vol. 506, Springer, Heidelberg, 1976.

9. M. Golomb and H. F. Weinberger, Optimal approximation and

Page 100: Optimal Estimation in Approximation Theory

COMPUTATIONAL ASPECTS OF OPTIMAL RECOVERY

error bounds, in "On numerical approximation", R. E. Langer ed., Univ. Wisconsin Press, Madison, WI., 1959, 117-190.

91

10. S. Karlin, Some variational problems on certain Sobolev spaces and perfect splines, Bull. Amer. Math. Soc. 79 (1973) 124-128.

11. S. Karlin and W. J. Studden, Tchebycheff systems: with applications in analysis and statistics, Interscience Publishers, John Wiley, New York, 1966.

12. H. G. Krein, The ideas of P. L. Chebyshev and A. A. Markov in the theory of limiting values of integrals and their further development, Uspekhi Mat. Nauk (n.S.) ~ (1951), 3-120; Engl. Transl.: Amer. Math. Soc. Transl. Series 2, 12 (1959) 1-122.

13. C. A. Micchelli, Best L1-approximation by weak Chebyshev systems and the uniqueness of interpolating perfect splines, J. Approximation Theory (to appear).

14. C. A. Micchelli & W. L. Miranker, High order search methods for finding roots, J. ACM 22 (1975) 51-60.

15. C. A. Micchelli, T. J. Rivlin and S. Winograd, The optimal recovery of smooth functions, Numer. Math. 26 (1976) 279-285.

Page 101: Optimal Estimation in Approximation Theory

INTERPOLATION OPERATORS AS OPTIMAL RECOVERY SCHEMES FOR CLASSES

OF ANALYTIC FUNCTIONS

Michael Golomb

Purdue University West Lafayette, Indiana 47907

1. INTRODUCTION

Suppose that the only information we have about a function f is that it belongs to a certain class ~ (usually a ball in a normed space) and that it takes on given values at some finitely many points xl•···•Xn of its domain (or that some other finitely many linear functionals t 1 , ••• ,tn have given values at f). Suppose we are to assign a value to f(x) where x does not belong to the set {x1 , ••• ,~} (or to t(f) where t is not in the span of {tl, ••• ,tn}). The value a* for f(x) is considered optimal if for any other assignment a there is some fa e ~with fa(Xi) = f(xi) (i = l, ••• ,n) for which the error Ia- fa(x)l is at least as large as Ia*- f(x)l. If moreover a function s* e ~ can be found such that s* evaluated at x gives the optimal value a* and this is so for every x in the domain of the functions f then s* is considered an optimal interpolant (or extrapolant) for these functions.

If .~ is a closed ball in a Hilbert space (and in some other cases) it is known that an optimal interpolant s* exists (see [1], [2]). The interpolant s* has several other extremal properties, and is in particular characterized uniquely by the property that among all functions in ~ which have the given values at x1 , .•. ,~, s* has minimum norm. This allows s* to be calculated from the variational equation and the interpolation conditions. Most of these interpolants are called splines of one kind or another.

More recently the problem of reconstructing in an optimal way the functions f of a class ~ from data f(x1), ••• ,f(xn) has

93

Page 102: Optimal Estimation in Approximation Theory

94 MICHAEL GOLOMB

been given a wider formulation. If Fn denotes the data vector {f(xl), ••• ,f(xn)} one asks for a mapping S (linear or non­linear) from the space of data vectors to the class ~. so that SFn is a good substitute for the only partially known function f. Each such mapping is called a recovery scheme and S* is said to be an optimal recovery scheme (with respect to a chosen norm or seminorm) if the norm of the largest error SFn - f that occurs in ~ is minimized by S* (for this concept and results based on it see [3], [4]).

It is not too surpr1s1ng that in the case where ~ is a ball in a Hilbert space the two problems lead to the same solution: S*Fn = s*. This is true, no matter what the norm or seminorm is that is used for measuring the error Ef = SFn - f. Those used most commonly in this paper are !Ef(x0 )1 for some fixed x0 in the domain of f E ~ , and supxED I Ef (x) for some subset D of the domain. The limitation imposed by the Hilbert space setting is compensated for by the simplicity of the recovery algorithm. The mapping S*:Fn·~ s* is, of course, linear, moreover the mapping fr+ s* is an orthogonal projection onto the n-dimensional subspace of spline interpolants with nodes at x1 , ... ,xn. These and other general results, which are hardly new, are presented in Sections 2-3 below. Practically all the Hilbert spaces of functions of use for numerical analysis have a reproducing kernel. This must be so because the place value functional x~ f(x) is necessarily continuous. This is, in particular, true of the spaces of analytic functions in one complex variable that are treated in detail in this paper. In Section 4 formulas for the optimal recovery operator and the optimal error in terms of the reproducing kernel are,presented.

The "splines" of our Hilbert spaces of analytic functions are themselves analytic functions. The higher order discontinuities -the "knots" of the usual polynomial splines - are replaced by poles outside of the domain of analyticity, which have a simple geometric relationship to the interpolation nodes. The spaces considered in Sections 5-8 have been selected so that they yield simple spline bases, which can be obtained without computation. This is not to imply that the methods are restricted to these examples. Simplicity is achieved by considering only simple domains (disk, annulus, strip, ellipse) and specially chosen norms. These norms, although not the most natural (or traditional) ones from the point of view of function theory, are topologically equivalent to them, and the metric difference is rather irrelevant to the numerical analyst. If, for example, in the case of the strip (with periodicity) the area integral of the square of the absolute value had been chosen for the norm, as was done in some recent papers for similar problems (see [5], [6]), the resulting splines would be elliptic functions rather than the elementary ones obtained by our choice of norm (this will be shown in a forthcoming note).

Page 103: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 95

After a basis for th~ splines has been found, it remains to solve the interpolation conditions. For the simplest of spaces these conditions are solved explicitly and simple expressions for the error supfe~IEf(x0 )l are established, which yield precise results on the asymptotic behavior of this error as n (= number of data) or R (= radius of the disk of analyticity tends to infinity (Section 5). The error sup e~IAEfl for an arbitrary continuous linear functional A is also evaluated explicitly and is then minimized with respect to the location of the interpolation nodes. This results in "hyperoptimal A-rules", which are generali­zations of the quadrature formulas of Gauss and Wilf (Section 6).

In the less simple spaces the interpolation conditions require, in general, numerical matrix inversion. We choose specially simple arrays for the interpolation nodes so that the interpolation conditions can be solved analytically (for a similar method applied to periodic functions of a real variable see [7]). This results in simple explicit formulas for the spline interpolants (optimal recovery schemes), which are well suited for programmed computa­tion. They make it also possible to obtain rather precise bounds and explicit asymptotic limits for the error. In Section 9 the error in optimal recovery from n data is compared to the n-width of the class ~ with respect to a coarser than the initial Hilbert space norm, for which ~ is compact. In most cases the asymptotic unity of the error and the n-width are the same. The upper bounds for the recovery error also provide some new upper bounds for n-widths.

From the error bounds given it follows that the splines sn (n = number of data) converge to f uniformly in any compact part of the domain ~ (at an exponential rate). The sequence {sn(z)} allows analytic continuation of f from the arc containing the data points to the arbitrary z e ~. Several numerical analysts have pointed out that analytic continuation of functions only given by their values in some infinite compact set is not a well­posed problem since small errors in the data can produce arbitrarily large errors of the continuation. Methods have been proposed to cope with this problem (see, e.g., ~1], [12]). Some of our examples, reported in the Appendix, show that even the values of a simple elementary function, computed with ordinary precision, are "noisy" enough to render useless the higher order approximations. In Section 10 a simple modification of the splines is given which achieves stability of the error. The device makes use of two parameters: a known upper bound £ for the magnitude of the noise and a known upper bound ~ for the function f. The procedure is similar to but not identical with the one used by Keith Miller in [12].

Page 104: Optimal Estimation in Approximation Theory

96 MICHAEL GOLOMB

2. INTERPOLATING~- SPLINES

Suppose !fl is a Hilbert space of complex-valued functions which has a reproducing kernel k. Thus if ~ is the domain of the functions, ( •, • )!§ is the inner product in if/, then

(2.1) f(x) = (f, k( ,x))b( , Vf e !:II , Vx e ~ •

Clearly llk(•,x)II.J/= k(x,x) and k(•,x0 )/k(x0 ,x0 ) is the unique function in fll ~at minimizes II f I~ among all the functions f e 1fl for which f (x0 ) = 1. In fact if g e ~ and g (x0 ) = 0 then

(2.2) (g,k(•,x0)~= 0, llk(•,x0 )/k(x0 ,x0 ) + gll:a-

11 k( • ,x0 ) /k(xo ,x0 I~+ II g IIF4 •

More generally, given f 0 e if/ and n distinct points xl•···•Xn in q], the unique function in !fl that minimizes llf II = II f lib( among all the functions f e if/ for which

(2.3)

is

(2.4a)

with the

(2.4b)

n s = ) y j k ( • , xj)

J=l

i = l, ... ,n

so chosen that the interpolation conditions

i = l, ... ,n

are satisfied. That there is a unique such interpolant follows from the fact that the matrix (k(xi,x·)) is the Gramian of the linearly independent functions k(. ,xl3' ••• ,k(. ·~). If g e !(/ ~ g(xi) = 0 (i = 1, ... ,n) then (g,s)b(= 0, II s + gll2 = II s 112 + hll •

Since the above minimizing property is characteristic for many classes of splines that have been defined in the past, it is justified to call s, as defined by (2.4a,b), an interpolating

!if-spline; xl•···•Xn are the interpolation nodes of the spline. The n-dimensional subspace 8 of if/ spanned by the splines k(•,xl), ••• ,k(•,xn) is the spline space of Jfl based on (xl•···•Xn>·

Let the set of distinct interpolation nodes xl•···•xn be fixed. Then if f is any function r!J-+ G: we write F for the n-tuple (f(xl), ••• ,f(xp)) (F0 for (f0 (xl), ••• ,f0 (xn)), G for (g(xl), ••• ,g(xn)), etc.) We denote the spline S defined by

Page 105: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS

(2.4a,b) by either Sf0 or SF0 • Clearly F~ SF is a linear operator from c:n to c§.

97

The interpolation conditions (2.4b) may also be interpreted in another way. Let Pc§ denote the orthogonal projector onto c§. Since f 0 - Sf0 vanishes at the points x1 , ••• ,Xn, f 0 - Sf0 e &~. hence P ¢ f 0 = Sf0 • We formulate this well known result as

Theorem 2.1. The~spline Sf that interpolates f e ~ at the points xl•···•Xn is the same as the orthogonal projection of f onto the spline space based on xl•···•Xn·

There is another minimizing property of the interpolating spline Sf, which is of crucial importance for its use in optimal recovery. Let ~ denote the unit ball in !if:

(2.5) ~ = {f edt: llfll ~ 1}.

We assert: Lemma 2.1. For any x0 e ~ (x0 + x1 , ••• ,Xu), n e C:, f 0 e~

Proof. Let hx be an element of ~ of norm 1 which vanishes 0

at x1 , ••• ,Xn and for which

(2.6) sup lf{x0 )l fe~,F=O

Put for a e R

(2. 7) f a Sfo + (1 - ~ Sfo II>~ eiahx

0

Then since we have

(2.8)

thus (fa -

11~1 = 1,

fa is in the com~eting class. By definition Sf0 Xxo> = (1-l Sf0 II )~eiahXo (x0 ), hence for any

max I fa (xo) - n I ~ (1 - II Sfo ll 2>~x (xo) 0 S a ~ 2'1T 0

and, a fortiori,

sup lf(x0 ) - nl fe~,F=F0

n e t:

Q.E.D.

Page 106: Optimal Estimation in Approximation Theory

98 MICHAEL GOLOMB

The content of Lemma 2.1 is that Sf0 (x0 ) is the "center" of the set {f(x0)}fe~F=Fo C ~. The proof also gives us the

radius and extremal elements of this set. Lemma 2.2. Let ex be the~spline with interpolating

0

nodes x0 ,xl•···•Xn• for which Cx (x0 ) = 1, c (x·) = 0 ( i • 1, .•• , n) • Then ° Xo 1

inf sup Jf(x0 ) - nl nee f~F=F0

Jlcx 11-1 (1- !SF0 f)~ ~ Jlcx JJ-l · 0 0

The only functions f* E ~ for which J f*(x0 ) - Sf*(x0 ) J

are the functions eincx /JJ ex JJ , a e R. 0 0

Proof. From the minimizing property of the spline by (2.5) we have

and

(2.10) JJ ex JJ = inf ---#-\1 = 1/ sup J f (x0 ) J = 1/hx (x0 ). 0 fe!fi ,F=O ~ f~,F=O o

For every at xi(i

f E ~. (1 - ~Sf JJ 2) -~(f - Sf) is in ~ and vanishes l, ... ,n), hence by (2.10),

(2.11) Jf(x0 )- Sf(x0 )J ~ lcx0 ll-1 (1 -JJsfJJ2 )~

On the other hand, by (2.9)

(2.12) sup Jf(x0 )- Sf(x0 )J ~ JJc JJ-1 (1 -1Sf0 f)~ fe~,F=F0 xo

(2.11) and (2.12) prove the first assertion of the Lemma. Now f* = eincx /I ex JJ (a E R) is in ~ and satisfies

0 0

Jf*(x0 ) - Sf*(x0 ) J = Jlcx 11-1 , and if g is any other function

in ~ then (1 - JJ SgJJ2)~~(g Sg) is in ~ and vanishes at xi (i = l, ..• ,n), hence

(1- JJsgJJ 2)-~Jg(x0 ) - Sg(x0 ) J < sup Jf(x0 ) J fe~,F=O

so that the second part of the Lemma is also proved.

Lemmas 2.1 by applying the easily extended

Lemma 2.3. which is not in f 0 e ~

and 2.2 deal with the set in C that is obtained functional ~ : <S,c (f) = f (x0 ) to ~. They are to the case of0 a geHeral linear functional. For every continuous linear functional A on ~

the span of ox , ••• ,ox , and for each n E ~. 1 n

Page 107: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS

sup IAf - nl ~ f€~,F=F0

sup IA(f) - A(Sf0 )l fE~,F=F0

If cA e ~ min~mJ.Zes II f II among all f € ~ for which f(xl) = ••• = f(xn) = 0, A(f) = 1, then

inf sup IH- nl = llcAII-1(1- ~SF0 f)~ ~ lc,_ll-l . nell: fe.~,F=F0

99

The only functio~s f" € ~ for which I Af* - !.Sf* I = II cA Il-l are the functions e~a.cA/ 11c~ II, a. E lR.

Proof. The proof ~s identical with that of Lemmas 2.1 and 2.2 except that hx is replaced by hA E !1/ of norm 1 which vanishes at xl,···~xn and satisfies sup IAfl = Ah,_.

fe.~,F=O

3. OPTIMAL RECOVERY SCHEMES FOR A BALL IN Ji/.

Let ~=~~denote the class of mappings. (not necessarily linear) from 4:n to ~. called recovery schemes. R € !Je is to associate an element RF E ~ with certain elements F e. en. F is interpreted as then-tuple (f(x1), .•• ,f(xn)), where f E~ is the unknown function to be recovered. The subset (f(xl), ••• ,f(Xo))fe.~C G:n is the domain of each R E~ ~. X denotes the set (xl•····~), the domain of the data. '

Definition. Rx €~~is an-optimal recovery scheme of~ at x 0 e.q if 0

Ex ~ = inf sup lf(x0 ) - RF(x0 ) I = sup lf(x) - Rx0 F(x0 ) I· 0 RE!Je telA f~ 0

The function x >+ E ~ is the minimal error function for recovery. An eleme~t fx e.~ such that

0

Ex0 ~ = I fx0 (xo) - Rx0 Fx0 (xo) I

is an extremal for recovery at x 0 • R* €~.~ is an optimal recovery-scheme of ~ if

sup sup lt(x) - R*F(x)l XEq} fE~ sup II f - R*FII fE~ 00

where lloo denotes the sup-norm for maps q} + C. E* ~ the minimal error for recovery of ~ from data on X.

is

With these definitions and the results of Section 2 we have the following:

Page 108: Optimal Estimation in Approximation Theory

100 MICHAEL GOLOMB

Theorem 3.1. The linear spline projector S which associates with the n-tuple F (or the function f) the inter­polating bf-spline SF (or Sf), with set of interpolation nodes (xl, .•. ,xn), is the unique optimal recovery scheme of !;j at every x e fiJ, x 'f x1, ••. •Xn· The minimal error function for recovery is x~ llcxll-1, where Cx is the interpolating ~-spline for which cx(x) = 1, cx(xi) = 0 (i = l, ... ,n) .. The only extremals for recovery at x are the functions e~CJ.cx/ llcx II, (a. e lR) • The minimal error in recovery of !;j is

Proof. The proof is a direct consequence of Lemmas 2.1, 2.2 and the definitions.

In place of recovery of !j at x0 we may ask for recovery at A where A is any continuous linear functional. Using Lemma 2.3 we have

Corollary 3.1. The unique optimal recovery scheme of !;j at the continuous linear functional A not in the span of ox1 , .•. ,oXn is A(SF). The minimal error in recovery at A is ~cAII-1 , where cA is as defined in Lemma 2.3. The only extremals for recovery at A are the functions eia.cA/11 cA II, a. e R.

There is another expression for the minimal error function that we find useful in many cases. In the formulation of this result we use the notation kx for k(•,x0 ) to avoid possible confusion. 0

~T~h~e~o~r~e~m=-3~·~2. Suppose kx (x). The minimal error fr8m data on (xl•···•Xn)

~ has the reproducing kernel k(x,x0 )

for recovery Of the ball {j C!f/ at X 0

is

Ex 00 = II kx - Skx II = [kx (x0 ) - Skx (x0 ) ]~ 0 0 0 0 0

where S is the projector onto the bf-spline space ~ based on (xl•···•Xn)·

(3 .1)

Proof. By Theorem 3.1

sup lf(x0 ) Sf(x0 )l fe!;j

sup I (f,kx) - (Sf,kx )I fe§i o o

sup I (f,kx - Skx )I fe!j o o

likx - Skx II 0 0

Page 109: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 101

Since S is an orthogonal projector, we also have

Remark 3.1. It is well known that )tk is the reproducing kernel for the subspace.,.( C !:fl if P..J, is the orthogonal projector onto .,.(. Therefore, Theorem 3.2 says that Exo~ is the norm of

the reproducing kernel at x0 of the space cE J. C d/ .

4. SPACES OF ANALYTIC FUNCTIONS

In the following we consider Hilbert spaces of functions that are analytic in some region ~C [ and have a reproducing kernel k(z,z0 ) with the property that the function zr+ k(z,z) is bounded on every compact subset of ~. If !:fl is such a space and {~v}v=l,2, ••• is a complete orthogonal system in b( then k(•,z0 ) = E ~v(z0)~v for every z0 e~, the series being con-

v vergent in the sense of 1i/. But it is also true that

co

(4 .1) k(z,z0 ) = E ~v(z0) ~v(z), v=l

in the sense of pointwise convergence. Moreover the convergence is uniform in any compact subset of ~ x ~. In particular

(4.2) Vz e f!J

and the sum is uniformly bounded in every compact subset of f!J. Conversely if b( is a Hilbert space of functions analytic in a domain f!JCt which has a complete orthogonal system {~v}v=l 2 such that El.~y(z)l 2 is uniformly bounded on every compact su~s~t·· of f!J then ~ has a reproducing kernel of the above kind. Indeed the kernel is given by the series (4.1). We observe that if f e!:fl then

(4.3) f(m)(zo) =

In general, if A any f e Jll

(4.4)

am ----;(k(•,z0 ),f)b(, Vz 0 e f!J; m = 0,1, •.• azo

is any element of the dual space ~ then for

Let Z = (z1 , ••• ,zn) be a fixed n-tuple of distinct points in 9J. As pointed out in Section 2, an interpolating b(-spline

Page 110: Optimal Estimation in Approximation Theory

102 MICHAEL GOLOMB

with nodes (zl•···•zn) is of the form

n (4 .5) s = .E y~k(•,zi)

~=1 ...

with y. E ~. By Section 3, the optimal recovery scheme of the unit bah~C~ from data F = (f(z1), ... ,f(zn)) is given by the spline

(4. 6) n

SF = E Yi k ( • , z i) i=l

where the yi are determined from the system

n (4.7) i~lYi~(zj,zi) = f(zj),

The minimal error Ez 0 (£I) obtained in terms of the kernel be determined from the system

(4.8) n E n·k(z z~) = ooJ· i=O ~ j ...

Then by Theorem 3.1

j = 1, ... ,n.

in recovery at z 0 can also be k. Let the numbers n0 , nl•···•nn

j 0, l, ... ,n.

(4.9) Ez0~ = lliEo nik(·,zi)ll~1 = (iJ=o k(zj,zi)njni)-~. From this expression we obtain bounds on the error. They are

valid for any Hilbert space of functions with reproducing kernel. Theorem 4.1. Suppose ~ is a Hilbert space of functions

fi) + G:, with the reproducing kernel k, and ~ is the unit ball in ~. Then the minimal error Ez0 (£~) in recovery of _CA at z 0 from data at the distinct points zl•···•Zn satisfies

~ < ~~ ft.~ Amin- Ez0~1 ~ max

where Amin• Amax are the smallest and largest eigenvalues of the matrix (k(zi,ZJ·))~ J"=O 1 n · ... , ,, .•• , +1

Proof. Let u0 be the vector (1,0, ••. ,0) in a;n , e the vector (n£,nl•···•nn) defined by (4.8), K the linear operator a;n+l + tn+ defined by the matrix (k(zi,Zj)). K is hermitian positive-definite, let K~ be its positive square root. By (4.8),

L -!.: Ke = u0 , hence K~e = K ~u0 • By (4.9)

(4.10)

where lvl denotes the Euclidean norm of the vector v E a;n+l. The assertion follows from (4.10).

Remark 4.1. Up to this point we have assumed that the inter­polation nodes Zl•···•Zn are distinct. In more general recovery problems we admit equality of some of the zi, say z 1 = •.• = Zm· With the usual convention, the data vector F then stands for

Page 111: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 103

(f(zl),f'(z1 , ••• ,f(m-l)(zl), ••• ) ("Hermite data"). To simplify the exposition assume m = n. We assert that thebf-spline interpolating Hermite data has the form

~ n-1 (i) s = i~O Yik (•,zl)

where k(i)(•,z) = aik(·,z)/azi. In fact if g g{i)(zl) = 0 for i = O,l, ••• ,n-1 then since g(i)(zl) = 0

e bf is such that (g,k(i) (• ,zl))~ =

Thus, s has the same minimizing property with respect to Hermite data that s in (2.4a) has with respect to Lagrange data (2.4b). We write again SF or Sf for s and conclude that the spline operator S is the optimal recovery scheme of ~ from Hermite interpolation data.

5. ~(DR)-SPLINES FROM INTERPOLATION DATA

Assume R > 1 fixed and let DR {z e C: I z I < R}. We denotj by ~(Dj) analytic in DR for which lzl 2 1dz in 0 ~ r < R. We express this latter

(5.1} I lfl2 < ® •

anR

be the open disk the space of functions

is a bounded function of condition as

f

This is f(rei8) locally limit.

justified because under the above condition the limit of as r ~ R - 0 exists for almost all e e R and is

square-integrable. f in (5.1) may be thought of as this

Using the integral (5.1) we define the norm

(5.2) (__!__ f lfl2>~ 21TR an

R

on ~DR) and the corresponding inner product (•,•) Q((DR)· ~(DR) is known to be a Hilbert space.

Let ev(z)

(5 .3)

ev(v = 0,1, ••• ) zV. Clearly

denote the function defined by

and the sequence {R-Vev}v=O,l, ••• forms a compl:;/.e orthonormal system in bf(DR). If f = L avev for some f e ~(DR) then f(z) = L avzv is the usual power series expansion of f.

r

Page 112: Optimal Estimation in Approximation Theory

104 MICHAEL GOLOMB

Condition (5.1) is equivalent to

(5.4)

and if g

(5.5)

~DR)

(5. 6)

Thus, the Zl•••••Zn

(5. 7)

00

R2vJetvl2 < E 00

v=O = E Bvev then

00

R2VCL S (f,g) bf(DR) = v~o \) \)

has the reproducing kernel

k0 (z,z 0 ) E R-2vzvzv = 1 v=O o 1 zz0 /R2

spline space 60 C bf(DR), based on the distinct nodes in DR, is spanned by the functions

1 sD,i(z) = 1 - zzi/R2 i 1, ... ,n

s0 i is, except for a constant factor, characterized as the rational function which vanishes at oo and has a simple pole at R2/zi. This point, which is the inversion of the node zi at the circle aDR, may be called the knot of the spline Si generated by the node zi. The knots of the usual polynomial splines, which are discontinuities of high-order derivatives, are replaced here by poles of the analytic continuation of the spline to the exterior of DR. We may state: The spline space 6 D C M(DR) , based on the distinct nodes z1, ... ,zn , is the span of rational functions that vanish at oo and have simple poles at the points Rz1- Rz;-zl' ... ' Zn.

We derive an explicit formula for the ~DR)-spline Szf that inter~olates f on Z = (zl, •. ,zn)· Since Szf has simple poles at R /zi (i = l, ... ,n) arid vanishes at oo,

TI (z - z·) TI (z· - R2/z1.) i#j l. i J (5.9) Sj(Z): Szoj (z) TI (z - RZ/z·) .IT. (z. - z·)

i l. l.oFJ J l.

is the spline for the data vector oj = (oij)(i=l, ••. ,n)" There­fore,

(5.10)

is an explicit formula for Szf. It is seen that, as R + oo, Szf converges, uniformly on every bounded subset of [, to the Lagrange interpolation polynomial

Page 113: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 105

(5.11)

The minimal error Ez (~) 0

explicit form. The spline cz0

at z0 e DR can also be given in (see Lemma 2.2) is

(5 .12) (z 0 - R2/z0 ) ~ (z - zi) ¥ (z0 - R2/zi)

Cz 0 (z) = (z - R2/z0 ) ~ (z - RZ/zi) ~ (z0 - zi)

hence by Theorem 3.1

Us'ing z = Rei6 in the integral, then and 11 - zz0 /R212 R-2(R2 - 2 Rr0 cos

(5.13)

IIi lz- zii/III1-zzi/R21 =Rn, e + r~), hence

lzo - zil II i 11 - z0 zi/R2 1

This is a remarkable simple expression for the minimal error in the general case. One of its consequences is the following limit relation. We write ~ for ~ to indicate the dependence on R.

(5.14)

Thus

(5.15)

More detailed results are obtained for the special case of symmetrically distributed interpolation nodes. Suppose 0 < r 0 < R

(5 .16) j = 0,1, ... ,n - 1

and write Z* for this special set of nodes. A simple calculation shows that (5.10) becomes

Page 114: Optimal Estimation in Approximation Theory

106 MICHAEL GOLOMB

(5 .17a) SZ f(z) *

where Lz f is the Lagrange interpolation polynomial *

1 n-1 f(zj) zn - rn (5.17b) Lz f(z) =- l: 0

* n j=O n-1 z - Zj zj

or also n-1

fj (z/r0 )j 1 n-1 (5.17c) Lz f(z) l: f. =- l: f(r e:k)E:-jk

J n k=O 0 n n * j=O

These formulas allow for extremely simple computation of Sz*f (see Appendix). The fractional factor in (5.17a) is, in most cases, so close to 1 that it can be ignored.

We also derive an explicit expression for the partial fraction expansion of the rational function Sz f. It has simple poles at the points ~j = (R2/r0 )e:j (j = 0, .•• ~ n- 1) and

res Sz f ~j *

Therefore,

1 - (r /R)2n 0

-n~~I Lz*f(~j) • J

(5 .18) 1 ~.Lz f(~.)

Sz f(z) = [1 - (r /R)2n] - J * J * o n ~j - z

This formula is, except for the factor [1- (r0 /R)2n ], the discrete analogue (with the integral evaluated by the trapezoidal rule) of the Cauchy integral for Lz*f along the circle 1~1 = R2/r0 • Like (5.17), it is easily programmed for computation. Moreover, it leads to particular simple expressions for the optimal recovery of the expansion coefficients of f at some z0 e DR (and of other linear functionals, see Corollary 3.1). If

co

we set f(z) = ~O av(z 0 )(z- z0 )V and Sz f(z) = co v- * v~O av*(z0 )(z - z0 )v then

(5.19) 1 n-1

a" (z0 ) = (1 - (r /R)2n)- l: v* o n j=O ( ) v+l'

~j - zo v = 0,1, •••

zo For the integral f f(~)d~ we find the optimal quadrature

formula z 0

f o 2 1 n-1 (5.20) Sz f = (1 - (r0 /R) n)0 .E ~jLz f(~j)log(l-z 0hj).

0 * J=O *

Page 115: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS

For the Fourier coefficient the optimal value

(5. 21)

107

v = 0,1, ... ; v' = v(n), 0 ~ v' s n- 1.

Let the minimal error at z0 be denoted as indicate the dependence on n as well as on R.

to

(5.22)

We note the two limit relations

(5.23) lim RnE~n) (~R) lzg - rgl R+oo 0

I zol /R (5.24) njE(n) (~ ) lim

n+oo z0 R

r 0 /R

l.. f I I zo > ro

if lzol < ro

If lz0 l = r 0 , but z~ + rg for any n (i.e. z0 never coincides with an interpolation node) then

(5.25) 2

We also note convenient bounds for the case • 1 1

(ro/lzol)n < 3, (rolzoi/R2)n < 3 :

1 <lzoi/R)n

2 (1 - lzol2fR2)3/4 (5.26)

6. ~(DR)-SPLINES SATISFYING INITIAL CONDITIONS. HYPEROPTIMAL >.-RULES.

and

In the first _part of this section we obtain an explicit formula for the bf(DR)-spline operator for optimal recovery of the ball in l#'(DR) from Hermite data

(6.1)

Let the spline be called Sz1f. By Remark 4.1

(6.2a) Sz f nEl y·k(i)(• z1) 1 i=O 1 ,

Page 116: Optimal Estimation in Approximation Theory

108 MICHAEL GOLOMB

where k is the reproducing kernel of bf(Dr) (see 5.6) and the Yi must be determined from the conditions

(6.2b)

(6 .3)

i = O,l, ••• ,n- 1

is a rational function with one pole of order

vanishing at ®. Hence we may set

n-1 k k~O l3k(z - zl)

Szlf(z) (1 - zzl/R2)n

n at

To determine the l3k we observe that Snf is analytic at z1 , hence

(6.4)

From (6.2b) we know the first n coefficients:

(6.5) v = O,l, ••• ,n- 1 .

Comparing coefficients in (6.3) and (6.4), we find the l3k· With some simplifications the result is

R2 - ZlZl n-1 -Sz1f(z) = k

RZ - zzl kEo l3k(z - zl)

(6.6) k ( -z ) . f(j) (Z]) ak E (n ) 1 k-J

k-j R2 - zzl j=O . ' J.

These formulas are easily programmed for computation.

The spline cz0 (see Lemma 2.2) is

(6. 7) z - z1 n (1 - z0 z1/R2) n cz (z) = ( ) 1 2

0 Zo - Zl 1 - zzl R

1 - z0 z0 /R2 1 - zz0 /R2

From this the minimal error for recovery is calculated by Theorem 3.1:

(6.8) E{n) (~R) R-n

11 zo - zl In

zo (1 - lzolZfR2)3/4 - zozl/RZ

We note the limit relations

(6.9)

Page 117: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 109

(6.10) lim n-+oo

In the second part of this section we return to the case where the data vector is (f(zl), ••• ,f(zn}), with Zl•···•Zn distinct points in DR· We first establish a convenient formula for the minimal error EA in recovery of the unit ball ~ at the functional A € !f/( DR)' and then minimize EA by choice of the data points zi. This inquiry was stimulated by work of Wilf's, Eckhardt's and Engels' on optimal quadrature formulas [13-17]* •

If S is the spline projector of (5.10) then

(6.11) EA : = sup IAf - ASfl f€~

Let Si (i = l, .•• ,n) data oil•···•oin at

be the spline (5.9), which interpolates the zl•···•Zn, and set cri = ASi. Then

(6.12)

is the "optimal A-rule" for computing Af, given the data f(zi), f € ~. Using the kernel k of (5.6) and the conjugate functional A*, defined by

A*(f) = i{f),

we have

(6.13) f(z) = (f,k(•,z)),

(The subscript Q((DR) is omitted from the inner product symbol, and the subscript z in A: indicates that A* acts on the function z~ k(•,z).) By (6.12) and (6.13)

(6 .14) - n -Af = (f, E crik(•,zi))

i=l

A simple calculation now gives

(6.15) AuA~k(u,v)

AuA~k(u,v)

n -- E cr·cr·k(zi z·) i,j=l ~ J ' J

n * - .r OiAzk(zi,z) , ~=1

*The author wishes to express his thanks to Dr. U. Eckhardt for calling his attention to these papers and for providing him with preprints of the unpublished ones.

Page 118: Optimal Estimation in Approximation Theory

110 MICHAEL GOLOMB

where we have made use of the fact that k(•,zj) is a spline, Sk( •, z j) k(•,zj), hence by (6.12)

n A~k(ZifZ)

n Ojk(zi,Zj) • (6.16) A.zk(z,zj) E Ojk(zi,Zj), E

i=l j=l

Using the function

(6.17) K= A.~k(•,z),

(6.15) can also be written in the compact form

(6.18)

which says that EX_ is equal to the error in computing A.K by the approximation A.K.

We now assume that A. is a real functional, by which we mean

A.(f) ER if f(x) em. for xeJR(')DR .

We also assume that there are distinct real nodes Zi = Xi (i = l, ... ,n) in DR that make Eft. a minimum, so that

(6.19) 0 l, ... ,n .

Then the numbers cri A.si and A.~k(xi,z) = A.zk(xi,z) (i = l, ..• ,n) are real. If condition (6.19) is satisfied, we say the optimal A.-rule

(6.20) ~f A. Sf

is hyperoptimal provided none of the cri is 0. The proviso is natural since if some cri vanish the corresponding terms in (6.20) should be dropped and n reduced. In the remainder of this section we establish conditions equivalent to (6.19), from which the nodes Xi for a hyperoptimal A.-rule can be determined.

In the following k1, k2 denote the derivatives of the functions zt-+ k(z,•), zt-+ k(•,z), and Oij (i,j = l, ... ,n) denotes the derivative of the function xj~-+ cri at xj. Using (6.15), condition (6.19) becomes

(6.21) 0' R. l, ... ,n .

From the identity (6.16) one obtains

Page 119: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS

(6.22)

- l..zk2 (z,xi)

From (6.21) and (6.22) it follows that n

cri[i~lcrik2 (xi,xi) - t..zk2 (z,xi)]

and since cri # 0, by (6.12):

0 ' i,i

0

(6.23) i=l, ... ,n.

111

1, ... , n

This equation says that if /.. is a hyperoptimal /..-rule then it evaluates the numbers /..k2(•,xi) (and, of course, also Ak(•,x.Q.)) (i = l, ..• ,n), accurately. The path from (6.19) to (6.23) can be reversed, hence ~ is a hyperoptimal /..-rule if and only if the X.Q. satisfy equations (6.23). Let X denote then-tuple (xl•···•xn) and Qx the linear space of rational functions which vanish at oo and whose only singularities are poles of order ~ 2 at the points R2/xi (i = l, .•• ,n). Qx is spanne~ by the functions k(•,xi), k2(•,xi) (i = l, ... ,n), hence A is a hyper­optimal A-rule if and only if

(6.24) Ag = t..g, g e Qx

space Thus,

The functions Qx· If ~

si, sr (i = l, ••• ,n) also form a basis of the is hyperoptimal then t..sr ~sr Ejcrjsi<xj) = cri.

(6.25) i=l, ... ,n

is still another necessary and sufficient condition that the /..-rule /.. with the weights cri be hyperoptimal. From (6.25) also follows that if /.. is not only a real, but a positive functional, i.e.

H :::. 0 if f (x) ~ 0 for x e 1R (') DR ,

then the weights cri of a hyperoptimal rule are positive.

Still another characterization of hyperoptimality results from the following considerations. Let Hf denote the ~(DR)­spline that interpolates f on the set (zl,zl,z2,z2,···•zn,zn). The data vector is then (f(zl),f'(zl),f(z2),f'(z2), ••. ,f(zn),f'(zn)) (see Remark 4.1). Hf is a linear combination of the splines k(•,zj), k1 (·,zj) (j = l, ••• ,n), thus Hf is a rational function

Page 120: Optimal Estimation in Approximation Theory

112 MICHAEL GOLOMB

that vanishes at oo and whose only singularities are poles of order~ 2 at the points R2/zi (i = l, .•• ,n). Let Qz denote the linear space of these rational functions. The functions st(z), (z- Zi)s!_(z): = ti (z)(i = l, ... ,n) also form a basis of Qz. Using this fact and the interpolation conditions

(6.26) 1, ... ,n

it follows that

(6. 27) (Hf)(z)

and that

(6.28) ~f: AHf .¥ [A(s~)f(z;) - 2s!(zi)A(ti)f(zi) ~=1 ~ .... ~

+ A(ti)f'(zi)]

is the optimal A-rule for computing Af, given the data f(zi), f'(zi), f e.~.

Now assume _zi =xi (i = l, ... ,n) are the nodes of a hyper­optimal A-rule A with the weights cri = A(si) = A(s!_). Then A(ti) = ~(ti) 0 by (6.12), hence

(6.29) ~f

- -If conversely the nodes Zi = xi are such that Af Af for every f E b((DR) (or only for every f e. Qx) then (6.24) is satisfied, hence X is hyperoptimal.

We summarize some of these results in

Theorem 6.1. Suppose A e ~(DR)' is a real functional; xl•···•xn are real distinct points in DR; sl•···•sn are the splines (5.4) with Zi =xi; and cri = ASi + 0 ( i = l, •.. ,n). The optimal A-rule, given the data f E ~. f(xi) is

- n Af = .L crif(xi)

~=1

and the minimal error is 1

E = (AK - XK)~ , A

where K = A:k(•,z). The A-rule ~ is hyperoptimal (i.e. aEA/axi = 0, i = l, ..• ,n) if and only if Xg = Ag for every rational function g that vanishes at oo and has no other singularities but poles of order S 2 at the points R2/xi. This is the case if and only if Ati = 0, where ti(z) = (z- xi)sr(z).

Page 121: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 113

Also, X is hyperoptimal if and only if Xf = Xf for all f E ~(DR), where 'X is the optimal A.-rule for the data f e ~ f(xi), f'(xi)• If A. is a positive functional then the weights cri of the hyperoptimal A.-rule are positive.

Remark. The special case A.f = f1 f is treated in [13-17]. -1

For this functional the hyperoptimal A.-rule is the Wilf quadrature formula for R = 1 (really R + 1), and is the Gauss quadrature formula for R + oo •

1. Assume R > 1 {z E ~:R-1 < lzl < R}. f analytic in AR for

fixed and let AR be the open annulus We deno7e by ~A~) the space of functions which J !f(z)i2!dz! is a bounded

lzl=r function of r in rl < r < R. As in Section 5, we express this condition as

(7 .1) J ifl2 < 00 •

aAR

Under this condition the functions

(7.2) ~0 (8) = lim f(re2nie), ~i(e) r + R- 0

lim f(re2nie) r +R + 0

exist for almost all e e R and are locally square-integrable.

Suppose f E bf(AR) convergent for z E AR.

00

has the expansion Then

f(z) =

00 rl~ol 2 \)=-"" v=-oo

0

and (7.1) is equivalent to 00

(7.3) 00 •

\1=-00

We introduce the inner product norm

(7.4) i!fW E la.vi 2R2 1vl l#'(AR) v=-oo

00

r a.vz" ' \)=-""

and this makes ~(AR) a Hilbert space. This norm is not easily

expressed in terms of f~ !~0 ! 2 and J~~~il 2 •

Page 122: Optimal Estimation in Approximation Theory

114 MICHAEL GOLOMB

Clearly the system {R-Jvlev}v=O ±l, •.• forms a complete orthonormal system in ~(AR). It follows that this space has the reproducing kernel

(7.4) oo -2lvl v::v - 1 - 1 L R z zo - 1 - R-2zz0 1 - R2zz0 v=-oo

Thus, the spline space cE A C 1:1/(AR), based on the distinct nodes Zl•···•zn in Ar, is spanned by the functions

(7.5)

sA,i is, except for a constant factor, characterized as the rational function which vanishes at 0 and oo and has simple poles at R2/zi and R-2/zi , which are the inversions of the node zi at aAR. An equivalent condition to vanishing at 0 is that the quotient of the residues at R2/zi, R-2/zi is -R4 • The spline space cE A C !:1/(AR), based on the nodes z1 , •.• ,zn, is the space of rational functions that vanish at oo, have simple poles at the points RZ/zi, R 2fzi (i = l, .•. ,n) with residues whose quotient for each pair is -R4

The !:1/(AR) -spline that interpolates f on z is given by

n z (7.6a) Szf(z) = L Yi (z - R2/z.)(z - R -2/zi) i=l 1

where the Yi are determined from the conditions

(7.6b) j=l, •.. ,n.

This requires matrix inversion and no practical explicit inverse has been found for the general case.

For the special case of symmetrically distributed inter­polation nodes

(7. 7) Z* = (r Ej) _ , £ = e2~i/n R-1 < r < R o j-0,1, •.. ,n-1 ' o

we obtain simple explicit formulas for the spline interpolant and for the error function. We set

Sz.f(z) ~ :~~ Yi (1- R ~r0fzo1- 1- R~r~lzoi) nil y· E [<R-2r-lz£i)v + (R-2r z-1£-i)v+l] i=O 1 v=O 0 0

Page 123: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 115

Since Sz*f(r0 €k) = f(r0 £k) we obtain

1 n-1 n-1 fJ. =- E f(r €k)€-kj = ~· E y.£ij n k=O 0 J i=O 1

(7 .9) E (r0 /R)2v + E (r0 R)-2v , j = O,l, .•. ,n-1 v=:j (n) v=-j (n)

(R2n-2i -R2i-2n)r~i + (R2i _ R-2i)r~(i-n) (RZn + R-Zn) - (r~n + r 02n)

n-1 With E Yi€ji = fj/~j substituted in (7.8), one finds

i=O

(7.10)

= (R2n +r2n) _ (r~n +r;;2n) n-1 A a; (z) Sz*f(z) (RZn +R Zn) - ((r0 z)n + (r0 z)-n) j~O fj aj (r0 )

aj(z) = (R2n-2j - R2j-2n)(r0 z)j + (R2j - R-2j)(r0 z)j-n

A 1 n-1 k k" fj = - r f(r € )€- J n k=O o

j = O,l, •.. ,n-1 •

Formula (7.10) is easily programmed for computing. Except for the fractional factor (which is indistinguishable from 1 in most cases), Sz f(z) equals

* (7.11) Lz f(z)

* which is the unique linear combination of the simple functions ao. al·····an-1 that interpolates f on z*.

We also give the explicit partial fractions e~ansion of Sz f. There are simple poles at the points ~j = R /r0 £j and ~j*= R-2/r0 £j (j = O,l, ••. ,n-1) with residues

(7.12)

Therefore

res Sz f c;j *

Page 124: Optimal Estimation in Approximation Theory

116

(7.13)

Sz f(z) *

(R2n + R-2n) _ (r~n + r~2n) R2n _ R-2n

MICHAEL GOLOMB

This formula is, except for the constant factor in front, the discrete analogue of the Cauchy integral for Lz on the boundary of the annulus R-2/r0 < lzl < R2/r0 containing*the knots ~j and ~1(j = O,l, ••• ,n-1). From it we obtain a simple formula for the optimal recovery of the Taylor expansion coefficients of f at some z0 € AR. If f(z) = vio av(z0 )(z- z0 )v and Snf(z) =

co E0 av (z0 )(z - z0 )V then

V"" *

(7.14)

2. Let ~R be the unit ball in bf(AR). The minimal error in recovery of ~ at z0 € AR is, by Theorem 3.1,

(7.15)

where cz0 is the bf(AR)-spline which interpolates 0 on Z* and 1 at z0 • By the characterization of the ~AR)-splines given above it follows that

(7.16)

where y is a constant and q is a polynomial in z-1 of degree n:

(7.17) q(z) = ¥ ~.(zz )-R. R.=O "' o

The aR. are determined (except for a common factor) from the conditions

and y is then found from Cz (z0 ) 1. 0

k 0,1, ••• ,n-1

Page 125: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 117

The remainder of this formidable calculation is carried out only for the case r 0 = 1, i.e., Z* = (Ej). (7.18) gives

(7.19)

k = O,l, .•. ,n- 1

Set

(7.20) Rj+rj=rj

Multiplying (7.19) by Ejk and summing over k gives

j

(7 .21) rn-481- rn+28o- rn-28nzo-n = 0

By choosing 81 as (rnrn-2rn-4)-l one finds

l, ... ,n-1

j = l, ... ,n-1

(7.22)

Thus q

(7.23)

in (7.16) is explicitly determined, and so is y:

y = (r2n- z~- z~n)(z0 - R2/z0 )/z0 (zg- l)q(z0 )

We use the result to calculate the asymptotic value of Ez (~R) for R ~ ®, For this we use

0

then

Lemma 7.1. If fR € ~(AR) for each R > 0, fR • f _1 and R

lim R-KifiRei6)1 = K(6) R~®

lim RAifR(Rei6)1 = L(6) R~O

(i) lim r2K II fR ~ 2 R ~ ® !f/(Ar)

lim r2A II fR 12 R ~ ® 1:1/(AR)

(ii)

(iii)

if K > A

if K < A

if K = A

0,

Page 126: Optimal Estimation in Approximation Theory

118 MICHAEL GOLOMB

co Proof. Suppose fR(z) = E av(R)zV for z € AR. From the

hypothesis it follows that v=-co

Lim r2K L J21T IE av(R)Rveiv8 12de = L J21T K2de , R -+ co 21T 0 21T 0

hence

(7 .24)

Similarly,

lim R2A Elav(R)I2R2v = !_ f21TL2de , R -+ 0 21T 0

hence since av(R) = av(R-1 ) ,

r2A Elav(R)I2r2v 1 r1T (7 .25) lim = .- L2de • R-+co 21T 0

If K > A then (7.25) gives

(7.26) lim R-2KEiav(R)I2R-2v 0 R-+co

and adding this to (7.24) gives

lim R-2K Elav(R)I2(R2v + R-2v) = ~1T f201TK2de ,

R-+co hence also

(7. 2 7) lim R-2KII fR 112 = lim r 2K E I a\) (R) 12R2 1 \)I R -+ co ~AR) R -+ co

= L Jh K2de • 21T 0

Here we have used lim R-2Kia0 (R)I2 = 0, which follows from R-+co

(7.26). Thus (i) is proved and (ii) follows similarly. If K =A then we arrive at

lim R-2K E lav(R)I2(R2v + R-2\1) = ~1T f20

1T(K2 + L2)d8 , R-+co

which implies

(7.28)

Page 127: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS

This proves (iii) since a 0 (R) = (2~i)-l I z-lf(z)dz. lzl=l

Using this lemma we now prove

119

Theorem 7.1. If Ez~~.r) denotes the minimal error for recovery of the ball ~ L~(AR) at z0 + 0 from data on the set ( e2~ik/n)k=O 1 _1 then

' ' ... 'n

lim Rn/ZEz0 (~R) = z-~lzol-n/Zizg- 11 . R+oo Proof. For simplicity we assume n even, n = 2m. Let qR

be the function defined in (7.14), with the Si given in (7.22). One sees readily that

lim Rm+41qR(Rei6) I R+oo

lim Rm+41SmR-mzo-ml = t lzol-m R+oo

(7.29) lim Rm-4 1qR(Rei6>1 lim Rm-4 1SmR-mz 0 -ml = t lzol-m R+O R+O

lim R4 1qR(z0 )l =lim R4 1smz0 - 2ml = t lzol-n R+oo R+oo

If YR denotes the constant of (7.23) then, using (7.29) one finds

(7 .30) lim R2n+61Yill = t lzol lzg- ll lzol-n R+oo

Finally, define

Then using (7.29), one obtains

(7.31)

lim R3m+61fR(Rei6>1 = t lz0 l-m+l R+oo

lim R-3m-61fR(Rei6) I = t lzol-m+l R+O

fR satisfies the hypotheses of Lemma 7.1. By (7.31), Case (iii) applies and since a0 (R) = 0 we conclude

(7.32) lim R-3m-61 f II = 2~ lz ~-m+l II R ~AR) o R+oo

By (7.15), (7.16), (7.30) and (7.32)

lim RmEz (~ R) R+oo 0

lim Rmhi1 1 II fRII~/. R + oo .w(AR)

z-~lzol-mlzg- ll '

Page 128: Optimal Estimation in Approximation Theory

120 MICHAEL GOLOMB

which proves the theorem.

3. We now investigate the dependence of the minimal error on n, the number of data ~oints. To indicate this dependence I denote the error as E~n)~R>· Although we do have an explicit

0 expression for E~n)(~R) I did not find it useful for the study of the asymptotic behavior of the error as n ~ oo. Indeed I use the result of Theorem 3.2

(7. 33) E~n) (~R) = [kz (z ) - Snkz (z0 ) ]~ , 0 0 ° 0

where the reproducing kernel is that of (7.4) and Sn is the spline projector denoted as Sz* above, for the set of nodes Z* = (Ej). We use the series expansions

(7.34)

where

(7.35)

Clearly,

(7.36)

Snev = A~l E z~ v ~=v(n) o

E ~=v(n)

r21~1

R-2v + R2v-2n 1 - R-Zn if v = O,l, ... ,n-1 •

These formulas follow easily from (7.10) by using ev for f.

For the error in approximating ev by Snev we have

zV0 - b"(z0 ) = A~l E (zv - z~)R-2 1~1 v v ~=v o o (7.37)

where we have we assume n bound E(n)

zo

A~l[z~(l- zg)R-zlv+nl + z~(l- z~n)R-2Iv-nl]+ 8v

introduced the remainder term 8v• For simplicity is even, n =2m, and temporarily, lzol ~ 1. To

from above we consider first the terms with lvl ~ m.

I. If lvl ~ m then by (7.35), Av ~ R-2v and one obtains for contribution of the 8v terms to the difference kz0 (z0 ) - Snkz0 (z0 ), with some work:

Page 129: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 121

(7.38) m I I v -21 vI I In -n R2 - R-2 _r. Zo R !:J.\1 < 2 Zo R R2 +r2- I 12 - I Zo I =-2 on v --m+l zo

where

[ R-2n <lzai/R)2n] on= max 1- <lzoi/RZ)n ' 1 - R 2n , hence

(7.39) lim on = 0 .

n+oo

For the principal terms in (7.34) one finds similarly

(7 .40)

Thus,

(7. 41) I I n R2 - R-2

< 2 ( zo /R) 2 2 I 12 I I 2 (1 + on) • R + R- - z0 - z0 -

II. For terms with lvl > m the error is treated as due to truncation. One finds

I r. zvR-2Ivl (zV- b (z )I < I r. lz 12vR-2Ivl lvl2:,m o o v o - lvl~m o

(7.42) 2 21 I I Zo I n R2

+ lvf~m lzol lbv(zo)IR- v < 6(-R-) R2 -lzolv

From (7.41) and (7.42) we obtain

(7.43) ( ) 2 R2 - R-2

Ezn (!;iR) <(I zi/R)n2(l+on) R2+r2=jzl2 -lzl-2 +

Putting lzl = r and observing that

1 + (r/R) 2 1 - (r/R)2'

Page 130: Optimal Estimation in Approximation Theory

122 MICHAEL GOLOMB

we have the slightly less accurate bound

(7 .44) [ (n) 12 1 + (r/R) 2 Ez (~R)J < 8(1 + on)(r/R)n 1- (r/R)Z '

R > lzl = r ~ 1.

Next, we determine a lower bound for the error. the point z1 = re-~i/n (or re(2k+l)~i/n, k e E). Theorem 3.2

(7 .45)

We choose By

The coefficient of zm (m = n/2) in the Laurent expansion of kz - Snkz is, by (7.34), R-2m(zy- bm(z1)). Because of the or!hogonalfty of the terms in this expansion it follows that

(7.46) E~~)<~r> > R-mlzi- bm(z1>1

= R-mA-ll<zm- z -m)R-2m + - k (zml- z~l)R-21~1 I . m 1 1 ~=m~n)

~~~~3m The infinite sum in (7.46) is treated again as a truncation error. One finds since Aml > R-2m :

(7 .47) E~~)(~R) > 2(r/R)m(l - nn) ,

2(r/R2)n nn < 1 _ (r/RZ)n

where

lim nn 0 hence n+<X>

Putting (7.44) and (7.47) together we have

(7.48) (r/R)n/22(1- nn) < sup E~n)(~R) lzl = r

< (r/R)n/2[8(1

1 ~ r < R •

A symmetry argument shows that these bounds are valid with r replaced by r-1 if R-1 < r s 1. In particular, the error bounds are the same for lzl = r and lzl = r-1. By the maximum principle it follows that these bounds also hold for

(7.49) 1 ~ r < R •

In particular, we have

Page 131: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 123

Theorem 7.2. If E(n)(~R) denotes the maximum of the minimal error for recovefy of the ball ~ C ~AR) in the annulus {r-1 s lzl S r}, 1 ~ r < R, from data on the set (e2~ik/n)k=O 1 _ then , , •.• ,n 1

(7.50) (r/R)~ n+co

8. ~(AR),- .~'(Sb)-AND ~(ER)-SPLINES

1. By ~(AR) we denote the subspace of ~(AR) of Section 7 whose elements are functions f with the symmetry property

(8 .1) f(z) = f(z-1),

The orthogonal complement of this space consists of functions with the symmetry property g(z) = -g(z-1), The functions

(8.2) zk + z-k

2 k = 0,1, ...

g

form a complete orthogonal system in ~(AR), and their norms are

leo II..,/_ = 1 • llekll..,/ _ = 2-~Rk , k = 1,2, ... ~ (AR) .w (AR)

We consider again the special case of symmetrically dis­tributed interpolation nodes on the unit circle.

(8.3)

(8.4)

The spline Sz f *

(see (7.10)) now becomes

n-1 A aj (z) E fj j=o aj (1)

aj = (R2n-2j - R2j-2n)ej + (R2j - R-2j)ej-n

1 n-1 f = - L f(Ek)E-kj , j n k=O

j 0,1, ... ,n - 1.

A more suitable form for computation is: 1 R2n + R-2n - 2 n-1 A _ _

Sz*f = 2 R2n + R-2n _ 2~ j~O fj[(l+Aj)ej + (1-Aj)en-jl

(8.5) 1 + r2n 1 _ R4j-2n Aj = 1 _ R-2n • l + R3j-2n j O,l, ••• ,n-1 •

Page 132: Optimal Estimation in Approximation Theory

124 MICHAEL GOLOMB

Clearly, the sum in (8.5) could be written with only half the number of terms since fj = fn-j and Aj = -An-j· It follows that Sz*f is the product of the rat1onal function (R~n + R-2n - 2)/(R2n + R-2n - 2en) (whose value, in most cases is indistinguishable from 1) and a linear combination of the 1 + [n/2] functions (1 + Aj)ai + (1- Aj)an-j (j = O,l,:··•[n/2]; 1 + [n/2] is also the number of the independent data f(eJ) in this case.)

In analogy with (7.11) we set n-1 A

(8.6) Lz*f(z) = j~O fj aj(z)/aj(l) ,

and use the knots ~j = R2e-j, ~! = R-2ej. We observe that Lz f(~j) = Lz f(~j). Thus, we obtain the following partial fr~ctions exp~nsion of Sz*f from that in (7.13):

(8. 7)

The error bounds for recovery of the ball ~R C M'(AR) , in P.articular (7.48) and (7.50), are also valid for the ball ~R Cb((AR)· Although in this case the number of independent data is O»lY 1 + [n/2], this is compensated for by the symmetry property of ~R·

2. The main interest in the space bf(AR) derives from the fact that we obtain from it, by changes of the independent variable, other important spaces. We set first

(8.8)

The annulus AR of the z-plane (covered infinitely often) trans­forms into the strip

(8.9)

of the w-plane, the unit circle carrying the data points e2nik/n becomes the unit interval 0 ~ Rew ~ 1, Imw 0, with the equi­distant nodes k/n (k0 = O,l, ••. ,n-1). If f e M'(AR) then the transformed function f:

(8.10)

is analytic in the strip Sb, has period 1, is even, and has an expansion

00

(8.11) f(w) v=-oo

a e2nivw with \1

00

v=-oo

Page 133: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 125

Let b((Sb) denote the Hilbert space of functions f analytic in the strip Sb, of period 1, even, which have expan~ions of the form (8.11), with norm

(8.12) II f II ~(Sbl - c~- I"" 12R21" I).. • R - e2Wb

We remark tyat the same class of functions would be obtained if the functions f were required to have boundary values on ash which are in Lz(O,l).

0

Ogtimal recovery of the unit ball ~ of this space from data f(k/n) (k = 06l, ••• ,n-l) (only [n/2Y + 1 of these are independent since f(k/n) = !(1 - k/n)) is effected by the spline Snf, which is explicitely given by (see (8.5) )

S f = l cosh 4~bn - 1 n 2 cosh 4~bn - Cn

Aj coth(2~bn)tanh 2~b(n- 2j) , cj(w) cos 2~jw

o = 1 n-;:1 f(~)e:-kj fj '-' j = 0,1, ••• , n - 1. n k=O n

3. We now go from the w-plane of the preceding paragraph to the z = x + iy plane by the transformation

(8.13) z = cos 2~ •

The strip , Sb transforms into the interior of the ellipse:

(8.14)

which has its foci at z = ±1 and semiaxes whose sum is e2~b R. The equidistant data points k/n of the w-plane become the "Chebyshev nodes"

(8.15)

each of n = 2m

(8.16)

xk = cos(2~/n) k = O,l, ••• ,n- 1

them being cov~red twice, is even. If f e bf(Sb)

0 0 1 f(z) = f(w) = f( 2 ~ arc

except x0 1 and Xm = -1 if then the transformed function f:

cos z)

is analytic in the ellipse ER and has an expansion co co

lavi2R2v (8.17) f a0 + 2 I: avTv with I: < co v=l v=O

Here Tv is the Chebyshev polynomial of degree k:

Page 134: Optimal Estimation in Approximation Theory

126

(8 .17)

Let ~(ER) the ellipse norm

(8 .19)

MICHAEL GOLOMB

Tv(z) = cos(v arc cos z), v = 0,1, ••••

denote the Hilbert space of functions f analytic in ER, which have expansions of the form (8.17), with

Let ~R again be the unit ball in this space.

Optimal recovery of ~ from data f(xk) at the Chebyshev nodes (there are [n/2] + 1 of them) is effected by the spline

(8.20) 1 + R-2n • 1 _ R4j-2n 1 - RZn 1 + R4j-2n

lJ· = .!_ ~l f (Xk)E -kj n k=O

= .!. rf(l) + 2 [nrl/2] f(cos 2uk/n)cos 2ukj/n n t k=l

+' (-l)R.f{-1)]

(+' is to indicate that the last term is missing if n is odd.)

Observe that (8.20) is a formula for numerical analytic continuation from the interval [ -1,1] to the region ER. Sn f is a rational function of degree (n-l,n) for [n/2] + 1 data. The polynomial

(8.21)

interpolates f at the Chebyshev nodes xk.

We also record the partial fraction expansion of Snf for this important case. From {8.7) one obtains

(8.22) 1 - R-2n 1 n-1 /~j - 1 1 + r2n "n j~O ~j - z Lnf(~j)

~. = cos(i + 2bi)2u • J n

The poles ~j lie on the ellipse ER2• From (8.22) we obtain the

Page 135: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 127

formula for the optimal recovery of the Taylor expansion of f at some z0 € ER· If f(z) = E av(z0 )(z- z0 )V

coefficients and

oo ~o Snf(z) E av (z0 )(z - z0 )v then

v=o *

(8.23) /r,2 - 1

j

We also have the strange optimal quadrature formula:

(8.24)

for any z0 € ER.

The asymptotic values of the minimal error of Theorem 7.1 and 7.2 are readily applied to this case. For the error of recovery of ~ at any z0 € C we have

(8.25)

n n

lim Rz Et) (~R) = 2-~~ z0 + /z~ - 1!2J (z0 + /z~ - l)n - lJ. R-+oo

For the maximum of the minimal error in the ellipse Ep with p < R the result is

(8.26) lim n/E~n)(~R) (p /R) ~ n-+oo

9. COMPARISON WITH n-WIDTHS

The concept of n-width of a set and its usefulness in approximation theory are by now well known (for survey articles see [8] and [9].) If ~ is a compact set in a normed vector space !!£ then the n-width (n 0,1,2, .•. ) of ~ with respect to g[ is defined as follows:

(9 .1)

Here inf

g € M

d (n) (~) = inf sup inf Jlf - gJI. '!!£ M €~n f € ~ g € M

~n is the set of n-dimensional subspaces If- gU is, of course, the distance of f

of Bl, from the subspace

M, and by taking the supremum of these distances for all f €~we get the "deviation" of ~from M. dW) (~) measures the minimal deviation that ~ can have f~ an n-dimensional subspace of !!£. This is clearly a lower bound for the error in approximating f by linear combinations of any n elements in !!£, where the choice of them does not depend on f, but only on the class ~.

Page 136: Optimal Estimation in Approximation Theory

128 MICHAEL GOLOMB

If the n-width of ~ is attained for some M* e ~n then M* is said to be an optimal manifold for n-dimensional approximation of :£in !!£.

We will show that the spline spaces considered in Sections 5-8 come close to being optimal manifolds for certain function classes. Also, the error bounds found for the recovery schemes supply useful upper bounds for some n-widths that are difficult to calculate.

1. We consider first the space bf(DR) of Section 5 with the unit ball ~R• Given r, 0 < r < R, we define two new topologies on ~DR), which are coarser than the initial one. One is that of bf(Dr), with r replacing R; the other one is an Y~-norm:

llfllu (D) = sup lf(z) I . ..Zco r z e Dr

(9.2)

In either topology the ball ~R is compact, hence the n-widths dSf) (~R) and d (n) (D ) (~R) are finite. Since 9JR can be ~ror> ~ r characterized by the quadratic inequality

(9.3) ~R = {f = ~ a e : ~ Ia 12(R/r)2vr2v ~ 1} , v=O v v v

where r"' is the bf<Dr)-norm of ev , it follows from Kolmogorov's theory of n-widths of "ellipsoids" (see [10], [9, Chapter 9.4]) that

(9 .4)

and that ~n) =

9JR in bf(Dr).

(9.4) also gives

(9.5)

sp(e0 , e1 , ... ,en-1) is an optimal manifold for Since II f llyco(Dr) 1!: If ~~~(Dr) for any f e ~DR), a lower bound for the other n-width:

In (5.22) we found the error E~n)(~R) for recovery of ~R at z from n data values on the circle {lzl r 0 }. The maximum of E~n)(~R) for {lzl = r} must be an upper bound for (9.5), and this is true for any r 0 • Using the limit as r 0 ~ 0 we obtain

(9 .6) (n) (r/R)n (r/R)n ~ ~(Dr)<9Jr) S [l _ (r/R)Z]3/4

Clearly, the bounds of (9.6) are rather sharp for the ~(Dr) n-width of 9JR. As far as I know the true value has not been found ret. If we denote the maximum of the recovery error E(n)(9JR) for {lzl = r} from data on {jzl = r 0 } by E(~)(9JR) the~ we have by (5.24)

Page 137: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 129

(9. 7) lim n/E$n)(~R) r/R if r > r 0 n + oo

Thus, by (9.4)-(9.7), if r > r 0 :

(9.8) lim n + oo

= r/R

This means that optimal recovery of the unit ball in ~(DR) from data on the set {r0 e2nlk/n}k=O 1 ... n-l is also an asymptotically optimal linear approximation if, r > r0 • This is not so if r < r 0 •

Equation (6.10) shows that (9.8) also holds if E£n)(~R) refers to the error of recovery from Hermite data at the point 0.

2. We next consider the space ~(AR) of Section 7. Again we introduce two coarser topologies: that of ~(Ar) for 1 ~ r < R and that of 5t00 (Ar) with norm:

(9.9) I fll!£. (A) = sup lf(z) I· oo r lzl €Ar

As before, the ball ~R is compact in either topology. Proceeding as above one finds easily the ~(Ar)- width:

(9.10) d(n) (~R) = (r/R)m for both n = 2m and n = 2m+ 1. t:f/(Ar)

An optimal manifold for the case n = 2m + 1 is

(9 .11)

zk). If f

II f 11 2 00

la.vl2r21vl 00

+ r-2v) E < E :~'<Ar) v=-oo v=-oo

la.vl2<r2v

1 f lf(z)l21dzl +-1- f lf(z) 12 1dzl = 2nr 27Tr lzl=r lzl,.r-1

5: sup lf(z)l 2 + sup lf(z)l 2 lzl= r I z I r-1

< 2 sup lf(z)l 2 • -r-1 5:lzl=>-r

Therefore, I fi~(Ar) 2::. z-~llfi~Ar) bound for the ~(~) n-width:

and (9.10) supplies a lower

Page 138: Optimal Estimation in Approximation Theory

130 MICHAEL GOLOMB

An upper bound is given by the upper bound (7.44) for the spline errors. Thus

(9.12) 2~(r/R)(n/2] < d(n) (~) < r8(l+o )1 + (r/R)2l\r/R)n/2 ~(Ax-) R ~ n 1 - (r/R)2J

For the asymptotic values we have, by (9.10)-(9.12) and (7.50):

(9 .13) lim njE~n) (~R) = lim nj~~A ) (~R) n -+ "" n -+ "" "" ·-r

= 2[n/2] /d(n) (~ ) = (r/R)~ t ti/(Ax-) R

Thus: Optimal recovery of the unit ball in ~AR) from data on the set {e2~1k/n}k- is also an asymptotically optimal li . . -0,1, ..• ,n-1

near approx1mat1on.

3. Little needs to be changed if ~(AR) is replaced by the space ~AR) of Section 8. Using the above terminology we find in place of (9.10)

(9.14) (r/R)n-1

An optimal manifold is in this case

( (n) - {- - - } 9.15) M* - sp e0 , e1 , .•• ,en-1

(ek(z) = t<zk + z-k)). (9.17) is replaced by

lim n/E?n-1) ~R) n+oo

n-l~(n) (~R) = r/R , 1 s r < R ~(~)

Consistent with previous notation E~2n-l)(~R) refers to the maximum on 1Ar of the error of recovery of ~R from data on the set {e2~ik 2n-l}k=O,l, .•. , 2n_2 , of which there are only n independent ones.

For the space ~(ER) of functions analytic in the ellipse ER the corresponding results are:

(9.17)

Page 139: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS

Here r and R refer to the sums of semiaxes in the confocal e~~fp~es Er and ER, respectively. An optimal manifold is M~ - sp(e0 ,el•···•eu-l). Also

(9.19) lim njE~2n-l)(~R) = n+oo

lim n/d32,(Er) (~R)

n-1/d (n) (~R) r/R. M'(Er)

131

It should be observed that, whereas the optimal manifold of dimension n for ~R in M'(Er) is the space of polynomials of degree ~ n - 1, the n-dimensional spline space for optimal recovery is a space of rational functions of degree (2n - 2, 2n 1) (see (8.20)), with fixed denominators, and numerators that are spanned by 1 and fixed linear combinations of the Chebyshev polynomials Tj and T2n-l-j (j = l, ..• ,n- 1).

10. STABILIZATION OF THE RECOVERY SCHEMES

It has been pointed out by numerical analysts (see, for example [11], [12]) that analytic continuation of functions only given by their values in some infinite compact set is not a well­posed problem. Small errors in the data can produce arbitrarily large errors of the continuation. If in the problem of Section S, where f €~ is to be recovered in a disk Dr, the data f(r0 e2nik/n) are replaced by fE(r 0 e2nik/n), where fE(z) = f(z) + dz/r0 )N, with E > 0 sufficiently small so that fE e~R• then the change in the data is of absolute value E, but the change in the continuation of the function at z, lzl = r1 > r 0 , is of absolute value E(rl/r0 )N, which becomes arbitrarily large with N. Moreover, this is so no matter how large the number n of data is.

The sequence of spline values {Snf(z)}n=l,2, .•. represents a valid numerical method of analytic continuation from the data set to z e Dr provided it is true that

(10.1) lim sup lf(z) - Snf(z)l = 0. n+oo zEDr

Indeed this convergence follows immediately from the error bound (5.24). However the proof requires that f e~R (more generally that f e M~ for some M < co) and consequently ~ Snf II ~ M for all n. In general, given discrete data do not allow this conclusion. Numerical data are necessarily of limited accuracy, if for no other reason but that they are given as finite decimals. One cannot be sure that the sequence of data vectors {Fn} inter­polate a function f e~. It is more reasonable to expect that

Page 140: Optimal Estimation in Approximation Theory

132 MICHAEL GOLOMB

they interpolate a bounded measurable function g that is uniformly close to the function f which is to be continued. But then the data must include some information that counteracts the instability of the continuation process. An explicit a priori bound for the function f (that is, ltl61~ ~. ~ a known number) is sufficient for this purpose and is often available. We show how, under these conditions, the splines can be modified so as to achieve stable analytic continuation (for a similar, but not identical procedure see [12]). We carry out the details only for the problem treated in Section 5. The function class to be recovered is the unit ball ~R in ~DR)! the data are given in the set {r0 £g, ... ,r0 £g-l} where en = eZTI~/n, and the approximation is to be uniform over the disk Dr with r 0 < r < R.

Specifically, we assume: and a function g € ~oo<lzl

(10.2) !til M(Dr)

Given are positive numbers r 0 ) such that

(10.3) sup lf(z)- g(z)l ~ ~ lzl = ro

£' ~

The function f is to be recovered in the disk Dr, r 0 < r < R, from the data vectors Gn {g(r0 £g), ... ,g(r0£~-1 )} (n = 1,2, .•. ).

We define

(10.4)

1 - (ro/R)2n

1 - (r0 z/RZ)n

It is seen that the denominator in Sng (see 5.17). We prove

Theorem 10.1.

lim sup lsn,£(z) - Snf(z) I 0 £+0 z€Dr

uniformly in n.

distinguishes sn,£ from

The proof is based on a few lemmas. We write for the ~(Dp) norm more simply

(10.5) lltllp = {2;P f lt(z)l 2 ldzl}~ lzl= P

Page 141: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 133

and write e for E/~.

Lemma 10.1. If (r/R) 2n ~ 1- (1 + e)-2 and (10.2), (10.3) hold then

(10.6)

(10. 7)

lsn,E - Snf~r < E 0

~sn,E - Snf~R < M = ~ + E/2.

Proof. kBy the definition of sn,E and Snf, and writing zk for r 0 En:

1 _ (r /R)2n n-1 li s - S fll < 0 II I: (g. - l·)e·/rj 11 n,E n ro - 1 - (ro/R)Zn j=O J J J o ~o

(10.8)

By (10.3):

{10.9)

n-1 I A A l2}~ { I: gJ. - fj j=O

n-1 . (1/n){ I: (1 + €(R/r 0 )J)-2

j=O

·lnrl [g(zk)-f(zk)(l+E(R/r0)j]E~jkl 2 }~ k=O

~ (1/n) nrl (l+£(R/ro)j)-2 1 nrl[g(zk)-f(zk)]E~jkl 2 }~ j=O k=O n-1 n-1 L + (1/n) I: (l+E(R/r0 )j)-2€(R/r 0 ) 2jl I: f(zk)E~jkl 2 }~ j=O k=O

= I + II.

00

If f(z) I: avzv and v=O

Ev=j(n)avr~, hence

~ I: Ia Rvl2 I: (r /R)2v v=j(n) v v=j(n) 0

= (ro/R)2j[l- (ro/R)2n]-l I: lavRvl2 • v=j (n)

Page 142: Optimal Estimation in Approximation Theory

134

Thus, if the condition of the Lemma is satisfied

(10.10)

~elf~ ~t. ~(DR)

By (10.8)- (10.10), (10.6) is proved.

To prove (10.7) we proceed as above and obtain

(10.11)

MICHAEL GOLOMB

where III and IV are the terms corresponding to I and II above. Using the above inequalities and the conditions of the Lemma, we have

(10.12) III

< c./2:£ = )l/2.

n-1 2 ~ IV s { E (1 + €(R/r 0 )j)-2€2(R/r 0 )2j E lavRvl }

j=O v=j(n)

x [1 - (ro/R)2n]-~

< (1 + £){v!olavRvl2}~ = (1 + £)~fii~(DR) s (\l + c.)/2.

By (10.11)- (10.13) we obtain (10.7), so the Lemma is proved.

From (10.6) and (10.7) we derive now that llsn- Snfllr is small with £. For this we use an L2-version of Hadamard's 3-Circle Theorem.

Lemma 10.2. Suppose 0 < rt < r2 < r3. If in the closed annulus {z: r 1 ~ zl s r3} then

(10.14)

f is analytic

00 2 Proof. If f(z) = E avzv then II f II~ = E lavl r 2v. The

function '' defined by v;(;) = E lavl 2z2V is also analytic in v=-oo

the above annulus, hence by Hadamard's Theorem

2 log(r3/r1) 2 log(r3/r2) 2 log(r2/r1) M2 S M1 M3

Page 143: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 135

where M~ = I slp I Hz>l • But Mk = llf llrk (k = 1,2 ,3), hence the z "'rk Lemma is prove •

The above proof shows that the Lemma remainr true if f analytic in the open annulus and the limits of lfl21dzl

lzl-r r -+ r1 + 0, r -+ r3 - 0 exist. -

is as

If Lemma (10.2) is applied to (10.6) and (10.7), one obtains log(R/r) log(r/r0 )

(10.15) II II log(R/r0 ) M log(R/r0 ) sn,£ - Snf r < £ if r 0 ~ r < R

or, if we set

r/r0 = (R/r0 )a ,

(10.16) I sn,£ - Snfllr < £1-aMa

This inequality shows that the ~(Dr)-norm of sn,£ - Snf becomes small with £. We now show that this is also true for the sup norm over Dr· For this purpose we prove

Lemma 10.3. Suppose 0 < r 0 < r < r1 , and f is analytic in the (open) disk Dr1 with If llr1 < co Then for every T, 0 < T < 1

log(r]/r7 ) log(r,/r0 )

(10.17) suplf<z>l ~ (_!:!__ + -L_r_) lfll log(rllro)l!fl' log(rlho> r1-r 1--r r1-r r 0 lr1 Z€Dr

where r, = -rr + (1 - -r)rl·

Proof: For lzl r we have

lf(z)l = l2!i J ~(~? zl lr,;l=r,

~ L!r f lf(r,;) l 2 ldr,;lf1{~ f lr,; - zl-2 ldr,;l}~ lr,;l=r, lr,;1=r,

S. If llr r: T r, r

Using Lemma 10.2 with r1 r 0 , r2 = r, , r3 = r 1 , we get (10.17).

We return to the proof of Theorem 10.1. If Lemma 10.3 is applied to the function sn,£ - Snf , with r1 = R , and (10.6) and (10.7) are used, one finds

Page 144: Optimal Estimation in Approximation Theory

136 MICHAEL GOLOMB

(10.18)

sup lsn,E(z) - Snf(z)l z € Dr

log(R/rr) :;;, (-R __ + _T __ r_)E log(R/r0 )

R - r 1 - T R - r

rT Tr + (1 - T)R.

If we set

Tr + (1 - T)R ro

then (10.18) takes the simpler form

log(rx/ro) M log(R/Ro)

( R T r ) l-ST BT sup lsn,E(z) - Snf(z)l S ~ + ~ R _ r M ·

z € Dr (10.19)

Each T between 0 and 1 provides an upper bound for the differ­ence on the left.

By taking the limit of (10.19) for E 7 0, we seem to obtain the conclusion of Theorem 10.1. However, this argument ignores the condition (see Lemma 10.1)

(10. 20)

By taking n sufficiently large Pn can be made smaller than 1 + E/~ for a given E > 0, but not for all E > 0. Yet this difficulty is easily overcome. Condition (10.20) was used only in (10.10) and (10.12). Without it (10.10) and (10.12) give

II < E -2xL__ IV < ~ p 21+€' 2 n

and Lemma 10.1 remains true with E, M replaced by

(10. 21) E ( --..e.n_ ~ El = z 1 + l + s ) , M = z(l + Pn) .

Clearly li~+oo Pn = 1. If N is so chosen Pn < 1.1 for n ~ N then El < l.OSE and inequality (10.20) holds for all n ~ N if the slightly larger El, M1. This completes Theorem 10.1.

that, for example, M1 < l.OSM and the E,M are replaced by the proof of

Page 145: Optimal Estimation in Approximation Theory

OPTIMAL RECOVERY SCHEMES FOR ANALYTIC FUNCTIONS 137

REFERENCES

[1] Golomb, M. and H. F. Weinberger. Optimal Approximation and Error Bounds. Symposium on Numerical Approximation, R. E. Langer ed., Madison 1959, pp. 117-190.

[2] deBoor, C. R., and R. E. Lynch. On splines and their minimum properties, J. Math Mech., 15(1966), 953-969.

[3] Miccheli, C. A., T. J. Rivlin and S. Winograd. The optimal recovery of smooth functions, Numer. Math. 26(1976), 19-

[4] Micchelli, C. A. and Allan Pinkus. On a best estimator for the class ~r using only function values. MRC Technical Summary Report, February 1976.

[5] Knauff, W. and R. Kress. Optimale Approximation linearer Funktionale auf periodischen Funktionen. Numer. Math. ~(1974).

[6] Knauff, W. and R. Kress, Optimale Approximation mit Nebenbedingungen an lineare Funktionale auf period­ischen Funktionen. Numer. Math.

[7] Golomb, M., Approximation by periodic spline interpolants on uniform meshes. J. Approximation Theory !(1968), 26-65.

[8] Tikhomirov, V. M. Widths of sets in function spaces and the theory of best approximation (Russian). Uspekhi Math. Nauk, t.XV, 3(93) (1960), 81-120.

[9] Lorentz, G. G., Approximation of Functions. New York, Chicago, San Francisco, Toronto, London, 1966.

[10] Kolmogorov, A., Ueber die beste Annaeherung von Funktionen einer gegebenen Funktionenklasse. Annals of Math. 37 (1936)' 107-111.

[11] Douglas, J., A numerical method for analytic continuation. Boundary Problems in Differential Equations, Madison, 1960, pp. 179-186.

[12] Miller, Keith. Least squares method for ill-posed problems with a prescribed bound. SIAM J. Math. Anal. 1(1970), 52-74.

[13] Wilf, H. S., Exactness conditions in numerical quadrature. Numer. Math.~ (1964), 315-319.

Page 146: Optimal Estimation in Approximation Theory

138 MICHAEL GOLOMB

[14] Eckhardt, U., Einige Eigenschaften Wilfscher Quadraturformeln. Numer. Math. 12(1968), 1-7.

[15] Engels, H., tiber allgemeine Gauss'sche Quadraturformeln. Computing 10(1972), 83-95.

[16] , Eine Familie interpolatorischer Quadraturformeln mit ableitungsfreien Fehlerschranken. Author's preprint.

[17] Engels, H. and Eckhardt, U., The determination of nodes and weights in Wilf quadrature formulas. To appear in Abhandlungen Math. Sem. Un. Hamburg.

Page 147: Optimal Estimation in Approximation Theory

OPTIMAL DEGREE OF APPROXIMATION BY SPLINES

Karl Scherer

University of Bonn, Institute of Applied Mathematics

Wegelerstrasse 6, 5300 Bonn

I . THE PROBLEM

A description of spline - or piecewise polynomial functions -involves three characteristics: the nature of their pieces, their location and the smoothness at the connections. Here the following classes of spline functions are considered: Given a Partition rr = {x.}~ of knots x. of the interval [a,b]

1 1=0 1

(I) a= x0 <x1< ••• <~ = b,

and integers n,k with -I<k<n-2 we set

(2) Sp(ll,n,k) = {s(x): s(x)l ( )£P, s(k)(x) continuous}, xi,xi+I n

where Pn denotes the space of all polynomials of degree <n-1

(In case k = -1 a spline function s(x) may be discontinuous at the knots its value there being defined as the limit from the right.).

Correspondingly we define the best approximation of integrable functions f

L (a,b) p

(3) Ep(f;IT,n,k) in£{ II f-s II : se:Sp(IT,n,k) }, l.::_p~_,,., p

In case p = oo we consider the class of continuous functions on [a,b]. The norm I I· I lp denotes the usual Lp-metric.

139

Page 148: Optimal Estimation in Approximation Theory

140 KARL SCHERER

Now, given a sequence {IT 1 }~=o of partitions, we say the degree

of (best) approximation of fe:L (a,b) p

is a(>o) if

for some constant C>o where for a partition IT we have set

(5) IT= max(x.+1-x.) , IT= min(x.+ 1-x.). . 1 1 . 1 1 1 1

There are well known estimates in spline theory of the form

C = C(n,p) being independent of IT and f, if f belongs to

the Sobolev space Wj(a,b) of functions with absolutely continuous p (j-1)-th derivative and j-th derivative in L (a,b), j=l, •.. ,n.

p These estimates show that the degree of (best) approximation for

the classes Wj(a,b) is j. They may indeed be realized by a p bounded linear projection from L (a,b) to Sp(IT,n,n-2) (cf. [2] p for p = oo and [7] for 12P<oo).

One may now ask after the sharpness or optimality of such estimates. More precisely we want to discuss the following questions

I. For a given a>o what is the optimal class K of functions which have this degree of best approximation?

II. Is there an optimal degree a of best approximation for all classes of functions?

The best way to solve these questions is to try to prove some sort of inverse estimates to (4), (6), i.e. estimates from above by the best approximation (3). This will be shown in the next section where the known inverse theorems are presented. We shall see that the answer depends on the nature of the sequence

{IT 1 }~=o of partitions as well as on the parameters n and k. But

before doing so we must refine the estimate (6) somewhat. To this end we introduce the n-th modules of continuity in the L -metric

p (o<t<oo)

Page 149: Optimal Estimation in Approximation Theory

OPTIMAL DEGREE OF APPROXIMATION BY SPLINES

w (f;t) n p

SUJ?

o<jhj<t

SUJ?

o<Jhj<t

{ f x,x+nh E:(a,b)

sup J L'l~f(x) I x,x+nh E:(a,b)

141

' p 00

where is the n-th forward difference with increment h.

Then one can prove (cf. [7])

THEOREM 1: There is a constant C>o independent of f and IT such that

(7) Ep(f;IT,n,n-2) < Cw (f;IT) - n p

This result generalizes and refines estimate (6) as one can see by the introduction of the following Lipschitz-classes of functions

(8) Lip(a,n;p) {feL (a,b): w (f;t) = O(ta),t~+}. p n p

Note that in view of the well known fact

(9) -n t W (f;t)~,t~+ ~ fEP n n

only the Lipschitz-classes Lip(a,n;p) with o<a<n are meaningful. For these we know by Theorem I

where o<a~n, -1~~n-2.

But in view of

for j = 1, ••. ,n the class Lip(j,n;p) is larger than W~(a,b), hence (10) is stronger than (6).

It will turn out below that within certain ranges of a asser­tion (10) is invertible so that the classes Lip(a,n;p) give the answer to problem I.

Page 150: Optimal Estimation in Approximation Theory

142 KARL SCHERER

2. INVERSE THEOREMS

Basic for the following inverse theorems is the following observation: Suppose x and x+nh lie in one segment (xi,xi+I)

of a partition IT of [a,b], then for any SESp(IT,n,-1)

since then s takes the values of one polynomial EP for In . n d . x+rh, r=o, ••• ,n. For nh _ small enough th1s almost woul g1ve

the inverse of (7) (for p=oo) if the argument would not fail in a neighbourhood of the knots. This motivates the following (cf. [BJ)

DEFINITION: A sequence {IT1};=l of partitions is called mixed

if there exists a constant d>o such that

(12) sup dist(x,rr1) ~ dllk l>k

uniformly in xE[a,bi and k,k+oo.

Thus "mixed" guarantees that for each k,k+oo, and each xE[a,bJ one can find a partition rr1 , l=l(x)~k (low 1 would be

of no use in (II)) such that x, x+nh lie in

provided h is small enough, namely nh~dTik.

ly leads to

dTik n w (f;---) < 2 sup E (f;n1 ,n-I),

n n oo - l>k oo

one segment of rr1 ,

Hence (II) immediate-

and after some elementary manipulation to the well known result (e.g. [8])

THEOREM 2: Assume the sequence {TI1 };=l is mixed and Til+l ~ rr1 .

Then there holds

for some constant C independent of f (= continuous in [a,b]) and t.

Page 151: Optimal Estimation in Approximation Theory

OPTIMAL DEGREE OF APPROXIMATION BY SPLINES

It is not hard to show that the sequence of equidistant partitions satisfies the conditions of this theorem which allows now an answer to the initial questions I and II:

COROLLARY I: Under the assumption of Theorem 2 one has for -I ~2._n-2, o<a2._n

a)

b)

143

Here we have only used definition (8) of the Lipschitz-spaces and property (9).

In the case l~<oo the proof of a theorem similar to Theorem 2 is not so simple. I take the opportunity here to present a slight generalization of the result in [13}

THEOREM 3: Let the sequence {rr 1 }~=l of partitions, rr1

have uniformly bounded mesh ratios rr1 1~1 and suppose it satisfies

the following L-mixing condition:

( 14) \ N d' ( (k) rr )np+I > e-rr np+I \ L u.l 1St xi ' 1 - k ' l. al <oo

l>k l>k

uniformly in i .and k, k+oo.

holds

Then for f£L (a,b),I~<oo, p

w (f;Tik) < C sup E (f;rr1 ,n,-I) n p - l>k p

with C being independent of f and k·

there

For the proof one needs the following (concerning a proof see (9])

LEMMA: For any s£Sp(IT,n,-I) and £>o

(IS) w (s;£) < C(I+£/IT)I-I/p£!/p n p- -

there holds

for some

s(j)(x.+) 1

C only depending on p and n, where [ s] ~ 1

s(j)(x.-) denotes the jump of the j-th derivative of 1

s at the knot x. of the partition IT. 1

Page 152: Optimal Estimation in Approximation Theory

14.4 KARL SCHERER

Proof of Theorem 3: By familiar properties of the moduli of con­tinuity one has

By (15) it follows

w (sk;nk) < C max {? ~p+l I [sk-s 1}J1: lp}l/p n p - o<k<n-1 1.

for any s 1£Sp(rr1 ,n,-l) and dist(xik) ,rr 1) = ~i(k,l)>o. Applying

Markov's inequality on the intervals (xi (k)±~i (k,l), xik)) to

I s~j) (xik) ±) - s~j) (xik)) I and denoting by II II p, i the Lp-

norm with respect to the interval gives

(k) (k) (x. - ~.(k,l),x. + ~.(k,l))

l. l. l. l.

and hence by (14)

n rr:ip+ I . k l.

-jp-1 I I P 1/p ~. (k, 1) I sk-sl I ·} · l. p,l.

which, together with (16), establishes Theorem 3 (for simplicity each constant is denoted by C).

It follows that for sequences of partitions with uniformly bounded mesh ratio and the mixing property (14) we can answer questions I and II for the L -metric, l<p<oo, as in Corollary I.

p -Again the sequence of equidistant part1.t1.ons is an example for which this applies. In fact, by a lemma in [6] it can be shown that (14) holds with

(ll {1 ~1: t~ .:_1 _gk 0 , otherwise

Page 153: Optimal Estimation in Approximation Theory

OPTIMAL DEGREE OF APPROXIMATION BY SPLINES 145

However, there are obviously sequences of partitions, which cannot be mixed in any sense at all, namely when rr1+ 1 c rr 1 , for

1 = 1,2, ... In this case one can prove

THEOREM 4: Let {II 1 }~=l be a sequence of nested partitions, i.e.

= 1,2, ... and II = {a,b}. For each 0

f£Lp(a,b),l2P~oo, and -I~~n-2,n>o, we have

j+l+l/p ~ -j-1-1/p . . } (17) w0 (f;~)P~cnk 1~ 1~1 {EP(f;rr1 ,n,J)+EP(f;II1_ 1,n,J)

where C is a constant only depending on n and p.

Sketch of the proof: In view of Sp(II1 ,n,k) c: Sp(II1+1,nk) we apply

the classical Bernstein-argument for We choose s 1£Sp(II1 ,n,k) such that

inverse approximation theorems. llf-s1 ilp = Ep(f;II1 ,n,j),

k write f = f-sk + L (s1-s1_ 1) + s

1=1 ° and obtain

(18) ( ) { ( ") ki_rrj+ 1+1/plls(j+ 1)1ioo} w f;~- < C E f;IIk,n,J + k l

n ---,.. P - P 1=1

Then we apply a Markov-type inequality to (j+1) (j+I)

sl - sl-1 (on a

segment of rr 1) in order to pass over to the difference s 1-s1_1

and by the triangle inequality to (17).

As a consequence we obtain an answer for question I.

COROLLARY 2: Under the assumptions of Theorem 4 let {II1} have a

uniformly, bounded mesh ratio

f£Lp (a, b), 129~00 , and o<a < j+1+1 /p:

Thus the answer is the same as in Corollary 1 but with the restric­tion o<ct<j+l+1/p. Without this restriction the corollary becomes false as is demonstrated by examples in [9). Concerning question II the answer is negative since, according to the theorem of Bernstein, in view of Sp(II1 ,n,k) C Sp(IIl+l'n,k) there exist non-trivial func­tions with an aroitrarily higfi degree of approximation.

Page 154: Optimal Estimation in Approximation Theory

146 KARL SCHERER

What can we say about arbitrary sequences of partitions? In [9] it is shown that the answer is the same as in the case of se­quences with nested partitions. This follows from (cf. [9]).

THEOREM 5: Let {IT 1 }~=l be a sequence of partitions with

and IT = {a,b}. 0

Then the statement of Theorem 4 remains true.

The proof involves a combination of the arguments in Theorem 3 and 4. Instead of (18) one uses

the lemma from above and successive application for 1 of

09> . max {~lnrt-j I rs 11~lp}t/p < J+l~r~n-1 1

+ max {'lrrr-l-j[s I~lp}l/p . ~ -1 1-1 1 J+l~r~n-1 1

k, •.. , I

Besides other technical details the key step of the proof in the simplest ca~e j = n-2 of (19) is the observation that

(20)

where T1 = s 1 - s 1_ 1 and the ai,bi are right end or left end

points, respectively, of open intervals of length ~1 /4 which

are free of knots of rr 1 and rr1_1• This allows then an appropriate

application of a Markov type inequality to the terms of (20). Con­cerning further details the reader is referred to [9] .

Theorem 5 does not give an answer to question II after a possible optimal (or saturation) order of approximation. The above discussion showed that this may depend from whether the sequence of partitions is mixed or nested. Nevertheless there is the following result which shows that at least for smooth functions there is a limit in the degree of approximation.

THEOREM 6: Let the continuous on [a,b]

(n-1)-th derivation of f be absolutely and its n-th derivation be in L (a,b),l<p<oo

p -

Page 155: Optimal Estimation in Approximation Theory

OPTIMAL DEGREE OF APPROXIMATION BY SPLINES 147

or continuous if p = oo.

{Ill}~=! Then

-n Ep (f ;rr1 ,n,-1 ~.!!_1 -+ o,l -+ oo for a

implies that sequence of partitions with lim Ill = 0

1-+oo f is a polynomial of degree n-1.

Proof: I take the occasion to give a corrected version of the proof in [13] restricted to the case of polynomial splines. Inequality (15) and the same argumentation as in (16) give for any o<e:<~

(21) w (f;Ilk) < 2~ (f;Ilk,n,-1) + n - p- p

n 1/p + C (I + l!k /e: ) e: max

o_:j2_n-1

Now let u.(x) be the Taylor-polynomial of l. degree n-1 or degree n Then denoting by II· II · p,l.

for the

1312_oo or L -norm on

p

Jlf - u.JI · < l. p,l. (

e:n llf(n)ll · p,l.

e:n4l (e:)

p

f at x~k) - e: of l.

oo, respectively. (x.-e:,x.+e:) we have l. l.

, l3'<oo

,p=oo

for some function 4l(x) with lim Hx) 0. x-+o+

Again a Markov-type inequality gives

(22)

for l3'<00 , and for p = oo

(23) max i

e:j I [sk]J1: I 2_ C max IJsk-u.ll . i l. oo,l.

Page 156: Optimal Estimation in Approximation Theory

148 KARL SCHERER

Inserting (22), (23) into (21) yields

(24) w (f;II,_) n -.-. p

< c [(l+I4/£)nEP(f;IIk,n-I)+ll~{ ~lif(n) il~,i}l/p,p<oo

(l+~/£)nE00 (f;IIk,n,-l)+~~(£) ,p = oo

Now by assumption Ep(f;IIk,n,-1)

positive function with lim ~(x) x+o+

we have for k+oo

~~(llk) where ~(x) is some

0. Setting £ = ~~(llk)I/ 2n

n 1/2 (1+~/£) Ep(f;IIk,n,-1) .::._C~(]Jk) ~

Since furthermore meas { U (x~k) -£ ,x~k) +£)} . 1 1 1

k+oo the result follows from (24).

We conclude with the remark that most of the above results have been carried over to the more general case of spline functions which are pieced together by null-solutions of a differential operator

n di A = I a. (x)--.- with a (x)>a>o (cf. [ 10], (13]). Inverse theorems

1=o 1 dx1 n -have also been obtained in case of best approximation by splines when the knots are not fixed with respect to the functions to be approximated but only their number (see [12], [5], [3], [1]). Also the multidimensional case has been treated (cf. [II}, [4]).

REFERENCES

(I) J.BERGH- J.PEETRE: On the spaces Vp(o<p<oo). Boll.Un.Mat. Ital. IV Ser 10 (1974), 632-648

[2] C.de BOOR: On uniform approximation by splines. J.Approx. Theory I (1968), 219-262

[3} Ju.A.BRUDNYI: Spline Approximation and functions of bounded variation. Soviet Math.Dokl. Vol.I5 (1974), No.2

[41 Ju.A.BRUDNYI: Piecewise polynomial approximation, embedding theorem and rational approximation. Proc.Colloq. ~pproximation Theory, Bonn 1976, Springer Lecture Notes 1n Math. No.554

Page 157: Optimal Estimation in Approximation Theory

OPTIMAL DEGREE OF APPROXIMATION BY SPLINES 149

[5]

{6]

J7)

[ 8]

[9]

[!OJ

[ I I ]

H.G.BURCHARD: On the degree of convergence of piecewise polynomial approximation on optimal meshes. Preprint 1974, to appear

G.BUTLER- F.RICHARDS: An L -saturation theorem for splines. Canad.J .Math. 24 (1972)~ 957-966

R.DEVORE: Degree of approximation. To appear in Proc.Symp. Approximation Theory, University of Texas, Austin, 1976

R.DEVORE - F.RICHARDS: Saturation and inverse theorems for spline approximation. Spline Functions Approx.Theory, Proc.Symp.Univ.Alberta, Edmonton 1972, ISNM 21 (1973), 73-82

R.DEVORE - K.SCHERER: A constructive theory for approximation by splines with an arbitrary sequence of knot sets. Proc.Colloq.Approximation Theory, Bonn 1976, Springer Lecture Notes in Math. No. 554

H.JOHNEN - K.SCHERER: Direct and inverse theorems for best approximation by A-splines. Proc.Symp. Spline Functions, Karlsruhe 1975, Springer Lecture Notes in Math. 501, pp. 116-131

M.J.MUNTEANU- L.L.SCHUMAKER: Direct and inverse theorems for multi-dimensional spline approximation. Indiana Univ.Math.J. 23 (1973), 461-470

[12] V.A.POPOV: Compt. rend. Acad. bulg. Sci. 26 (1973), 10,1297

[13] K.SCHERER: Some inverse theorems for best approximation by A-splines. To appear in Proc.Symp.Approximation Theory, University of Texas, Austin 1976

Page 158: Optimal Estimation in Approximation Theory

MINIMAL PROJECTIONS

Carlo Franchetti

University of Florence

Istituto di Matematica U. Dini - V. le Morgagni 67/A

1 - Let X be a normed space, Y a closed subspace of X and denote by A the set of all projections P : X+Y.. A projection Q is minimal if IIQII ~ liP II, PEA. If IIQII = 1, Q is minimal since for any P, II PI I :<!: 1. A may be empty, if not Y is called comple­mented, this is the case when the dimension or the codimension of Y is finite. When dim Y<~ there is always a minimal element in A, but codim Y<~ does not imply this conclusion (see [3]). We will discuss the problem of characterizing minimal element in A or in some proper subset. This problem is trivial when X is an Hilbert space or Y is !-dimensional.

2 - This paragraph is an attempt to motivate our problem in two ways: one from approximation theory and the other from the theory of Banach spaces.

Suppose we want to approximate elements of X with elements of a subspace Y. Projections give a natural procedure for con­structing approximations: if we regard Px as an approximation in Y of X€X then the error is I lx-Pxl I· We have

llx - Pxll = II (x-y) - P(x-y) II = II (I-P)(x-y) II ~

~ II I - P II II x - Y II hence

llx-Pxll ~ III-PI I infllx-yii .. III-P!Id(x,Y)~(l+IIPII )d(x.Y) y€Y

So in some sense the "best" projection PEA would be one for which III-PI I is a minimum, one for which IIPII is a minimum is only a "~ood" one. However it may happen that II-PI I is a minimum iff IIPII is a minimum which is in particular true when III-PII=l+IIPII·

151

Page 159: Optimal Estimation in Approximation Theory

152 CARLO FRANCHETTI

The last equation holds when X=C(Q), with Q Hausdorff compact with no isolated points, and dim Y<~; for a proof (see [5]).

To explain the second motivation we first give a short outline of the theory of projection constants.

A Banach space X is a GA-space, A~l, if

a) for every superspace Z of X there exists a projection P:Z+X with II p II $ A.

The (relative) projection constant of X in Z is defined by

A(X,Z) = inf {j jPj I : P projects Z onto X}

if X is not complemented in Z we put A(X,Z) = ~. Q is minimal in the set A of all projections from Z onto X iff I jQj I = A(X,Z).

The (absolute) projection constant of X is defined by

A(X) = sup {A(X,Z) , Z is a superspace of X} .

It is clear that l$A(X)::;~. To understand the relevance of this parameter A(X) we recall the property a) is equivalent to each of the following two properties: b) For every superspace Z of X, for every Banach space Y and for every linear operator T:X+Y there exists an extension T:Z+Y such that II T II $A II T II · c) For every paid of Banach spaces Z, Y, z~y and for every T:Y+X there exists an extension T:Z+X such that I ltl j::;Aj jTj I.

A Banach space X is a GA-space for every A>A(X). For a dis­cussion of GA-spaces see [7], [14].

If dim X=~, then X reflexive or separable =>A(X) =~ (Grothendieck).

If dim X= n, then A(X)::; ~ ([13], [11]); this asymptotic estimate for A(X) is the best possible, for ex. A(l1 (n))~ ~ ([12]).

Given a Banach space X, of special interest are those spaces Z, z~x such that A(X,Z) = A(X). For ex. this equation holds when Z is a G1-space or when dim X<~ and Z = C(Q) with Q Hausdorff com­pact (see [13]). So any X with dim X<~ is a GA-space also for A = A(X) since minimal projections exist.

If Z=C[a,b] and dim X<~, for a minimal projection Q we have:

II Q II = A (X, Z) = A (X) , (X is a G II Q 11-space).

II I - Q II = 1 +II Q II , ( II I - Q II is a minimum) •

Finally we note that for many Z the subspaces X for which A(X,Z)=l have been characterized. For ex. if Z=L (T,E,~) lSp<~

p then A(X,Z)=l iff X is isometric to a space L (T',E',~'), ([1]).

Investigating minimal projections from p Z onto a subspace X is very different in the two cases A(X,Z)=l, A(X,Z)>l; we will con­sider the second one.

Page 160: Optimal Estimation in Approximation Theory

MINIMAL PROJECTIONS 153

3 - We review here the few examples of (non trivial) minimal projections already known. Consider the Banach space C of continuous 2~-periodic functions with the usual sup-norm. The operator P which maps any X€C into the partial sum x of its Fourier series is na

n • projection from C onto the subspace V of trigonometrical polynomials

n whose degree does not exceed n. Minimality of P was established by Lozinski in 1948, unicity was proved in 1969 ([4]). We have A(V ,C)=A(V >=I IP I I = A , where A are the classical Lebesgue con-n n n n n stants. If everything is interpreted with the L,-norm we have II P II= 1, since P is orthogonal. n n .

If Z is one of the sequence spaces c or 1 and if X is an hyperplane in Z, then A(X,Z) and minimal p~ojecttons (when exist) can be explicitily computed ([3]): the same is done if Z=l1 (n) or Z=lco(n) ([9]).

4 - We discuss now some recently discovered minimal projections. i) The space X=L1 (T,r,~). Assume that (T,r,~) is a finite nonatomic measure space, that

Y is a finite dimensional and smooth subspace of X(Y is smooth iff yEY, ~0 => y(t) + 0 ~-almost everywhere on T). Minimal elements of A are characterized in [8]. If Y is also rotund we have unicity, if Y is the subspace of linear polynomials there is also a unique mini­mal projection which is explicitily computed. The main results de­pend upon the use of a suitable "Lebesgue function". Recall that if L is any operator mapping C(Q) into itself the classical Lebesgue function AL may be defined by AL (t)= II toL II , where t is the evalua-

tion functional at tEQ (t(x) = x(t), XEC(Q)); we have ALEC(Q) and

IIALII=IILII· A "Lebesgue function" for an operator M on an arbi­trary Banach space should be a function AM(•) such that II Mil =

IIAMIIco• In our L1-case

defined by

the Lebesgue function for a projection

P :X-+-Y is

where L of co

* A = sup IP vi , p V€V

* P is the adjoint operator of P and V is a norm-one elements such that sup(y,v)=l IYI I

V€V easily seen that A EL p co and that IIAPIIco=IIPII·

The following result is of interest:

countable set in for any yEY. It is

THEOREM - If P is a minimal projection onto Y (under the same hypotheses), then the Lebesgue function A is constant ~-a.e. on T.

ii) The sequence spaces c0 and 11 .P

Assume that X=c0 or X=l1 , Y is a subspace of X with dim Y<co and P is a projection onto Y. The Lebesgue functions are defined by:

Page 161: Optimal Estimation in Approximation Theory

154 CARLO FRANCHETTI

A (t)=lltoPII, if X= c ; A (t) = lltaP*II if X=l1 • p 0 p

Minimal projections are characterized ([9]) for any Y if X=c ; if X=l then Y satisfies a Haar type condition. For n-dimensional0

Haar s&bspaces of 11 the Lebesgue function of a minimal projection is constant with the exception of at most n-1 points.

iii) The Rademacher projection. n Consider the topological space T consisting of 2 disjoint in-

tervals. If I. is the j-th interval, then we define functions y1 , ••• ,yn in J C(T) by setting yi(t) = (-1)* on Ij with k =

[(j-1)2i-n]. Here l~i~n, l~j~2n and [•] denotes the greatest in­teger. These functions are the classical Rademacher functions suit­ably modified for our purpose. The yi are an orthonormal system in the L2-sense, that is

CallY the linear span of they. and define the "Rademacher pro­l. jection" P:C(T)+Y setting:

n Px = E <x,y.> yi

i=l l.

XEC(T) •

P is minimal in A([lO]). For the proof we use two results. GrUnbaum ([12]) proved in 1960 that the projection constant of 11 (n) is:

(l) A (l (n)) = 2m+l (2m) m = [n-1] 1 22m m 2

Rademacher proved in 1922 ( [16]) that the Lebesgue function of this ~,>roj ection is constantly equal to the number in (1) , so II P II= I IAPI I = A.(l1 (n)). Now remark that Y is isometric to 11 (n), hence

A.(Y,C(T)) = A.(T) = A.(l1 (n)) = I IPI I and therefore Pis minimal.

5 - We have seen that the Fourier and the Rademacher projection are orthogonal in the L2-sense, also both possess a constant Lebesgue function. We consider here the role of orthogonality in the problem of minimal projections. Assume that X=C(Q), with Q Hausdorff compact; we say that a projection P from X onto a closed subspace Y is ~rthogonal ([2]) if there exists a nonnegative nonzero functional fEX such that

(2) f[(x-Px)] = 0 y€Y XEX •

If we put <a,b>f = f(ab), then <•,•>f is a pseudo inner product on X, it becomes a true inner product if x ~ 0 implies f(x2)>0 •

The set N of all orthogonal projections from X onto Y has a minimal element ([2]). For the general problem of existence of

Page 162: Optimal Estimation in Approximation Theory

MINIMAL PROJECTIONS 155

minimal projections in restricted classes see [2] and also [15]. Be­sides the classical orthogonal projections defined using polynomials orthogonal with respect a positive weight w (f(x) = J0x(t)w(t)dt), N includes also Lagrange interpolatory projections and many other projections. 2

Given a subspace Y of X with dim Y<oo let us call Y the linear span of all products y1y2 with yi€Y. For a positive functional

* f€X assume that <•,•>f defines a true inner product on Y

(y€Y, y~O => f(y2)>0). Let us take in Y an orthonormal basis {y1 , •.• ,yn} (with respect to <•,•>f); the projection Pf defined by

n n (3) Pfx = ~ <x;yi>fyi ~ f(xyi)yi

i=l i=l is orth~gonal in the sense that (2) is ~atisfied. Observe that to any f€X such that f~O; y€Y, y~O => f(y )>0 it corresponds one and only one projection which satisfies {3), this (orthogonal) projection is given by (3). One can see that the Lebesgue function of Pf

(A (t) = lltopfll> is given by" (t) = f(l~iyi(t)yil) hence pf pf

IIPfll =sup f(l~.y.{t)y.l). t 1 1 1

* Suppose now that a positive functional g€X is such that 2 2 g/Y = f/Y , then <•,•>f = <. •>

' g on Y; in other words f and g induce

on Y the same Hilbert space structure, P can be represented in the g

form Pgx = ~g(xyi)yi, where {y1 , •.• ,yn} is a <•,•>£-orthonormal basis in Y. All this suggests the DEFINITION ((10]). zet f be a positive functional with the property that y€Y, ~0 => f(y )>0. Af denotes the set

* 2 2 Af = {Pg:g€X ' g ~ 0 ' g/Y = f/Y } •

Af is a subset of N which has a minimal element. Generally is a rather "large" set, but in some special cases it reduces to a singleton; the structure of Af is investigated in [10] where, in particular, a simple characterization theorem for minimal elements is given. Statements on minimal elements of Af depend upon the critical

set Kf of the Lebesgue function (Kf = {t€Q:A (t) = I IPfl I>; when f pf

is supported on all Q (when x ~ 0 => f(x2)>0) we have special results. As an example of the theory we refer here a necessary and a suffi­cient condition for minimality in Af.

THEOREM 1 Assume that Q is infinite, l€Y, space of C(Q), supp f=Q, I IPfl 1>1 and Pf minimal

infinite.

y 2 is an Haar sub­in Af then Kf is

Page 163: Optimal Estimation in Approximation Theory

156 CARLO FRANCHETTI

THEOREM 2 Assume that 1EY2 and that A is constant, then Pf is minimal in Af. pf

Theorem 2 shows that the Fourier and the Rademacher projections are minimal in their classes A (actually both are minimal in the total class). Let us now constder for n~2 the Fourier-Chebyshev projection S (S is obtained taking the Chebyshev orthogonal poly-

n n nomials). In [6] it is proved that S is not minimal in the total

n class since AS has only two critical points; theorem 1 implies the

n nonminimality of Sn in Af.

We hope that the use of orthogonal projections can make easier to understand the nature of the (absolute) minimal projections, for ex. from C[a,b] onto the polynomial subspaces TI (even for n=2

n

minimal projections are unknown). A progress would be to prove or disprove that among minimal projections there is one which is finitely supported (representable in the form Px = ~x(ti)yi, where y,ETI and t.E[a,b]).

1 n 1

REFERENCES [1] S.J. Bernau- H.E. Lacey. The range of a contractive pro­

jection on an L -space. Pacif.J.Math. 53(1974), 21-41. p

[2] J. Blatter - E.W. Cheney. On the existence of extremal pro­jections. J. of Approximation Theory 6 (1972), 72-79.

[3] J. Blatter- E.W. Cheney. Minimal projections on hyperplanes in sequence spaces. Ann. Mat. Pura e Appl. (IV) Vol. 101 (1974), 215-227.

[4] E.W. Cheney, C.R. Hobby, P.D. Morris, F. Schurer, D.E. Wulbert. On the minimal property of the Fourier projection. Trans.Am. Math. Soc. 143 (1969), 249-258.

[5] E.W. Cheney - K.H. Price. Minimal projections, in Approximation Theory, A. Talbot ed. Academic Press, New York (1970),261-289.

[6] E.W. Cheney- T.J. Rivlin. Some polynomials approximation operators. Math. Z. 145 (1975), 33-42.

[7] M. Day. Normed linear spaces. Academic Press, New York (1962).

[8] c. Franchetti - E.W. Cheney. Minimal projections in ~-spaces. Duke Math. Journ. 43 (1976).

Page 164: Optimal Estimation in Approximation Theory

MINIMAL PROJECTIONS 157

[9] C. Franchetti - E.W. Cheney. Minimal projections of finite rank in sequence spaces. To appear.

[10] C. Franchetti - E.W. Cheney. Ort~ogonal projections in spaces of continuous functions. To appear.

[11] D.J.H. Garling- Y. Gordon. Relations between some constants associated with finite dimensional Banach spaces. Israel J. Math. 9 (1971), 346-361.

[12] B. Grlinbaum. Projection constants. Trans. Am. Math. Soc. 95 (1960) 451-465.

[13] M.I. Kadets -M.G. Snobar. Some functionals over a compact Minkowskii space. Math. Notes 19 (1971), 694-696.

[14] J. Lindenstrauss. Extension of compact operators. Memoirs of the A.M.S. 48 (1964).

[15] P.D. Morris - E.W. Cheney. On the existence and characteriza­tion of minimal projections. J. reine angew. Math. 270 (1974), 61-76.

[16] H. Rademacher. Einige Satze Uber Reihen von allgemeinen Ortho­gonalfunktionen. Math. Annalen 87 (1922), 112-138.

Page 165: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN CRYSTALLOGRAPHY

Robert Schaback

Lehrstlihle fur Numerische und Angewandte Mathematik

3400 Gottingen, Lotzestr. 16-18

One of the most important problems in crystallography is the determination of the unit cell of a crystalline substance and the relative positions of atoms within that cell. This contribution is mainly concerned with the estimation of unit cell dimensions from data obtained by x-ray diffraction of a polycrystalline substance. Oompared to the number of parameters to be estimated and to the de­sired accuracy, the given information turns out to be rather limi­ted and relatively noisy. Therefore some deeper insight into the underlying problem is necessary. This reveals a nonlinear mixed­integer approximation problem and gives some hints for the numeri­cal solution of the problem. By combination of common approximation methods and combinatorial strategies of branch- and bound-type a numerical estimation method for a high-speed computer can be deve­loped. In practice this method provides the user with a set of esti­mates which fit into the noise limits of the data, as is shown by a series of test examples. A reduction of the number of possible estimates can be made by introducing additional restrictions invol­ving more crystallographic information. The purpose of this contri­bution is not to give a purely crystallographical or a purely mathe­matical description of the topic but to try to arouse some interest for crystallographical questions among mathematicians working with high-speed computers.

159

Page 166: Optimal Estimation in Approximation Theory

160 ROBERT SCHABACK

1. INTRODUCTION AND PROBLEM FORMULATION

As in every branch of the natural sciences, there are of course lots of estimation problems arising in crystallography. In this contribution we want to concentrate on the most important one, i.e. the estimation of the parameters describing the geometrical structure of crystals and their complicated inner lattice of atoms.

Fig.1: Crystal of snow [42]

For sake of preciseness we should first note that a crystal may be defined as a solid composed of atoms, ions or molecules arranged ~n a patte~n periodic in three d~~ensions [17 ,42]. S~nce in practice the per~ods are of length ~10 em, one can cons~der the pattern as being infinite in any direction. The periodicity implies that there are three linear independent vectors ai, i = 1,2,3, such that the transformations

X -+ X + a. ~

= 3 , i 1,2,3, xElR

map the pattern onto itself, and such that the lengths of the a. . ~ can not be shortened. Then the parallelep~ped spanned by the a. is called a unit cell of the crystal. By periodic repetition ~ of the unit cell we get the point lattice of the crystal.

Page 167: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN CRYSTALLOGRAPHY 161

Fig.3: Point lattice [42] Fig.4: Unit cells [42]

Fig.4 shows that the unit cell is a purely mathematical tool for describing the domain of periodicity of the arrangement of atoms in the crystal. We now can formulate

Estimation Problem 1. Determine the shape of a unit cell of a given crystalline substance.

Before we turn to methods for solving this problem (which will be the main part of this paper) we should look upon what fol­lows a solution of the problem. The obvious next step is

Estimation Problem 2. Determine the sort and the number of atoms in the unit cell.

Whereas the first part of this is a chemical task which usu­ally is completed before the crysLallographer starts his work, the second part consists of a simple calculation:

gives the sum S of the atomic weights of the atoms in the uni~ cell as a function of the density p of the substance (in g/cm ) and the volume V of the unit cell (in 13). From the chemical for­mula of the sUbstance one gets the sum s of atom weights of atoms in the chemical unit. Then S/s must be approximately equal to an integer number giving the number of chemical units in the unit cell (e.g. 4 for Na Cl, 8 for diamond). We note for further use that (1) provides a check for correctness of unit cell volumes.

One now knows how many atom positions (say n) within the unit cell have to be estimated. Since n may be fairly large (see Fig.6) and there are 3n unknown real parameters, one first tries tore­duce the number of variables by

Page 168: Optimal Estimation in Approximation Theory

162 ROBERT SCHABACK

Estimation Problem 3. Determine the symmetries of the contents of the unit cell.

This task consists of finding the correct group of symmetry transformations out of 230 possibilities [6,42]. The choice depends on the shape of the unit cell (e.g. a cUbic cell~ admit many more symmetry transformations than a general triclinic celll,on the num­ber and the sorts of atoms in the unit cell (e.g. symmetry transfor­mations must map atoms of the same sort onto each other), and on cer­tain pecuiiari ties arising in the data from x-ray diffraction of specimens of the sUbstance (i.e. systematic extinctions of diffrac­ted beams due to inner symmetries of the unit cell [42]). Even if all the above information is available, there usually are several possibilities left. The problem of finding the correct symmetry group (which is, like problem 2, more a decision problem than an estima­tion problem) thus comes out to be one of ruling out impossible al­ternatives and testing the remaining possibilities with respect to the outcome of

Estimation Problem 4. Determine the relative positions of atoms (or ions or molecules) within the unit cell.

This problem is usually tackled by assuming a certain structure, calculating its x-ray diffraction properties and modifYing the struc­ture until its calculated x-ray diffraction properties fit the ob­served x-r~ diffraction data. We shall discuss this method in the third section of this paper.

The total problem of determining the structUre of a crystal usually is considered as being solved if the above problems have led to a structure wh1ch is in good agreement with all available measurements [10]. Since problems 1,3 and 4 need not have unique so­lutions (in the sense of mathematics) some doubts to this argument seem to be appropriate.

2. DETERMINATION OF UNIT CELL DIMENSIONS

The most important means for measuring crystallographic data is x-r~ diffraction [3,10,17,39,42,69,74]. These methods either use a single-crystal specimen (von Laue, rotating-crystal, Weissenberg, divergent-beam and retigraph methods) or a specimen in powder form (Debye-Scherrer method, improved by Seeman & Bohlin and Guinier). In case a single crystal is available, one of the above methods will give complete information on the unit cell in a comparatively easy way. If not, one can only try to analyze the data supplied by a Debye-Scherrer-type method, which will be done the sequel.

A powder specimen diffracts a monochromatic beam of x-rays into a(practically finite) number of cones of r~s (see e.g. [3,10,17,39, 42,69]). The relation of the (measured) angles 2. ai of the

Page 169: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN CRYSTALLOGRAPHY 163

cones to the unit cell (spanned by vectors a,b,c) of the crystal is the following:

There are integer numbers h.,k.,l. such that J. J. J.

q. = (4. sin2 e.)/">..2 = llh.a* + k.b* + ll..c*ll 2 J. J. J. J.

where the Euclidean norm is taken and a*,b*,c* are "reciprocal" to a,b,c:

a*= (b x c)/V, b* = (c x a)/V, c* = (ax b)/V,

where V = (a x b,c) denotes the volume of the unit cell. It is easy to see that the determination of the unit cell of the given crystal is equivalent to the determination of the "reciprocal" unit cell spanned by a*,b*,c*. If, in addition, we introduce A as the Gram matrix corresponding to a*,b*,c*, we finally arrive at the follo­wing

Estimation Problem 1' . Given a vector q E JRN with positive components, determine a real 3 x 3 -matrix A and an integer 3 x N -matrix M such that

(2) . T the dJ.agonal of M A M equals q,

3 i.e. qk = L a .. •m1.k•mJ.k if A=(a .. ), M=(m.k), q=(qk),

i,j=1 l.J l.J J.

(3) A is positive definite.

Because of periodicity, A contains all information on the (recipro­cal) unit cell in the form of 6 parameters (3 vector lenghts and 3 angles). As was already pointed out in [61], the above problem will have many symmetrical solutions, if any, and because one has to allow for noise in qk (about 0.1%, but unevenly distributed) it has always solutions fit~ing into arbitrarily small noise limits. Thus an additional restriction is needed:

(4) lmjkl S c (c chosen between 4 and 18 in practice),

which eliminates practically impossible cases of diffraction of ar­bitrarily high order or by extremely improbable sets of parallel lattice planes.

To account for random noise we reformulate the task to get

Estimation Problem 1". Given a positive vector q EJRN, find a real 3 x 3 - matrix A and an integer 3 x N - matrix M such that

Page 170: Optimal Estimation in Approximation Theory

164 ROBERT SCHABACK

(5) :r(A,M) = II q - diag (MT AM) II~ 2

is minimized sUbject to (3) and (4).

Since this problem consists o:r the simultaneous estimation o:r six real. parameters and 3 • N integer parameters occurring nonlinearly in a given object :function, one is led to try to separate the two sets o:r variables and to estimate them iteratively:

Algorithm 1. Start with an estimate AO :for A.

Iteration. Step 1: Given Ai, minimize :r(Ai,M) with respect to M under the restricion (4).

Step 2: Given a solution ~+ 1 o:r Step 1, minimize ( i+1) . :r A,M w1th respect to A under the re-

striction (3). Re-iterate with the result Ai+1.

The algorithm is discussed 1n detail in [61], where we proved the :following

Theorem. r:r all ~ have rank 3, then

a) Step 2 always has a unique solution. b) Step 1 always has a solution c) The algorithm is :finite if' the result o:r Step 1 ip calculated

deterministically :from A1 (i.e. equal A's yield equal results).

A practical way o:r handling Step 1 (Step 2 is a standard linear least-squares problem) is described in detail in [61]. Since the publication o:r [61] E.Larisch has carried out a large number o:r test runs with Algorithm 1. We shall comment on two observations.

First, i:r one starts the algorithm with "true" cell parameters :from the standard re:ference :file issued by the American Society :for Testing Materials (ASTM) one can :frequently observe that the ite­ration stops with a di:r:rerent set o:r cell parameters and a reaso­nable improvement o:r the least-squares sum :r(A,M). We :round out later (see below) that this e:r:rect occurred i:r the noise in the data set was larger than the critical noise level.

We there:rore conclude that serious doUbts are appropriate con­cerning the use o:r the ASTM :file as a standard re:ference (see also [ 17] , p. 386: 11 •• • the user must decide which pattern in the :rile is the most reliable").

Page 171: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN CRYSTALLOGRAPHY 165

Second, the algorithm may give a different "solution" if a perturbation of 0.5% in the initial estimate is applied ([45]). This implies that the algorithm needs fairly precise starting parameters and performs only a refinement of an estimate.

Therefore a good routine for getting started is needed. Since the testing of many candidates must be expected to get the "correct" estimates, the algorithm should make as sure as possible that the "correct" estimate is not overlooked. This is done by carefully placing the intersection between a statistical and a deterministic part of the computation in order to allow as little randomness as possible and to keep deterministic time-consuming calculations within reasonable bounds. After a series of tests of different approaches E.Larisch [45] finally developed the following

Algorithm 2. Step 1 :

Step 2:

Determine all 2-dimensional sublattices (planes) which can be formed out of the first n 1 data values and calculate their "fitting quality" with respect to the first ~2 data v':ll!-es. This "zone-findin~" step 1s very s1m1lar to the correspond1ng me­thod in [38,66,70,71].

Scan the output of Step 1 for higher symme­tries of the crystal, i.e. test whether special assumptions on cell parameters lead to reasonably good fittings. Calculate such estimates and their fitting quality with respect to all data values.

Step 3: Take the m best planes from Step 1. For each pair of planes, estimate the two remai­ning parameters by checking the distribu­tion of the parameters when some of the unfitted data values are assumed to stem from systematically varied grid points. Calculate the fitting quality of the esti­mates with respect to all data values.

Step 4: Perform some additional checks on the out­put of steps 1 and 2 (e.g. test for evi­dence of parallel planes to the already constructed planes).

The output of this algorithm may consist of up to 5000 estimates for the cell parameters, all of which are refined by a sUbsequent run of Algorithm 1. The CPU time of Algorithm 2 on the IBM 370/165 of the Regionales Hochschulrechenzentrum Bonn varied between 50 seconds and 30 minutes with an average of 6.4 minutes whereas the refinement of

Page 172: Optimal Estimation in Approximation Theory

166 ROBERT SCHABACK

all possibilities by Algorithm 1 varied between 6 seconds and 18 minutes with an average of 4.4 minutes. Since the length of the cal­culation can by no means be estimated beforehand, one has to incor­porate a restart facility into the program.

The test runs in [45] provided a series of examples throwing some light on the dependence of the result on the noise in the data. The first and unexpected observation is that in any case, even for arbitrarily manipulated data, the algorithm is able to give (usually several) estimates with an error not exceeding 0.5 • 10-7 but only rarely lower than 0.35 • 10-8 (the error is measured by a weighted sum of squares, the weights of which are related to the experimental situation and the quality of the observations; since the weights are normalized to give a sum of 1, the error can be used for comparison between different test examples). This implies that the noise level in the data (measured in the same way as the error with respect to the true solution) must be lower than 0.5 • 10-7 in order to make a better estimate from the data by any method whatsoever. On the

other hand, only estimates with error less than 0.35 • 10-8 can be considered as dependent on the data.

The outcome of the test runs can be simply grouped by the mag­ni tude of the error E0 of the "known solution" taken as a measure for the noise 1n the data.

Group 1. E S 0.2 • 10-7 (i.e. somewhat better than maXlmum 0

noise level). In all nine examples tested the correct cell was found (among others), except for equivalence transformations of the unit cell. In seven case~ the optimal solution had an error below 0.35 • 10- . Al­ternative estimates with equally low errors differed considerably with respect to the unit cell volume.

Group 2. 0.2 • 107 < E < 0.5 • 10-7 (at noise level). The 0

two examples in this group were not solved. The first example had an error reduction from 0.34 • 10-7 to 0.79 • 1o-8, but the resulting estimates were false; the second example has a "known solution" which fits seven data values extremely badly.

Group 3. E ~ 0.5 • 10-7 (worse than the critical no1se level). 0

In 4 of the 10 examples in this group the correct cell was found, but this must be contributed to luck.

At first glance this result seems rather positive; nevertheless it must be borne in mind that even for the examples in Group 1 the correct solution may occur between estimates with equally low errors. At this stage it can only be recommended to concentrate on cases with

Page 173: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN CRYSTALLOGRAPHY 167

-8 noise level E0 S 0.35 • 10 and to ignore estimates with larger errors.

On the other hand, one should try to achieve some reduction of the noise by major modifications of the problem. This can be done by taking more data into account or by introducing more restric­tions on the estimates.

A first possibility is to use (1) for the elimination of esti­mates giving impossible cell volumes. A second possibility arises from the observation that often the correct cell belongs to a spe­cial class of cells (monoclinic, orthorhombic etc.) but occurs among general cell estimates with equally good or lower errors or it has an error within the above noise level, which was obtained for gene­ral cells. Therefore the error of special estimates should be comr pared with the (yet unknown) noise level of estimates belonging to the corresponding class of cells. This would lead to higher levels of admittable noise and the significance of special estimates would be increased. There is some experimental evidence [46] that for data with noise below these levels the check on the cell volume using (1) will finally cut down the number of reasonable estimates to very few, including the correct solution.

A third possibility arises in conjunction with the second pos­sibility: for some special classes of cells there may be sUbclasses having systematic absences of certain (m1k,m2k,m3k) triplets depen-

ding on the interior symmetries of the unit cell (see e.g. [42]). For instance, if a,b,c are mutually perpendicular (this implies that the cell belongs to the rhombic, tetragonal, or ctibic class) and the transformation x + x + (a+b+c)/2 maps the crystal pattern onto itself, the pattern is called "body-centered" and only trip­lets with m1k + ~ + m3k even are possible. It should be tried to

evaluate noise levels for these sUbclasses as well as for the clas­ses themselves in order to gain a further increase in admittable noise.

A fourth possibility is to use the rema1~ng estimates as a starting point for further investigations of the inner structure of the crystal. This means that estimates should be discarded if they turn out to be incapable of fitting additional diffraction da­ta, mainly the observed intensities of diffracted beams when trying to solve Estimation Problem 4.

For comparative testing of other methods [4,5,37,38,49,56,64, 65,70,71], and especially the computer-oriented methods of Visser ([66], based on [56,38,70,71]) and Haendler & Cooney ([8], based on [38]) we unfortunately had not enough common examples. The strategy described above uses elements of both methods in Step 1 of

Page 174: Optimal Estimation in Approximation Theory

168 ROBERT SCHABACK

Algorithm 2, but enlarges their scope by taking more possibilities into account. Therefore it can be assumed that Algorithms 1 and 2 have a smaller probability of missing the correct solutions (of course at the cost of considerably increased computing time).

3. ESTIMATION OF THE RELATIVE POSITIONS OF ATOMS

WITHIN THE UNIT CELL

The theory of 3-dimensional point lattices (see e.g.[42]) imr plies that to each triplet m = (m1~~) of integers there corres-

ponds a set of parallel lattice planes in the lattice. The diffrac­tion of an x-ray beam by that set of planes occurs at an angle go­verned by Bragg's law (which gives the data used in section 2 of this paper) and with the (complex) amplitudes

(6) Fm = J p(x) e2~i(x,m)dx ' c

(called the structure factor corresponding to the triplet m), where p(x) is the (periodical) electron density within the crystal and the domain of integration is the unit cell C(e.g.[10],[39],[42]). This means that F ~s the Fourier transform of p and we have

m

(7) p(x) = l ~ F e-2~i(x,m) v J.z3 m '

where V is the volume of the unit cell C. If the complex numbers F could be estimated, one would get p and the atom-positions as

m extrema of p immediately from (7), (see Fig.8).

So we have

Estimation Problem 4 1 • Determine F for mE z3 • m

The data for Estimation Problem 1 consists of measurements of diffraction angles ek for a series of diffracted beams; by solving Problem 1 via (2) it is known which beam belongs to which triplet ~ = (m1k'~'m3k) and its intensity must therefore be the square

of the absolute value IF I of the complex amplitude F . Thus ( • • 11\t • . -~ ) after certa~n correct~ons due to exper~mental cond~t~ons the mea-surements of intensities yield estimates for IF I which can be con­sidered as data for Problem 4•. The need of inf~rmation on the pha-

ses ' of F = IF lei'm is called the phase problem in the deter-m m m mination of crystal structures [10,11]. We arrive at

Page 175: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN CRYSTALLOGRAPHY

Estimation Problem 4". Determine <!> for mE z3 . m

169

Unfortunately one can not expect in general to have more pure­ly mathematical information for solving Estimation Problem 4 than the information about the unit cell C and the absolute values IF I, which leaves no direct information about the phases <P . Thus m one needs additional assumptions from crystallographym to arrive at mathematically reasonable estimation problems.

The simplest way of making additional assumptions is to consider p(x) as being a superposition of weighted Dirac measures located at the positions x., j = 1, ••• ,n of then atoms in the unit cell. This implies J

n ( 8) L

j=1

where the atomic scattering factors f, can be estimated and some of the coordinates of x. may already J be known from symmetry con­siderations. In general~ for N observed diffracted beams, ( 8) pro­vides a set of N nonlinear equations for maximally 3 • n unknowns. Thus it seems reasonable to try either to solve (8) by a standard method for equation solving or to approximate directly the vector of IF I values by the absolute values of the right-hand sides of (8) min the 12 sense, varying the xj• The well-known nonunique­ness and stability problems arising in such approaches [7-9] make it necessary to start the process with a reasonably good estimate for the x., obtained by one of the methods described below.

J

Another iterative method for getting the atom positions x· is to start with an estimate for the x.'s, determine an estimaie for the~ 's from (8) and calculate J new estimates for the x.'s

m J by moving them towards the extrema of p(x), where p(x)is calculated

from (7) with coefficients IF lei~m using the measured value IF I m m rather than the value given by (8). Within these procedures one can start with a smaller number of heavier atoms and add the lighter atoms later. Because the phases frequently are dominated by the contributions of the heavier atoms, this yields only smaller chan­ges of the right-hand side of (8). Estimates for the positions of smaller atoms may be derived from new extrema showing up in plots of p calculated as in the second method. Another possible method 1 s the global minimization (in the 12 sense) of the function

o(x) = L (IF I - IF' I) e21Ti(m,x)+i~m mEz3 m m

over the unit cell C, where IF I is taken from the observations -~ m

and IF le 1 miscalculated by evaluating the right -hand side of (8). m

Page 176: Optimal Estimation in Approximation Theory

170 ROBERT SCHABACK

This implies that o depends only on the relative atom positions x .. The shape of o may also be used as a final check on the calcula- J ted estimates, because it resembles the errors in the IFml and the ~m·

The 12 methods described above (and many other versions) have been used frequently for refinement of estimates in crystallogra­phic research [1,2,10,12,16 22,23,31,44,50,55,59,60]. Since no fail-safe standard method has yet emerged, a thorough mathemA.tical investigation by means of approximation theory seems worthwhile. In addition, numerical difficulties are great because of the large number of free parameters (sometimes more than 2000). Computing times may be enormous (50 minutes per iteration cycle with 1603 phase variables on an IBM 360/91[60]),

Fig.5: Structure of Carbobenzoxy-1-lencyl-p-nitrophenyl Ester [15]

Like in section 2 of this paper the iterative processes avai­lable for solving Estimation Problem 4 tend to calculate only local optima. So we finally must comment on corresponding global methods, which are of independent interest because of their close relation­ship to Fourier series theory.

The Patterson function P(y) ~s defined by the convolution

( 9 ) P ( y) = J P ( x) • p ( x+y) dx v

Page 177: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN CRYSTALLOGRAPHY 171

and has the series representation

( 10) P(y) = --1 L IF 12 cos (2n(m,y)) v2 mEz3 m

which can be evaluated directly from the observed data. It is easi­ly seen from (9) that P(y) is large if there are peaks of p having the distance vector y. This implies that distance vectors of atoms may be read off a plot of P(y), as shown in Fig.[.

Fig.6: Structure of KH 2Po4 [42]

(2-dimensional projection)

Fig.7: Correspondent plot of the Patterson function [42]

Refinements of this approach have been given by various authors [1,2,11,24,47,54,55,58,72], including computer programs and inter­active graphical methods, but no generally satisfactory automatic method has yet been designed to calculate reliable global estimates for atomic distances from the intensities IF 12 via the Patterson function alone. The second and perhaps most m important class of methods of estimating phases is based on inherent relations bet­~een_several values of Fl!! for varying m. One of the simplest [57] 1s g1ven by the observat1.on that the shape of p 2 should be very similar to that of p. This implies the convolution relation

F ~..!. \' F F m V L 3 n m-n nEZ

( 11 )

and the influence of the absolute values on the phases is apparent. For instance, if IF • F I dominates the whole sum, then

n m-n 0 0

Page 178: Optimal Estimation in Approximation Theory

172

or equivalently, if IFn F I is large m

ROBERT SCHABACK

Also, if the larger values of IF I have correct phase estima­tes, (11) will estimate the phases n ofF with smaller absolute values. By separating (11) into real and m imaginary parts one gets the equation

( 12)

\ IF IIF I sin(cp +'cp + ) L 3 n n+m n n m tan cp ~ .;.;n""E.;.;Z __________ _

m L 3 IFniiFn-ml nEZ

cos ( cp +cp + ) n n m

known among c.,·y!3tallographers as "tangent formula".

In addition, there are formulae describing the probability distribution for esti~ting phases by (11) and (12), provided the probabilities of already assigned phases on the right-hand side are known (see e.g.[1,13,14,29,30,58]).

Of course there are problems arising when initial phase gues­ses have to be made. But as the choice of origin of the atom coor­dinate system influences the cp values, one can fix the ori~in by assigning arbitrary phases tomthe three largest values of IF I having linear independent vectors m. Because of symmetry m there may be additional possibilities of fixing certain phases, but a major tool for reducing the number of free parameters is the sym­bolic additionproc.edure [41,73]: One assigns formal names to cer­tain phases and uses origin-independent linear relations ("struc­ture invariants") like

( 13) 1Pm1 + 1Pm2 + cpm3 = const for m1 + m2 + m3 = o

to generate formal phases for a number of other IFml values. The more symmetry a crystal has, the more relations like (13) are valid. Now several possibilities for the values of the formal names are tested; in each case (11) or (12) is taken to try to assign more phases as long as certain probability measures are satisfactory. Thus a number of sets of phase estimates is generated; for any set (before 12-refinement is made) a Fourier synthesis (7) is made by plotting p and practically impossible estimates are sorted out by visual inspection.

Page 179: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN CRYSTALLOGRAPHY 173

There is a vast literature on the above strategy (see e.g. [1,2,13,14,18,19,21,25-27,29,30,33-36,40,41,67,68,72,73] and many other publications) and there are mainly two important program sy­stems based on it (X-RAY-subroutines, [1,62,63], and MULTAN [20,43, 48,51,52]), which already solved a large number of moderately com­plex structure problems.

Of course, the scope of the method is limited by the applica­bility of relations based on (11), which has led to new efforts in developing phase estimates and to combining phase estimation with L2-methods [e.g. 1,22,23,33,34,48]. Since the chemical components occuring in today's microbiological research still mainly are be­yond the scope of this method, some further investigation of the above estimation problems seems worthwhile, both from the mathema­tical and the crystallographical point of view.

D I I I

Fig.8: Electron density p(x) of phthalocyanine [39]

Page 180: Optimal Estimation in Approximation Theory

174 ROBERTSCHABACK

REFERENCES

1. Ahmed,F.R.(ed.): Crystallographic Computing, Copenhagen: Munks­gaard 1970

2. Ahmed,F.R.(ed.): Crystallographic Computing II, Copenhagen: Munksgaard (to appear)

3. Azaroff, L.V. and Buerger,M.J.: The Powder Method ~n X-ray Cry­stallography, New York: Me Graw-Hill 1958

4. Babenko,N.F. and Brusentsev,F.A.: Indexing X-Ray Powder Patterns for a Crystal of an Arbitrary System, Kristallografiya 19(1974), 506-510

5. Barabash,I.A. and Dayydov,G.V.: On the Indexing of Powder Pat­terns for Polycrystalline Materials of the Orthorhombic System, Acta Cryst.A 24(1968),608

6. Bradley,C.J. and Cracknell,A.P.: The mathematical theory of symmetry in solids. Representation theory for point groups and space groups. Oxford: Clarendon Press 1972

7. Braess,D.: Die Konstruktion der Tschebyscheff-Approximierenden bei der Anpassung mit Exponentialsummen, J. of Approx.Th.3(1970), 261-273

8. Braess,D.: Nichlineare Anpassung von Me~daten nach der Methode der kleinsten Quadrate, Schriftenreihe des Rechenzentrums der Universitat Mlinster 6(1974)

9. Braess,D.: Nonlinear Approximation in Lorentz,G.G.(ed.): Appro­ximation Theory, London - New York: Academic Press (to appear)

10. Buerger,M.J.: Crystal-Structure Analysis, New York- London: Wiley & Sons 1960

11. Buerger,M.J.: Vector Space and its Application in Crystal-Struc­ture Investigation, New York: Wiley & Sons 1959

12. Busing,W.R., Martin,K.O. and Leyy,H.A.: ORFLS/ORFFE Oak Ridge National Laboratory Reports ORNL-TM-305/306, 1962/1964

13. Cochran,W.: Relations between the phases of structure factors, Acta Cryst.8(1955),473-478

14. Cochran,W. and Woolfson,M.M.: The theory of sign relations bet­ween structure factors, Acta Cryst. 8(1955),1-12

Page 181: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN CRYSTALLOGRAPHY 175

15. Coiro,V.M. and Mazza,F.: The Crystal Structure of the Confor­mational Analysis of Carbobenzoxy - L - lencyl - p - nitro­phenyl Ester, Acta Cryst.B 30(1974),2607

16. Cruickshank,D.W.J., Pilling,D.E., Bujosa,A., Lovell,F.M. and Truter,M.R.: Computing Methods and the Phase Problem in X-ray Crystal Analysis, Oxford: Pergamon Press 1961

17. Cullity,B.D.: Elements of X-ray Diff'raction, Reading, Mass .• : Addison - Wesley 1956

18. Debaerdemaeker,T. and Woolfson,M.M.: On the Application of Phase Relationships to Complex Structures IV: The Coincidence Method Applied to General Phases, Acta Cryst.A 28(1972),477

19. Debaerdemaeker,T. and Woolfson,M.M.: The determination of struc­ture invariants II, Acta Cryst.A 31(1975),401

20. Declerq,J.P., Germain,G., Main,P. and Woolfson,M.M.: On the application of Phase Relationships to Complex Structures V: Finding the solution, Acta Cryst.A 29(1973),231

21. Declerq,J.P. and Woolfson,M.M.: On the Application of Phase Relationships to Complex Struct.ures VIII: An extension to the magic-integer approach, Acta Cryst.A 31(1975),367

22. Gassmann,J.: Least-squares Refinement of Phases in Direct and Reciprocal Spaces, Acta Cryst.A 32(1976),274

23. Gassfllann ,J. and Zechmeister ,K. : Limits of Phase Expansion in Direct Methods, Acta Cryst.A 28(1972),270

24. Germain,G. and Woolfson,M.M.: Some Ideas on the Deconvolution of the Patterson Function, Acta Cryst.A 21(1966),845

25. Germain,G. and Woolfson,M.M.: On the Application of Phase Re­lationships to Complex Structures, Acta Cryst.B 24(1968),91

26. Germain,G., Main,P. and Woolfson,M.M.: On the Application of Phase Relationships to Complex Structures II: Getting a Good Start, Acta Cryst.B 26(1970),274

27. Germain,G., Main,P. and Woolfson,M.M.: On the Application of Phase Realtionships to Complex Structures III: The Optimum Use of Phase Relationships, Acta Cryst.A 27(1971),368

28. Haendler,H. and Cooney,W.: Computer Determination of Unit Cell from Powder-Diffraction Data, Acta Cryst.16(1963),1243

Page 182: Optimal Estimation in Approximation Theory

176

29.

30.

31.

32.

33.

34.

35.

36.

37.

38.

39.

40.

41.

42.

ROBERT SCHABACK

Hauptman ,H. and Karle ,J.: The solution of the phase problem I: The centrosymmetrical crystal, Amer.Cryst.Assoc.Monograph No.3, Ann Arbor: Edwards Brothers 1953

Hauptman ,H. and Karle ,J.: Phase determination from new joint probability distributions: Space group P1, Acta Cryst.11(1958), 149-157

Hauptman ,H., Fisher,J. and Weeks,C.M.: Phase Determination by Least-Squares Analysis of Structure Invariants: Discussion of This Method as Applied on Two Androstane Derivations, Acta Cryst.B 27(1971),1550

Hauptman ,H.: Crystal structure determination, New York: Ple­num Press 1972

Hauptman ,H.: A joint probability distribution of seven struc­ture factors, Acta Cryst.A 31(1975),671

Hauptman ,H.: A new method in the probabilistic theory of the structure invariants, Acta Cryst.A 31(1975),680

Hoppe,W.: Phasenbestimmung durch Quadrierung der Elektronen­dichte im Bereich von 2 A- bis 1,5 A- Auflosung, Acta Cryst. A 15(1962),13

Hoppe,W. and Gassmann,J.: Phase Correction, a New Method to Solve Partially Known Structures, Acta Cryst.B 24(1968),97

Ishida,T. and Watanabe,Y.: Analysis of Powder Diffraction Patterns of Monoclinic and Triclinic Crystals, J.Appl.Cryst.4 (1971),311

Ito,T.: X-ray Studies on Polymorphism, Tokyo: Maruzen 1950

Jeffery,J.W.: Methods in X-ray Crystallography, London- New York: Academic Press 1971

Karle,J. and Hauptmann,H.: Phase determination from new joint probability distributions: Space group P1, Acta Cryst.11(1958), 264-269

Karle,J. and Karle,I.L.: The Symbolic Addition Procedure for Phase Determination for Centrosymmetric and Noncentrosymmetric crystals, Acta Cryst.21(1966),849

Kleber,W.: Einflihrung in die Kristallographie, 10.Auflage, Berlin: VEB Verlag Technik 1969

Page 183: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN CRYSTALLOGRAPHY 177

43. Koch,M.H.J.: On the Application of Phase Relationships to Com­plex Structures VI: Automatic Interpretation of Electron-Den­sity Maps for Organic Structures, Acta Cryst.B 30(1974),67

44. Konnert,J.H.: A Restrained-Parameter Structure-Factor Least­Squares Refinement Procedure for Large Asymmetric Units, Acta Cryst.A 32(1976),614

45. Larisch,E.: Algorithmus zur Indizierung von Rontgen-Pulverauf­nahmen, Diplomarbeit Munster 1975

46. Larisch,E.: Private communication, 1976

47. Lenstra,A.T.H. and Schoone,J.C.: An Automatic Deconvolution of the Patterson Synthesis by Means of a Modified Vector-Verifi­cation Method, Acta Cryst.A 29(1973),419

48. Lessinger,L.: On the Application of Phase Relationships to Complex Structures IX, MULTAN Failures, Acta Cryst.A 32(1976), 538

49. Louer,D. and Louer,M.: Methode d' Essais et Erreurs pour 1' Indexation Automatique des Diagrammes de Poudre, J.Appl.Cryst. 5( 1972) ,271

50. Lynch,M.F., Harrison,J.M., Town,W.G. and Ash,J.E.: Computer Handling of Chemical Structure Information, London: Mac Donald, New York: American Elsevier 1971

51. Main,P. ,Woolfson,M.M. and Germain,G.: MULTAN: A Computer Pro­gram for the Automatic Solution of Crystal Structures, Univ. of York 1971

52. Main,P., Woolfson,M.M., Lessinger,L., Germain,G. and Declerq, J.P.: MULTAN 74: A system of Computer Programmes for the Auto­matic Solution of Crystal Structures from X-R~ Diffraction Data 1974

53. Pawley,G.S.: Advances in Structure Research by Diffraction Methods, New York: Pergamon Press

54. Rae,A.D.: The Phase Problem and its Implications in the Least­Squares Refinement of Crystal Structures, Acta Cryst.A 30(1974), 761

55. Rollett,J.S.(ed.): Computing Methods 1n Crystallography, Ox­ford: Pergamon Press 1965

56. Runge,C.: Die Bestimmung eines Kristallsystems durch Rontgen­strahlen, Physik. Z. 18 ( 1917) ,509-515

Page 184: Optimal Estimation in Approximation Theory

178 ROBERT SCHABACK

57. Seyre,D.: The Squaring Method: A New Method for Phase Deter­mination, Acta Cryst.5(1952),60

58. Seyre,D.: The double Patterso~ .function, Acta Cryst.6(1953), 430

59. Seyre,D.: On Least-Squares Refinement of the Phases of Crystal­lographic Structure Factors, Acta Cryst.A 28(1972),210

60. Seyre,D.: Least-Squares-Refinement II: High-Resolution Phasing of a Small Protein,Acta Cryst.A 30(1974),180

61. Schaback ,R. : Ein Optimierungsproblem aus der Kristallographie, in Collatz,L. und Wetterling,W.(ed.): Numerische Methoden bei Optimierungsaufgaben, ISNM 23, Basel - Stuttgart: Birkhauser 1974,113-123

62. Stewart,J.M., Kundell,F.A. and Baldwin,J.C.: The X-RAY 70 system, Computer Science Center, Univ.of Maryland , College Park, Maryland 1970

63. Stewart,J.M., Kruger,G.J., Ammon,H.L., Dickinson,C. and Hall, S.R.: X-RAY system, Report TR-192, Computer Science Center, Univ.of Maryland, College Park, Maryland 1972

64. Vand,V. and Johnson,G.G.: Indexing of X-Rey Powder Patterns, Part I. The Theory of the Triclinic Case, Acta Cryst.A 24(1968), 543

65. Viswanathan,K.: A Systematic Approach to Indexing Powder Pat­terns of Lower Symmetry Using De Wolff's Principles, The Ame­rican Mineralogist 53(1968),2047

66. Visser, J.W.: A Fully Automatic Program for Finding the Unit Cell from Powder Data, J.Appl.Cryst.2(1969),89-95

67. Viterbo,D. and Woolfson,M.M.: The Determination of Structure Invariants I. Quadrupoles and Their Uses, Acta Cryst.A 29(1973), 205

68. White,P.S. and Woolfson,M.M.: The Application of Phase Rela­tionships to Complex Structures VII: Magic integers, Acta Cryst.A 31(1975),53

69. Wilson,A.J.C.: Mathematical Theory of X-Rey Powder Diffracto­metry, Eindhoven: Philips Technical Library 1963

70. de Wolff,P.M.: On the Determination of Unit-cell Dimensions from Powder Diffraction Patterns, Acta Cryst.10(1957),590-595

Page 185: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS-IN CRYSTALLOGRAPHY 179

71. de Wolff,P.M.: Indexing of Powder Diffraction Patterns, Advan­ces in X-r~ Analysis 6,1-17

72. WoGlfson,M.M.: Direct methods in Crystallography, Oxford: Per­gamon Press 1961

73. Zachariasen,W.H.: A new analytical method for solving complex crystal structures, Acta Cryst.5(1952),68-73

74. Zachariasen,W.H.: Interpretation of Monoclinic X-r~ Diffrac­tion Patterns, Acta Cryst.16(1963),784

This work was supported by the Deutsche Forschungsgemeinschaft within the projects "Indizierung von Rontgen - Pulveraufnahmen" at the University of Miinster and "Approximation alrl' DV - Anlagen" in the Sonderforschungsbereich 72 "Approximation und Optimierung" at the University of Bonn. Calculations were done on the IBM 360/50 of the Rechenzentrum of the University of Miinster and the IBM 360/ 165 and IBM 360/168 of the Regionales Hochschulrechenzentrum Bonn.

Page 186: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN DATA-TRANSMISSION SYSTEMS

G. Ungerboeck

IBM Zurich Research Laboratory

Saumerstrasse 4, 8803 Rlischlikon, Switzerland

ABSTRACT

The transfer of information over noisy and dispersive media has traditionally been, and still represents, an important subject of applied estimation and approximation theory. In this paper the specific problems encountered in synchronous data-transmission systems are reviewed. A first set of problems arises in timing­and carrier-phase tracking and in adaptive equalization, where continuous-valued parameters are to be estimated which may change slowly over time. A second set of problems deals with recovering the transmitted information from received noisy signals. It is shown that in the presence of severe signal distortion and/or redundant sequence coding the optimum receiver has to solve a dynamic programming problem.

I. INTRODUCTION

The field of information transfer over noisy and dispersive media has traditionally been closely related to estimation and approximation theory. It is primarily the receiver functions that follow in quite a direct manner from these areas. This is well illustrated by standard texts on communication theory, e.g., Van Trees,l and by many articles in the literature. In this paper the specific estimation and approximation problems encountered in the receiver section of synchronous data-transmission systems are reviewed. Today, this form of transmission is rapidly growing in importance, owing to increasing needs for efficient data communi­cation over various media and to the expected massive introduction

181

Page 187: Optimal Estimation in Approximation Theory

182 G.UNGERBOECK

of PCM principles in telephony.

In synchronous data transmission, information is encoded in discrete signal values ~ which are transmitted at constant rate 1/T. There is little doubt that this is most efficiently accomplished by linear pulse-modulation schemes which, depending on the channel available, are used in baseband or linear carrier-modulation form. For this class of modulation schemes the signal at the receiver in­put can be written as

where

X (t) c

a n

h(t) T

Re { L n

a h(t - T n s

- nT) e jw t +

c + w (t),

c

discrete signal values (pulse amplitudes), signal element (elementary pulse waveform) signal spacing (1/T .. signaling rate), timing phase, carrier frequency in radians (w = 2rrf ),

c c carrier phase, additive channel noise,

(1)

and the term linear pulse modulation is due to in the ~·

xc(t) being linear

Each of the above quant1t1es may a priori be unknown in the receiver, which ultimately must determine from xc(t) the values ~ that have been transmitted. Usually, however, at least VT and we can be assumed to be known with sufficient precision so that uncertainities with respect to them can be absorbed in slow var1-ations of T and ~ .

s c

In baseband (BB) modulation (we = 0, ~c = 0) an and h(t) = hr(t) are real-valued. In carrier modulation (we ~ 0) one must distinguish between vestigial-sideband (VSB) modulation, where ~ is real-valued and h(t) = hr(t) ± j hi(t) [hi(t) ~ Hilbert transform of hr(t)]* and double-sideband (DSB) modulation, where ~ .may be real or complex-valued depending on whether one is dealing with amplitude modulation (AM), phase modulation (PM), or a combination thereof (AM/PM), and h(t) ~ hr(t). These aspects are further summarized in Fig. 1 (see also Refs. 2-4).

* Half of the spectrum of hr(t) is cancelled by adding 1±1 j hi(t), since Hi(f) ~ - j sign(f) Hr(f) leads to H(f) ~ 2 Hr(f) for f\~JO and to H(f) ~ 0 for flSJO, with an appropriate spectral roll-off across f = 0 which on one side of f = 0 leaves a "vestigial sideband".

Page 188: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN DATA-TRANSMISSION SYSTEMS 183

MODULATION an h(f) di {Re [h(f) e iwc'n TYPE

88 REAL hr (f) l I f'H(f~ f 0 1/zT

VS8 REAL hr(t11! 1ihi (I) n I J=}H(f~f 0 fc fc+'lzT

<t,- 'h Tl (fcl

DSB- AM REAL

DSB-PM I COMPLEX I = 1 "' hr(t) l I 1 I l I 'YH(f~ f DSB- AM/PM COMPLEX 0 fc-'lzT fc fc+'ltT

Fig. 1. Overview over modulation forms.

The first function to be performed by the receiver in the case of VSB or DSB, is transposing the signal into baseband* which leads from (1) to

x(t) j q>c

L a h(t - • - nT) e + w(t) n s (2)

n

In the following we shall regard the received signal only in this form which exhibits the least differences between the various modulation forms. Note that for VSB and DSB, x(t) is complex. In implementation this means that one deals concurrently with two real signals representing the real and imaginary parts of x(t) (also called in-phase and quadrature components).

Ideally, •s and q>c would be known, and depending on whether ~ is real-valued (BB, VSB, DSB-AM) or complex-valued (DSB-PM, DSB-AM/PM) h(t) would satisfy

BB,VSB,DSB-AM:

h (JI.T) r

* -jw t E.g., by multiplying xc(t) bye c, and cancelling signal components around f = -2fc by low-pass filtering.

(3a)

Page 189: Optimal Estimation in Approximation Theory

184

DSB-PM,DSB-AM/PM:

h(R.T) h • 0 at <===> I H (f + ~) = h0 /T

t

G.UNGERBOECK

(3b)

(o~ ••• Kronecker delta). This is the condition for zero inter­symbol interference (ISI) between consecutive pulses at sampling instants. The frequency-domain equivalent of this condition is well-known as the first Nyquist criterion and explains the two­sided bandwidth requirement of 1/T or 2/T Hz, plus spectral roll­offs, necessary for transmission of real or complex values 8n• respectively. In the absence of noise the values of 8n could simply be determined by sampling x(t) e-j~c at times nT + T

s

a n

(Re) 1 -jql c

ho x(t)e I t nT + T

s

(4)

However, the fact that none of the above ideal suppos1t1ons is satisfied under real conditions, requires the following estimation functions to be performed in a receiver for synchronous data signals:

(a) Estimation of Ts:

(b) Estimation of !llc:

(c) Estimation of h(t)

(d) Estimation of {8n}

redundant coding:

"Timing-phase tracking"

"Carrier-phase tracking"

and corrective action in case of non­zero lSI: "Adaptive equalization"

in the presence of noise and lSI and/or

"Signal detection".

These functions will be discussed in Sections III to V. In Section II we establish a basis for these discussions by indicating the general mathematical solution to the entire receiver problem.

II. GENERAL SOLUTION BY MEANS OF A SINGLE LIKELIHOOD FUNCTION

Suppose that the signal x(t) is observed within the time interval [0, NT], where N is sufficiently large, and let XN denote a realization of x(t) within this interval. Assuming w(t) is stationa~y additive white Gaussian noise (AWGN) with real and, in the case of VSB and DSB, imaginary parts being independent and each having two-sided spectral noise power density No, one obtains the likelihood function (see, e.g., Refs. 1, 5, 6)

Page 190: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN DATA-TRANSMISSION SYSTEMS 185

p(~-l{an}' T , ~) ~ A(~-l{a }, T , ~) -~ s c -~ n s c

(5)

where {~}., Ts and ~c are estimates of these respective quanti­ties. The condition of white noise is no restriction since it can always be induced by an appropriate whitening filter. The assump­tion of Gaussian noise is, however, somewhat artificial; it must be made in order to make the problem mathematically tractable. In (5), h(t) is assumed to be known, but it could also be replaced by an estimate h(t). Defining

(h

and

y (T ) n s = I h(-t) x(nT + T s

- t) dt

... conjugate complex of h) and

s = ~

I h(-t) h(~T - t) ~t s_~

taking the logarithm of (5)' yields

A(X i{a },T .~) -~ n s c

1 NT 2 --I lx<t> 1 dt +

2No o

(6)

(7)

N -j~ N N + - 1-{ L 2Re[y (T )e c.~ ] - L L ~-s._ka } + Max. (8)

2N0 N=O n s n i=O k=O 1 1 k

The derivation of the sample values Yn(Ts) from x(t) is illus­trated in Fig. 2.

These sample values represent, for statistiasl for estimating {~} from

Ts Ts, a set of suffiaient x(t). According to the con-

_x_(_t) -t MATCHED FILTER

{Yn(Ts)} ~--~~--------·

h (-t)

Fig. 2. Derivation of sample values Yn(Ts) by means of a matched filter and sampling.

Page 191: Optimal Estimation in Approximation Theory

186 G.UNGERBOECK

cept of maximum likelihood (ML) an optimum receiver must jointly determine the quantities {in}, Ts and ~c [and perhaps also h(t)] which maximize (8).

Formulation of the receiver problem in this general form is, however, impractical for solving in one step. It is more appropriate to approach the various problems individually. For timing- and carrier-phase estimation this will be discussed in the next section.

III. ESTIMATION OF TIMING PHASE AND CARRIER PHASE

Assume that the time-continuous output signal of the matched filter (MF) is given by

y(t) j'Pc I a s(t - T - nT) e + r(t) n s

n (9)

and that s(t) satisfies the condition for zero ISI: s(~T) = 0 for~~ 0 (for VSB and DSB-AM Re[s(~T)] = 0,~ ~ 0 is suffici­ent). In the following one must distinguish between various degrees of knowledge about the transmitted values of ~· The simplest case is when the values of ~ are already known (e.g., as a result of decisions made by the receiver).

a) {an} known. According to (8), the best estimates of Ts and 'Pc are those which maximize

N A(X_I{a },T .~) ~ I Re[y (T) -~ n s c n=O n s

-j<Pc - J e a

n + Max

Differentiating with respect to Ts and - yields 'Pc

ClA N -j~ ;J + 0 I Re ( y (T ) e c

--~ ClT n=O n s s

N ·-ClA -Jq> c

;n] + 0 --~ I rm(y(T) e a-q>C n=O n s

(10)

(11)

(12)

and replacing summation over n by expectation indicates that the receiver must solve the regression problemS

-j~ E{Re[yn(Ts)e c;n]}+o, (13)

Page 192: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN DATA-TRANSMISSION SYSTEMS 187

-j~ E {Im[yn (rs) e c ~:J} + 0 , (14)

which may be accomplished by the simultaneous stochastic approxi­mation5,7 algorithms

- (n+l) T

s - (n) T

s (n)

a > 0, (15)

ft> (n+l) c

- (n) - 8 (n) Im[y ( i ) ~c n s

-jtp e c ~n] ' 8 (n) > 0. (16)

This approach is referred to as decision-aided timing- and carrier­phase tracking.

A slight modification of these algorithms is obtained if in­stead of maximizing (10) one minimizes the mean-square value of the decision error en:

(17)

It is easy to show that in (15) and (16), ~ must then be replaced by -en. This has the effect of accelerating convergence consider­atly. The careful reader will observe that criterion (17) in fact is a combination of ML criterion (10) with ML criterion (20) or (21), discussed further below.

It is sometimes desirable to avoid the sampled time deviative in (15). Let E{aiak} = oi-k· For estimating 's one may then exploit that s(-T + ~Ts) - s(T + ~Ts) ~ 0 for ~Ts = Ts - 's ~ 0, which suggests the approximation algorithmS

A (n+ 1) T

s A (n) T

s

-j~ a (n) Re(e c (y 1 (T )~

n- s n

Expectation:

y (T ) ~ 1)]' n s n-(n)

a > 0 ,

(18)

The convergence properties of decision-aided algorithms for re­covery of 's have been analyzed in Refs. 8 and 9.

Page 193: Optimal Estimation in Approximation Theory

188 G.UNGERBOECK

b) {an} unknown, noPmaLLy distributed. Here one must refer back to (5) and integrate

f p(~_l{a },T .~) dP({a }) -~ n s c n (19)

in order to eliminate {an}, which for real ~ (BB: <l>c 0, VSB, DSB-AM) yields

N e-j~c]12 >.(~iT .~ ) 'V I IRe[y (T ) s c n=O n s

l N N -j2ij) 2 I IY (T >1 2 + ~ Re[ I y2(T) e c] +Max, (20)

n=O n s n=O n s

and for complex an (DSB-AM/PM)

N

>.(X IT ,ij))"' I IY (T >1 2 +Max . -~ s c n=O n s

(21)

The ML criterion (21) indicates that for DSB-AM/PM, with the assumption of {an} being normally and identically distributed in its real and imaginary parts, carrier-phase estimation is not possible, whereas from (20) it follows that for VSB and DSB-AM, <pc can be estimated only with an ambiguity of w. A further obser­vation is that substituting in (20) the optimum value of <l>c would make (20) and (21) identical.

The traditional timing-phase tracking method without prior knowledge of {a~} has been squaring the received signal, i.e., computing iy(t)l2, and extracting a spectral line at fT = 1/T Hz by means of a narrow bandpass filter ("timing tank"). The maxima or, with a phase shift of ± T/4, the zero crossings of the so ob­tained sinosoidal signal determine the sampling instants nT + Ts· For BB modulation this direct method has been analyzed in Ref. 10, and for DSB-AM/PM the method is described in Ref. 11. Its resemblance to the ML approach suggested by (21) is rather apparent, if the band­pass-filter function is recognized as a sliding summation over T­spaced signal values.

Differentiating (20) with respect to <Pc yields

~ Re(y (T ) e -j~c) • rm(y (T ) e -j~c) + 0 ~o n s n s

(22)

Page 194: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN DATA-TRANSMISSION SYSTEMS

which in essence describes the control term used in Costa's loop for carrier-phase tracking in VSB and DSB-AM systems (Ref. 8,

189

p. 184). A direct method for deriving a carrier signal with ap­propriate phase from the received-signal, when fc >> 1/T, con­sists in squaring the received signal, extracting a sino~oidal component at f = 2fc by a narrow bandpass filter, and frequency­dividing this signal by two (with phase ambiguity kn), (Ref. 8, p. 182). The output signal of the bandpass filter can be written as

[ 2 j2wct] [ j2wct + j[arc l<t) ILPrl] Re y ( t) I LPF • e "' Re e •

Since the ML phase estimate from (20) is given by

if follows that the above direct method is essentially equivalent to ML estimation of ~c·

It can be shownl2,13 that the variances of the maximum-likeli­hood estimates of Ts and ~c derived from (20) and (21) ({~} un­known), depend critically on the spectral roll-off characteristics of S(f) = ar{s(t)}. Roughly, Var(Ts) is proportional to (N • A)-1, whereas for VSB and DSB-AM, Var(~c) is proportional to (N • B)-1, where A and B are defined in Fig. 3. With very steep roll-offs around f = ± l/2T, A approaches zero and hence timing­phase estimation becomes impossible. With respect to carrier-phase estimation one sees that VSB must be considerable inferior to DSB-AM since BvsB << BDSB· For single-sideband (SSB) modulation, B is zero and hence carrier-phase estimation with normally distri­buted values of {a } cannot work. In SSB systems pilot tones have therefore often beeR used to provide carrier-phase information.

Fig. 3. Definition of values A and B.

-t

VSB

DSB

Page 195: Optimal Estimation in Approximation Theory

190 G.UNGERBOECK

In not yet published work13 Cramer-Rao bounding techniques (Ref . 1, Part I) have been used to compute efficient lower bounds for the ML estimates of Ts and ~c for the various modulation schemes and for {an} known or unknown (normally distributed). The results have been expressed in terms of the mean-square error, de­fined in (17), that results from noise and the uncertainities in estimating Ts and ~c· Thus bounds of the form

E{l e (T ,(j) ) 12} > E{l e (T .~ ) 1

2 } • (Q/N + 1) , n s c - n s c (23)

have been obtained, where N.T is the length of the observation interval and Q represents an inverse quality factor. For VSB, DSB-AM and DSB-AM/PM this factor is shown in Fig. 4 as a function of the spectral roll-offs and for {an} known and unknown. (For carrier-phase estimation in DSB-AM/PM {an} must always be known, i.e., a decision-aided scheme must be used.) In (23) equality is well approached for large N, say N > 10. The small values of Q

0 S (f): LINEAR ROLL -OFFS

EfT EfT

1I I JI 4.0

-Vz T 0 1f2T DSB

0 ljzT VSB

3.0 (o} { 0 n} KNOWN (DECISION AIDED)

(b) { 0 n} UNKNOWN (NORMALLY DISTRIBUTED!

2.0

1.0

oL-~~-L-L~~~~~~-L---­

o 0.2 0.4 0.6 0.8 1.0 £

Fig. 4. Inverse quality factor Q as a function of the linear roll-off quantity £ for various modulat ion forms with T and ~ being known or unknown.

s c

Page 196: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN DATA-TRANSMISSION SYSTEMS

indicate then a remarkable capability for fast recovery, except when the roll-offs in DSB became smaller than 15% and in VSB smaller than 40%, and {an} is not known. Clearly, this finding indicates a considerable weakness of VSB which explains why DSB-AM/PM and not VSB became the preferred carrier-modulation form for high-speed data-transmission systems.

191

Much more could be said about timing- and carrier-phase tracking. In particular, a performance comparison should be of interest between direct ML estimation of 's and ~ by means of continuous-time processing, which in good approximation is realized in existing systems by signal squaring and bandpass filtering, and estimating 's and ~c iteratively from T-spaced signal values by stochastic approximation. The latter technique becomes important, when digital realizations are envisaged. It also represents the more natural approach when knowledge about {~} is available.

It has so far been assumed that {~} is either known or unknown and normally distributed. In between are cases where {an} is not known, but advantage can be taken from knowing its possible discrete values. For example, if {~} is binary and BB modulation is used, from (19) the likelihood functionl4

(24)

is obtained. A full-wave linear rectifier instead of a square-law device, followed by bandpass filtering, will therefore in this case be more appropriate for timing-phase estimation. This subject is still open for further investigation and extension.

Tracking carrier phase in the presence of significant time variations of ~c• termed carrier-phase jitter, which occurs on certain telephone channels, is another estimation problem which we shall not discuss here in further detail. Finally, the effect of signal distortion on the performance of the various timing- and carrier-phase tracking schemes must be considered as a further topic worth closer examination.

IV. ADAPTIVE EQUALIZATION

Suppose that 's and ~c have already been determined, but signal distortion still causes a significant amount of intersymbol interference (lSI). This situation occurs especially in data trans­mission over telephone channels where, owing to a priori unknown

Page 197: Optimal Estimation in Approximation Theory

192 G.UNGERBOECK

and slowly changing channel characteristics, a fixed equalizer cannot adequately correct distortion. This led in the mid 1960's to the development of adaptive equalizers4 which to some extent find now application also in data transmission over other channels. Let, after preconditioning the received signal by a fixed filter (possibly approximating a matched filter), the carrier-phase cor­rected signal at sampling instant nT + T be

s

X n L hna n + W 9- ._ n-._ n

(25)

where h9. I 0 for £ I 0 causes ISI. A linear adaptive equalizer is usually realized in the form of a transversal filter with T­spaced taps and adjustable tap gains ci, 0 2 i < N, as shown in Fig. 5. The output signal of the equalizer

z n

N

I i=O

C. X • ~ n-~

T C X

-n (26)

should approximate ~ as closely as possible. Using the mean­square decision-error criterion, one has

(27)

where R is the (N+l) x (N+l) autocorrelation matrix

Fig. 5. Adaptive transversal equalizer.

Page 198: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN DATA-TRANSMISSION SYSTEMS

R E(x -n

T • X )

-n

and b is the (N+l) vector

b E (x • a ) • n n

The tap gains that minimize (27)

c -opt

15 are

193

(28)

(29)

(30)

The estimation problem consists in estimating ~o~t from {~n}

and {en}, assuming that {~} = {an} is known w~th already suf­ficiently low probability of error. The stochastic approximation algorithm4,15,16

(n+l) c

ale 12 (n) 1 n

- a 2 (n)

c CJc

(n) (n) c - a e x

n -n

(n) a > 0 ,

solves this problem in most existing systems.

(31)

It has been shown that for the convergence of (31) the eigen­values of R and the·number of N play a significant role,l6,ll and that fa~test initial convergence of (31) is generally obtained by using the step-size parameterl7

a opt (32)

Therewith, ~2.3 N iterations are required to decrease E{lenl 2} by one order of ma~nitude, provided the eigenvalue spread of R is not excessive.l

In half-duplex and multi-point systems, fast receiver set-up is very important, and many attempts have therefore been made to shorten equalization time. A significant new concept has been cyclic equalization, 18 where a known cyclic data sequence with period equal to the length of the equalizer delay line is used for equali­zer training. This concept permits operation of the stochastic approximation algorithm at iteration rates above 1/T with precisely known values {an} from the very beginning. Note that the cyclic nature of the problem also permits discrete Fourier transform techniques to be used for estimating c . Another new concept

-opt

Page 199: Optimal Estimation in Approximation Theory

194 G.UNGERBOECK

has been equalization with tap spacing somewhat smaller than T (fractional tap spacing) which makes equalization possible for arbitrary sample-timing phases.l9

A theoretically interesting approach uses the Kalman filter algorithm for estimating the "states" c from20

-opt

(n+l) (n) ("state equation") c c -opt -opt

(33)

T ("measurement equation") a c X - e n -opt --n opt n

(34)

Convergence is considerably faster than with the usual stochastic approximation algorithm (31), but the algorithm is too complex for practical implementation. In effect, the Kalman filter algorithm can be regarded as a method for computing

(n) c

n -1

( I ~k~] · k=O

n

L (~k a.) k=O 1

[cf. (30)],

without ever inverting a matrix. The same can be achieved by computing

[ I - -T]-1 k=O~~

(35)

. . . . . 21,22 h' h 1terat1vely by means of the matr1x 1nvers1on lemma, w 1c leads to an algorithm still much more complicated than (31).

Further related topics are adaptive decision-feedback equalization23 and adaptive matched filtering.6 From an estimation point of view, these subjects are treated in much the same way as for linear equalization.

V. SIGNAL DETECTION

In the preceding two sections continuous-valued parameters had to be estimated which may change only slowly. The estimation of these parameters plays only an auxiliary role for the main function of the receiver: detecting the discrete signal values {an}· Let 's and ~c again be already decided, but ISI not necessarily be removed from the signal,

Page 200: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN DATA-TRANSMISSION SYSTEMS

X n

195

(36)

where L is the lSI-memory order of the transmission system. Let {~} be an aribitrary sequence of values from the finite alpha­bet ~ = {aO, al, •• aM-1}. If L = 0 (zero lSI), independent decisions can be made by determining ~ = ai : lxn- ho ail 2 +Min. This leads to the usual symbol-by-symbol detection approach. It is sometimes, however, very reasonable not to eliminate lSI completely, especially when the signal is strongly attenuated at the band edges and linear equalization would lead to severe noise enhancement.

In decision-feedback equalization only leading lSI is elimi­nated by linear equalization, whereas trailing lSI is removed with­out noise enhancement by using previous independent signal de­cisions to compute

X n

a + w n-JI. n

L

2 Jl.=l

a n-JI. (37)

Decision-feedback equalization leads to improved error performance, but still involves the problem of error propagation. To obtain the optimum ML receiver one must recognize that at the output of the linear section of an optimal decision-feedback equalizer24 (=Forney's whitened matched filter25) the noise samples are in­dependent.26 Assuming Gaussian noise, from (36) the likelihood function

(38)

is obtained. Hence, the optimum receiver must find the discrete sequence {an} which minimizes (38) in an unbounded interval. This sequence can be determined efficiently by a dynamic pro­gramming algorithm, called the Viterbi algorithm.25,27 To explain this algorithm let

0 n ~ {an-L+l ••• an} (39)

be one of ML states in a finite-state machine~ and let (cr 1 , a ) be a possible transition. Then (38) can be written as n-

n

A(~-1 {an})~ 2 lx - F(8 1,8 >1 2 , -~ n n- n (40)

n

Page 201: Optimal Estimation in Approximation Theory

196 G.UNGERBOECK

where

F(8 1,8 ) n- n a n-R. (41)

and minimizing (38) is equivalent to determining a sequence of states {8n} in the state transition diagram ("trellis diagram") of the finite-state machine. Suppose now that up to time n - 1 the best paths through the trellis diagram are known that terminate individually in the ML states 8n-l· Call

{ .. n-1

Min I I~- F(8k-l'8k)J 2 8 2 } k=-oo n-

8n-l• and let the sequence { ..• crn-2}, minimum, be the associated path history.

(42)

the survivor metric at which in (42) achieves To extend from n - 1 to n, the following recursion must be used:

J(8 ) n

Min {J(8 _1) + Jx - F(o 1 ,8 )J 2} (cr -+cr ) n n n- n

n-1 n

(43)

where minimization goes over all states crn-l that have a tran­S1t1on to On· Proceeding in this manner and updating path histories accordingly will leave, in the trellis diagram, a common path which represents the best estimate of {an} ~ {crn}. For M = 2 and L = 1, this is illustrated in Fig. 6.

An equivalent algorithm with different metric calculation can be obtained direct from (8), if gne considers S = s_R, and SR, = 0 for L > 0, and therewith writes R.

A(~J{an}) "' I [ 2 Re (y ~ ) - G(cr 1 ,8 ) ] n n n- n (44) n

where L

G(8 1,a ) a so a + 2 Re (~ I SR, a R.) . n- n n n n R.=l

n- (45)

The new recursion then becomes

J'(cr) = 2 Re (y n n

~ ) + Max {J' (cr _1) - G(o 1 ,a )J. (46) n (o -+cr ) n n- n

n-1 n

Page 202: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN DATA-TRANSMISSION SYSTEMS

(T =+1oL ~ n-2

(T = -1 010----oo---o n-2 n-2 n-1

0

n n+1 n+2

n+2 COMMON PATH

~--~-~l

n+3

n+3

197

Fig. 6. Illustration of the Viterbi algorithm for M = 2 and L 1.

The different metric calculation results from the fact that instead of the output Xn of a whitened matched filter or linear section of a decision-feedback equalizer, the output Yn of a matched filter is used direct, where in the case of non-zero lSI, the noise is correlated. Assuming that the quantities F and G are in both cases precomputed, the new metric calculation apparently offers computational advantages.

In Ref. 6 it has been shown that the algorithm is very robust against lSI in that·ISI has essentially no influence on error pro­bability if

(47)

If (47) is satisfied with equality, there will normally be a spectral null in the signal spectrum. Equalization to remove lSI is then no longer possible. The Viterbi algorithm, however, still persuits signal detection ~ith almost no degradation in error performance.

The Viterbi algorithm was originally develo~ed for optimum decoding of convolutionally encoded signals, 28 ' 2 where a different

Page 203: Optimal Estimation in Approximation Theory

198 G.UNGERBOECK

form of state-oriented dependence between successive signal values exists, but otherwise the same discrete optimization problem must be solved.

VI. CONCLUSION

In this paper the attempt has been made to demonstrate the various estimation and approximation problems arising in synchronous data-transmission systems. The presentation had necessarily to re­main incomplete in many respects. It is hoped, however, that it may serve as a useful introduction into this field. For a more thorough understanding the reader is referred to the original literature.

VII. REFERENCES

1H.L. Van Trees, Detection, Estimation and ModuZation Theory, Parts I, II and III, New York: Wiley 1968, 1971.

2 W.R. Bennett and J.R. Davey, Data Transmission, New York:

McGraw-Hill, 1965.

3 J.M. Wozencraft and I.M. Jacobs, PrincipZes of Communications Engineering, New York: Wiley, 1965.

4 R.W. Lucky, J. Salz, and E.J. Weldon, Jr., PrincipZes of Data Communicaton, New York: McGraw-Hill, 1968.

5H. Kobayashi, "Simultaneous adaptive estimation and decision algorithm for carrier-modulated data-transmission sys.tems," IEEE Trans. Commun. TechnoZ., vol. COM-19, pp. 268-280, June 1971.

6G. Ungerboeck, "Adaptive maximum-likelihood receiver for carrier-modulated data-transmission systems," IEEE Trans. Commun., vol. COM-22, pp. 624-636, May 1974.

7H. Robbins and S. Monro, "A stochastic approximation method," Ann. Math. Stat., pp. 400-407, 1951.

8 K.H. Mueller and M. Mueller, "Timing recovery in digital synchronous data receivers," IEEE Trans. Commun., vol. COM-24, pp. 516-530, May 1976.

9D. Maiwald, "On the performance of decision-aided timing re­covery," IBM Research Report, RZ 749, December 1975.

Page 204: Optimal Estimation in Approximation Theory

ESTIMATION PROBLEMS IN DATA-TRANSMISSION SYSTEMS 199

10L.E. Franks and J.P. Bubrouski, "Statistical properties of timing jitter in PAM timing recovery scheme," IEEE Trans. Commun., vol. COM-22, pp. 913-920, July 1974.

11D.L. Lyon, "Timing recovery in synchronous equalized data cotmllunication," IEEE Trans. Commun., vol. COM-23, pp. 269-274, February 1975.

12 k 11 • • • f • d • • II L.E. Fran s, Acqu1s1t1on o carr1er an t1m1ng data - I, presentation at NATO Advanced Study Institute on New Directions in Signal Processing, in Cotmllunications and Control, Darlington, U.K., August 1974.

13 G. Ungerboeck, unpublished work.

14P.A. Wirtz and E.J. Luecke, "Performance of optimum and suboptimum synchronizers," IEEE Trans. Commun. Teahnol., vol. COM-17, pp. 380-389, June 1969.

15 A. Gersho, "Adaptive equalization of highly dispersive channels for data transmission," Bell System Teah. J., vol. 48, pp. 55-70, January 1969.

16K. Moehrmann, "Einige Verfahren zur adaptiven Einstellung von Entzerrern fUr die schuelle Dateniibertragung," Naahriahten­teahnisahe Zeitsahrift, vol. 24, pp. 18-24, January 1971.

17 G. Ungerboeck, "Theory on the speed of convergence in adaptive equalizers for digital cotmllunications," IBM J. Res. Develop., vol. 16, pp. 546-555, November 1972.

18K.H. Mueller and D.A. Spaudling, "Cyclic equalization - A new rapidly converging adaptive equalization technique for syn­chronous data cotmllunication," Bell System Teah. J., vol. 54, pp. 369-406, February 1975.

19G. Ungerboeck, "Fractional tap-spacing equalizer and con­sequences for clock recovery in data modems," IEEE Trans. Commun., vol. COM-24, pp. 856-864, August 1976.

"20 D. Godard, "Channel equalization using a Kalman filter for

fast data transmission," IBM J. Res. Develop., vol. 18, pp. 267-273, May 1974.

21A.P. Sage, Optimum Systems Control, Prentice-Hall, Englewood Cliffs, N.J., 1968.

Page 205: Optimal Estimation in Approximation Theory

200

22 . h. . d k b bl. h d R.D. G1t 1n an F.R. Magee, Jr., wor to e pu 1s e •

23P. Monsen, "Feedback equalization for fading dispersive channels," IEEE Trans. Info. Theory_, vol. IT-17, pp. 56-64, January 1971.

24J. Salz, "Optimum mean-square decision-feedback equalization," Bell System Tech. J._, vol. 52, pp. 1341-1373, October 1973.

25G.D. Forney, "Maximum-likelihood sequence estimation of digital sequences in the presence of intersymbol interference," IEEE Trans. Info. Theory_, vol. IT-18, pp. 363-378, May 1972.

26R. Price, "Nonlinearly feedback-equalized PAM vs. capacity for noisy filter channels," Conference Record ICC 1972_, Philadelphia, pp. 22-12/16, June 1972.

27G.D. Forney, "The Viterbi algorithm," Proc. IEEE_, vol. 61, pp. 268-278, March 1973.

28A.J. Viterbi, "Error bounds for convolutional codes and an asymptotically optimum decoding algorithm," IEEE Trans. Info. Theory_, vol. IT-13, pp. 260-69, April 1967.

29A.J. Viterbi, "Convolutional codes and their performance 1n communication systems," IEEE Trans. Commun. Technol._, vol. COM-19, pp. 751-772, October 1971.

Page 206: Optimal Estimation in Approximation Theory

OPTIMAL APPROXIMATION IN AUTOMATED CARTOGRAPHY

Wigand Weber

Institut for Angewandte Geodiisie, Frankfurt a. M.

ABSTRACT

For several years, attempts are being made in many countries of the world to apply electronic data processing techniques in the production and updating of maps. Therein, cartographic generalization still is a vital problem.

Generalization in cartography is the process of transforming the contents of a map into a form appropriate for a map of smaller scale; it consists of several partial processes (as selection, geometric combination, qualitative summarization, simplification and dis­placement of map objects) which are interdependent in a non-hierarchic way.

It is shown that cartographic generalization is a problem of optimal approximation and that it may be described by a model of mathematical optimization. The basis of such a model is the calculation of semantic information content of the individual map objects in a way corresponding to the thesises of information theory and to the way people (most·probably) read maps. Prediction techniques play an essential part in this respect.

The generalization model is subsequently used as a scheme for the classification and judgement of the most important contemporary partial solutions of automated map generalization.

Concluding, three other problems in automated cartography are described which are solved by using methods of approximation theory. These are the reduction of the quantity of digital cartographic data to a minimal amount allowing the 'exact' recon­struction of a line, the Helrnert Transformation (which is a conformal transformation - optimal in the sense of the least-squares method) for the minimization of distortions in geometric data digitized from a distorted map sheet, and finally the calculation of splines for the reconstruction of lines from a limited number of given points for output on flatbed plotters.

201

Page 207: Optimal Estimation in Approximation Theory

202 WIGAND WEBER

1. INTRODUCI'ION

For several years many institutions all over the world try to apply techniques of electronic data processing also to the production and updating of topographic maps. The reason is not only curiosity as usual with scientists and engineers, but also and above all the necessity to offer a greater variety of different maps for a more and more specialized human community and economy. These maps must keep pace with the many changes in our surroundings in a way much quicker than it is usual up to now. There is no doubt, that such high demands cannot be met only by the usual manual methods, especially because a lack of highly specialized cartographic craftsmen has to be expected.

Automated cartography may be divided into four subtasks:

A c qui sit ion of graphic (analog) original data in digital form by suitable 'digitizers' from air photographs or already existing maps.

Storage of these digital data (essentially coordinates and quality information) in a data base with a suitable data structure and with possibilities of direct access and detailed data selection.

M a n i p u 1 a t i o n of the stored data in batch or interactive mode due to the topographic changes in the map area or during the derivation of small-scale maps from large-scale maps.

0 u t p u t of the digital data from the data base in analog (graphic) form using computer-controlled plotters.

Experiences gained so far show that data output is the best known subtask in the whole process; data storage may also be considered optimistically. Data acquisiton, however, is actually the bottleneck in the system, and with data manipulation -especially with the derivation of small-scale maps from large-scale maps human interaction via interactive graphic displays is still widely necessary.

At this point we have touched the problem of cartographic g e n e r a 1 i z a t i o n which is central in this paper. Preceding a more precise definition of this term as given in section 3, 'generalization' may roughly be described as 'the reduction of a map's information content due to a reduction in scale'. Generalization, of course, takes already place when initially portraying the natural topography on a map. In this process the cartographer's task is to approximate the natural topography by his map ' a s w e 11 a s p o s s i b 1 e ' . Thus it becomes evident that cartographic generaliza­tion belongs to the field of approximation theory: Cartographic generalization must be done in such a way that from the limited amount of information shown on the map, natural topography may 'optimally' be recovered.

Within generalization the following p a r t i a 1 p r o c e s s e s may be discerned; all of them are due to the reduction in the map surface caused by the reduction of scale:

Page 208: Optimal Estimation in Approximation Theory

OPTIMAL APPROXIMATION IN AUTOMATED CARTOGRAPHY 203

S e I e c t i o n of objects to be omitted.

S imp I if i cat i o n of lines (also area outlines).

C o m b in a t i o n of several objects into one object (mostly represented by en­larged symbolization).

S u m m a r i z a t i o n of several object qualities into one.

D i s p 1 a c e m e n t of objects because the symbols of neighbouring objects re­quire relatively more place at smaller scales.

The process of generalization is essentially complicated by the fact, that most of its partial processes depend on others (e. g. displacement depends on selection and vice versa). Although this dependence is not always hierarchic, attempts have been made to establish such a hierarchy for purposes of automated cartography [1]. This hierarchy corresponds to the sequence of the partial processes mentioned above.

The publications concerning automated generalization nearly exclusively deal either with the partial process of selection or simplification - mostly starting from quite different premises (see section 5). The lack of a comprehensive mathematical model of cartographic generalization is often deplored, and S t e w a r d is r4t}tt, when he states [ 2] that the profit of such models is based on the conceptual clarification of existing relations which allows for a more reliable categorization and evaluation of preliminary partial solutions, - even though the model as a whole could not be converted into a software package under present conditions.

On the following pages a contribution is made to such a model, using the nomenclature of mathematical optimization, information theory, and statistics. Stimulations were also taken from publications [3], [4], [5], [19], [20] and [21].

2. DETERMINATION OF INFORMATION CONTENT

As during generalization a map's information content is changed in a systematic man­ner, the first task is to fmd a method of its determination. In information theory, 'information' is generally defmed as the 'diminuation of uncertainty'; therefore, the information content of some character is assumed to be the bigger the greater the improbability of its occurrence is. According to this defmition, the information con­tent of a town shown on a map is not necessarily high, only because it has many inhabitants: rather depends its information content on whether it is situated in a region with many big towns (i.e. where one really 'expects' big towns), or in a lonely steppe, where the existence of a big town is very 'surprising'. Similarly a mountain rising abruptly from a plain must be depicted on a map, whereas the same mountain has only a low information content in a mountaineous region and could therefore be omitted.

The information content I of an appearance which has the probability p is expressed by the formula

Page 209: Optimal Estimation in Approximation Theory

204

I= ld .!_ (bit). p

WIGAND WEBER

(1)

In this formula 'I' means either the 'syntactic' or the 'semantic' information content depending on whether the probability p refers to the appearance of the characters used for display or to the meaning they transmit. Investigations published so far deal nearly exclusively with the syntactic information content of maps and mostly even with an average value only which refers to a whole map sheet and is called entropy. The entropy E is calculated from

E = ~ p· · ld _!_ i I Pi

for discrete appearances and for continuous appearances from

+oo 1 E = J P( ) • ld - • dx

-oo X P(x)

supposing stochastic independence.

For generalization, however, the values calculated in this way are of little use, as only semantic information content is relevant to the map user - and this not for the whole map sheet, but for any individual cartographic object shown on it: only in this way can a decision be made, which object is to be canceled or combined with others during generalization without diminishing the map's total information content more than inevitable.

With a cartographic object (e. g. the image of a town, a forest or a road) several types of semantic information are to be distinguished.

Some of these types refer to the p r o p e r t i e s of the single objects. These are: their geometric information (position, length, area), their quality information (e. g. 'town', 'forest' etc.), and finally their numerical supplementary information (e. g. the number of inhabitants of a town). ·

Other types of information result from r e 1 a t i o n s between several objects, that is from sets of pairs or triples (etc.) of objects: The most important relations shown on a map are the connection relation in its traffic network, the neighbourhood relation of objects (as the 'over' or 'under' in the crossing of roads with railways or the adjacency or intersection of different objects), and finally the topological incidence relation in the planar graph describing a network of area-borderlines (e. g. of communities) or in the tree-graph describing a river-network.

Subsequently, a method of determining the semantic information content of individual cartographic objects is proposed, assuming the information content of an object to be a function of the difference between the data attached to it and the data to be predicted there from its surroundings according to some prediction strategy using statistical relations. (The data used for the prediction may be of equal or different kind as those to be predicted so far, as there is only a sufficiently strong interrelation between them.) This assumption corresponds both to the definition of 'information content' in

Page 210: Optimal Estimation in Approximation Theory

OPTIMAL APPROXIMATION IN AUTOMATED CARTOGRAPHY 205

information theory as cited above and imitates the behaviour of a map user, to which evidendy only those appearances of a map are of higher interest which he is not able to deduce already from the general 'information trend' of the map or from his empirical knowledge about the interdependences of topographiC facts.

The prediction techniques to be applied with properties are different, depending on whether data of equal or different type (compared to the data to be predicted) are used and whether these data are (quasi) continuous or discrete (see paragraphs a) through c)); prediction techniques with relations are dealt with in paragraph d):

a) As to the prediction of continuous data from data of equal type, its technique considers the fact, that the 'information trend' mentioned above is conveyed to the map user by the low frequencies of the data which can be imagined to be a surface above the map plane, whereas the information content is contained in the higher frequencies. Beside by a Fourier Transformation, the smoothed surface of informa­tion trend can also be determined by regression using polynoms of limited order, 'least-squares-interpolation' according to the theory of linear ftlters of Wiener­Kolmogoroff, or by applying the 'gliding weighted arithmetic mean'-technique (also called 'convolution', in which the points of the original surface are replaced by the mean of the values in their surroun~s after weighting them according to their horizontal distance). The distances of data points from this trend-surface are ftnally the basis for the determination of the information content of the appertaining cartographic objects. - In this way the geometric information and the numerical supplementary information mentioned above may be treated*).

b) Another approach is to be taken, when the types of predicting and predicted data are different. E. g. there is a strong correlation between the position of rivers and the curvature of the earth's surface (as expressed in the contour lines) and mostly also vice versa. The strength of the (linear) statistical depe!ldence between the two variables can be judged after the calculation of their covariance and the correlation coefficient to be deduced from it**). If its amount is not too far from 1, one variable may be used for the prediction of the other one, using a linear formula. The dependence of one variable of one or more other variables may also be found by regression.

*) As an example the geometric information of the rivers of a map is used: At Hrst the map is subdivided by a raster of quadratic meshes of suitable width, and the total length of the rivers in each mesh is taken as z-coordinate of "a control point to be placed above the center of the mesh. These control points are then used for the calculation of the trend-surface. The vertical distances of the control points from the trend-surface defme the information content (using e. g. formula (2) to be given below), which is then distributed to the rivers (or river portions) of each mesh proportionally to their length. Similarly point and areal objects may be treated either with respect to their geometric information or to their numerical supplementary information as numbers of inhabitants, height values etc.

**) Correlation coeffl.cients may also be calculated by 'factor analysis' as is shown e. g. in [7].

Page 211: Optimal Estimation in Approximation Theory

206 WIGAND WEBER

As in paragraph a) the differences between the calculated (predicted) value and the actual value are the basis for the determination of the information content of the appertaining cartographic object.

c) When trying to predict the quality (e. g. 'river', 'town', etc.) of a cartographic object on the background of its surroundings, one has to consider the fact that it is of discrete nature - in contrast to the information types dealt with in the para­graphs a) and b) above. Therefore, the conception of 'conditional probability' is used which is known from statistics. The conditional probability p (B/A) of an event B with reference to event A means the probability of the occurrence of B, if A has occurred and can be expressed by the formula

(B/A} = p (A n B) . p p(A)

E. g. there is a greater probability that a farm exists in a field than in a forest.

These conditional probabilities have to be determined empirically - and that separately for every distinct type of landscape.

d) As stated above, information in cartographic objects can also result from relations established by them, e. g. by roads in a traffic network. In such cases, predictions as to the existence of a connection (edge) between two nodes can be made. For this purpose, the methods of Dijkstra (for the calculation of the shortest path b~tween two nodes in a labeled graph) or of Ford and Fulkerson (for the determination of maximal flows in networks) could be applied. -The differences between the path­length or the flows in the networks, containing and lacking the edge under con­sideration, is then used for the calculation of its probability.

In the above paragraphs (except c)) only differences between a predicted and an actual value of some information variable have been calculated instead of probabilities p as they are required for formula (1). For the transformation of differences into proba­bilities an empirical density distribution or in the lump the Gaussian law of errors can be used:

1 (2) P(e) = m ...;z:if • e

Herein m is the mean square error and de the tolerance of e. From the formula it can be seen that the probability of some information about a cartographic object becomes the lower, the greater the difference between the predicted and the actual value is; consequently, its information content - to be calculated from formula (1) - will attain a high value.

A difficulty arises from the fact that (as already mentioned above) a cartographic object may have several types of semantic information for which predictions are made and probabilities are calculated separately. Theoretically, one could try to modify formula (2) by introducing a multi-dimensional density distribution or to have a sepa­rate object function for each of the different information types instead of only one for

Page 212: Optimal Estimation in Approximation Theory

OPTIMAL APPROXIMATION IN AUTOMATED CARTOGRAPHY 207

all together according to formula (3) (see below). However, for practice it seems to be sufficient, simply to add for each object the information contents of the different information types after appropriate weighting according to the multiplication law of statistics neglecting stochastic dependences.

The method of determination of information content described so far, still implies some subjective decisions to be made: among them are the choice of the degree of smoothing of the trend-surface, of the mesh width of the raster used with some of the predictions, and of the weights, when adding the different types of information con­tents. Also there is still some uncertainty about the question, which of the many interdependences in map information are really unconsciously perceived by map read­ers and used for their 'optimal estimation' of how the natural topography may look like.

Having determined the semantic information content, we dispose of an essential pre­requisite for the generalization model.

3. A MODEL OF CARTOGRAPHIC GENERALIZATION

In the model of cartographic generalization described subsequently, the terminology of mathematical optimization calculus is used. As already mentioned in section 1, topo­graphy has to be approximated during generalization 'as well as possible' - that is o p t i m a 11 y in a sense to be defmed more exactly hereinafter.

Since the development of the simplex algorithm for linear optimization some 40 years ago, mathematical optimization has made considerable progress, especially with respect to nonlinear and integer optimization. The importance of mathematical optimization for geodesy and cartography was most recently emphasized by the 'International Union for Geodesy and Geophysics' in Resolution No. 22 of its General Assembly 197 5 in Grenoble.

Within mathematical optimization the unknowns of a problem are to be determined in such a way that a so-called object-function is maximized or minimized. In this process the unknowns or functions of them have to fulfd supplementary conditions (restri­ctions), given either as equations or inequalities. Recent developments aim at a gener­alized optimization containing more than one object function.

During cartographic generalization, one tries to keep the total information content of the result as high as possible. Therefore, the object function of a map sheet is defined as follows:

(3)

I; is the semantic information content of the ;th cartographic object in the initial map as determined accord~ to section 2; gk is a constant weight with which the members of the k th category of cartographic objects of a pragmatic evaluation scale contribute·

Page 213: Optimal Estimation in Approximation Theory

208 WIGAND WEBER

to the total information content*). ei, Zil and Zim belong to the unknowns of the optimization model: ej is a 0-1 variable ('selection variable') and indicates, whether or not the ph ring of combinable objects of the initial map containing the object i will appear on the generalized map; z;z and z;m are the 'combination variables' of object i with reference to its neighbouring objects land m**). The combination variables are e. g. allowed to vary between the values Zmin = 1/3 (with full combination) and Zmax = 1/2 (with no combination). The assumption of different degrees of combina­tions is in agreement with practice, where combined objects are symbolized by areas of more or less extent.

Now the restrictions of the optimization model will be specified. They are typical for cartographic generalization and responsible for the fact that normally the maximum of the object function will not reach the information content of the original map:

S h i f t r e s t r i c t i o n s : In order to guarantee a certain positional accuracy, shifts of objects or parts of them caused by displacement (see above) may not exceed a certain maximum. If parts of objects are shifted differently, additional steadiness conditions have to be defined for the shift vectors.

D i s t a n c e r e s t r i c t i o n s : Caused by the dimensions of map symbols and resolving power of the human eye, the distance of adjacent map objects may not fall below certain minimal values D. The left side of a distance restriction, referring to the objects i and k, is represented by an expression for the distance of these objects after shifting them by still unknown shift vectors; their right sides are given by the following reference value R which depends on both the selection and combination variables of the objects i and k:

Zjk - Zmin R = • D + (1 - e· • ez) • N,

Zmax -Zmin 1

where N is a negative constant of suitable value and ej, ez are the selection variables of the rings of combinable objects, in which the oojects i and k respectively are contained.

D e n s i t/ r e s t r i c t i o n s : Owing to the distance restrictions, it is already guarantee to a certain degree that the generalized map remains readable after scale reduction. However, additional restrictions concerning maximal density of symbols per square unit are inevitable, if legibility of the map shall be maintained. Because symbol density will often change throughout a map sheet, several such restrictions have to be defmed. Of course, these restrictions will contain both the selection and the combination variables of the map objects.

S t r u c t u r e r e s t r i c t i o n s : By using these restrictions, it shall be assured that after displacement typical geometrical relations are maintained: e. g. that a straight row of houses remains straight.

*) Thus g1e transforms semantic information content into what information theory calls 'pragmatic' information content, i. e. the information content in which a s p e c i a 1 user is interested in.

**) In agreement with the 'assignment problem' of operations research, with each object exactly two references to (an) other object(s) have been presumed.

Page 214: Optimal Estimation in Approximation Theory

OPTIMAL APPROXIMATION IN AUTOMATED CARTOGRAPHY 209

Sup p 1 em en tar y restrictions : They comprise the 0-1 condition for the selection variables and conditions forcing the values of combination variables into the range between Zmin and Zmax·

In the generalization model described above the partial processes of selection, combi­nation, and displacement are taken into consideration, - summarization and simplifica­tion have not yet been mentioned. As to the latter, it can be treated in advance as a separate optimization problem, because the 'micro-information-content', exclusively changed by simplification, may be considered to be independent of the 'macro-infor­mation-content' used in our model. As to summarization, it has to be done in accord­ance with a rule which is uniform for a certain scale change and not only for the generalization of an individual map sheet, thus ftguring as a constant in our model. Though these rules have been developed and standardized in a long period of carto­graphic practice, scientists try to find an objective basis also in this area using methods of operation research [8].

4. EVALUATION OF THE GENERALIZATION MODEL

Theoretically the model of section 3 refutes the thesis that the numerous and not only hierarchic interdependences of the partial processes of cartographic generalization could not be imitated by sequential algorithms as characteristic for EDP-machines. However, experts in mathematical optimization will notice that the numerical system, necessary for the generalization of only a rather small part of a map sheet, will (inspite of some simplifying neglections*)), be so large that optimization algorithms of today would have to fail with respect to time and numerical stability. This impression is intensified by the fact that 0-1 variables, non-linearities, and even non-convexities are implied.

In addition, it is to be taken into account that the different restrictions could pre­sumably be written down by a human being looking at the map- that, however, for a computer the same task causes serious recognition problems: he must e. g. during the formation of the distance restrictions, be able to decide, which objects are adjacent. The same is true with respect to the definition of combination variables. Perhaps these problems could be overcome by using an adequate data structure in the data base**) or by solving the 'assignment problem' of operations research. However, when forming the structure restrictions, the very complex area of pattern recognition is already touched.

But what is the model then good for at all? - In my opinion, its use consists in the clarification of the relations between information content, degree of generalization, and 'optimality', and of the complex interdependence of the various partial processes of generalization. Additionally, such a model permits a more detailed judgement on

*) E. g. the assumption of the information content of map objects in the optimization model as being constant.

**) E. g. by representing the 'intersection graph' (which describes the neighbourhood relations and intersections of map objects by its 'cliques') in a data structure with n-array relations [9].

Page 215: Optimal Estimation in Approximation Theory

210 WIGAND WEBER

character, suitableness, and interrelations of partial solutions already devdoped. Some of these partial solutions will be described in the following section. So far, they have not yet been integrated into an integral mechanism of generalization.

5. SOME AVAILABLE PARTIAL SOLUTIONS OF AUTOMATED CARTOGR.A.PHIC GENERALIZATION

S e 1 e c t i o n : In automated cartography the so-called law of Topfer is often used. It was found empirically and serves the calculation of the number np of objects on the generalized map from the number nA of objects at the original scale:

np=nA • ~; mA, mp are the reciprocal scales of the original and the generalized map. Supplemen­tary coefficients are often introduced additionally to adapt the law of Topjer to the special characteristics of individual object types. The law does not say anything about whether an individual object should be retained or cancelled. As to this question, until now 'rank descriptors' ( 1 0] are widely used which are at best determined by a weighted summation of different data types attached to an object and which, by their value, decide on the selection - independently of the information trend in the object's surroundings.

S i m p 1 if i c a t i o n : Automatic methods of simplification are applied to lines, surfaces, and point agglomerations:

a) L i n e s : At present, lines are mostly smoothed by applying a 'gliding weighted arithmetic mean', independently of whether they are given in point- or raster re­presentation. The method was taken from the fllter theory and is known there as 'convolution' in the context of low-pass flltering; it was already described before. The diminuation of information content is achieved by diminating the high fre­quency portions in the Fourier spectrum of the line- that is the very one equipped with the highest information content. As a measure of the amount of diminuation of the information content of the line, the increase of its 'sampling distance' (i.e. the smallest wavdength contained in the line) is taken. - Other smoothing methods apply regression polynoms (piecewise) or Fourier series and assume the line's in­formation content to be proportional to their degree. - Still others assume the information content of a line to be expressed by the number of its relative curvature extrema and reduce them systematically during smoothing according to the law of Topfer*).

Typical for all these methods is the fact that for the determination of the informa­tion content, the aspect of probability (according to (1)) is replaced by other more heuristic quantities, and that the demand for maximal information content after generalization is replaced by the principle of its reduction by a fixed percentage according to the law of Topjer.

*)For more details of all these methods see (4].

Page 216: Optimal Estimation in Approximation Theory

OPTIMAL APPROXIMATION IN AUTOMATED CARTOGRAPHY 211

b) S u r fa c e s : Surfaces are simplified, i. e. smoothed, very similarly to lines. Addi­tionally, 'multiquadratic functions' were introduced for smoothing interpolation [11], [12]. Also 'least squares interpolation' (sometimes called 'linear prediction') [5], [13] and 'gliding slant planes' as local regression functions [14] have been applied.

c) Point a g glomera t ions: Point agglomerations (namely the houses of settlements) have been smoothed using the method of convolution already men­tioned [15]. A measure for the diminuation of information content was not deter­mined in this case. Points with a great information content (e. g. a lonely restaurant in a forest) are, of course, lost with this method.

Displacement: As far as I know, until now only the publications [16] and [17] deal with cartographic displacement. In the methods described there, maximal shifts and areas of shift influence are taken into account. However, conflicts in displacement, caused by lack of place, are not solved by selection or combination processes or displacement of more distant objects.

In this context, a remark of B o y l e may be of interest, as it considers our problem from quite a different (rather Anglo-Saxon pragmatical) point of view: "Maps should be matched to the new capabilities. Generalization problems could, for instance, be simplified and map accuracy improved, if symbolization or variation of line thickness could be used in place of feature displacement".

6. OPTIMAL APPROXIMATION IN OTHER DOMAINS OF AUTOMATED CARTOGRAPHY

After having described cartographic generalization as a problem of optimal approxima­tion, some of its other applications in different domains of automated cartographr shall be mentioned here more briefly: especially, the topics of 'data reduction', 'He­mert transformation', and the construction of splines are concerned.

In d a t a r e d u c t i o n , the problem is to find a minimal set of data sufficient for the 'exact' reconstruction of a line. This set may consist of point coordinates or coefficients of mathematical functions. Again the problem belongs to the domain of mathematical optimization. However, in contrast to generalization, here the amount of data, necessary for the coding of a line, is to be minimized (object function) with the restriction that either its information content remains constant or the reconstructed line differs from the original line only by predefined maximal amounts. The problem has been tackled until now using either the least-squares-method [ 4] or the Tscheby­schef approximation [ 18].

The H e l m e r t t r a n s f o r m a t i o n is often used for the rectification of coordi­nates, digitized from distorted images, or as a quick approximative transformation of one type of planar representation of the earth's surface into another one. It is a conformal projection minimizing the sum of squares of the transformation residuals in the control points.

As to the calculation of s p l i n e s please refer to other papers in this volume.

Page 217: Optimal Estimation in Approximation Theory

212 WIGAND WEBER

BIBLIOGRAPHY

[1] Christ, F.; Schmidt, E.; Uhrig, H.: Untersuchung der Generalisierung der topo­graphischen Obersichtskarte 1 : 200 000 auf mogliche Gesetzma.Bigkeiten. -Nachrichten aus dem Karten- und Vermessungswesen, Series I, Nr. 51, Frank­furt a. M. 1972.

[ 2] Steward, H. ]. : Cartographic generalization, some concepts and explanation. -Cartographica, supplement No. 1 to the Canadian Cartographer, Vol. 11, 197 4.

[3] Hake, G.: Der Informationsgehalt der Karte- Merkmale und Ma.Be.- In: Grund­satzfragen der Kartographie, Osterr. Geograph. Gesellschaft, Wien 1970, pp. 119-131.

[ 4] Gottschalk, H.-].: Versuche zur Definition des Informationsgehaltes gekri.immter kartographischer Linienelemente und zur Generalisierung. - Deutsche Geoda­tische Kommission bei der Bayer. Akademie der Wissenschaften, Series B, No. 189, Frankfurt a.M. 1971.

[5] Lauer, S.: Anwendung der skalaren Pradiktion auf das Problem des digitalen Gelandemodells.- Nachrichten aus dem Karten- und Vermessungswesen, Series I, No. 51, Frankfurt a. M. 1972.

[ 6] Bollmann, ]. : Kartographische Zeichenverarbeitung - Grundlagen und Verfahren zur Quantifizierung syntaktischer Zeicheninformation. - Diplomarbeit, Freie Universitat Berlin, 1976.

[7] Mesenburg, K. P.: Ein Beitrag zur Anwendung der Faktorenanalyse auf Genera­lisierungsprobleme topographischer Karten. - Dissertation, Universitat Bonn, 1973.

[8] Monmonier, M.S.: Analogs between class-interval selection and location - allo­cation models.- The Canadian Cartographer, Vol. 10, 1973, pp. 123-131.

[9] Weber, W.: Datenstruktur eines digitalen Gelandemodells. - CAD-Bericht KFK-CAD 11 'Das digitale Gelandemodell', Gesellschaft fi.ir Kernforschung, Karlsruhe, September 1976.

[ 1 0] .Kadmon, N.: Automated selection of settlements in map generalization. - Car­togr. Journal, 1972, pp. 93-98.

[11] H~rdy, R. L.: Multiquadratic equations of topography and other irregular sur­faces.- JournalofGeophys. Res., Vol. 76, No.8, 1971, pp.1905-1915.

[12] Gottschalk, H.-].: Die Generalisierung von Isolinien als Ergebnis der Generali­sierung von Flachen.- Zeitschrift fur Vermessungswesen, 1972, pp. 489-494.

[13) Kraus, K.: Interpolation nach kleinsten Quadraten in der Photogrammetrie. Bildmessung und Luftbildwesen 40, 1972, pp. 3-8.

Page 218: Optimal Estimation in Approximation Theory

OPTIMAL APPROXIMATION IN AUTOMATED CARTOGRAPHY 213

[14] Koch, K. R.: Hoheninterpolation mittels gleitender Schragebenen und Prii.diktion. - Vermessung, Photogrammetrie, Kulturtechnik, Mitteilungsblatt 71. ann. set, 1973, pp. 229-232.

[15] Gottschalk, H.-].: Bin Rechnerprogramm zur Berechnung von Flii.chen aus flii.­chenartig strukturierten Punktmengen mit Hilfe der Bearbeitung von Binii.rbil­dern. - Deutsche Geodii.tische Kommission, Series B, No. 205, Frankfurt a. M. 1974.

[16] Gottschalk, H.-].: Bin Modell zur automatischen Durchfiihrung der Verdrii.ngung bei der Generalisierung. - Nachrichten aus dem Karten- und Vermessungswesen, Series I, No. 58, Frankfurt a. M. 1972.

[ 17] Schittenhelm, R.: The problem of displacement in cartographic generalization. -Attempting a computer-assisted solution. - Paper VIntli International Carto­graphic Conference, Moscou 1976.

[18] Tost, R.: Mathematische Methoden zur Datenreduktion digitalisierter Linien. -Nachrichten aus dem Karten- und Vermessungswesen, Series 1, No. 56, Frank­furt a. M. 1972.

[19] Pfaltz, ]. L.: Pattern recognition VI - MANS, a map analysis system. - Univer­sity of Maryland, Techn. Report TR-6 7-4 2, 196 7.

[20] Freeman,].: The modelling of spatial relations. -University of Maryland, Techn. Report TR-281, 1973.

[21] KMb, L; Karsay, F.; Lakos, L.; Agfalvi, M.: A practical method for estimation of map information content. -In: Automation the new trend in cartography, Final report on the ICA Commission III (Automation in cartography) scientific work­ing session, August 1973 in Budapest, Budapest 1974.

Page 219: Optimal Estimation in Approximation Theory

RECONSTRUCTION FROM X-RAYS

K.T. Smith, S.L. Wagner and R.B. Guenther

Oregon State University, Corvallis, Oregon 97331 D.C. Solman State University of New York at Buffalo, New York 14240

Introduction. According to T.J. Rivlin (these Proc.) "The problem of optimal recovery is that of approximating as effectively as possible a given map of any function known to belong to a certain class from limited and possibly error-contaminated information about it." More precisely, in the scheme envisioned by Rivlin there are given spaces X of possible objects, Y of possible data, and Z of possible reconstructions of certain features of the objects. Also maps are given between the spaces as follows:

F

(0.1)

objects X~ Z reconstructed features

data D i ~R For any given object 0, D(0), is the data coming from 0, and F(O) is the selected feature of 0; for any given set of data d, R(d) is the reconstruction of the selected feature provided from the data d by the particular reconstruction method R in use. A metric p is given on the space Z, and the number

(0.2) e(R) = sup p(F(0),RD(0)) OEX

is taken as a measure of the error in the reconstruction method R. The problem of optimal recovery is to minimize the error function e.

This research has been supported by the National Science Founda­tion under grant No. DCR72-03758 AOl

215

Page 220: Optimal Estimation in Approximation Theory

216 K. T. SMITH, S. L. WAGNER, AND R. B. GUENTHER

In this article we shall discuss three medical reconstruction problems from the point of view of this abstract formulation. These problems, which involve reconstructions from x-ray data, are quite different in nature from most problems that have been considered in optimal recovery. Therefore, it can be expected that there will be open questions and new points of view in both domains.

In particular, it is useful to broaden the notion of optimal recovery a little. Suppose given, in addition to the above, a posi­tive integer p and a norm on the space Rp satisfying:

If 0 ::;; x. ::;; y. , i = 1, .•. , n, then II x II ::;; II Y II · 1 1 --

Set X xP,y = yP,z = zP, and let F, D, and R be the obvious maps. The norm on Rp and metric on Z induce a natural metric on zP: if z = (z1 , ••• ,zp) and w = (w1 , ••. ,wp), then

p(z,w) = llp(z1 ,w1), •.• ,p(zp,wp) II Suppose given a subset XT c X for testing, and define

(0.3) e(R) = sup ii (F (D) ,RD(D)) . 0EXT

In some problems XT may be equal to x, but in medical problems XT is always finite.

Of course this seemingly broader scheme fits into the original, but then objects become p-tuples of the actual objects, etc., which is rather unnatural.

Table of Contents

1. Mathematical generalities.

2. Computerized Axial Tomography.

3. Discrimination between cancer and fibrocystic disease of the breast.

4. Noninvasive angiography.

Page 221: Optimal Estimation in Approximation Theory

RECONSTRUCTION FROM X-RAYS 217

1. MATHEMATICAL GENERALITIES

Since dimensions 2 and 3 are both important in x-ray work, it is convenient to give the statements and formulas for the general dimension n. If 0 is an object in Rn and e is a direction (i.e. a

n-1 point on the unit sphereS ), then the information provided by an x-ray from the direction e is effectively the total mass of 0 along each of the lines with direction e. Thus, if f is the density function of 0, this information is the function

00

(1.1) Pef(x) = f f(x+te)dt -00

.L for xe:e ,

where e.L is the n-1 dimensional subspace orthogonal to e. P8f is called the radiograph of O(or of f) from the direction e. Tfie reconstruction problem is that of recovering certain selected fea­tures of an unknown object 0 from (primarily) the knowledge of a certain finite number of radiographs. Since the radiographs are determined by the density function, the first basic question is the extent to which the density function itself can be recovered from the radiographs. (The density function does not tell the full story, however, even in radiology, for new radiographs can be taken after the object 0 has been modified in known ways, and the new radiographs can be compared with the old ones.) The answer to the question of the determination of the density function is contained in the simple formula relating the Fourie-r transforms of f and P ef [ 3] :

(1.2)

where

(1. 3)

-n

.L for i;e:e ,

~(/;) = (2n) 2 f e-i<x,l;> f(x) dx •

This shows that ~. and therefore f itself, are uniquely determined by the Pef, and leads to two inversion formulas. [3]

Fourier inversion formula: Given the Pef, set

~(0 (1/IZn) A .L (Pef) (1;) for ~;e:e ;

then -n

(1.4) f(x) = (h)2 f ei<x,/;>~(1;) dl;

Radon inversion formula:

(1.5)

where 1 ski orthogonal fined by

(1.6)

-11 n-111 n-21 f f(x) = (2n) s s n~lPef(Eex) de, . s k

is the k-dimensional area of.L the sphere S , Ee is the projection on the subspace e , and A is the operator de-

Page 222: Optimal Estimation in Approximation Theory

218 K. T. SMITH, S. L. WAGNER, AND R. B. GUENTHER

The operator A can be expressed directly (i.e. without Fourier transforms) in the following way. If

(17) R.g(x) l.

xi-yi J -- +1 g(y) dy, I In x-y

the integral being taken in the sense of the Cauchy principal value, then

(1. 8) Ag

Formula (1.5), with A defined by (1.8), is not quite the classical formula of Radon, for the Radon transform involves integration over t~e n-1 planes perpendicu~ar.to a direction e, rather than over the 1·1.nes parallel to 8, but 1. t 1.s the same sort of thing.

The Fourier and Radon inversion formulas presuppose that the radiographs Pef are known for all directions e. Since all real ob­jects have finite extent, i.e. the density functions have bounded support, the Fourier transforms are real analytic functions. This gives the following result. [3]

Theorem 1.9 The density function of any object is completely determined by the radiographs from any infinite set of directions.

Nevertheless, the practical case is the one in which only a finite number of radiographs are known, and in this case the theorem is rather unsettling. [3]

Theorem 1.10 Suppose given any object 0 with density function f. its radiographs from any finite number of directions. a totally arbitrary density function f', and an arbitrary compact set Kin the interior of the support of f. Then there is a new object 0' with the same shape and the same radiographs as 0, and the density func~ tion f' on the compact set K.

For technical reasons it is assumed in this theorem that the space X consists of all objects having infinitely differentiable density functions with compact support. It might appear more real­istic to consider finite dimensional spaces of objects. In this case the theorem is as follows. [3]

Theorem 1.11 Let X be a finite dimensional space of objects 0

with density function basis_ f 1 , ••• ,fN. Let V be the set of

directions e such that at least two objects in X have the same 0

radiograph from the direction e. Then V is an algebraic variety

inSn-l and there are polynomials q1 , ••• ,qN such that if e€V, then

Page 223: Optimal Estimation in Approximation Theory

RECONSTRUCTION FROM X-RAYS

(1.12) N E qj(e)fj<~> = o

j=l for all ~E9.L •

Since an algebraic variety is either the whole sphere or a very

219

small subset, and since condition (1.12) almost never holds on the whole sphere, the theorem shows that for almost any finite dimension­al space it is possible to distinguish between objects in the space by a single radiograph from almost any direction.

The most commonly used finite dimensional space X in recon­struction work is the space of step functions on a gri8. Let

Q = {x:lx. 1~1} be the unit cube in Rn, choose a number N, and parti­J

tion Q by the planes x. = k./N, where k. is an integer and J J J

-N~k.~N. Then X is the space of functions which are constant on J 0

each of the little cubes in the partition. Theorem 1.11 can be used to show easily that the only directions which do not distinguish between objects in X are those determined by the points k/N with lkji~2N. The 2 dimeRsional case of this was established by Mercereau

and Oppenheim [2].

This is a case where the infinite dimensional theorem reflects nature much more faithfully than the finite dimensional one. In­deed, it is evident that in practice an infinitesimal change in the x-ray direction cannot improve the information on a radiograph, while, on the other hand, experience shows that for practical pur­poses the infinite dimensional theorem is correct. Even with a great many x-ray directions, reconstruction methods do lead to medi­cally feasible but wrong objects with the correct radiographs. This problem can be lessened by incorporation into the method of addi­tional information about either the object or the reconstruction method - i.e. information not coming from the x-rays. A couple of examples are given in [3,5], but much work remains to be done.

The three reconstruction methods in current use for recovering the density function are the following (or modifications thereof).

1. Fourier Method. This is a numerical implementation of the Fourier Inversion Formula.

2. Convolution Method. This is a numerical implementation of the Radon Inversion Formula.

3. Kacmarz or ART Method. This is an iterative procedure which goes as follows. Fix a compact set Kin Rn containing all objects under consideration, and regard the density functions as

2 points in L (K). Fix the x-ray directions e1 , ••• ,9M' and let Nj be

Page 224: Optimal Estimation in Approximation Theory

220 K. T. SMITH, S. L. WAGNER. AND R. B. GUENTHER

the null space of P0 . and Qj be the orthogonal projection in L2 (K) J

on f+N., where f is the unknown true object. It is easily seen that

N. con;ists of the functions in L2 (K)which are zonstant on the lfnes with direction e., so that for any gin L (K), Q.g is comput-

J J able even though f is unknown. It can be shown that

(1.13)

Q being the orthogonal projection on f+N, where N is the intersection of theN .• For any g, Qg is an object with the correct radiographs

J from all the given directions. The Kacmarz or ART method consists in choosing an initial guess g and in taking the left side of (1.13) with a suitable mas an approximation to the unknown f. A proof of the convergence in (1.13) is given in [3] and an explicit description of the rate of convergence is given in (7].

In the present situation where the objective is the recovery of an approximation to the density function, the optimal re~overy scheme looks as follows. The space X of objects is the space L (K), where K is a compact set in Rn containing all of the objects under con­sideration - i.e. X is the space of density functions of all objects under consideration. If M x-rays are taken, the space Y of data con­sists of all M-tuples of x-ray films iMor, since each film is read at a finite number m of points, Y = R . Normally, the space Z is the space of step functions on a grid, i.e. the space X described above, which means that the true density functions are ~pproximated by step functions on the grid.

The fundamental initial problem in the program of minimizing the error function is that of defining a relevant metric on the space z. In most significant medical problems, e.g. the distinction between normal and abnormal tissue, the density differences are extremely small. On the other hand, bone, for instance, is something like a honeycomb, so that the irrelevant difference between bone at one point and bone at another is likely to be very large. Consider, for example, the case of a head with a tumor, and two reconstruction methods, one of which finds the tumor but fuzzes up the skull a bit, while the other recovers a perfect skull, but does not find the tu­mor. Since the tumor is rather small and differs by very little from normal tissue, almost any natural metric on Z will show the second method to be very good - while in fact it is a disaster. Since the skull is rather large and very different in density from tissue, almost any natural metric on Z will show the first method to be a disaster - while in fact it is very good.

In this case some very subtle metric is needed, which first locates the skull and then gives very little weight to differences nearby.

Page 225: Optimal Estimation in Approximation Theory

RECONSTRUCTION FROM X-RAYS 221

2. COMPUTERIZED AXIAL TOMOGRAPHY

For reasons of computational simplicity, the recovery of the density function of an object is usually achieved two dimensionally with the recovery of cross sections. When the cross sections and x-ray directions are perpendicular to a fixed axis the procedure is called Computerized Axial Tomography, or CAT.

CAT came to medical prominence, although it had been in the air for some time, and had actually been used in other fields, about 1973 with the invention by G.N. Hounsfield of the EMI scanner, which reconstructed cross sections of the head by means of the Kacmarz method. The performance of the EMI scanner in pinpointing tumors, hemorrhages, blood clots, etc. was so dramatic as to revolutionize radiology. Indeed, this scanner was called the greatest advance in radiology since Roentgen's discovery of x-rays. In the interim scanners have proliferated. There are scanners using the Fourier and Convolution Methods in addition to those using the Kacmarz method - the trend seems to be toward the Convolution Method; and there are scanners designed for other specific parts of the body, such as breasts or forearms, and scanners designed for all parts of the body at once. Since the scanners are very expensive (with an average initial cost of about $500,000 and rather heavy maintenance charges), and since they have certain inherent disadvantages (some of which are discussed in [5]), it seemed worthwhile to see what could be done with standard hospital x-ray equipment commonly avail­able even in towns and villages.

The results of our research and experiments are described in [1,3,5], and explicit quantative comparisons with the results of the EMI scanner are given in [5].

In this problem the optimization set up is that described at the end of §1, with n=2. The space X of objects is (or can be taken to be, since the r~covery of the density function is what is de-2 sired) the space L (K) of density functions on a compact set KcR which contains all cross sections ~ the objects under consideration. The space Y of data is the space R , where M is the number of x-ray directions and m is the number of points at which each x-ray is read. The space Z is the space of step functions on an N x N grid, or, equivalently, the space of N x N matrices, which are called re­construction matrices. For a given object OEX, F(O) (see (0.1)) is the step function whose value on a square of the grid is the average density of 0 on this square, and D(O) is the mM tuple obtained by reading the M radiographs of 0 at the prescribed m points. At pre­sent, at least, the reconstruction map R is one of the three de­scribed: the Fourier Method, the Convolution Method, or the Kacmarz method - or a modification of one of these.

Page 226: Optimal Estimation in Approximation Theory

222 K. T. SMITH, S. L. WAGNER, AND R. B. GUENTHER

For the reason described at the end of §1, the optimization of recovery cannot even be studied, much less carried through: namely, no medically relevant metric on the space Z is known. Nevertheless, various quantitative limits to the accuracy of the reconstruction can be given in terms of the size N of the reconstruction matrix, the number M of tolerable x-rays, the number m of points at which they are read and the positive thickness of a real cross section as opposed to the zero thickness of a theoretical one. These limits and relationships between them are discussed in [5]. The other major factors affecting the accuracy of reconstructions are the In­determinacy Theorem (1.10) and the noise in the data. These are discussed, and certain remedies are discussed, from a much less quantitative point of view in [5].

The lessons to be learned from this particular problem appear to be the following. The problem fits admirably into the set up of Rivlin. The spaces X, Y, and Z, as well as the maps F, D, and R, are known. However, despite considerable knowledge, both quantita­tive and qualitative, on the factors limited the accuracy of recon­structions, nothing whatever is known regarding a truly relevant metric on the space z. For this reason the program of optimal re­covery breaks down. This is not a defect in the program, but a defect in our present practical knowledge.

3. DISCRIMINATION BETWEEN CANCER AND FIBROCYSTIC DISEASE IN THE BREAST

Early and reliable diagnosis of breast cancer is one of the most pressing needs in medicine today. Each year, in the United States alone, some 450,000 women are discovered to have abnormali­ties in the breast, about 90,000 of these are found to have cancer, and about 33,000 women die of breast cancer.

The problem divides naturally into two parts: 1) The development of a large scale screening procedure for the early detection of ab­normalities; 2) The development of a reliable and noninvasive pro­cedure for discriminating between malignant and non-malignant ab­normalities. (At the present time surgery is the only sufficiently reliable procedure for making this discrimination.)

We have begun research on the second problem and have developed a simple, cheap, and noninvasive procedure which, in the thirty-three cases studied so far, discriminates correctly between adenocarcinoma (the most common of the breast cancers) and fibrocystic disease (the most common non-malignant abnormality). Briefly it is as follows: details are given in [4,6,8].

On the mammogram (breast x-ray film) a line through the lesion

Page 227: Optimal Estimation in Approximation Theory

RECONSTRUCTION FROM X-RAYS 223

is read in steps of .1 mm. Suppose that D(x) is the reading at the point x. The edges a and b of the lesion on this line are located by examination of the readings themselves, a graph of the readings, and the film. The linear function L with values D(a) at a and D(b) at b is taken as an approximatio~ to the readings that would be pro­duced by normal tissue, and the difference

(3.1) A(x) = L(x)-D(x)

is called the abnormality function. A rectangle is circumscribed around the graph of A, and the ratio of the area of this rectangle to the area under the graph of A is called the linear mass ratio.

linear mass ratio

area of rectangle area under graph of A

The original readings D are highly dependent on both the film and the exposure. The abnormality function A is independent of the exposure, but depends upon the film. The linear mass ratio has the interesting property of being independent of both.

The linear mass ratios from the thirty-three cases studied are as follows:

1.29 1.32 1.34 1.35 1.39 1.40

Cancer

1.4. 1.46 1.48 1.51 1.51

1.53 1.56 1.57 1.59 1.59

Fibrocystic Disease

1.66 1.68 1. 75 1. 78 1.80 1.81

1.88 1.91 1.91 1.92 1.99 2.02

2.06 2.12 2.15 2.17 2.26

The reason for the success of the linear mass ratio is still obscure. It is not (as might be expected) due to a greater average density in cancer than in fibrocystic disease. Indeed, the linear mass ratio is largely independent of the average density, for any round lesion with constant density would produce a ratio 4/n. The linear mass ratio depends upon the internal distribution of density within the lesion, not upon the average density.

Onthe other hand, the average density seems to be of relatively low diagnostic importance in this problem. In the cases of this study, particularly in the typical cases where the lesion was not neatly circumscribed, but consisted of a primary lesion embedded in other abnormalities, no correlation could be found between the

Page 228: Optimal Estimation in Approximation Theory

224 K. T. SMITH, S. L. WAGNER, AND R. B. GUENTHER

average density and the surgical diagnosis.

Ifit is correct that the distinction between cancer and fibre­cystic disease lies in the internal density distribution rather than the average density, then the poor success of breast scanners is easy to explain. A 128 x 128 reconstruction matrix is being used to represent a cross section about 30 em. in diameter, so that a fairly large 1 em. lesion is represented by a 4 x 4 square in the matrix. A 4 x 4 square should give a fairly accurate picture of the average density, but no information whatever about the internal distribution.

In this problem the optimization set up as as follows. The space X of objects is the space of breasts, or of density functions of breasts. The spaceY of data is Rm, where m is the number of points at which each mammogram is read; and the data map D takes each breast, or density function, into the m-tuple of readings. The space Z can be one of two: z1 = real numbers, or z2 = {0,1}.

If Z = z1 , then the reconstruction map R carries a given set of data

to the linear mass ratio, while if Z = z2 , then R carries a given

set of data into 0 = cancer if the linear mass ratio is less than or equal to 1.63, and into 1 = fibrocystic disease if the ratio is larger than 1.63.

In the first case, where Z = z1 = real numbers, the feature map F: X+Z is, for practical purposes, uncomputable. Even with biopsies, autopsies, etc., it is not practical to compute total masses along lines spaced at .1 mm. intervals.

In the second case, where Z = z2 = {0,1}, the feature map is computable by biopsy. There is only one possible metric on z2 (up to a constant multiple), namely that with d(O,l) = 1. Thus e(R) = 0 if the reconstruction method R is perfect (on the class of test objects XT, as indicated in §2), and otherwise e(R) = 1.

Obviously this is a case where the optimization (0.3) must be used rather than the original (0.2). If, for example, p = 100 and Rp is given the ~l norm, then e(R) is the maximum number of errors

in any of the groups of 100 forming the test objects in XT. This would seem to be a relevant measurement of error.

It is apparent that in most, if not perhaps all, cases where Z is a finite set the optimization (0.3) should be used.

Page 229: Optimal Estimation in Approximation Theory

RECONSTRUCTION FROM X-RAYS 225

4. NONINVASIVE ANGIOGRAPHY

The usual procedure in radiology for vizualizing blood vessels is called angiography. It uses the injection of a large amount of iodine contrast dye into the vessel in question directly at the site of the examination, which can be painful and dangerous and always calls for hospitalization. The reason for the injection of a large amount of dye directly into the vessel at the site of the examination is that unless the dye is present in high concentration, nothing is visible on the radiograph.

The scanning microdensitometer used in this work to read the x-ray films is capable of discriminating between 16,000 shades of gray. (The film is not capable of recording that many, but in some of the work about 1000 shades were needed.) With such sensitivity, iodine dye, even in very low concentrations is easily detected. However, the changes in the readings due to the dye cannot be inter­preted, for they are lost in the changes due to changes in tissue from one point to another, changes in the thickness of the object, etc.

If one x-ray is taken before the dye injection, and others at various times afterwards, then, since the differences between the objects at these times are due solely to the dye, the differences between the densitometer readings should reveal the progress of the dye through the vessels and tissues.

This procedure was carried out in an examination of the blood vessels of the neck. A 50 cc. bolus of 28% iodine solution was in­jected into the principal (antecubital) vein of the arm. One x-ray was taken before the injection. The injection itself took 10 sec­onds. A second x-ray was taken just at the end, a third 5 seconds after the end, and a fourth 10 seconds after the end.

It turned out that the blood circulated much more rapidly than we had anticipated. By the time of x-ray 2 the dye was already widespread throughout the tissues and veins. At the time of x-ray 4 it remained in the tissues and veins, but had left the arteries. As might be expected, the difference between x-ray 1 and x-ray 4 revealed the veins and the thyroid gland, which has a very heavy blood supply. The difference between x-ray 2 and x-ray 4 revealed the arteries. The veins and tissues effectively cancelled because the dye was present at both times. (The difference between x-ray 1 and x-ray 2 did not reveal much of anything except widespread dye, which, before we appreciated how fast the blood had circulated, led us to believe the experiment had been a failure. X-ray 3 was not used.)

Page 230: Optimal Estimation in Approximation Theory

226 K. T. SMITH, S. L. WAGNER, AND R. B. GUENTHER

Although this is certainly a reconstruction problem, namely, to reconstruct the blood vessels in the neck, we do not see how to fit it into the optimal recovery scheme. In this case, the space X must be the space of actual necks, not of density functions. If there are two films, one before and one after, and if m lines are read on each film with n readings per line, then Y = M xM , where

11111 11111

M is the space of m x n matrices. Finally, Z = M • The data 11111 11111

map D is obvious, and the reconstruction map R is, as described, the difference. The feature map F, however, is unreachable. Even in theory there is no way to pass from a given neck to a matrix of numbers representing total masses along lines before and after dye injection.

In such cases a less refined feature space must be selected, i.e. a space Z' with a map r: Z+Z', for which a computable feature map F':X+Z' is available. Thus

For example, if the medically relevant question is whether the carotid artery is obstructed, then Z' = {0,1}. The map F' is com­putable either by standard angiography or by surgery, and the map R' is the diagnosis obtained by looking at the reconstruction.

If, however, the problemis to visualize the blood flow through the various vessels, we have no idea whether a suitable Z' can be found. Radioactive scans are used for this purpose, but they are much too crude to be of any use in measuring the error in this procedure.

REFERENCES

1. R. Guenther, C. Kerber, E. Killian, K. Smith, and S. Wagner, "Reconstruction of Objects. from Radiographs and the Location of Brain Tumors", Proc. Natl. Acad. Sci. USA 71:4884-4886 (1974).

2. R. Mersereau and A. Oppenheim, "Digital Reconstruction of Multi­dimensional Signals from their Projections", Proc. IEEE 62:1319-1338 (1974).

3. K. Smith, D. Solmon, and S. Wagner, "Practical and Mathematical Aspects of the Problem of Reconstructing Objects from Radio­graphs",

Page 231: Optimal Estimation in Approximation Theory

RECONSTRUCTION FROM X-RAYS 227

Bull. Amer. Math. Soc. (To appear).

4. K. Smith, S. Wagner, R. Guenther, and D. Soltnon, "The Diagnosis of Breast Cancer in Mammograms by the Evaluation of Density Patterns" (To appear).

5. K. Smith, S. Wagner, R. Guenther, and D. Solmon, "Computerized Axial Tomography from Ordinary Radiographs - An Alternative to Scanners" (To appear).

6. D. Solmon, K. Smith, S. Wagner, and R. Guenther, "Breast Cancer Diagnosis from Density Studies of Mamograms", International Conference on Cybernetics and Society, IEEE, November 1-3, 1976, Washington, D. C.

7. D. Solmon and C. Hamaker, "The Rate of Convergence of the Kacmarz Method", J. Math. Anal. & Appl. (To appear).

8. S. Wagner, K. Smith, R. Guenther, and D. Solmon, "Computer As­sisted Densitometric Detection of Breast Cancer", Application of Optical Instrumentation in Medicine V, SPIE 96: 418-422

(1976).

Page 232: Optimal Estimation in Approximation Theory

PLANNING OF RADIATION TREATMENT

Udo Ebert

Computing Center, University of MUnster

West Germany

ABSTRACT

The aim of radiotherapy is to destroy a malignant tumour. In accomplishing this goal one has to spare the healthy tissue and organs as much as possible since otherwise they could perish. Thus, the total irradiation time should be short.

These considerations lead to a mathematical model in which constraints reflect the restrictions on the dosage -lower bounds in the tumour and upper bounds in the healthy tissue. The time of irradiation becomes the objective function of the optimization problem. By fixing some of the possible parameters of the treatment one gets a model in which the velocity of the source of radiation can be determined. This solution is approximated, solving a linear and quadratic or parametric programming problem. The model is implemented as a programming system for radiation-treatment planning. An example is given showing its application to a kidney tumour.

0. INTRODUCTION

In this paper a mathematical model of radiation-treatment planning is presented. It is possible to determine within this model optimal treatment plans: The source of radi­ation can move along certain orbits around the patient and, as a result, an approximation of the optimal velocity can be computed, taking into account additional restrict­ions.

229

Page 233: Optimal Estimation in Approximation Theory

230 UDO EBERT

Although the model discussed here is different from the F-model [2] (in which fixed positions for the source of radiation are determined) it is possible to extend the usual programming system for radiation-treatment planning to this optimizing model too [1,3].

1. FOUNDATIONS OF RADIOTHERAPY

The aim of every radiation treatment program is to destroy malignant tissue within a patient's body. When a tumour is irradiated, the healthy tissue which is confronted with the high-energy rays may also be damaged. The task of radiotherapy is to determine a treatment plan which destroys the tumour and injures the healthy tissue as little as possible. Moreover, one has to take into account that in some parts of the body and some organs (e.g. the liver or spinal card) the dosage must not exceed a certain amount since otherwise these parts may perish. Therefore, a treatment plan has to satisfy the following conditions: In some parts of the patient's body minimal doses must be achieved (e.g. in the tumour) and in some parts (e.g. in the liver) maximal doses must not be exceeded.

These restrictions are reflected in a mathematical model by constraints on the dosage in the body. The goal of minimal injury of healthy tissue is reached by mini­mization of the total irradiation time.

The dosage at a point of the body depends on the source of radiation, the condition at the patient, and the time of irradiation. Essential points are the type of the source of radiation, its position, its shape and its alignment. In the patient's body different tissues absorb radiation in different manners. Thus the dosage at a point P is described by

( 1. 1)

t denotes the irradiation time and p 1 , ••• ,p designate the remaining parameters. We may assume tha~ D is con­tinuous.

The central ray of the cone of radiation is named the central axis.

If the source irradiates the patient for a fixed time from a fixed location with constant parameters, one calls it a fixed field, or simply field.

In the following, a model with a moving source of radiation is developed.

Page 234: Optimal Estimation in Approximation Theory

PLANNING OF RADIATION TREATMENT 231

2. MODEL FOR SPATIAL DISTRIBUTION OF FIELDS

2.1 DESCRIPTION

The model for radiation treatment which shall be examined here tries to distribute the directions in space from which the patient is irradiated in as uniform a man­ner as possible. It is called the S-model (model for spatial distribution of fields).

The underlying idea is easily explained: if one tries to destroy a tumour with one field then certainly the tissue lying between the tumour and the treatment unit will perish, too. Because of the absorption of radiation by the patient's body the dose there has to be even higher than in the tumour. As the doses of separate fields add up in every place it is more advantageous to use more than one field. These fields should come from very diffe­rent directions and overlap only in the tumour. This concept is generalized in the S-model: it is assumed that the source of radiation moves along one or more orbits around the patient. For each orbit the central axis of the field has to go through a fixed point within the patient, the target point, which is situated in the tumour. The remaining parameters of the source are also constant and are prescribed in advance. The orbits are chosen according to anatomical and clinical factors, for instance by providing a plane and a constant distance between the source and the point of entry of the central axis into the patient.

Then the s-model determines the velocity with which the source moves along the orbit: v [em/sec] or its in­verse w=l/v [sec/em], the marginal "staying" rate at a point of the orbit. The objective function of the optimi­zation problem is the total irradiation time, and it is to be minimized. As a result, advantageous parts of the orbits are used, i.e. the source of radiation will move slowly there. Disadvantageous parts will be avoided, i.e. the source will move very rapidly or not at all over these parts.

2.2 MATHEMATICAL MODEL

For an analytical representation of the model it is as­sumed that there are n orbits 0 .• Each orbit o. can be characterized by the continuously differentiable func­tions

(2. 1 )

Page 235: Optimal Estimation in Approximation Theory

232 UDO EBERT

These describe the coordinates of the orbit within a fixed global coordinate system (the patient coordinate system). We can choose the parametric representation so that u £ [0,1] for all orbits. c. denotes the proposed constant parameters of the i-th Brbit. Let wi(u) be the marginal staying rate on Oi; wi(u) is assumed to be Lebesgue-integrable on [0,1}. Let P be a point of the body B, a compact connected subset of 3-dimensional space, and define w=(w1•···•wn). Then we have the total dose

( 2. 2) D (P,w)= ~ 1Jo(P,x. (u) ,y. (u) ,z. (u) ,c. )w. (u) • s i=1 0 l l l l l

fx ~ (u) 2+y ~ (u) 2+z ~ (u) 2 du l l l

and the total time of irradiation

n 1J I 2 2 2 (2.3) F (w)= L: w. (u) v'x~ (u) +y~ (u) +z ~ (u) du. s i=1 0 l l l l

For P€B there exist constraints: the minimal dose M1 (P) and the maximal dose ~ (P). We may suppose the t-~1 and Mu are continuous functions. For medical reasons it 1s convenient to have a maximum marginal staying rate M, i.e. 1/M is a minimum velocity of the source; there should be no quasi-fixed fields.

So the system for the S-model can be expressed 1n the form

( 2 . 4 ) M l ( P ) ..::_ D s ( P , w) ..::_ Mu ( P ) P £. B

(2.5) o..::. wi(u) < M ut- [0,1], i=1, ... ,n

(2.6) wt. L[0,1] { (w1 (u), ••• ,wn (u)) I for i=1, ..• ,n

wi(u) Lebesgue-integrable on [0,1]}

(2.7) minimize Fs(w).

We have Theorem 1: If there exists at least one w satisiying (2.4)-(2.6) then there is an optimal solution w to the problem (2.4)-(2.7). Proof: For the exact details about the underlying (to­pological) assumptions and the proofs of this and the following theorems see [1].

Page 236: Optimal Estimation in Approximation Theory

PLANNING OF RADIATION TREATMENT 233

2.3 DISCRETIZATION OF THE MODEL

The relationships (2.4)-(2.7) describe the S-model as an ideal one. In practice we do not have complete infor­mation about the patient's body, only an approximation. Moreover, the weight function, w, should lend itself to practical application. This leads to the following con­siderations: 1) The constraint (2.4) is weakened. It is replaced by a finite number of inequalities

(2.4') M1 (P.) < D (P.,w) < M (P.) j=1, .•• ,m. J - s J - u J

If we choose P. suitably, (2.4) and (2.4') are almost equivalent: J Theorem 2: For every e>O we can find a natural number m-m(e) and m points P1, .•• ,Pm such that from (2.4'), (2.5), (2.6) follows

M1 (P)-e < D (P,w) < M (P)+e for all P£B and wfL(0,1], - s - u satisfying (2.5).

2) The function class (2.6) is restricted to a finite dimensional subspace of L[o,1]. Here polygonal functions are considered (analogous results can be shown for step functions) • Let ~ be a partition of [0,1] into r=r(1X) subintervals with O=u0 <u1< ••• <ur=1. Then we have the weight functions

u E' [uk-1 ,uk] k=1, ••• ,r~i=1, ••• ,n.

In the case of a closed orbit we suppose w~ = wf for i=1, ••• ,n. For this restricted problem we obtain an analogous result: Theorem 1': If there exists at least one w~ satisfying* (2.4), (2.5), (2.6') then there is an optimal solution w1lt to the problem (2.4), (2.5), (2.6'), (2.7). The restriction to polygonal functions does not lead to a considerably worse solution~ indeed, one can show Theorem 3: Let~ be a sequence of partitions of the interval [0,1} satisfying~~ Clt~+ 1 and

lim . max I ui -ui_1 I = 0 then -.r+oo J.=1, .•• ,r('tt)

yo

inf .,..eN

F s (wu ) ., * = F8 (w ) •

Page 237: Optimal Estimation in Approximation Theory

234 UDO EBERT

3) The above reflections prove that the S-model can be considered in the discretized form (2.4 1 ), (2.5), (2.6 1 ),

(2.7). Combining (2.4 1 ) and (2.6 1 ) we have

n + 0 r-1 + _ k (2. 4 1 1 ) M1 (P.) < I: (a. 0 (P.) w. + I: (a. k (P.) +a. k (P.)) w.

J - i=1 1 J 1 k=1 1 J 1 J 1

where

+ aik(P)

and

j=1, ••• ,m

uk+1J uk+1-u = D(P,x. (u) ,y. (u) ,z. (u) ,c.) uk 1 1 1 1 uk+1-uk

/x ~ (u) 2+y ~ (u) 2+z! (u) 2 du 1 1 1

These integrals can be computed without knowing thew .• The objective.function (2.7) also turns out to be li- 1

near in the wJ. So the S-model can be reduced to a prob­lem in linear1 prograrnrning.

2.4 SOLUTION

To solve this problem we can use any of the known methods of linear programming [11]. Applying one of these methods we find a basic solution (a vertex or extremal point of the polytope of the feasible solutions). If, because of symmetry, there is more than one optimal solution there will exist optimal solutions which are not basic solu­tions. Some of these will distribute the directions of radiation better than most other optimal solutions (e.g. the average of all optima~ basic solutions).

Page 238: Optimal Estimation in Approximation Theory

PLANNING OF RADIATION TREATMENT 235

2.5 IMPROVING THE SOLUTION

In such cases we can improve the solution with the help of one of the following two methods. For the sake of an easier description a new and simpler notation is intro­duced. The S-model can be represented in the following form

(2. 8) Ax < b

(2.9) 0 < xi < M

(2.10) minimize c'x

i=1, ••• ,r

where A is a matrix of dimensions qxr, b=(b1 , ••• ,bq) ', x=(x1, ••• ,xr>' ~ 0 are*column-vectors, and c=(c1, ••• ,cr) is a row-vector. Let x be an optimal solution. We want to find among all optimal solutions a most ap­propriate one. Therefore we add to (2.8), (2.9) the constraint

(2. 11) c'x * = c'x

and replace (2.10) by a different objective.

1) Modifying (2.10) we get r 2

(2.10') minimize E cixi • i=1

Substantially this means seeking a time-optimal solution of such a kind that its functional values have as little variation as possible. Then this solution will diJtribute the directions of radiation more uniformly than x • It is difficult to solve this quadratic problem for large q,r. A slight variation leads to a simpler problem: In the first step we seek all basic solutions x1, ••• ,xl on the hyperplane (2.11) (cf. [11]). Generally there will be less than r. In the second step we find a linear convex combination of x1, ••• ,xl,

1 i i.e. x = E A.X satisfying (2.11).

i=1 l.

Thus we must solve the essentially simpler quadratic problem:

(2. 12)

(2.13)

r E A. = 1

i=1 l.

A. > 0 l. -

i=1, ••• ,1

Page 239: Optimal Estimation in Approximation Theory

236

(2.14) n

minimize E i=1

UDO EBERT

1 1 E E qJ.kAJ. Ak

j=1 k=1 where Q = (qjk) is a symmetric positive definite matrix. 2) The second method requires less computation. To get a uniform distribution here we try to force down the maxi­mum of the weight function as much as possible. We choose as the objective

(2.10'') minimizeM.

The problem (2.8), (2.9), (2.10"), (2.11) is solved by parametric programming. Remarks: In general the results of both methods do not agree. Also it is convenient to replace (2.11) by

(2.11') * c'x < c'x +e:

with small e:>O and to similarly apply both methods.

3) Example The model described above has been implemented. The dose function is approximated with the help of Sterling's field equations [12,13,14,15]. In the optimizing part the S-model uses the MPS (Mathe­matical Programming System [4,5]) for the linear ~ro­gramming and a PL/I-version of Lemke's algorithm L7,10] for the quadratic programming. The output part makes use of a program written by Pudlatz [9] for the graphic re­presentation of the result. All programs are written in PL/I (6].

The following example describes the irradiation of a kidney tumour. It has already grown very extensively and the lymphatic ganglions situated in front of the spinal cord must also be irradiated (fig. 1).

/1}> ~0

Figure 1

tumour

spinal cord

Page 240: Optimal Estimation in Approximation Theory

PLANNING OF RADIATION TREATMENT 237

i ------------ -------- ---- -------- ---- ------------------- --------- -----· · · ... ··-· -·········. · --- --- -------···· ..... ··------- -- .. i I ••• ··· ·· ·· ''' •••••••<<<<< ••• I 1 .••••• • ••• ••• •• ••••• tccc c.ta::c~cc: ••• I I •• .. . . - -...... . . ..... .. • • • • • • •• • • • • • • • ••• c c < t < c c.,; c c c.~ c. cc < c c c ~ c < o •• I

I • • , . . . ...... , ..... . I • • • • • • • • ••• • • •• •• • •• • • t • • • •• •••• • • • t'""' "' t c < c c c c < c c c c ~ c c c ~ c c c c c c c c < C<:; • t

1 ' ········· ····· ······ · ·• •·· ··········· • • .• • •• :tttlltlt t t<<<<c:cccc:cc c ccc cccc ccccc:cc: • 1 I ......... ' . .. .. - . . . . . . . . . . . . . . . . . ; : ... .. , , I,,, ' I , I t ( ~ ( ' ( ( c c ( ( ( ( c ( ( ' ~ ( ( c ( ~ ( ( ( c ( ( ( ( • I

1 • ········•·•···· .. •• • • • •••••••••••• ,,,, • • · ·• · ·· ' " '""' '''~ ",;,;ccccc<.cc<ccccccccc~,;,;~cccc c ~c • I

1 ' ·· •· ·· • ·· ····· • · · ···· • ······· • · ·· ·· ·· • · :t l ••.. . J •• "I' ' ' • ' ' ''~c.;sc:c~;<: !t cccccc ~ ·ccc:'c' s '~'!cc! • 1 I • . . . . . . . • • • . • . . . . • • . . . . . . . . . . . . . . • . • • . • . . • . •••• I I II .I I c (: : c' c :' '.:: '.:: : (~:: ( t ~ ·=: ~ c c': '.:: ( (;:: ::: : : . • I

t • ·· · ··•· ••••••·· ···· ·· ·· ·········•···· .. ,, , ,,.,,, ,,,,,,,,,,,,:c cccc :c c ttcccc:ccc:ccc:cc:ccc;:c: • 1

I • , , , , , , , 0 , 0 , • , • , •• , , , , , , , , .. , , , , ,. , , , , , , , , , :I I I I Icc C C::: (CCC C (' C C: c' c.: c C C l C t C C:: C ICC C C (.:: ~; c c c c • I

1 • ·······•· ••··· .. ·· ···· ··-- 1 cccccccc:ccccccc::cccccc:cctcc.:c c • 1

I • ······ ··• • ·• · • • ·· • • ·· ·· · ' ···· · ·· ' ·•·· · il: • • I I • . - - .... .. .•... ' .... . . . - . - . . . . . . . . . • . ( ( ( (' ( (c.:' ( (.: c c c: ~ ~ ( c ' ( (.: C( c c . I

1 • · ··· ·· · ·· • ••• • • •••••••··· ···· :c,:ccccc;c•:cc:.::c:~~=~~=~==• • I 1 • • • • .. • • • • • • • • • .. • • • • • • • . . • • • . . • c c c: c c c c c c c c.c c c c c c :, c ~ c c c c c c t • I I ,, . ,,.,,, ,,,, ,,,,. ,,,. cc:ccccccccccc.:cc::cccccccccc • I

1 . . ..... .. 1...... . .. :~ccc , cc:c~-c:<: Sc: : -:: c c s ccl'''" ' t I .: c :: (: c ( ~ c ( c;;: ;.:: .::; ; :.: : :I,, , I''' . I c c c ' ( c 'c c ( c c c ( : ~ c : : ( ( I I I I I I' ' 'I • •

I c c c c c c c c c c c c c c cc c: ~ " ' ,' ,,,,'',. o •

1 :<ccc.:: ) ccccc:c:,;: "'''"" ' -'111 .. J,, •

I I

Figure 2. Result of the S-model -----5000 rd isodose

.. ; .. -~·

~4000 rd isodose

Page 241: Optimal Estimation in Approximation Theory

238 UDO EBERT

In addition to the tumour the spinal cord has been drawn because it contains relatively sensitive nerve fibres. It should receive not more than 4000 rd since otherwise it could be irreparably damaged. The tumour itself must get as a minimum 5000 rd to be destroyed. Moreover, there are restrictions on the dosage at the surface (4500 rd).

There are two admissible orbits for the source of ra­diation; each has a fixed target point (crosses 1 and 2 in fig. 1). The orbits are selected so that the central axis of the fields always penetrates the target point, and so that the resulting source-skin-distance is con­stant. The number of orbits and the choice of target points suggest themselves by the shape of the target area. Moreover, a fixed field size on each orbit was de­termined in advance (for the orbit around target point 1 it is smaller than for orbit 2). The weight function (marginal staying-rate) has the form of a polygonal function. An improvement of the solution was obtained using parametric programming. Figure 2 shows the result.

ACKNOWLEDGEMENT

The auther wishes to thank Prof. Dr. H. Werner for suggesting the problem and for his advice, and Prof. Dr. E. Schnepper and Dr. H. Terwort for explaining the prob­lems of radiation-treatment planning.

Page 242: Optimal Estimation in Approximation Theory

PLANNING OF RADIATION TREATMENT 239

REFERENCES

[1] U.Ebert: Optimale Auslegung von Bestrahlungsplanen, Schriftenreihe des Rechenzentrums der Universitat MUnster, Nr. 19, 1976

[2} U.Ebert: Computation of Optimal Radiation Treatment plans, to appear

[3] U.Ebert: A System for Calculating Optimal Radiation Treatment Plans, to appear

[ 4]

[ 5]

[6}

[7]

[8]

[9}

[ 10]

[ 11]

[ 1 2]

[ 13]

IBM Application Program, Mathematical Programming System /360: Control Language, User's Manual, IBM-Form H20-0290-1

IBM Application Program, Mathematical Programming System /360 Version 2: Linear and Separable Pro­gramming - User's Manual, IBM-Form GH20-0476-2

IBM System /360 Operating System, PL/I(F): Language Reference Manual, IBM-Form GC28-8201-3

C.E.Lemke: Bimatrix Equilibrium Points and Mathema­tical Programming, Management Science 11 (1965), 681-689

S.Matschke, J.Richter, K.Welker: Physikalische und technische Grundlagen der Bestrahlungsplanung, Leipzig 1968

H.Pudlatz: GEOMAP - ein FORTRAN-Programm zur Erzeu­gung von Choroplethen- und Isolinienkarten auf dem Schnelldrucker, Schriftenreihe des Rechenzentrums der Universitat MUnster, Nr. 16, 1976

A.Ravindran: A Computer Routine for Quadratic and Linear Programming Problems (alg.431), Communi­cations of the ACM 15 (1972), 818-820

M.Simmonard: Linear Programming, Englewoods Cliffs, 1966

T.D.Sterling, H.Perry, L.Katz: Automation of Radia­tion Treatment Planning, IV. Derivation of a mathe­matical expression for the per cent depth dose sur­face of cobalt 60 beams and visualisation of mul­tiple field dose distributions, British Journal of Radiology 37 (1964), 554-550

T.D.Sterling, H.Perry, J.J.Weinkam: Automation of Radiation Treatment Planning, V. Calculation and visualisation of the total treatment volume, Bri­tish Journal of Radiology 38 (1965), 906-913

Page 243: Optimal Estimation in Approximation Theory

240 UDO EBERT

[14] T.D.Sterling, H.Perry, J.J.Weinkam: Automation of Radiation Treatment Planning, VI. A general field equation to calculate per cent depth dose in the irradiated volume of a cobalt 60 beam, British Journal of Radiology 40 (1967), 463-468

[15] J.J.Weinkam, A.Kolde: Radiation Treatment Planning System Manual, Saint Louis

Page 244: Optimal Estimation in Approximation Theory

SOME ASPECTS OF THE MATHEMATICS OF LIMULUS

K.P.Hadeler

Lehrstuhl Biomathematik, Univ. TUbingen

Auf der Morgenstelle 28, 74 TUbingen,W.Germany

Abstract The compound eyes of the horse-shoe crab exhibit one of the few nervous networks which are well understood, mainly because of its simple and repetitive structure. The theory of this network poses various mathematical questions such as existence and uniqueness of stationary solutions, stabi­lity problems and, at present of greatest interest, the existence of periodic solutions of differential equations with retarded argument.

1. Facts from biology

The horseshoe crab Limulus is a marine arthropode from the coastal waters of the north american east coast. It is one of the five recent species of the ancient order Xiphosura, a group more closely related to spiders than to true crabs.

The animal is up to 60 em long. The main exterior fea­ture is a huge shield covering the entire body and almost the legs, and a tail-rod. The shield consists of two parts, connected by a hinge. Inserted in the anterior part of the shield are two rather large compound eyes.

241

Page 245: Optimal Estimation in Approximation Theory

242 K. P. HADELER

These eyes have been introduced as an object of investi­gation by Hartline in 1928 ~6) and have since then proved to be very useful, so much that at present "Limulus" is a field of nervous physiology as well as "Drosophila" is a field in genetics.

It is somewhat astonishing that up to now the role the eyes play in the life of Limulus has not been completely clarified, their purpose seems somewhat doubtful at least in adult animals.

The nervous system of the Limulus eye represents a rather regular network, with each ommatidium connected to its neighbors. These connections can be schematically depic­ted as in fig.1. Connections are not restricted to nearest neighbors.

Fig.1: Scheme of retina

1 i gh t ommatidia

lateral connections axons

First we consider a single ommatidium. For simplicity we neglect that fact that each ommatidium contains several retinal cells. Light of intensity x is falling from the exterior into the. retinal cell and generates a potential (generator pot~ntial) y. The potential_y is a monotone function of x 1n the form of a saturat1on curve. In the following we shall forget about x and consider y as a measure of the excitation of the ommatidium.

If the excitation is below a certain threshold y nothing will happen. If y exceeds y then the nerve cell releases nervous impulses (spikes) with a frequency z proportional to the excess potential. Thus, under stationary conditions, the function of the ommatidium can be described as

z=b J (y-y),

where J. is the function defined by

( 1)

Page 246: Optimal Estimation in Approximation Theory

SOME ASPECTS OF THE MATHEMATICS OF LIMULUS 243

J (x)=max (x,O)= i (x+lxl). ( 2)

Experiments have shown that ommatidia act inhibiting upon each other, i.e. a large excitation in one ommatidium tends to decrease the excitation in the neighbors. The inhibition acts through the nervous connections exhibited in fig. 1. The inhibitory effects decrease with distance. We assume that the inhibition acts linear. The relations of x,y, and z are represented in fig.2.

Fig.2:Transitionx~ y~z

The inhibitory effect of the k-th ommatidium on the j-th ommatidium is measured by the inhibition coefficient ~jk~ 0. Self-inhibition may be present, thus pjj need not be zero. Then a compound eye with n ommatidia is described by the equations

1'\

z . = b . .& ( y . - y-.-:L pJ. k z k) , j = 1, •..• n. J J J J k=1

( 3)

Equations (3) are known as the Hartline-Ratliff model. The model describes a stationary situation. It has been discussed by Reichardt and McGinitie (23] , Varju [24], Walter (25].

A possible "purpose" of lateral inhibition is compensation of a loss in ~cuity of image during perception. Suppose the animal is shown a spatial step function (in fig.3 we consider one space dimension).

Page 247: Optimal Estimation in Approximation Theory

244 K. P. HADELER

X y

Fig.3: Regain of contrast

Since each ommatidium will perceive light from various points of the exposed pattern, the excitation, as a func­tion of space, is some smoothened step function, and the result of lateral inhibition is a regain in contrast, leading to the formation of Mach bands. (Mach bands are a famous effect in human psychophysics: In a pattern of black and white stripes the white adjacent to black appears very white such that in the middle of white stripes some greyish stripes seem to appear).

2. The model

If the eye is exposed to patterns varying in time then the yj are functions of time, and so are the zj. As appropriate state variables we choose the reduced exci­tations

1\

v.=b.(y.-y.- L ~J·kzk). J J J J k=l

( 4)

The quantity Vj is the actual excitation which generates the nervous impulses, taking all inhibitory effects into account, multiplied by the factor bj.

We remark that in the work of Coleman and Renninger ( [3]- U-3]) the term "reduced excitation" denotes the exci­tation that would occur in an ommatidium disconnected from all others, i.e.

Page 248: Optimal Estimation in Approximation Theory

SOME ASPECTS OF THE MATHEMATICS OF LIMULUS

b.(y.-y.-R .. z.) J J J I~JJ J

From (3), (4) follows

z.=~(v.), J J

j=l, ... ,n.

245

( 5)

( 6)

Under non-stationary conditions the reduced excitation will not immediately respond to a change of the input (yj)' but adapt to a change with a time constant d,

n. rv.+v.=b.(y.-y.- 2: ~-kzk)

J J J J J k=l J

In this equation we can replace zk by !1-(vk), thus

n tv . + v . = b . ( y . - y . - L:. ~ . k ~ ( v k ) ) • ( 7 )

J J J J J k=l J

Such models have been considered by Morishita and Yajima [20] . Recently it has been found that inhibiton acts w i t h a de 1 ay T w hi c h i s i n dependent of the d i stance of the ommatidia. There is no such delay in self-inhi­bition. If this delay is incorporated into the model then the equations read

II"V . ( t ) + v . ( t ) = b . ( y . ( t ) - y . - R . . !}-( v . ( t ) ) ( 8 ) o J J J J J rJJ J

n - L: ~J·k !}-(vk(t-1:)) 1 j=l, ... ,n.

k=l k=tj

The model can be adapted more closely to reality by assuming that the coefficient ~j" depends explicitely o n z k a n d t h a t i n h i b i t i on a c t s wi t h a t h r e s h o 1 d , t o o .

Indeed there is some indication that inhibition increases with excitation, such that ~jk should be a non-decreasing function of zk.

Page 249: Optimal Estimation in Approximation Theory

246 K. P. HADELER

With such modifications the model becomes

'i"y . ( t } + V • ( t } = b . ( y . ( t } - y . ( t } -y . -R . J. (3( V • ( t ) ) ~ ( V • ( t ) - Z .. ) o J J J J J J rJ J J JJ

n (9) -I: ~ . k (,9-( v k ( t -T) (.:)( v k ( t-T)- zJ. k) , k= 1 J ktj j=1, ... ,n.

However, to obtain mathematically tractable equations we shall stay with the model (8). By introducing new varia­bles

( 10)

we arrive at

tv j ( t > + v j ( t > = y j ( t > -p j j'J( v j ( t > >-5 ~ j ~( v k ( t -tJ > •

ktj (11) j=1, ... ,n.

With vector notation

v=(vj), y=(yj),

D=(Pjj~jk), B=(~jk),

equations (11) take the form

ty ( t } + v ( t ) = y ( t } -D .5-( v ( t ) ) - ( B - D )!} ( v ( t- 'L) } . ( 12 )

3. The equilibrium states

Consider equation (12) with a constant input y(t}sy. A stationary solution v satisfies the equation

v=y-s3'(v). (13)

We show that this equation is equivalent to the stationary Hartline-Ratliff model

Page 250: Optimal Estimation in Approximation Theory

SOME ASPECTS OF THE MATHEMATICS OF LIMULUS 247

z= ~(y-Bz) (14)

in the sense that either both equations have no solution or there is a one-to-one correspondence between solution sets.

If v is a solution of (13), then z=,(v) satisfies z=)(y-B~v))=~(y-Bz). On the other hand, if z solves (14) then v=y-Bz fulfils ~(v)=9(y-Bz)=z, thus v=v-~(v).

If v< 1>, v< 2> are two solutions of (1) and z( 1)=)(v< 1>),

z< 2 > =~(v( 2 )) then from z< 1>=z( 2) follows that v< 1>•v< 2> coincide in their positive components, and from (13)

follows v< 1 >=v< 2 >. On the other hand, if z( 1), z( 2) are

two solutions of (14) and v< 1>=y-Bz( 1), v< 2>=y-Bz( 2) then

from v< 1>=v( 2) follows Bz( 1)=Bz( 2) and thus z( 1)=z( 2) from (14).

To equation (14) we relate the mapping T: mn--+ mn,

Tz= !T(y-Bz ), (15)

solutions of equation (14) correspond to fixed points of

T. T maps mn into m~, and T maps m! into the compact convex domain

(16)

By Brouwer's fixed point theorem the set M contains at least one fixed point; and all fixed points are con­tained in M.

The corresponding stationary states of equation (8), i.e. solutions of equation (13), are contained in

(17)

If the spectral radius 9(B) of the matrix B is less than 1 then T is a contraction with respect to an appropriate

Page 251: Optimal Estimation in Approximation Theory

248 K. P. HADELER

norm

1'1. llxU = I: d.. I X • I ,

j = 1 J J

where the ~j are the components of a positive left eigenvector of the matrix B=(~jk+f), E.>O sufficiently small. Thus ~(B)<1 implies uniqueness of the stationary state ( (lt] ) .

In general (i.e. for ~(B)~1) the stationary state will not be unique. This is important insofar as in the Limulus eye normally inhibition is as strong as ~(B)> 1.

4. Oscillating solutions

Comparably recent results (Barlow,Adler, see (3] - [8]) indicate that even with a constant excitation the nervous impulses do not appear equally spaced in time, but in "bursts". The impulse frequency is oscillating or periodic. Apparently Coleman and Renninger ( (3]- [8]) were the first explaining these bursts as a consequence of the delay with which the inhibition acts.

To investigate these effects we simplify the model by assuming that the compound eye is homogeneous and that it is exposed to a spatially homogeneous excitation. To avoid at present any complications related to the geo­metry of the eye we assume that the inhibition coeffi­cients ~jk' jtk, do not depend on j,k, furthermore we neglect self-inhibition. Then, if only spatially homo­geneous solutions are considered, each component of such a solution satisfies a scalar difference-differential equation

rv(t)+v{t)=y(t)-~~(v(t-L)) . {18)

By rescaling the time variable we achieve

rv(t)+v(t)=y-p~(v(t-1)).

The stationary state is v=y/(8+1), and by way of the substitution v=v+x we shift the stationary state into

Page 252: Optimal Estimation in Approximation Theory

SOME ASPECTS OF THE MATHEMATICS OF LIMULUS

zero

or

where

X+jx =- 3-<}pv+f~x)+if3v

x(t}+Yx(t)=-('(a+bx(t-1))-aj

\1 =} , a=ipv, b=~ p

249

( 1 9)

(20)

For some special cases Coleman and Renninger DJ were able to exhibit the existence of periodic solutions. We shall prove the existence of such solutions for equation (19) for all cases where the zero solution is unstable.

This equation is a special case of the equation

x(t}+Yx(t)=-f(x(t-1)) (21)

whereYis a non-negative parameter and the continuous function f: (R~ rR has the properties

f (X) X ) 0 fij r X t 0, (22)

f is differentiable at x = 0 and f'(O)=~> 0. (23)

The linearized equation corresponding to (21) is

x(t}+Yx(t)~x(t-1)=0, (24)

and the characteristic equation is

(25)

For Y ~ 0 let ~ybe the smallest positive solution of the equation

,{ 2 2' " + ~cos v~ - Y = 0 {26}

One can show (see V,3]): Fo rcl.<~v the cha racte ri st i c equation has no roots in the right half-plane, and the zero solution of (24) is stable. Forct>~~the characteris­tic equation (25) has a root>.with Re~~ 0, and the zero solution of (24) is unstable.

Page 253: Optimal Estimation in Approximation Theory

250 K. P. HADELER

If~ increases from 0 to infinity then the solutions of equation (21) in the neighborhood of 0 become oscillatory before the stationary solution 0 looses its stability property, namely (see [).3} ):

Suppose there is c\,. 0 and c.,. Yl(e" -1)(c> 1 if ')1=0) such that

lf(x)l ~ c lxl for lxl ~ ~. (27)

Let"){ be the cone of all functions tp(f C [-1,0] with the properties

1) e"t~t) is a non-decreasing function oft, where -1 :!!: t ~ 0.

2) <'p( -1) =0.

Let 1{0 =1{- {o) Let x(t)=x(t,f) be the solution of equation (21) genera­ted from the 1nitial function CfC: K0 • Then:

1) x has denumerably many zeros z 1< z 2 < z 3 < ... , and

2) z 1> 0, z 1-z >1 for n=1,2,3, ... n+ n

3) x(z2k-1)< 0, x(z2k) > 0 for k=1,2, ..

4) The function eYt x(t) is non-increasing on each inter­val (z 2k_ 1 ,z 2 k_ 1+1) and non-increasing on each inter-

val (z 2k, z 2k+1), k=1,2,3, ...

5) For every M > 0 there is a constant CM> 0 such that from IICfl ~ M follows z 2 ~ Cw

These properties of the solution enable one to define a mapping tf: X. ~c [0, 1] by

"f(~)=x(z 2 (f)+1+t,y), -1~t~O, for f*O (28)

Cf(O)=O;

Page 254: Optimal Estimation in Approximation Theory

SOME ASPECTS OF THE MATHEMATICS OF LIMULUS 251

Fig.4: The operator ~

F i g . 4 shows how T acts . 0 n e can show .1){c .:7{ , T i s con­tinuous and compact. If, in addition, f satisfies the condition There is a numberee> 0 such that f(x)~-~for all x (29)

then :Tmaps the closed, bounded ·convex set of infinite dimensionS =~'fE'~ :l'f'K ~~}into itself. Therefore Schauder's theorem would immediately yield the existence of a fixed point ofT and thus to a periodic solution, if it were not forT(O)=O. . Following the proof of R. Nussbaum [21] for the case\)=0, we apply the Browder-Horn theorem on ejective fixed points. A fixed point~ of~ is ejective, if there is an open neighborhood u,~ ,offiU , such that for~ ~very ~U, 'f1t 1, t h e r e i s a n i n t e q e r n = n ( 'f ) s u c h t h a t ~·> 'f ~ U . T h e Browder-Horn theorem [1] states that a continuous compact mapping of a closed bounded convex set of infinite di­mension has a non-ejective fixed point. This theorem yields the following result (see (13] ): Ify =0 is given and the function f has properties (22), ( 2 3 ) , w h e r e «. > Cl("' , a n d ( 2 9 ) , t h e n e q u a t i o n ( 2 1 ) h a s a non-constant periodic solution.

Similar but weaker results have been shown by Pesin ~2] by an immediate ap2lication of Schauder's theorem, and K a p 1 an an d Yo r k e l18] have proved the e xi s ten c e of peri o­die solutions for the general equation

x(t)=-f(x(t),x(t-1))

Page 255: Optimal Estimation in Approximation Theory

252 K. P. HADELER

by a phase plane method. Again, when restricted to equation (21), their result is somewhat weaker than the above. See also Chow ~] and the earlier work by Jones [17] and Grafton [10] .

The equation (19) satisfies all conditions required for a periodic solution if b>O.v·

5. The vector system

Consider again equation (12) with constant input

lv(t)+v(t)=y-D~(v(t))-(B-D)~(v(t-r)). (12)

A stationary solution V satisfies

(B+1)v=y . ( 30)

We assume that equation (30) has a positive solution v and that the non-negative matrix B is non-singular. In (12) we substitute

v=v+Bx and obtain

~(t)+x(t)=V-B- 1 D,{v+Bx(t))-(I-B- 1 D)j(v+Bx(t-~)). (31)

If we exclude self-inhibition (0=0) then equation (31) simplifies to

rx(t)+x{t)=-fJ(v+Bx(t-~~-v] or, with a="CV/( , B= T:B/0 , v= "tl(

x(t}+Yx(t)=-(3(a+Bx(t-1))-a] (32)

similarly to equation (19). For x sufficiently small equation (32) is linear

x(t)+Yx(t)+Bx(t-1)=0 (33)

With x(t)="iexp (At) we obtain the characteristic eigen­value problem

Page 256: Optimal Estimation in Approximation Theory

SOME ASPECTS OF THE MATHEMATICS OF LIMULUS 253

Now the possible complex or negative eigenvalues con­siderably effect the stability of equation (33). If A =~+icf is a characteristic root, then-(~+ll)?' =)J is an eigenvalue of the matrix B. In general nothing will be known about )l except y=c:A.exp (i'f") with O~d.~~(B). Then

)J+I3+ o( e- ~cos (r-'1') =0, '{- cle -f3s in ((-~) =0 .

These equations have a solution with ~>0 for an appro­p r i ate 'f" i f f Y "~-

Therefore we have the following situation: Let v be fixed, for y(B)<~ all characteristic exponents have negative real parts, for f(B)>d.v there is at least one characteristic exponent w1th positive real part. Thus the eigenvalues of B which are not real and positive tend to cause instabilities.

6. Related problems from ecology

One of the classical ecological models is the Verhulst equation, in normalized form

u(t)= d_u(t)(1-u(t)). ( 34)

For any initial data u(O)> 0 the solution u(t) approxi­mates the carrying capacity u=1. All solutions are mono­tone. Equation (34) is equivalent with

x(t)=-~X(t) ( 1+x(t)). (35)

If the growth rate depends on history we arrive at Hutchinsons's equation [15] (see (9],(?4] for biological details)

u(t)=~u(t)(1-u(t-1)) (36)

or,equivalently,

x(t)=- ~x(t-1)(1+x(t)). ( 37)

Page 257: Optimal Estimation in Approximation Theory

254 K. P. HADELER

These equations have oscillating solutions for suffi­ciently large ol. The characteristic equation

a. -" "+ole = 0

has a zero). with Re~>O iff ck > T/2, and it is known that equations (36) and (37) have periodic solutions if cl>1T/2. Dunkel (9] , Halbach [14] proposed the more general equation

0 X ( t ) =- t~.f X ( t + s ) d a- ( s ) ( 1 +X ( t ) ) •

-1 (38)

where~~O and O"""is a function, monotone and continuous from the left, with 0\-1-)=0, ~(0+)=1. The corresponding characteristic equation is

o "s I'M ?\ + cJ..J e d '-'\ s) =0. (39) -1

A serious obstacle on the way to the proof of existence of periodic solutions is the location of the~zeros of the characteristic equation. For which cr,el are there zeros in the right half-plane? Obviously there are no ;zeros in the right half-plane ifcrhas a step of height 1 at s=O. Hence we conjecture that there is a zero in the right half-plane at least for large d. if the "mass" ofCT is mainly concentrated in the past. The two results so far known support this conjecture: H.O.Walther D~ showed that with

0 { l sin 1rs l dtT(s) > 1/11'

-1 (40)

t h e r e i s a z e r o f o r c:l s u f f i c i en t 1 y 1 a r g e . T h e a u t h o r fi 2) showed that ifO""(s)=1 for s>-t then there is a zero in the right half-plane for

( 41)

Page 258: Optimal Estimation in Approximation Theory

SOME ASPECTS OF THE MATHEMATICS OF LIMULUS 255

Walther [27] succeeded to prove the existence of periodic solutions of

-1 x(t)=- elf x(t+s}der(s}(1+x(t))

-T (42)

forc:i>'Jr/2 and T-1 sufficiently small. It would be de­sirable to treat problems like

-1 x(t}+l'x(t)=- «.f x(t+s}dcr(s)(1+x(t))

-T (43)

but approaches so far known lead to tremendous diffi­culties.

Re fe ren ces

~] Browder,F.E., A further generalization of the Schauder fixed point theorem, Duke Math.J.32, 575-578 (1965)

~] Chow,S.-N., Existence of periodic solutions of auto­nomous functional differential equations, J.Diff. Eq. 15, 350-378 (1974)

pJ Coleman,B.D., and Renninger,G.H., Theory of delayed lateral inhibition in the compound eye of Limulus, Proceedings Nat.Acad.Sc. 71, 2887-2891 (1974)

~] Coleman,B.D., and Renninger, G.H., Consequences of delayed lateral inhibition in the retina of Limulus I. Elementary theory of spatially uniform fields, J. Theor. Biol. 51, 243-265 (1975)

~] Coleman,B.D., and Renninger,G.H., Consequences of delayed lateral inhibition in the retina of Limulus. II.Theory of spatially uniform fields, assuming the 4-point property, J. Theor. Biol. 51,267-291 (1975)

Page 259: Optimal Estimation in Approximation Theory

256

[7]

[a]

[u]

[12]

[13]

K. P. HADELER

Coleman,B.D., and Renninger,G.H., Periodic solutions of certain nonlinear integral equations with a time­lag, SIAM J.Appl.Math. 31, 111-120 (1976)

Coleman,B.D., and Renninger,G.H., Periodic solutions of a nonlinear functional equation describing neural action, Istituto Lombardo Acad:Sci.Lett.Rend.A, 109, 91-111 (1975)

Coleman,B.D., and Renninger,G.H., Theory of the res­ponse of the Limulus retina to periodic excitation, J.of Math.Bio1.3, 103-119 (1976)

Dunkel ,G., Single species model for population growth depending on past history in: Seminar on Differential euqations and dynamical systems, Lecture Notes in Mathematics Vol. 60, Berlin Springer 1968.

Grafton,R.B., A periodicity theorem for autonomous functional differential equations, J.Diff.Equ.6, 87-109 (1969)

Hadeler,K.P., On the theory of lateral inhibition, Kybernetik 14, 161-165 (1974)

Hadeler,K.P., On the stability of the stationary state of a population growth equation with time-lag, J.Math. Biol. 3, 197-201 (1976)

Hadeler,K.P., and Tomiuk,J., Periodic solutions of difference-differential equations, Arch.Rat.Mech.Anal. to appear.

Halbach,K. und Burkhard,H.J., Sind einfache Zeitver­zogerungen die Ursachen fUr periodische Populations­schwankungen? Oecologia (Berlin) 9, 215-222 (1972).

Hutchinson,G.E., Circular causal systems in ecology, Ann.N.Y. Acad.Sci. 50, 221-246 (1948)

Hartline,H.K., Studies on excitation and inhibition in the retina, Chapman and Hall, London (1974)

[1 7] J o n e s , G . S . , T h e e x i s t e n c e o f p e r i o d i c s o 1 u t i o n s o f f'(x)=-2f(x-1)(1+f(x)), J.Math.Anal.Appl.5,435-450 (1962).

[18] Kaplan,J.L. and Yorke,J.A., Existence and stability of periodic solutions of x'(t)=-f(x(t),x(t-1)), in:

Page 260: Optimal Estimation in Approximation Theory

SOME ASPECTS OF THE MATHEMATICS OF LIMULUS 257

[19]

~o]

[21]

[22]

Cesari Lamberto (ed.) Dynamical Systems I,II Providence (1974)

May,R., Stability and complexity in model ecosystems, Pop.Biol.6, Princeton Univ.Press (1974)

Morishita,!., and Yajima,A., Analysis and simula­tion of networks of mutually inhibiting neurons, Kyberneti~ 11, 154-165 (1972)

Nussbaum,R.D., Periodic solutions of some nonlinear autonomous functional differential equations, Annali di Matematica Ser.4, 51, 263-306(1974)

Pesin,J.B., On the behavior of a strongly nonlinear differential equation with retarded argument. Dif-ferentsial•nye uravnenija 10, lo25-lo36 (1974)

Reichardt,W., and MacGinitie,G., Zur Theorie der lateralen Inhibition, Kybernetik 1, 155-165(1962)

Varju,D., Vergleich zweier Modelle fUr laterale Inhibition, Kybernetik 1, 200-208 (1962)

Walter,H., Habilitationsschrift, SaarbrUcken 1971.

Walther,H.-0, Existence of a non-constant periodic solution of a nonlinear autonomous functional dif­ferential equation representing the growth of a single species population, J.Math.Biol.1, 227-240 (1975)

Walther,H.-0., On a transcendental equation in the stability analysis of a population growth model, J.of Math.Biol.3, 187-195(1976)

Page 261: Optimal Estimation in Approximation Theory

ANALYSIS OF DECAY PROCESSES AND APPROXIMATION

BY EXPONENTIALS

Dietrich Braess

Institut fUr Mathematik, Ruhr-Universitat

4630 Bochum, F.R. Germany

In several sciences, e.g. in physics, chemistry, biology,

and medicine, often decay processes have to be analyzed. Then

an empirical function f(t) is given, its domain may be an inter­

val or a discrete point set; and a sum of exponentials of the

form

n

I v=l

a e v

-A t v

is to be fitted to the function f. Here, n is the given number

of matters, while the a 's are the concentrations and the A 's v v are the decay constants. The latter 2n parameters are to be

determined by the fit.

During the last decade an involved theory for the mathe­

matical treatment of this problem has been developped, see e.g.

[ 1,4,6]. In this talk, however, we will restrict our attention

to those problems which arise when the theory is applied. Then

the following question seems to be crucial: How sensitive are

the different methods to the noise (pollution) of the empirical

data? Although there is a certain connection with the numerical

259

Page 262: Optimal Estimation in Approximation Theory

260 DIETRICH BRAESS

stability [ 31, this problem must be considered seperately.

Let us start our discussion with the "peeling-off-method"

(German: Abschalmethode). It is one of the first effective methods.

We do not know who has developped it. But it will not be dis­

cussed for historical reasons. At first glance the peeling-off­

method seems to be a poor algorithm from the mathematical point

of view. Yet it shows the limitation of any method, and therefore

it is probably superior to a highly developped mathematical

procedure which is used only in the black-box-manner.

The peeling-off-method proceeds in an inductive way. Note -A.t

that the parameters of a single exponential term a•e may be

determined by a simple interpolation at two points t 1 and t 2 :

I f(tl), ---·log---t -t f(t )

2 I 2 (I)

Moreover, if the decay constants are ordered: A1<A 2< ... <Am'

then all terms die away faster than the first one does. Therefore

one gets a reasonable estimate for the first·term by interpolating -A It f(t) at two large t's. The result a 1e is peeled of the curve,

1.e. it is subtracted from f(t). Then the process is repeated.

Now, let us look for the conditions under which the peeling­

off-method makes sense. They are obvious in the case when there is

only one term and when f is given at m points t 1<t 2< ... <tm. Then

we require

A. I

(2) ;, 5 t -t 1 m I

A. .;; 1 t3-tl (3)

The first condition says that the time must be so large that the

decay causes an observable decrease of the amplitude. The second

Page 263: Optimal Estimation in Approximation Theory

ANALYSIS OF DECAY PROCESSES 261

one says that the curve must not die out, before the proper measuring starts. According to my experience the second requirement seems to be a hard one. Often there are transients because of the switching-on effects of the apparatus. Then the first measured data have to be abandoned.

If there are two or more terms, there are even more restric­tive conditions. The decay constants must not be too close. Instead of (2) we have

I I A. -A. >-. ---i+I i~-' 5 t -t m I

Otherwise the terms cannot be distinguished within the observation time. Moreover, to be more precise we should insert only the time interval which is used for determining the i-th term, instead of referring to the total time (tm-t1). In any case, the conditions above seem to be intrinsic ones and they are independent of the applied numerical method.

The main drawback of the peeling-off-method results from the fact that generally the data are noisy. The interpolation (I) is very sensitive to a small change of the data. Since the formula refers only to two values, there 1s no smoothing, no cancelling of the statistical errors. Some methods, which make use of the solu­tion of certain difference equations, are even worse, c.f. [5, p. 278] •

Therefore the problem has to be considered and solved in the framework of approximation theory. Given f the sum of exponentials F is to be determined which minimizes the norm of the error function

llf-FII.

Page 264: Optimal Estimation in Approximation Theory

262 DIETRICH BRAESS

It is not a priori clear which norm has to be chosen. If the

function would depend linearly on the unknown parameters then the

least-squares fit is appropriate unter the assumption that the

errors are normally distributed. But we do not get a serious hint

from statistics, when the parameters enter into the ansatz in a

non-linear manner. Here, both the 2 2-norm and the uniform norm

have to be considered. Which is the better one for eliminating

the pollution of the data?

Let us discuss the 2 2-approximation first. Then the (square of

the) norm of the error function

U f-FII 2 m

I i=I

2 [f(t.)-F(t.)]

l. l.

depends explicitly on all the data f(t 1),f(t 2), ... ,f(tm). There

are no distinguished points. Consequently, the pollution of the

data is not as dangerous as the interpolation. This is an advantage

of 2 2-norm. Moreover, the best approximation is unique, if the

distance of the given function to the approximating one 1.s small

[ 7 ] •

But that means that the error of the data must be small. In

many actual situations the analysis has to be performed, though

the data are noisy. Then the uniqueness result above is of no use.

Moreover, another complication. seems to be a more severe drawback.

It is not sufficient to regard only the best approximations! When

computing a solution, one has to apply iterative procedures:

Newton-like algorithms or gradient methods. Starting with a

reasonable fit, the approximating function is successivily improved.

Then the local best approximations cause trouble. No descent

algorithm can distinguish between local and global minima. There­

fore, each local best approximation acts like a mouse trap at

least as long as you have not available a very good guess to start

Page 265: Optimal Estimation in Approximation Theory

ANALYSIS OF DECAY PROCESSES

the iteration. From this viewpoint Wolfe's result [7] on the

approximation by exponentials on an interval is discouraging.

263

THEOREM. Given an integer m there is a function f 0EL2 such that

each f in a neighborhood of f has at least m local best approxi­o

mations.

Now we turn to uniform approximation by sums of exponentials.

Then no trouble is caused by the fact that there may be local

best approximations which are not global minima. The number of

local solutions is bounded [2] and the bound is independent of

the function f. Moreover,

it 1

2

3 l 3 I

4 1~9

Table I. Maximal number of local best approximations

in the family of sums of exponentials with order n.

there is a numerical algorithm for computing all the local best

approximations (see below).

How sensitive is the best approximation to pollution? The

solution is characterized by an alternant of length j with

j~2n+l, i.e., there are j points t 1<t2< ••• <tj such that

(-l)i[f(t.)-F(t.)]=± Rf-FU, i 1 1

1,2, ••. ,j.

These m points are distinguished. At first glance the norm and

therefore also the solution seem to depend only on the data at

Page 266: Optimal Estimation in Approximation Theory

264 DIETRICH BRAESS

these points. But have the other points really no influence?

Indeed, small changes of those data which do not enter into

the alternant have no influence to the result. Nevertheless they

prevent that a large change of the solution occurs when the

distinguished data are perturbed. Therefore the situation is not

comparable with interpolation. Moreover, small perturbations of

data cannot cause that local solutions vanish or that new ones

appear. There is only one exception. The number of solutions is

not (locally) constant for a neighborhood of f, if a local best

approximation of f is degenerate, i.e., when one of the factors

a. vanishes. v

As a conclusion we proceed as follows when a reasonable fit

by a sum of exponentials is to be determined:

1. Compute the best uniform approximation to the given function

by sums with 1,2, ... ,n terms. Here the local solutions with

k-1 terms, k=2,3, ... ,n, are used as the starting"points for

the iteration at the k-th stage.

2. Use the best uniform approximation to start an iteration for

computing a (local) best ~ 2-approximation.

3. Check the solution with respect to the conditions for the

peeling-off-method. Moreover, check whether the degree of

approximation with n terms is significantly better when the

comparison with n-1 terms is performed. Otherwise return to

the result for n-1.

We conclude with a numerical example, where the data stem

for a reactor experiment.

Page 267: Optimal Estimation in Approximation Theory

ANALYSIS OF DECAY PROCESSES

Here, approximations with nonnegative factors a were required. v

265

The results for the uniform approximation are presented in Table

3. The restriction a ~0 implies, that the approximation cannot v be improved even when 4 or more terms are admitted.

Table 2. Data from a reactor experiment.

t f(t) t f (t)

0 86o 9 148

6o3 lo 128

2 491 I I lo9

3 4o7 12 96

4 341 13 82

5 288 14 71

6 242 15 62

7 2o5

8 174

Table 3. Results for data in Table 2.

n best approxUmation degree

814.7 -o.229x 45.2 e

2 2oo.5 e- 1· 36x+654.6 -o. 162x 4.87 e

3 152.8 e-2· 327+556.7 -o.216x o.994 e

+151.4 -o.o868x

e

Page 268: Optimal Estimation in Approximation Theory

266 DIETRICH BRAESS

REFERENCES

I. D.Braess, Uber Approximation mit EXponentialsummen. Computing 2, 3o9-321 (1967)

2. - Chebyshev approximation by ~-polynomials, III: On the num­ber of local solutions (submitted)

3. -, Zur numerischen Stabilitat des Newton-Verfahrens bei der nichtlinearen Tschebyscheff-Approximation. Proceedings of the "Kolloquium tiber Approximationstheory", Bonn, Juni 1976

4. D.W.Kammler, Existence of best approximations by sums of exponentials. J.Approximation Theory 9, 173-191 (1973)

5. C.Lanczos, Applied Analysis. Prentice Hall, Englewood Cliffs 1956.

6. E.Schmidt, Zur Kompaktheit bei Exponentialsummen. J.Approxi­mation Theory 3, 445-454 (197o)

7. J .M. Wolfe, .on the unicity of nonlinear approximation in smooth spaces. J.Approximation Theory 12, 165-181 (1974)

Page 269: Optimal Estimation in Approximation Theory

OPTUfAL STATE ESTIMATION AND ITS APPLICATION TO INERTIAL

NAVIGATION SYSTEM CONTROL

W. Hofmann

Deutsche Forschungs- und Versuchsanstalt flir Luft­und Raumfahrt e.V. (DFVLR)

D-8031 Oberpfaffenhofen I West Germany

ABSTRACT

The behaviour of most functions in practical situations can always -- at least approximately -- be described by a finite-dimen­sional system of differential equations. The theory of optimal state estimation deals with the minimum error-variance reconstruct­ion of the functions from a limited information in a noisy environ­ment. The solution approach, algorithm and stability properties for the linear Gaussian estimation problem are reviewed. A general and typical way for the design of a suboptimal estimation algorithm with special regard to system properties and realization require­ments is shown for the technical problem of the error estimation of an aided inertial navigation system. According to the closed­loop system requirement of high accuracy in a noisy environment the considerations about the estimator design are extended to the design of simple, (sub-) optimal controllers. A comparison to the optimal estimation accuracy shows the efficiency of proposed output feed­back controller design procedure in this technical example.

I . INTRODUCTION

In a practical situation a control engineer is faced with the problem to find control inputs for a physical system sue~ that the performance of the system behaves in a desired manner. Information about the system is available from measurement outputs. The be­haviour of the output and input functions can always -- at least approximately -- be described mathematically by a finite dimensional system of differential equations, the state equations of the physi­cal system and the controller. The modelling of physical and ran-

267

Page 270: Optimal Estimation in Approximation Theory

W.HOFMANN

dom disturbing states is discussed in chapter 2. For a broad class of controllers it is necessary to reconstruct the complete system state vector from the limited information of the output functions. The theory of optimal state estimation deals with the optimal e.g. minimum-error-variance reconstruction of the state in a noisy en­vironment. The solution approach, algorithm and stability proper­ties for the linear Gaussian estimation problem are reviewed in chapter 3. The technical problem of state estimation and platform error angle control of a Doppler-aided inertial navigation system is considered in chapter 4. A general and typical way for the de­sign of a suboptimal estimation algorithm is shown with special re­gard to system properties and realization requirements. According to the closed-loop system requirement of a most accurate performance a design procedure for simple, minimum-variance controllers is de­rived. The direct equivalence to the optimal estimation problem is shown, such that all results of estimation theory can be used di­rectly for the proposed output feedback controller design procedure. The procedure is applied to the design of simple feedback networks for the control of the platform error angles. The suboptimal accu­racy results are compared with the performance of the optimal (Kal­man-Bucy) filter, which gives the lowest reachable bound of accuracy for state vector feedback controllers.

2. MODELLING OF PHYSICAL AND RANDOM DISTURBING STATES

Since in general physical systems have the capability of stor­ing energy, they have to be treated as dynamical systems. To des­cribe the dynamical behaviour mathematically, differential equations have to be used. The mathematical model, which is at least only an approximation for the dynamical behaviour of the physical reality, has to be chosen to provide for an as good as necessary fitting in the input-output-behaviour. If the parameters of the physical sy­stem are lumped, the dynamical behaviour can be described by nonli­near time-varying differential equations [1]. Since the general treatment of nonlinear differential equations is rather difficult and solutions can often be obtained for a distinct class of equations only, it is advantageous to split the general control problem in two parts (fig. 2.1):

I. The design of a guidance, i.e. steering commands to keep de­sired nominal trajectories.

2. The design of a controller, i.e. control commands to keep the deviations from the nominal values small.

Whereas the first part involves the treatment of nonlinear differential equations, it is admissible in many practical problems to linearize the nonlinear model about the nominal trajectories by a first-order Taylor expansion [2]. The dynamical behaviour of the

Page 271: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION 269

nominal measurements

steering co;m:ds mE?OSurements ,~

Physical Guidance

~ ~;(+ ;~+

w "' inputs System -•

control/ Controller

...._ commands ...

mE?Osurement deviations

Fig. 2.1 Splitting of the General Control Problem

Page 272: Optimal Estimation in Approximation Theory

270 W. HOFMANN

deviations from the nominal values can be described by linear time­varying differential equations, the so-called state equations [2]. The deviation measurements are linear algebraic functions of the deviations.

The linear model of a dynamical system can always be presented by a linear time-varying first-order vector differential equation [2]

:X<t>

~(t)

~(t)

F(t) ~(t) + G(t) ~(t) + D(t) E_(t)

n-dimensional state vector with Gaussian initial conditions ~(t0 ) ~ N(Q, P0 ) +) ,

s-dimensional Gaussian white noise input vector with ~(t) ~ N(Q, Q(t)) ,

r-dimensional (control) input vector

and a linear measurement equation

~(t) H(t) ~(t) + ~(t)

~(t) m-dimensional (measurement) output vector,

~(t) m-dimensional Gaussian white noise output vector with ~(t) ~ N(Q, R(t)) •

(2.1)

(2. 2)

The system matrices are of appropriate dimension. The struc­ture of the mathematical model is illustrated in (fig. 2.2). The presentation by vector equations forms the basis for the application of state-space methods, a powerful tool in modern control theory. Especially with regard to the use of digital computers state-space methods provide for advantageous system analysis and control design procedures [2, 3].

In many practical applications it is possible to set up differential equations for the physical state vector ~1 (t) by the insight in the physical relations of the considered system, yielding a linear vec­tor differential equation of order n 1

(2. 3a)

and a measurement equation

+)

~(t) = H1 (t) ~(t) (2.3b)

This notation describes a Gaussian (normal) distribution with mean vector 0 and covariance matrix P0 [4].

Page 273: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION 271

~ (t) v(t)

!! (t)

Fig. 2.2 General Structure of a Linear Dynamical System

(""1~(1:) (j) physical state

(6) disturbance

--~----------_.------------~w

Fig. 2.3 Autocorrelation and Spectral Density Functions

Page 274: Optimal Estimation in Approximation Theory

272 W. HOFMANN

Sometimes the deterministic viewpoint is sufficient to des­cribe the dynamical behaviour of the physical system. But especial­ly in the case of high accuracy requirements in a noisy environment it is necessary to include random disturbances in the model

~I (t) = F II (t) ~I (t) + D I ~(t) + ~*(t) ; ~I (t) ~ N(£, P 10)

(2.4a)

~(t) = H1 (t) ~I (t) + ~*(t) (2.4b)

with the random disturbance vectors w*(t) and v*(t) which in general cannot be influenced by the control.- The modelling of stochastic disturbances often requires extensive experimental investigations. Due to the linear point of view it is often possible to treat the modelling problem within the framework of Gauss-Uarkov processes [4].

Complete statistics for a Gauss-Markov process E(t) are given by the mean-vector ~~(t) and the covariance matrix C~(t)

~(t) E{~(t)} (2.5)

E{(~(t)- ~(t)(~(t)- ~(t))T} (2.6)

Treating a single stationary [4] random process ~(t) two defi­nitions related to the second-order moment are introduced [5]:

the autocorrelation function

(2.7a)

and the spectral density function, which is the Fourier transform of the autocorrelation function

Some examples are given in (fig. 2.3, 2.4). The definitions (eq. 2.7) form the theoretical basis for several random disturbance ana­lysers.

It is convenient [4] to distinguish between time-uncorrelated (white) and time-correlated (coloured) disturbances, since the mo­delling of the latter requires an augmentation of the state vector. Disturbances, whose spectral densities are broadband relative to the system power spectrum (fig. 2.3), are usually modelled by Gauss­ian white noise, a fictitious mathematical model with frequency in­dependent spectral densitiy and no time correlation, which is ex­pressed by the delta-impulse autocorrelation function (fig. 4.3.).

Page 275: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION 273

If the time constants of the physical system cannot be neglected, the disturbances can be modelled by the outputs of linear and Gauss­ian white noise driven shaping filters (fig. 2.4), whose dynamical behaviour can be described by linear vector differential equations.

Following the previous discussion the disturbance vectors w*(t) and v*(t) in (eq. 2.4) can be modelled by a linear superposition of Gaussian white and coloured noise

w*(t)

~*(t)

F12 (t) ~2 (t) + G1(t) ~(t)

H2(t) ~2 (t) + ~(t)

(2. 8a)

(2.8b)

The dynamical behaviour of the coloured noise state vector ~~(t) of the n2-order shaping filters is given by the linear vector d1fferential equation

(2.8c)

Combining (eqs. 2.4, 2.8) leads to a system description accord­ing to (eqs. 2.1, 2.2) for the augmented state vector of order n = n 1 + n2 and a special structure of the system matrices

(2.9a)

F(t) G(t) (2.9b)

0

D(t) H(t)

0

The structure of the system including random disturbance modelling is pointed out in (fig. 2.5).

3. THE LINEAR, GAUSSIAN ESTIMATION PROBLEM

The application of several control schemes requires the exact or at least estimated knowledge about the complete state vector [2, 3]. The problem of estimating the state vector of a linear system with Gaussian disturbances (eq. 2.1) from noisy measurements (eq. 2.2) such that the estimation error-variance is minimized can be formulated as follows [4]:

Given a set Z(tf), t 0 ~ t ~ tf, of measurements ~(t) of the state vector ~(t) according to (eq. 2.2), where the dy-

Page 276: Optimal Estimation in Approximation Theory

274 W. HOFMANN

w (t)

white noise coloured noise

Fig. 2.4 Structure of a Shaping Filter

Yt(t)

Fig. 2.5 System Structure including Random Disturbance

Page 277: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION 275

namical behaviour of x(t) is described by the linear vector differential equation-(eq. 2.1). Find an estimate i<tf), such that the cost functional

(3. I)

is minimized with the estimation error vector

(3.2)

Since the exact derivation of the optimal state estimation procedure -- the famous Kalman-Bucy filter -- is well-known from the literature [4, 6], a more heuristic approach [7] is taken, which gives a deaper insight in the action of the optimal filter.

The structure of the state estimator (fig. 3.1) is assumed to be described by a linear vector differential equation

i(t) = F(t) i<t) + K(t)(~(t) - H(t) i(t)) + D(t) ~(t)

i(t ) = 0 - 0 -

(3.3)

The choice of the structure can be motivated by the following considerations:

I. The assumptions of linearity of system dynamics. measurements and state estimator and of the Gaussian distribution of the random disturbances implies a Gaussian conditional density

(3.4)

which contains all information about the state ~(tf) given the set of measurements Z(tf). The conditional density (eq. 3.4) is completely characterized by its first two moments [4], the con­ditional mean i<tf) and the conditional error-covariance matrix S(tf)

E{~(tf) I Z(tf)}

,..., ..Jf I E { ~(tf) ~ (tf) Z(tf) }

(3. Sa)

(3.5b)

In contrary to an arbitrary conditional density (eq. 3.4), e.g. in the nonlinear filtering case, the linear filtering problem involves the solution of a finite dimensional problem only.

2. The structure involves the simulation of the system dynamics, which are driven by the estimated output error (innovation [4]) v(t) = z(t) - H(t) x(t) and the known deterministic inputs. This assumption is meaningful, since the only information to compare the estimate i<t) is given by the measurements ~(t).

Page 278: Optimal Estimation in Approximation Theory

276 W.HOFMANN

dynamical system state estimator

Fig. 3.1 Structure of State Estimators for Linear Systems

Page 279: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION 277

Imaging the noise free case and exact state knowledge, the estimator would use the system dynamics and deterministic in­puts only to produce exact estimates. The structure of the state estimator is the same as the structure of observers in the deterministic observation problem [2, 3]. The state esti­mator works on the instantaneous measurements z(t) only, such that no storage of measurements is necessary. -

The gain matrix K(t) of the state estimator still needs to be determined. In contrary to the goal of observer design -- a de­sired stability behaviour -- the optimal filter gain matrix K(t) is found to minimize the cost functional (eq. 3. I). The error beha­viour of the linear state estimator (eq. 3.3) can be described by the error-covariance matrix S(t), which is solution of the matrix differential equation

S(t) = (F(t) - K(t) H(t)) S(t)

+ S(t) (F(t) - K(t) H(t))T

+ K(t) R(t) KT(t)

+ G(t) Q(t) GT(t) S(t ) = P . 0 0

(3. 6)

Using S(t) the cost functional (eq. 3.1) can be reformulated as

(3. 7)

which is no longer a conditional cost functional, since the error­covariance matrix S(t) is independent of a realization z(t), an im­portant property of the linear optimal filter. Using the matrix minimum principle [8], the gain matrix K*(t) minimizing (eq. 3.7) can be found as [4]

(3.8)

where the optimal error-covariance matrix P(t) is solution of the nonlinear matrix Riccati differential equation

i><t> F(t) P(t) + P(t) FT(t)

- P(t) HT(t) R- 1(t) H(t) P(t)

+ G(t) Q(t) GT(t) ; P(t0 ) = P0 (3.9)

From (eqs. 3.8, 3.9) it is to be seen that a necessary condition for the existence of the optimal filter is:

exists or R(t) > 0 (3. 10)

This requirement of the pos1t1ve definiteness of the output noise covariance matrix R(t) is equivalent to the assumption that

Page 280: Optimal Estimation in Approximation Theory

278 W.HOFMANN

all measurements contain white noise. In cases where measurements contain no or only coloured noise the singular problem of fil-tering -- a solution can be found by reducing the order of the sy­stem [9].

Some properties of the Kalman-Bucy filter are summarized [4]:

I. The optimal filter (eq. 3.3) is linear; No other estimation structure-- even nonlinear-- minimizes (eq. 3.1).

2. The optimal filter is an unbiased estimator

E{(~(t) - i(t)} = 0 or E{~(t)} = E{~(t)}. (3. II)

3. The optimal filter provides for optimality uniformly in the time t, i.e. it is independent of the terminal-time tf and the weight­ing matrix X(tf).

4. The optimal gain matrix K*(t) and the minimum error-covariance matrix P(t) {eqs. 3.8, 3.9) are independent of the measurements so that the filter parameters and the expected filter accuracy can be computed off-line.

One important question still remains to be discussed: the stability of the state estimator (eq. 3.3), i.e. the stability beha­viour of the filter dynamics (F(t) - K(t) H(t)). It is well-known that optimality does not imply stability. Moreover problems about the existence of a bounded, positive semidefinite solution for the error-covariance matrix P(t) arise, since it is solution of the non­linear matrix Riccati differential equation (eq. 3.9).

Tl: If the system (eqs. 2.1, 2.2) is completely observable and completely controllable by the noise input vector w(t) (a more adequate expression seems to be completely-distur­bable) and if P0 > O, then the optimal filter dynamics (F(t) - K(t) H(t)) are asymptotically stable and the posi­tive definite solution P(t) is bounded from above.

If these conditions are met, the solution P(t) is numerically stable, i.e. the numerical solution Pc(t) tends to the real solution P(t) in the presence of disturbances as e.g. round-off errors [II].

The previous summarized results can be specialized to the prac­tically important case of the constant gain filter design for sy­stems with time-invariant system matrices:

Assuming the system (eqs. 2.1, 2.2) to be time-invariant and the observation interval to be infinite, - oo < t < tf, the optimal filter is given by

~(t) F i(t) + K00 (~(t) - H ~{t)) + D ~(t) (3.12a)

Page 281: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION

with the constant filter gain

p HT R-1 K 00 00

279

(3.12b)

The stationary error-covariance matrix P00 is solution of the algebraic matrix Riccati equation

(3.12c)

For this case there are necessary and sufficient conditions for sta­bility and the uniqueness of P00 ~ 0 [12].

T2: The optimal filter is asymptotically stable, i.e. Re A(F - Koo H) < 0, and the positive semidefinite solution P00 of (eq. 3.12c) is unique, iff the system (eq. 2. 1, 2.2) is detectable and stabilizable by the input noise ~(t).

There is a final remark necessary on the computational solution of the Kalman-Bucy filter. Except in simple examples the solution of the Kalman-Bucy filter cannot be given analytically. It is ne­cessary to compute it pointwise on a digital computer. Efficient algorithms for the solution of the matrix Riccati differential or algebraic equation are presented in [13].

4. STATE ESTD1ATION AND PLATFORM ERROR ANGLE CONTROL OF A DOPPLER­AIDED INERTIAL NAVIGATION SYSTEM

The considerations about physical and random disturbing state modelling (chapter 2) are now applied to the error modelling of inertial navigation systems. To get an imagination of the at most reachable accuracy of state vector feedback based control schemes the behaviour of the Kalman-Bucy filter is investigated. Using system properties (observability and controllability) and regarding realization requirements (simple and reliable realization) a general and typical way for the design of a suboptimal estimation algorithm is shown. Finally an output feedback controller for the error angle control is proposed, which owns the properties of simple and reliable implementation and as accurate as possible behaviour of the closed­loop system. The results of optimal filtering can be used directly for this type of optimal control.

Inertial navigation systems (INS) are technical measuring units supporting [14] information about velocity and position of an arbi­trarily moving carrier (e.g. satellite, rocket, aircraft, ship) re­lative to a distinct -- inertially fixed or rotating -- coordinate system. They consist of (fig. 4.1a, 4.1c) gyros to keep the stable platform fixed in directions of the desired coordinate system, acce­lerometers, mounted on the platform to measure the accelerations of

Page 282: Optimal Estimation in Approximation Theory

280 W.HOFMANN

the carrier along the axis of the desired coordinate system, and a navigational computer producing velocity and position information by appropriate integration schemes and -- if necessary -- control commands to rotate the stable platform. A conventional coordinate system for earth-fixed carriers is represented by the local, ver­tical coordinate system (fig. 4.1b) with the three orthogonal axis in east, north and vertical direction [14]. In this case it is ne­cessary to keep the platform "horizontally aligned" by counnands de­pending on the earth's rotation and the carrier's motion. A very simple realization for one horizontal axis of a Schuler-tuned INS is indicated in (fig. 4.1c), where also the general structure of an INS is pointed out.

Due to error signals ar1s1ng from the nonideal properties of the gyros and of the accelerometers, e.g. nonorthogonal adjustment, the alignment of the stable platform no longer coincides with the desired coordinate system. The velocity and position information contains slowly varying (low-frequency) erros. The error behaviour of the platform error angles is shown in (fig. 4.2). It is governed by a drift due to constant (random) inputs and by an oscillation -­the Schuler oscillation -- due to the Schuler-tuned feedback for "horizontal alignment". It is obvious that it is necessary to use redundant, external information about velocity and/or position to get a better performance of the INS, especially for the long-time behaviour. In [15] the application of an external velocity infor­mation, supplied by a Doppler-radar, is investigated for the on-line INS-alignment in moving aircrafts. The Doppler-velocity information contains rapidly varying (broad-band) erros besides the true infor­mation. Defining a difference measurement (fig. 4.1c) between INS­velocity and Doppler-velocity the true velocity is eliminated. The difference measurement contains the errors of INS and Doppler-radar only. The control design problem can now be formulated:

Find a controller using the difference measurement only to compensate the random disturbances, such that the accuracy of the physical states of the closed-loop system behaves "best in the sense of their variances" and such that the realization of the controller is "as simple as possible" for reasons of reliable implementation.

Assuming only small deviations from the nominal values, i.e. from the "horizontal alignment" of the stable platform, and exact integration schemes of the accelerations, it is possible to set up a linear error model for the physical error states and the stocha­stic disturbances. It should be emphasized that this assumption is not severe because the goal of a satisfying controller design is just to keep the deviations as small as possible.

The physical error states of the Doppler-aided INS are the plat­form error angles of the east, north and vertical axis a, a andy,

Page 283: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION

s.m.

r. s.

Fig. 4. Ia

Fig. 4. lb

s_m. : servomotor r. s. : resolver ace.: accelerometer

north vertical east

281

Page 284: Optimal Estimation in Approximation Theory

282

W(t)

gyro

W. HOFMANN

navigational computer

Fig. 4.lc INS: Stable Platform (a), Coordinate System (b), and One-Axis Structure

d (CX) /deg d6()1 deg

d(~)/deg

0.25

1.0 t/h 0.5 1.0 t/h

Fig. 4.2 Standard Derivations of Platform Error Angles

Page 285: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION 283

the east and north. velocity error ~ve and ~vn and the east and north position error axe and ~. summarized in the physical state vector

T -x 1(t) = (a, e, y, ~V, ~V , ax, ~X) • e n e n (4. I)

Using physical relationships as linearized gyro-dynamics and rotational kinematics the dynamical behaviour can be described by a linear vector differential equation according to (eq. 2.4a)

i 1 (t) = F II (t) ~I (t) + ~(t) + ~*(t);

~I (to) ~ N(Q_, p 10) (4.2) +)

The dynamic matrix F 11 depends on the nominal flight velocity and the geographic latitude of the aircraft such that F11 = F11 (t) is a function of the independent variable t. Another important property is given by the equal number of independent control input components and of physical state variables, such that every state component can be influence directly either by control moments applied to the stable platform or control signals added in the navi­gational computer.

The velocity difference measurements for the east and north components Ze = ~Ve + VJ and zn = ~Vn + v2 with Doppler-radar velo­city errors v 1 and v2 can be summarized in a linear output vector equation according to (eq. 2.4b)

~(t) = H ~1 (t) + ~*(t) . (4.3)

Besides the modelling of the physical states ~1 (t) there still remains the more elaborate analysis of the random system and mea­surement disturbances w*(t) and v*(t). In many practical problems the modelling of stochastic disturbances cannot be performed by a physical - insight - based way, but is done by an experimental cor­relation function or spectral density analysis. The underlying basic disturbance property of these analysis methods is the assumpt­ion of stationary Gauss-Markov processes, the mathematical des­cription of which is already discussed in chapter 2. The reasons for this assumption are on one hand the necessity of formulating the disturbance identification problem as a -- at least approxi­mately -- finite dimensional problem. Uoreover it is often more promising to use simple, low-order disturbance models (shaping fil­ters) than sophisticated, high-order models. There is not only the greater effort for analysis, but also a higher bound of confidence in the identified model parameters. On the other hand it is known

+) For parameter values of the complete INS-error model see [15].

Page 286: Optimal Estimation in Approximation Theory

284 W. HOFMANN

by the central limit theorem [4] that the superpos1t1on of a large number of small independent random effects, regardless of their in­dividual distribution, causes the distribution of the sum of these effects to be -- at least approximately -- Gaussian. In addition, if a dynamical system with low-pass filter characteristic is driven by a disturbance with nearly arbitrary distribution, the distribut­ion of the output of the dynamical system is -- at least approxi­mately-- Gaussian [5].

Returning to the INS-disturbance modelling it is necessary to investigate the random disturbance inputs to the gyros, to the accelerometers and the Doppler-radar velocity error. Extensive analysis and a comparison of the influence of different effects show that the random disturbances can be modelled using the follow­ing (most simple) mathematical models.

I. Gaussian white noise Ew(t):

A random disturbance with negligible time-correlation (relative to the other time-constants of the system) can be modelled by Gaussian white noise, whose spectral density and correlation function are illustrated in (fig. 4.3)

Ew(t) ~ N(O, q)

2. Random constant (bias) ~b(t):

A random disturbance, which is constant -­stant during the considered time-interval! by a random constant (bias) (fig. 4.3)

2 Eb(t) = o ; ~b(O) ~ N(O, ob)

3. Exponentially correlated noise Ec(t):

(4.4a)

at least nearly con­can be modelled

(4.4b)

If the time-correlation of a random disturbance is not negli­gible, it can often be modelled by exponentially correlated noise (fig. 4.3)

E (t) c

!; (0) c

- 1/T E (t) + w(t) c c

N(O, o2) c

(4.4c)

Since stationary processes are assum2d it is necessary to choose the white noise variance qc = 2 · Oc /Tc such that the correlated noise variance Pc(t) = E{oc2(t)} as solution of

p (t) = - 2/T p (t) + q ; c c c c p (0) = o2 c c

(4.5)

Page 287: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION 285

autocorrl!lati on function spectral density function block diagram

white noise _j_. d= Ill 'i = q. 6t't'l St(w) ::q

4_. _Lw r; random bias ~b

~('l') = 0 5 (W) = Z'ffd; 6(c.u)

A I"'-w + 'fc

exponentially + correlated / noise "1' "' Tc zdf. Tc

C('t'l=d2-exp(-l'tlll I s (lll) = f c c ~ 1+W2/T2 c

Fig. 4.3 Description of Simple Random Processes

Page 288: Optimal Estimation in Approximation Theory

286 W.HOFMANN

is stationary by Pc(t) = Pc(O) = const.

Summarizing the disturbing states resulting from the shaping filters (eqs. 4.4b, 4.4c) in the disturbance state vector ~(t) of order n2 = 5 the findings of [15] can be expressed according to (eq. 2. 5)

i2(t)

~*(t)

~*(t)

F22 ~2(t) + G2 ~(t)

F12 ~2 (t) + G1 ~(t)

~(t)

with the statistics

~2(to) ~ N(Q, p20)

~(t) ~ N(Q, Q)

~(t) ~ N(Q, R)

(4.6a)

(4.6b)

Augmenting the physical state model (eq. 4.2) by the distur­bance state model (eq. 4.6) the complete error model of the Doppler­aided INS of order n = 12 is given by

i) (t) F II ( t) F 12 ~I (t) + Gl ~(t) + u(t)

i2(t) 0 F22 ~2(t) G2 0

~(t) [HI O] ~] (t) + ~(t) (4. 7)

~2(t)

which can be treated within the general framework of chapter 2.

Since several control schemes require the knowledge about the complete state or at least an optimal state estimate, the Kalman­Bucy filter (eq. 3.9) for the optimal error 3tate estimation of the Doppler-aided INS (eq. 3.7) is investigated for a 1.5 h, east-west­east flight. The estimation accuracy is a lower bound for the reachable closed-loop system accuracy [16]. Since especially the accuracy of the stable platform alignment is of importance, the standard deviations of the estimation errors of the platform error angles are plotted in (fig. 4.4). A comparison to the error beha­viour of the non-aided INS shows that the final uncertainty is re­duced by about 75% for the horizontal and by about 50% for the ver­tical error angles.

The solution of the matrix Riccati differential equation (3.9)

Page 289: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION

d la)/deg d (j3 )/deg

0.5 1.0

diYl/deg

0.1

t/h 0.5 1.0 tl h

Fig. 4.4 Optimally Platform Error Angle Standard Deviations

287

Page 290: Optimal Estimation in Approximation Theory

288 W.HOFMANN

for the optimal error-covariance matrix P(t), which is of dimension (12,12), requires the simultaneous solution of 78 nonlinear differen­tial equations. The (optimal) Kalman-Bucy filter cannot be suggested to be realized on an onboard-computer. A suboptimal, but much simpler estimation algorithm has to be developed. The main steps of a general and typical design procedure with special regard to system properties and realization requirements are summarized below. An extensive discussion can be found in [15].

The finally aspired structure of the suboptimal state estima­tor is given by

- ~

~(t) = A ~(t) + K(~(t) - C ~(t)) + DE ~(t) , (4.8)

an as low-order as possible (<< 12) filter with constant filter ma­trices. It requires the solution of a low-order time-invariant vector differential equation on the onboard-computer only. (Eq. 4.8) represents a realization rule for analogue computers directly. If there is available a digital computer, (eq. 4.8) can be solved by means of the transition matrix [2] yielding a vector difference equation.

To get the desired algorithm (eq. 4.8) the following steps are executed:

I. Assumptions on the system:

For the given INS-activity there is no prefered flight direct­ion leading to the assumption of zero flight velocity. This implies the system matrix Ftt(t) in the INS-error model (eq. 4.7) to be independent of the time and so the complete system to time-invariant.

2. Observability and (noise) controllability analysis:

From (chapter 3, T2) it is known that an asymptotically stable, constant gain filter can be designed by means of the algebraic Riccati equation only for the detectable and (noise) stabili­zable subsystem of a given system. An observability and con­trollability analysis, which can be performed either analyti­cally or numerically [2], shows the order of the detectable and (noise) stabilizable subsystem to be five. The transformed state variables of this subsystem are the first five physical states in (eq. 4.1), i.e. the three platform error angles a, a. y and the two INS-velocity errors 6ve and 6vn, which are li­nearly combined with the disturbing state ~2 (t).

The dynamics of the detectable and (noise) stabilizable subsy­stem can be described by a time-invariant 5th-order vector dif­ferential equation

Page 291: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION 289

~(t) = A ~(t) + B ~(t) + Df; ~(t) (4.9a)

together with a linear measurement vector equation

~(t) = c ~(t) + ~(t) (4.9b)

3. Constant gain filter design via Riccati equation:

Since the subsystem (eq. 4.9) is timeinvariant, detectable and (noise) stabilizable, the stationary Kalman-Bucy filter algo­rithm (eq. 3.12) can be applied to find a filter gain matrix K for the filter algorithm (eq. 4.8). It should be emphasized that the filter gain matrix K is optimal only under the assumpt­ions of zero flight velocity and an infinite measurement inter­vall. In all other cases it works suboptimal. The result of one typical simulation of the filter algorithm (eq. 4.8), estimating the platform error angles from noisy measurements for the above mentioned flight, is shown in (fig. 4.5).

4. Q-stabilization for better filter performance:

From (fig. 4.5) it is to be seen that the filter designed in step 3 using the nominal statistics of the INS-error model (eqs. 4.2, 4.6) is missing the capability of reducing the ini­tial platform error angles within the finite measurement inter­vall of I ,5 h.

This is a typical result for filters designed by the algebraic Riccati equation (eq. 3.12c) for systems with small system to measurement noise ratio and system eigenvalues near to the ima­ginary axis. The disadvantageous property can be explained by the assumed infinite measurement intervall. Loosely spoken the filter gain matrix K is such big that it keeps the filter to be "just stable", i.e. to reduce initial errors very slowly. On the other hand it is as small as possible to increase the error variances not by to much measurement noise driving the filter (eq. 4.8). Such designed filter tend to possess another unde­sirable property: divergence. This means that the true esti­mation error behaviour is not asymptotically stable in the rea­listic environment due to e.g. modelling errors [4].

To prevent these disadvantages the Q-stabilizing method [4] can be used. By an artificial increasing of the system noise cova­riance matrix Q the filter (eq. 4.8) tends to be more stable, i.e. to reduce initial errors faster. A more systematic but more sophisticated procedure for the inclusion of stability con­straints in the filter design is derived in [17].

Applying the Q-stabilizing method to the constant gain INS-fil­ter design -- a suitable Q-matrix was found by trial and error -­the results for the standard deviations of the estimation errors of the horizontal platform error angles are shown in (fig. 4.6).

Page 292: Optimal Estimation in Approximation Theory

290 W. HOFMANN

... y It I

Fig. 4.5 Simulation of Platform Angle Error Behaviour (Sub­optimal Estimation)

d{a)/deg d{~)/deg

0.5 1.0 t/h 0.5 1.0

Fig. 4.6 Suboptimally (Q-stabilized) Platform Error Angle Standard Devi ations

t/h

Page 293: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION 291

The vertical error accuracy does not differ from the optimal one (fig. 4.4) within the illustration accuracy. In comparison to the optimally reachable accuracy of the Kalman-Bucy filter (fig. 4.4) the loss of accuracy is negligible compared to the simple implementation of the filter algorithm (eq. 4.8).

If the Kalman-Bucy filter (eq. 3.3, 3.8, 3.9) or the subopti­mal estimation algorithm (eq. 4.8) is used as a state estimator to­geth~r with some control law [2, 3, 16] in a closed-loop system, the presented results (fig. 4.4, 4.6) are at most lower bounds for the reachable accuracy. It seems to be more adequate to attack the control problem with the requirement of most accurate closed-loop system behaviour directly by minimizing a cost functional, which gives regard to the closed-loop system variances. The following optimal control problem is considered:

A dynamical system is described by the linear vector differen­tial equation (eq. 2.1) and a linear output vector algebraic equation (eq. 2.2), where the state vector x(t) is partitioned according to (eq. 2.6a) and the system matrices are of the structure (eq. 2.6b). It will be assumed that the output noise covariance matrix R(t) and the control input matrix D1(t) are invertible.

Let an output vector feedback controller (fig. 4.7) be given by

~(t) =- K1(t) ~(t) + K3(t) ~2 (t) (4. lOa)

~2 (t) = F22 (t) i 2(t) + K2 (t)(~(t) - H2(t) i 2(t)) . (4. lOb)

Find the parameters of the feedback matrices K1(t), K2(t) and K3(t) such that the cost functional

~I (tf) }

~2(tf)

(4.11)

with the disturbance compensation error ~(t) = ~2 (t) - iz<t) and the terminal-time t 0 2 t ~ tf is minimized.

The choice of the controller structure (eq. 4.10) can be moti­vated by the following considerations:

I. The simplest structure of the controller would be the static out­put feedback K1(t) ~(t). By the assumption of an invertible con­trol input matrix D1(t) this type of control is sufficient for stabilizing the physical system.

2. The disturbance vector w*(t) (eq. 2.5b) contains information by the time-correlation of-~(t). It seems to b~ useful to con­struct an estimator for ~(t). The estimate ~(t) serves to com-

Page 294: Optimal Estimation in Approximation Theory

292 W. HOFMANN

1; (t)

Fig. 4.7 Structure of the Output Feedback Controller

Page 295: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION 293

pensate the correlated noise ~(t) in the mean and provides for a more accurate behaviour of the closed-loop system.

The solution of this parameter optimization problem is given by [ I8]

KI (t) -I *

= D I KI (t) (4.I2a)

K2(t) K;(t)

K3(t) - D ~I (t) FI 2(t) + KI(t) H2(t)

with

K*(t) K!(t) P*(t) HT R-I (t) (4. I2b)

K;(t)

where the error covariance matrix P*(t) is solution of the n-th or­der, n = ni + n2, matrix Riccati differential equation

P*(t) = F(t) P*(t) + P*(t) FT(t)

- P*(t) HT(t) R-I(t) H(t) P*(t)

+ G(t) Q(t) GT(t)

(4. I2c)

By the kind of solution it is to be expected that there exists an equivalence between linear optimal filtering and linear output vector feedback. This is shown and discussed in detail in [I8]. As a consequence properties and results of optimal filtering (chap­ter 3) can directly be applied to the proposed controller design procedure. Some main results are repeated:

I. The stability behaviour is governed by (chapter 3, TI) or in the constant gain design case by (chapter 3, T2).

2. The controller provides for optimality uniformly in t, i.e. it is independent of the terminal time tf and the weighting matrix IT(tf).

3. The accuracy behaviour of the closed-loop system can be judged off-line by inspection of the diagonal elements of the error covariance matrix P*(t).

By the equivalence it is also possible to use the already known numerical results from the above discussed filters. The optimal output vector feedback controller involves a dynamical compensator o£ order n2 -- the order of the disturbance model (eq. 4.6a) -- and

Page 296: Optimal Estimation in Approximation Theory

294 W. HOFMANN

time-varying controller gain matrices. The resulting accuracy of the closed-loop system -- equal to that of the Kalman-Bucy filter is to be seen from (fig. 4.4). According to the design conside­rations of the suboptimal filter algorithm the design of a subopti­mal controller can be performed yielding a static output vector feedback control

~(t) = K _!(t) (4.13)

The feedback matrix K is equal to the filter gain matrix in (eq. 4.8). The accuracy results of the suboptimally controlled sy­stem are to be seen from (fig. 4.6). The considerations about loss of accuracy and simplicity of realization are the same as in the filter case.

From these results it is to be seen that the proposed control­ler design procedure for systems of the assumed structure leads to simple controllers and most accurate closed-loop system behaviour. All results and experiences of linear optimal filtering can be applied directly. Moreover the procedure is independent of the weighting matrix of the cost functional such that the optimal solu­tion is maintained within one step and a satisfying optimal solution needs not to be determined by trial and error.

5. SUMMARY

The -- at least approximate -- mathematical description of dy­namical systems by means of linear vector differential equations is discussed with special regard to random system and measurement noise modelling. The derivation of the Kalman-Bucy filter using a parame­ter optimization approach and the properties of the linear optimal filter solution, especially the stability conditions, are reviewed. The practical problem of controlling a Doppler-aided inertial navi­gation system in an as accurate and as simple as possible way is in­vestigated. To provide for a lower bound of accuracy of state vec­tor feedback based controllers the linear optimal filter and a sub­optimal estimation algorithm, derived in a general and typical way regarding system properties and realization requirements, are exa­mined. An optimal output vector feedback controller is proposed which enables the design of low-order random disturbance compensa­tors with most accurate closed-loop system behaviour. By the equi­valence of linear optimal filtering and the proposed optimal output vector feedback control all properties and results of estimation can be applied directly to the control problem.

Page 297: Optimal Estimation in Approximation Theory

APPLICATION OF ESTIMATION TO NAVIGATION 295

[ 1] COURANT, R.' HILBERT, D.

[2] KWAKERNAAK, H. SIVAN, R.

[3] ANDERSON, B.D.O., MOORE, J.B.

[4] JAZWINSKI, A.H.

[5] PAPOULIS, A.

[6] GELB, A., e.a.

[7]

[8] ATHANS, M.

[9] HOFMANN, W., KOR~1, W.

[10] BROCKETT, R.W.

[ 11]

[12] KUCERA, V.

LITERATURE

"Methods of Mathematical Physics", Vol. I, II. Interscience Publishers, New York, 1962.

"Linear Optimal Control Systems". Wiley-Interscience, New York, 1972.

"Linear Optimal Control". Prentice-Hall, Englewood Cliffs, 1971.

"Stochastic Processes and Filtering Theory". Academic Press, New York, 1970.

"Probability, Random Variables, and Sto­chastic Processes". McGraw Hill, New York, 1965.

"Applied Optimal Estimation". The M.I.T. Press, Cambridge, 1974.

"Theory and Applications of Kalman Fil­tering". AGARDograph 139, 1970.

"The Matrix Minimum Principle". J. on Information and Control, vol. 11, p. 592-606, 1968.

"Reduced-Order State Estimators with App­lication in Rotational Dynamics". "Gyrodynamics", Springer-Verlag, Berlin, 1974.

"Finite Dimensional Linear Systems". John Wiley & Sons, New York, 1970.

"A Short Course on Kalman Filtering and Application". The Analytic Science Corporation, Reading ~lass.

"A Contribution to Matrix Quadratic Equations". IEEE Trans. on Automatic Control, vol. AC-17 (June 1972), pp. 344-347.

Page 298: Optimal Estimation in Approximation Theory

296

[13] VAUGHAN, D.R.

[14] MULLER, P.C.

[ I 5] HOFMANN, W.

[16] BRYSON, A.E., HO, Y.

[ 17] HOFHANN, W.

[18] HOFUANN, W.

W. HOFMANN

"A Negative Exponential Solution for the Ma­trix Riccati Equation", IEEE Trans. Automatic Control, vol. AC-14, (Feb. 1969), pp. 72-75.

"Special Problems of Gyrodynamics". International Centre of Mechanical Sciences, Courses and Lectures No. 63, Springer-Ver­lag, Udine, 1970.

"Die Anwendung der Kalman-Bucy-Filtertheorie beim Entwurf von Rilckkopplungsnetzwerken filr Dopplergestiltzte Tragheitsnavigationssysteme im Flug". DFVLR-IB Nr. 552-75/18.

"Applied Optimal Control". Ginn and Company, London, 1969.

"Application of loners to the Design of Op­timal State Estimators Under Stability Con­straints". Proceedings of the Conference on Information Sciences and Systems, Baltimore, 1976.

"Optimal Stochastic Disturbance Compensation by Output Vector Feedback". to appear as DLR-Forschungsbericht.

Page 299: Optimal Estimation in Approximation Theory

INDEX

algebraic variety 218,219 amplitude modulation (AM) 182,183,186,188-190 angiography 225,226 autocorrelation function 272 autocorrelation matrix 192 automated cartography 202,203,210 Banach space 151,153 bandlimited 55,56,62 bandpass filter 188,189,191 baseband modulation (BB) 182,183,188 best approximation 139,140,262,263 bounded mesh ratios 143-145 Bragg's Law 168 branch-and-bound-type 159 breast cancer 222 Brouwer's fixed point theorem 247 Brouwer-Horn theorem 251 B-spline 86,87 carrier phase tracking 181,189,191 cartographic generalizations 201-203,207-209 Chebyshev center 7,8,10 Chebyshev nodes 125,126 Chebyshev perfect spline 38 Chebyshev polynomial 125 Chebyshev radius 8 Chebyshev system 30,34 Chebyshev type inequalities 78 convolution method 219,221 Costas loop 189 covariance matrix 272,289 decay constants 259-261 decay processes 259 Dijkstra method 206 discrete Fourier transform 193 divided differences 29,79 double sideband modulation (DSB) 182-184, 186,188-191 Drosophila 242

~7

Page 300: Optimal Estimation in Approximation Theory

298

dynamic programming 181,195 eigenfunction 7,50,58,66 eigenvalue 49,58,66,102,193,289 ellipsoids 128 elliptic functions 94 EM! scanner 221 entropy 204 envelope construction 69,75 envelope theorem 36 envelope of smooth functions 28 fibrocystic disease 222-224 F-model 230 Ford-Fulkerson method 206 Fourier coefficients 48,107 Fourier inversion formula 217,218 Fourier method 219,221 Fourier series 210 Fourier transform 168,205,272 Fredholm determinants 66 Gauss elimination 82,87,88 Gauss-Markov process 272,183 Gaussian distribution 272 Gaussian estimation 267,268,273 Gaussian white noise 184,185,270,272,273,278 Gram matrix 5 Gramian 96 Haar subspace 154,155 Hadamard's 3-circle theorem 134 Hartline Ratliff model 243 Helmert transformation 211 Hermite data 103,107,129

INDEX

Hilbert space 5,10,12,23,48,49,57,93,94,96,101-103,113,126,151,155 Hutchinson's equation 253 hypercircle 8,12 hypercircle inequality 7,11 hypernormal A-rule 110,112,113 inertial navigation system (INS) 267,275,276,278-280 information content 203,204,206-211 integer approximation problem 159 integral equation 58,59,64 intersymbol interference (lSI) 184,186,194,195 inverse approximation theorems 145,148 Kacmarz (ART) method 219,220,221 Kalman-Bucy filter 194,268,275,288,289,291,294 Lagrange interpolation 103,104,106 likelihood function 184 Limulus 242,248 Lipshitz classes 141 local best approximation 262

Page 301: Optimal Estimation in Approximation Theory

INDEX 299

low pass filter 210 mammogram 224 Markoff-Krein inequalities 31 Markov-type inequality 145-147 matched filter 192,194,195,197 matrix Riccati differential equation 277-279,289 maximum likelihood (ML) 186-189 mixing property 144 modules of continuity 140 moment conditions 29 narrow bandpass filter 188,189 natural spline 6,7,50 n-dimensional subspace 49,56,61,64,66,127,131 Newton's method 82,262 nonlinear filtering 275 n-width 50,55,57,95,127,128 Nyquist criterion 184 ommatidium 242-245 optimal approximation 201 optimal estimation 13,207 optimal filter 277-279,293 optimal interpolation 55,87,89,90,93 optimal linear algorithm 20,23,36,49 optimal quadrature formula 106,109,127 optimal recovery 3,4,7,38,39,42,49,50,55,56,64,67,69,72,73,75,

78-80,93,97,102,103,107,116,125,126,129,215,216, 220

optimal A-rule 109,112,113 optimal sampling points 7 optimal state estimation 267,268,275 optimal subspace 50,56,57,59,61,66,128 Paley-Wiener 63,65 partitions, equidistant 143,144 partitions, mixed 142 partitions, nested 145 Patterson function 170,171 peeling-off~ethod (abschalmethode) 260,261 perfect splines 28,29,33,35,36,37,71,72,74,76,78 phase modulation (PM) 182-184,190 principal representation 31,78 programming, linear 229,234 programming, parametric 229,236 programming, quadratic 229,236 projection 5,63,64,140,151-153,155 projection constant 154 projection, orthogonal 155,156 projection, Fourier 154,156 projection, Lagrange 155 projection, Rademacher wave 154,156

Page 302: Optimal Estimation in Approximation Theory

300

prolate spheroidal wave equation 59 radiation-treatment planning 229,230 radiograph 217-219,221 radiotherapy 229,230 Radon inversion formula 217,218 Radon transform 218 reconstruction from x-ray 216,219 regression 186 reproducing kernel 26,63,65,94,102,108,114,120 restricted moment space 30,33,34,78 Rolle's theorem 31,32,71,74 saturation 146 Schauder's theorem 251 semantic information content 201,204,207 separation of convex sets 14 separation theorem 15 S-model 231-233,235 spectral radius 247 spline interpolation 36,38,83,90 spline projector 109,120

INDEX

splines 32,37,73,79,80,86,93,94,96,97,104,107,108,114,126,130,140, 148,201,211

stochastic approximation 187,191,194 Sturm-Liouville 59 synchronous data transmission 181,182,198 Taylor expansion 33,79,116,127,268 time limited 55,56 tomography 221 Topfer's Law 210 totally positive kernel 56,67 totally positive matrix 82,88 trellis diagram 196 trend surface 205,207 tumor 220,229,230,236,238 uncertainty relationship 56 Verhulst equation 253 vestigial sideband modulation (VSB) 182-184,186,189 Viterbi algorithm 195,197 weak Chebyshev system 29,30,34,78,80 Wiener-Kolmogorff 205 worst function 18,19 x-ray diffraction 159,162