Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic...

138
SPRINGER BRIEFS IN MATHEMATICS Jesús Martínez-Frutos Francisco Periago Esparza Optimal Control of PDEs under Uncertainty An Introduction with Application to Optimal Shape Design of Structures

Transcript of Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic...

Page 1: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

S P R I N G E R B R I E F S I N M AT H E M AT I C S

Jesús Martínez-Frutos Francisco Periago Esparza

Optimal Control of PDEs under UncertaintyAn Introduction with Application to Optimal Shape Design of Structures

Page 2: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

SpringerBriefs in Mathematics

Series editors

Nicola Bellomo, Torino, ItalyMichele Benzi, Atlanta, USAPalle Jorgensen, Iowa City, USATatsien Li, Shanghai, ChinaRoderick Melnik, Waterloo, CanadaLothar Reichel, Kent, USAOtmar Scherzer, Vienna, AustriaBenjamin Steinberg, New York, USAYuri Tschinkel, New York, USAGeorge Yin, Detroit, USAPing Zhang, Kalamazoo, USA

SpringerBriefs in Mathematics showcases expositions in all areas of mathematicsand applied mathematics. Manuscripts presenting new results or a single new resultin a classical field, new field, or an emerging topic, applications, or bridges betweennew results and already published works, are encouraged. The series is intended formathematicians and applied mathematicians.

[email protected]

Page 3: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

BCAM SpringerBriefsEditorial Board

Enrique ZuazuaDeusto TechUniversidad de DeustoBilbao, Spain

and

Departamento de MatemáticasUniversidad Autónoma de MadridCantoblanco, Madrid, Spain

Irene FonsecaCenter for Nonlinear AnalysisDepartment of Mathematical SciencesCarnegie Mellon UniversityPittsburgh, USA

Juan J. ManfrediDepartment of MathematicsUniversity of PittsburghPittsburgh, USA

Emmanuel TrélatLaboratoire Jacques-Louis LionsInstitut Universitaire de FranceUniversité Pierre et Marie CurieCNRS, UMR, Paris

Xu ZhangSchool of MathematicsSichuan UniversityChengdu, China

BCAM SpringerBriefs aims to publish contributions in the following disciplines: AppliedMathematics, Finance, Statistics and Computer Science. BCAM has appointed an EditorialBoard, who evaluate and review proposals.

Typical topics include: a timely report of state-of-the-art analytical techniques, bridgebetween new research results published in journal articles and a contextual literature review,a snapshot of a hot or emerging topic, a presentation of core concepts that students mustunderstand in order to make independent contributions.

Please submit your proposal to the Editorial Board or to Francesca Bonadei, Executive EditorMathematics, Statistics, and Engineering: [email protected].

More information about this series at http://www.springer.com/series/10030

[email protected]

Page 4: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Jesús Martínez-Frutos • Francisco Periago Esparza

Optimal Control of PDEsunder UncertaintyAn Introduction with Application to OptimalShape Design of Structures

123

[email protected]

Page 5: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Jesús Martínez-FrutosComputational Mechanics and ScientificComputing Group, Departmentof Structures and Construction

Technical University of CartagenaCartagena, Murcia, Spain

Francisco Periago EsparzaDepartment of Applied Mathematicsand Statistics

Technical University of CartagenaCartagena, Murcia, Spain

ISSN 2191-8198 ISSN 2191-8201 (electronic)SpringerBriefs in MathematicsISBN 978-3-319-98209-0 ISBN 978-3-319-98210-6 (eBook)https://doi.org/10.1007/978-3-319-98210-6

Library of Congress Control Number: 2018951210

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2018This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or partof the material is concerned, specifically the rights of translation, reprinting, reuse of illustrations,recitation, broadcasting, reproduction on microfilms or in any other physical way, and transmissionor information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilarmethodology now known or hereafter developed.The use of general descriptive names, registered names, trademarks, service marks, etc. in thispublication does not imply, even in the absence of a specific statement, that such names are exempt fromthe relevant protective laws and regulations and therefore free for general use.The publisher, the authors and the editors are safe to assume that the advice and information in thisbook are believed to be true and accurate at the date of publication. Neither the publisher nor theauthors or the editors give a warranty, express or implied, with respect to the material contained herein orfor any errors or omissions that may have been made. The publisher remains neutral with regard tojurisdictional claims in published maps and institutional affiliations.

This Springer imprint is published by the registered company Springer Nature Switzerland AGThe registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland

[email protected]

Page 6: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

To our families

[email protected]

Page 7: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Preface

Both the theory and the numerical resolution of optimal control problems of sys-tems governed by—deterministic—partial differential equations (PDEs) are verywell established, and several textbooks providing an introduction to the funda-mental concepts of the mathematical theory are available in the literature [11, 16].

All of these books as well as most of the researches done in this field assume thatthe model’s input data, e.g. coefficient in the principal part of the differentialoperator, right-hand side term of the equation, boundary conditions or location ofcontrols, among others, are perfectly known. However, this assumption is unreal-istic from the point of view of applications. In practice, input data of physicalmodels may only be estimated [5]. In addition, as will be illustrated in this book,solutions of optimal control problems may exhibit a dramatic dependence on thoseuncertain data. Consequently, more realistic optimal control problems shouldaccount for uncertainties in their mathematical formulations.

Uncertainty appears in many other optimization problems. For instance, instructural optimization, where one aims to design the shape of a structure thatminimizes (or maximizes) a given criterion (rigidity, stress, etc.). All physicalparameters of the model, e.g. the loads, which are applied to the structure, aresupposed to be known in advance. However, it is clear that these loads either varyin time or vary from one sample to another or are simply unpredictable.

Among others, the following two approaches are used to account for uncer-tainties in control problems:

• On the one hand, if there is no a priori information on the uncertain inputs shortof lower and upper bounds on their magnitudes, then one may use a worst-casescenario analysis, which is formulated as a min-max optimization problem.More precisely, the idea is to minimize the maximal value of the cost functionover a given set of admissible perturbations. The advantage of this formulationis that, in general, it does not require an exhaustive previous analysis of theuncertain parameters of the model, just the lower and upper bounds of those.However, this a priori advantage may become a disadvantage because if theupper and lower bounds on the uncertain data are relatively large, then optimal

vii

[email protected]

Page 8: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

solutions obtained from this approach may be too conservative, leading to apoor performance of the physical system under consideration [1].

• On the other hand, when statistical information on the random inputs is avail-able, it is natural to model uncertainties using probabilistic tools. In someapplications, the mean and the spatial correlation of these inputs are estimatedfrom experimental data. From this, assuming a Gaussian-type distribution ofsuch inputs, those may be modelled by using, for example, Karhunen–Loève-type expansions. More elaborated statistical methods may also be used torepresent input parameters [5, 15]. Due to the great interest in applications, veryefficient computational methods for solving PDEs with random inputs have beendeveloped during the last two decades (see, e.g. [2, 3, 6, 9]). These continuoustheoretical and numerical improvements in the analysis and resolution of ran-dom PDEs have enabled the development of control formulations within aprobabilistic setting (see, e.g. [7, 13, 14]). Probably, the fact that this proba-bilistic framework is fed with experimental data is the main drawback of thisapproach. On the other hand, optimal solutions obtained from these probabilisticmethods are, in general, better than those obtained from a worst-case scenario.

This book aims at introducing graduate students and researchers to the basictheoretical and numerical tools which are being used in the analysis and numericalresolution of optimal control problems constrained by PDEs with random inputs.Thus, a probabilistic description of those inputs is assumed to be known in advance.The text is confined to facilitating the transition from control of deterministic PDEsto control of PDEs with random inputs. Hence, a previous elementary knowledge ofoptimal control of deterministic PDEs and probability theory is assumed. Werecommend the monograph [16] for a basic and very complete introduction todeterministic optimal control. Concerning probability theory, any introductorytextbook on probabilities, e.g. [8], provides the bases which are required in thisbook. See also [12, Chap. 4] for a complete review of this material.

The book is composed of seven chapters:Chapter 1 is devoted to a very brief introduction to uncertainty modelling, by use

of random variables and random fields, in optimal control problems. To this end,three very simple but illustrative examples are introduced. These examples coverthe cases of elliptic and evolution (first and second order in time) PDEs. We willconsistently use these models, but especially the elliptic one, to illustrate (some of)the main methods in optimal control of random PDEs.

Chapter 2 introduces the mathematical and numerical tools which are required inthe remaining chapters. Special attention is paid to the numerical approximation ofrandom fields by means of Karhunen–Loève (KL) expansions.

Chapter 3 focusses on the mathematical analysis of optimal control problems forrandom PDEs. Firstly, well-posedness of such random PDEs is proved byextending the classical methods for deterministic PDEs (Lax–Milgram theorem forthe elliptic case and the Galerkin method for evolution equations) to a suitableprobabilistic functional framework. Secondly, the existence of solutions for robustand risk-averse optimal control problems is proved. The former formulation

viii Preface

[email protected]

Page 9: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

naturally arises when one is primarily interested in minimizing statistical quantitiesof the cost functions such as its expectation. In order to reduce the dispersion of theobjective function around its mean, it is reasonable to consider measures of sta-tistical dispersion such as standard deviation. This choice, usually referred to in theliterature as robust control [10], aims at obtaining optimal solutions less sensitive tovariations in the input data. However, if one aims to obtain optimal controls forunlike realizations of the random input data (which therefore have a very littleimpact on the mean and variance of the objective function but which could havecatastrophic consequences), then risk-averse models ought to be considered. In thissense, risk-averse control [4] minimizes a risk function that quantifies the expectedloss related to the damages caused by catastrophic failures. A typical examplewhich illustrates the main difference between robust and risk-averse control is thefollowing: imagine that our optimization problems amount to finding the optimalshape of a bridge. If one is interested in maximizing the rigidity of the bridge in theface of daily random loads (e.g. cars and people passing across the bridge), then arobust formulation of the problem is a suitable choice. However, if the interest is todesign the bridge to be able to withstand the effects of an earthquake, then one mustuse a risk-averse approach. In probabilistic terms, robust control focusses on thefirst two statistical moments of the distribution to be controlled. Risk-averse controlalso takes into account the tail of such distribution.

Chapter 4 is concerned with numerical methods for robust optimal controlproblems. Both gradient-based methods and methods based on the resolution of thefirst-order optimality conditions are described. Smoothness with respect to therandom parameters is a key ingredient in the problems considered in this chapter.Thus, we focus on stochastic collocation (SC) and stochastic Galerkin(SG) methods [2], which, in this context, are computationally more efficient thanthe classical brute force Monte Carlo (MC) method. An algorithmic expositionof these methods, focussing on the practical numerical implementation, is pre-sented. To this end, MATLAB computer codes for the examples studied in thischapter are provided in http://www.upct.es/mc3/en/book.

Chapter 5 deals with the numerical resolution of risk-averse optimal controlproblems. An adaptive gradient-based minimization algorithm is proposed for thenumerical approximation of this type of problem. An additional difficulty whichemerges when dealing with risk-averse-type cost functionals is the lack ofsmoothness with respect to the random parameters. A combination of a stochasticGalerkin method with a Monte Carlo sampling method is used for the numericalapproximation of the cost functional and of its sensitivity. As in the precedingchapter, an algorithmic approach supported with MATLAB computer codes forseveral illustrative examples is presented.

Chapter 6 aims to illustrate how the methods presented in the book may be usedto solve optimization problems of interest in real applications. To this end, robustand risk-averse formulations of the structural optimization problem are analysedfrom theoretical and numerical viewpoints.

Preface ix

[email protected]

Page 10: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Finally, Chap. 7 includes some concluding remarks and open problems.A list of references is included in each chapter. Also, links are provided to

websites hosting open-source codes related to the contents of the book.We have done our best in writing this book and very much hope that it will be

useful for readers. This would be our best reward. However, since the book isconcerned with uncertainties, very likely there are uncertainties in the text associ-ated with inaccuracies or simply errors. We would be very grateful to readers whocommunicate any such uncertainties to us. A list of possible errata and improve-ments will be updated on the website: http://www.upct.es/mc3/en/book/.

We would like to thank Enrique Zuazua for encouraging us to write this book.We also acknowledge the support of Francesca Bonadei, from Springer, and of theanonymous reviewers whose comments on the preliminary draft of the book led to abetter presentation of the material. Special thanks go to our closest collaborators(David Herrero, Mathieu Kessler, Francisco Javier Marín and Arnaud Münch)during these past years. Working with them on the topics covered by this texthas been a very fruitful and enriching experience for us. Finally, we acknowledgethe financial support of Fundación Séneca (Agencia de Ciencia y Tecnologíade la Región de Murcia, Spain) under project 19274/PI/14 and the SpanishMinisterio de Economía y Competitividad, under projects DPI2016-77538-R andMTM2017-83740-P.

Cartagena, Spain Jesús Martínez-FrutosJune 2018 Francisco Periago Esparza

References1. Allaire, A., Dapogny, Ch.: A deterministic approximation method in shape optimization under

random uncertainties. SMAI J. Comp. Math. 1, 83–143 (2015)2. Babuška, I., Novile, F., Tempone, R.: A Stochastic collocation method for elliptic partial

differential equations with random input data. SIAM Rev. 52 (2), 317–355 (2010)3. Cohen, A., DeVore, R.: Approximation of high-dimensional parametric PDEs. Acta Numer.

Graduate Studies in Mathematics, 1–159 (2015)4. Conti, S., Held, H., Pach, M., Rumpf, M., Schultz, R.: Risk averse shape optimization.

SIAM J. Control Optim. 49(3), 927–947 (2011)5. de Rocquigny, E., Devictor, N., Tarantola, S.: Uncertainty in industrial practice: a guide to

quantitative uncertainty management. Wiley, New Jersey (2008)6. Gunzburger, M. D., Webster, C., Zhang, G.: Stochastic finite element methods for partial

differential equations with random input data. Acta Numer. 23, 521–650 (2014)7. Gunzburger, M. D., Lee, H-C., Lee, J.: Error estimates of stochastic optimal Neumann

boundary control problems. SIAM J. Numer. Anal. 49 (4), 1532–1552 (2011)8. Jacod, J., Protter, P.: Probability essentials (Second Edition). Springer, Heidelberg (2003)9. Le Maître, O. P., Knio, O. M.: Spectral methods for uncertainty quantification. with appli-

cations to computational fluid dynamics. Springer, New York (2010)10. Lelievre, N., Beaurepaire, P., Mattrand, C., Gayton, N., Otsmane, A.: On the consideration of

uncertainty in design: optimization—reliability—robustness. Struct. Multidisc. Optim.(2016) https://doi.org/10.1007/s00158-016-1556-5

x Preface

[email protected]

Page 11: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

11. Lions, J. L.: Optimal control of systems governed by partial differential equations. Springer,Berlin (1971)

12. Lord, G. L., Powell, C. E., Shardlow, T.: An introduction to computational stochastic PDEs.Cambridge University Press, Cambridge (2014)

13. Lü, Q, Zuazua, E.: Averaged controllability for random evolution partial differential equa-tions. J. Math. Pures Appl. 105(3), 367–414 (2016)

14. Martínez-Frutos, J., Herrero-Pérez, D., Kessler, M., Periago, F.: Robust shape optimization ofcontinuous structures via the level set method. Comput. Methods Appl. Mech. Engrg. 305,271–291 (2016)

15. Smith, R. C. Uncertainty quantification. Theory, implementation, and applications.Computational Science & Engineering, Vol. 12. Society for Industrial and AppliedMathematics (SIAM), Philadelphia, PA (2014)

16. Tröltzsch, F.: Optimal control of partial differential equations: Theory, methods and appli-cations. Graduate Studies in Mathematics Vol. 112. AMS. Providence, Rhode Island (2010)

Preface xi

[email protected]

Page 12: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Contents

1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.2 Modelling Uncertainty in the Input Data. Illustrative

Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21.2.1 The Laplace-Poisson Equation . . . . . . . . . . . . . . . . . . . . . 31.2.2 The Heat Equation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41.2.3 The Bernoulli-Euler Beam Equation . . . . . . . . . . . . . . . . . 5

References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7

2 Mathematical Preliminaires . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92.1 Basic Definitions and Notations . . . . . . . . . . . . . . . . . . . . . . . . . . 92.2 Tensor Product of Hilbert Spaces . . . . . . . . . . . . . . . . . . . . . . . . 112.3 Numerical Approximation of Random Fields . . . . . . . . . . . . . . . . 15

2.3.1 Karhunen-Loève Expansion of a Random Field . . . . . . . . . 162.4 Notes and Related Software . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28

3 Mathematical Analysis of Optimal Control Problems UnderUncertainty . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 313.1 Variational Formulation of Random PDEs . . . . . . . . . . . . . . . . . . 31

3.1.1 The Laplace-Poisson Equation Revisited I . . . . . . . . . . . . . 323.1.2 The Heat Equation Revisited I . . . . . . . . . . . . . . . . . . . . . 343.1.3 The Bernoulli-Euler Beam Equation Revisited I . . . . . . . . 36

3.2 Existence of Optimal Controls Under Uncertainty . . . . . . . . . . . . . 383.2.1 Robust Optimal Control Problems . . . . . . . . . . . . . . . . . . 383.2.2 Risk Averse Optimal Control Problems . . . . . . . . . . . . . . . 40

3.3 Differences Between Robust and Risk-Averse Optimal Control . . . 413.4 Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43

xiii

[email protected]

Page 13: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4 Numerical Resolution of Robust Optimal Control Problems . . . . . . . 454.1 Finite-Dimensional Noise Assumption: From Random PDEs

to Deterministic PDEs with a Finite-Dimensional Parameter . . . . . 464.2 Gradient-Based Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

4.2.1 Computing Gradients of Functionals MeasuringRobustness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49

4.2.2 Numerical Approximation of Quantities of Interestin Robust Optimal Control Problems . . . . . . . . . . . . . . . . 52

4.2.3 Numerical Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . 614.3 Benefits and Drawbacks of the Cost Functionals . . . . . . . . . . . . . 704.4 One-Shot Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 704.5 Notes and Related Software . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77

5 Numerical Resolution of Risk Averse Optimal Control Problems . . . . . 795.1 An Adaptive, Gradient-Based, Minimization Algorithm . . . . . . . . 795.2 Computing Gradients of Functionals Measuring Risk Aversion . . . . . 825.3 Numerical Approximation of Quantities of Interest in Risk

Averse Optimal Control Problems . . . . . . . . . . . . . . . . . . . . . . . . 825.3.1 An Anisotropic, Non-intrusive, Stochastic Galerkin

Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 835.3.2 Adaptive Algorithm to Select the Level

of Approximation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 855.3.3 Choosing Monte Carlo Samples for Numerical

Integration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 855.4 Numerical Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 875.5 Notes and Related Software . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90

6 Structural Optimization Under Uncertainty . . . . . . . . . . . . . . . . . . . 916.1 Problem Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 916.2 Existence of Optimal Shapes . . . . . . . . . . . . . . . . . . . . . . . . . . . . 946.3 Numerical Approximation via the Level-Set Method . . . . . . . . . . . 98

6.3.1 Computing Gradients of Shape Functionals . . . . . . . . . . . . 986.3.2 Mise en Scène of the Level Set Method . . . . . . . . . . . . . . 101

6.4 Numerical Simulation Results . . . . . . . . . . . . . . . . . . . . . . . . . . . 1026.5 Notes and Related Software . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107

7 Miscellaneous Topics and Open Problems . . . . . . . . . . . . . . . . . . . . 1097.1 The Heat Equation Revisited II . . . . . . . . . . . . . . . . . . . . . . . . . . 1097.2 The Bernoulli-Euler Beam Equation Revisited II . . . . . . . . . . . . . 1147.3 Concluding Remarks and Some Open Problems . . . . . . . . . . . . . . 118References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 120

Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121

xiv Contents

[email protected]

Page 14: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

About the Authors

Jesús Martínez-Frutos obtained his Ph.D. from theTechnical University of Cartagena, Spain, in 2014 andis currently Assistant Professor in the Department ofStructures and Construction and a member of theComputational Mechanics and Scientific Computinggroup at the Technical University of Cartagena, Spain.

His research interests are in the fields of robustoptimal control, structural optimization under uncer-tainty, efficient methods for high-dimensional uncer-tainty propagation and high-performance computingusing GPUs, with special focus on industrial applica-tions. Further information is available at: http://www.upct.es/mc3/en/dr-jesus-martinez-frutos/.

Francisco Periago Esparza completed his Ph.D. atthe University of Valencia, Spain, in 1999. He iscurrently Associate Professor in the Department ofApplied Mathematics and Statistics and a memberof the Computational Mechanics and ScientificComputing group at the Technical University ofCartagena, Spain.

His main research interests include optimal control,optimal design and controllability. During his researchtrajectory, he has published a number of articlesaddressing not only theoretical advances of theseproblems but also their application to real-worldengineering problems. During recent years, his researchhas focussed on optimal control for random PDEs.Further information is available at: http://www.upct.es/mc3/en/dr-francisco-periago-esparza/.

xv

[email protected]

Page 15: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Acronyms and Initialisms

a.e. Almost everywherea.s. Almost sureANOVA Analysis of VarianceATD Anisotropic Total Degreee.g. for example (exempli gratia)etc. etceteraFEM Finite Element MethodHDMR High-Dimensional Model Representationi.e. that is (id est)i.i.d. independent and identically distributedKL Karhunen-LoèveMC Monte CarloPC Polynomial ChaosPDEs Partial Differential EquationsPDF Probability Density FunctionPOD Proper Orthogonal Decompositionr.h.s. right-hand sideRB Reduced BasisRPDEs Random Partial Differential EquationsSC Stochastic CollocationSG Stochastic GalerkinUQ Uncertainty Quantificationw.r.t. With respect to

xvii

[email protected]

Page 16: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Abstract

This book offers a direct and comprehensive introduction to the basic theoreticaland numerical concepts in the emergent field of optimal control of partial differ-ential equations (PDEs) under uncertainty. The main objective of the book is toprovide graduate students and researchers with a smooth transition from optimalcontrol of deterministic PDEs to optimal control of random PDEs. Coverageincludes uncertainty modelling in control problems, variational formulation ofPDEs with random inputs, robust and risk-averse formulations of optimal controlproblems, existence theory and numerical resolution methods. The exposition isfocussed on running the whole path starting from uncertainty modelling and endingin the practical implementation of numerical schemes for the numerical approxi-mation of the considered problems. To this end, a selected number of illustrativeexamples are analysed in detail along the book. Computer codes, written inMATLAB, for all these examples are provided. Finally, the methods presented inthe text are applied to the mathematical analysis and numerical resolution of the,very relevant in real-world applications, structural optimization problem.

Keywords Uncertainty quantification � Partial differential equations with randominputs � Stochastic expansion methods � Robust optimal controlRisk-averse optimization � Structural shape optimization under uncertainty

xix

[email protected]

Page 17: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Chapter 1Introduction

... in the small number of things which we are able to know withcertainty, even in the mathematical sciences themselves, theprincipal means for ascertaining—induction and analogy—arebased on probabilities; so that the entire system of humanknowledge is connected with the theory set forth in this essay..

Pierre Simon de Laplace.A Philosophical Essay on Probability, 1816.

1.1 Motivation

Mathematical models (and therefore numerical simulation results obtained fromthem) of physical, biological and economical systems always involve errors whichlead to discrepancies between simulation results and the results of real systems. Inparticular, simplifications of the mathematical model are often performed to facili-tate its resolution but these generate model errors, the numerical resolution methodintroduces numerical errors, and a limited knowledge of the system’s parameters,such as its geometry, initial and/or boundary conditions, external forces and mate-rial properties (diffusion coefficients, elasticity modulus, etc.), induces additionalerrors, called data errors. Some of the above errors may be reduced, e.g., by buildingmore accurate models, by improving numerical resolution methods, by additionalmeasurements, etc. This type of errors or uncertainties that can be reduced is usu-ally called epistemic or systematic uncertainty. However, there are other sources ofrandomness that are intrinsic to the system itself and hence cannot be reduced. It isthe so-called aleatoric or statistical uncertainty. A classical example of this type isUncertainty Principle of Quantum Mechanics.

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2018J. Martínez-Frutos and F. Periago Esparza, Optimal Control of PDEs under Uncertainty,SpringerBriefs in Mathematics, https://doi.org/10.1007/978-3-319-98210-6_1

1

[email protected]

Page 18: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

2 1 Introduction

Quantification ofsources of uncertainty

Random variables

Model of the system Uncertainty Propagation

PDE Statistical moments

Probability of failure

Output PDF

Fig. 1.1 Uncertainty Quantification in PDEs-based mathematical models

For the above reasons, it is clear that if one aims at obtainingmore reliable numeri-cal predictions frommathematical models, then these should account for uncertainty.Aleatoric uncertainty is naturally defined in a probabilistic framework. In principle,the use of probabilistic tools to model epistemic uncertainty is not a natural choice.However, there is a tendency in Uncertainty Quantification (UQ) to reformulate epis-temic uncertainty as aleatoric uncertainty where probabilistic analysis is applicable[2]. In addition, from the point of view of numerical methods that will be proposedin this text, there is no difference between the above two types of uncertainty. Hence,a probabilistic framework for modelling and analysis of uncertainties is adoptedthroughout the book.

The termUncertainty Quantification is broadly used to define a variety ofmethod-ologies. In [5, p. 1] it is defined as “the science of identifying, quantifying, andreducing uncertainties associated with models, numerical algorithms, experiments,and predicted outcomes or quantities of interest”. Following [4, p. 525], in this textthe termUncertaintyQuantification is used to describe “the task of determining infor-mation about the uncertainty in an output of interest that depends on the solution ofa PDE, given information about the uncertainties in the system inputs”. This idea isillustrated in Fig. 1.1.

Remark 1.1 When statistical information on uncertainties is not available, othertechniques, such as worst case scenario analysis or fuzzy logic, may be used to dealwith uncertainty. These tools are outside the scope of this text.

1.2 Modelling Uncertainty in the Input Data. IllustrativeExamples

Among the different sources of uncertainty described in the preceding section, thistext is focussed on optimal control problems whose state variable is a solution ofa Random Partial Differential Equation (RPDE). The term random means that thePDE’s input data (coefficients, forcing terms, boundary conditions, etc.) are uncertainand are described as random variables or random fields.

In the following, we present several illustrative examples that will be analysed indetail along the book.

[email protected]

Page 19: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

1.2 Modelling Uncertainty in the Input Data. Illustrative Examples 3

1.2.1 The Laplace-Poisson Equation

Let D ⊂ Rd , d = 1, 2 or 3 in applications, be a bounded domain and consider the

following control system for Laplace-Poisson’s equation

{−div (a∇ y) = 1O u, in Dy = 0, on ∂ D,

(1.1)

where y = y(x), with x = (x1, . . . , xd) ∈ D, is the state variable and u = u(x) isthe control function, which acts on the spatial region O ⊂ D. As usual, 1O standsfor the characteristic function of O , i.e.,

1O (x) ={1, x ∈ O0, otherwise.

Here and throughout the text, both the divergence (div) and the gradient (∇) operatorsinvolve only derivatives with respect to the spatial variable x ∈ D. Recall that

∇ y =(

∂y

∂x1, . . . ,

∂y

∂xd

)

and

div (a∇ y) =d∑

j=1

∂x j

(a

∂y

∂x j

).

Equation (1.1) is amodel for a variety of physical phenomena (deformation of elas-ticmembranes, steady-state heat conduction, electrostatics, steady-state groundwaterflow, etc.). For instance, in the case of heat conduction, y denotes the temperature ofa solid which occupies the spatial domain D, a is the material thermal conductivity,and u represents an internal heat source, which heats/cools the body (e.g., by elec-tromagnetic induction). For some nonhomogeneous materials, such as functionallygraded materials, large random spatial variabilities in the material properties havebeen observed [1].

Another example is the modelling of a steady-state, single-phase fluid flow. In thiscase y represents the hydraulic head, u is a source/sink term, and a is the saturatedhydraulic conductivity. In most aquifers, a is highly variable and never perfectlyknown (see [6, Chap. 1] where some experimental data for a are reported).

It is then natural to consider the coefficient a as a random function a = a (x,ω),where x ∈ D and ω denotes an elementary random event. As a consequence, thetemperature y, solution of (1.1), becomes a random space function y = y (x,ω).Hereafter, the probability distribution of a is assumed to be known and thus anunderlying abstract probability space (Ω,F ,P) is given. The sample space Ω isthe set of all possible outcomes (e.g., all possible measures of thermal conductivity),

[email protected]

Page 20: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4 1 Introduction

F is the σ—algebra of events (hence, subsets of Ω to which probabilities may beassigned), and P : F → [0, 1] is a probability measure.

A typical optimal control problem that arises in this context amounts to finding acontrol u = u (x) such that its associated temperature y = y (u) is the best approx-imation (in a least-square sense) to a desired stationary temperature yd (x) in D.Hence, the cost functional

1

2

∫D

|y (x,ω) − yd (x) |2 dx + γ

2

∫O

u2 (x) dx, (1.2)

with γ ≥ 0, is introduced. The second term in (1.2) is a measure of the energy costneeded to implement the control u. It is observed that (1.2) depends on the realizationω ∈ Ω and hence a control u that minimizes (1.2) also depends on ω. Of course,one is typically interested in a control u, independent of ω, which minimizes (insome sense) the distance between y (x,ω) and yd (x). At first glance, it is naturalto consider the averaged distance, with respect to ω, of the functional (1.2). Theproblem then is formulated as:

⎧⎪⎪⎪⎪⎨⎪⎪⎪⎪⎩

Minimize in u : J (u) = 12

∫Ω

∫D |y (x, ω) − yd (x) |2 dx dP (ω) + γ

2

∫O u2 (x) dx

subject to−div (a (x, ω)∇ y (x, ω)) = 1Ou (x) , (x, ω) ∈ D × Ω

y (x, ω) = 0, (x, ω) ∈ ∂ D × Ω

u ∈ L2 (O) .

(1.3)

Following the usual terminology in Control Theory, throughout this text yd =yd(x) is the so-called target function, y = y(x,ω) is the state variable and u = u(x)

the control. To avoid awful notations, we shall make explicit only the variablesof interest that different functions depend on. For instance, to make explicit thedependence of the state variable y on the control u we will write y = y (u).

1.2.2 The Heat Equation

Consider the following control system for the transient heat equation

⎧⎪⎪⎨⎪⎪⎩

yt − div (a∇ y) = 0, in (0, T ) × D × Ω

a∇ y · n = 0, on (0, T ) × ∂ D0 × Ω

a∇ y · n = α (u − y) , on (0, T ) × ∂ D1 × Ω,

y (0, x,ω) = y0(x,ω) in D × Ω,

(1.4)

where the boundary of the spatial domain D ⊂ Rd is decomposed into two disjoint

parts ∂ D = ∂ D0 ∪ ∂ D1, and n is the unit outward normal vector to ∂ D. In system(1.4), yt denotes the partial derivative of y w.r.t. t ∈ (0, T ), and the control functionu = u (t, x) is the outside temperature, which is applied on the boundary region ∂ D1.

[email protected]

Page 21: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

1.2 Modelling Uncertainty in the Input Data. Illustrative Examples 5

As reported in [1], in addition to randomness in the thermal conductivity coeffi-cient a = a (x,ω), the initial temperature y0 and the convective heat transfer coeffi-cientα are very difficult tomeasure in practice and hence both are affected of a certainamount of uncertainty, i.e., y0 = y0 (x,ω) and α = α (x,ω). In real applications α

may also depend on the time variable t ∈ [0, T ] but, for the sake of simplicity, hereit is assumed to be stationary.

As in the preceding example, the state variable y = y (t, x,ω) of system (1.4),which represents the temperature of the material point x ∈ D at time t ∈ [0, T ] forthe random event ω ∈ Ω , is a random function.

Assume that a desired temperature yd (x) is given and that it is aimed to choosethe control u whichmust be applied to ∂ D1 in order to be as closer as possible to yd attime T . Similarly to problem (1.3), one may consider the averaged distance betweeny (T, x,ω) and yd (x) as the cost functional to be minimized. Another possibility isto minimize the distance between the mean temperature of the body occupying theregion D and the target temperature yd , i.e.,

∫D

(∫Ω

y(T, x,ω) dP(ω) − yd(x)

)2

dx .

Since only the mean of y (T ) is considered, there is no control on the dispersionof y (T ). Consequently, if the dispersion of y (T ) is large, then minimizing theexpectation of y (T ) is useless because for a specific random event ω, the probabilityof y (T,ω) of being close to its average is small. It is then convenient to minimizenot only the expectation but also a measure of dispersion such as the variance. Thus,the optimal control problem reads as:

⎧⎪⎨⎪⎩Minimize in u : J (u) = ∫

D(∫

Ω y(T, x,ω) dP(ω) − yd (x))2 dx + γ

2

∫D Var (y(T, x)) dx

subject to

y = y (u) solves (1.4), u ∈ L2(0, T ; L2 (∂ D1)

),

(1.5)where γ ≥ 0 is a weighting parameter, and

Var (y(T, x)) =∫

Ω

y2(T, x,ω) dP(ω) −(∫

Ω

y(T, x,ω) dP(ω)

)2

is the variance of y (T, x, ·).

1.2.3 The Bernoulli-Euler Beam Equation

Under Bernoulli-Euler hypotheses, the small random vibrations of a thin, uniform,hinged beam of length L , driven by a piezoelectric actuator located along the randominterval (x0 (ω) , x1 (ω)) ⊂ D := (0, L) are described by the system:

[email protected]

Page 22: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

6 1 Introduction

⎧⎪⎪⎨⎪⎪⎩

ytt + (ayxx )xx = v[δx1(ω) − δx0(ω)

]x , in (0, T ) × D × Ω

y(t, 0,ω) = yxx (t, 0,ω) = 0, on (0, T ) × Ω

y(t, L ,ω) = yxx (t, L ,ω) = 0, on (0, T ) × Ω

y(0, x,ω) = y0(x,ω), yt (0, x,ω) = y1(x,ω), in D × Ω,

(1.6)

where the beam flexural stiffness a = E I is assumed to depend on both x ∈ D andω ∈ Ω . i.e. a = a(x,ω). As usual E denotes Young’s modulus and I is the areamoment of inertia of the beam’s cross-section. Randomness in a is associated to anumber of random phenomena, e.g. variabilities of conditions during the manufac-turing process of the material’s beam (see [3] and the references therein).

In system (1.6), x0(ω) and x1(ω) stand for the end points of the actuator. Thedependence of these two points on a random event ω indicates that there is someuncertainty in the location of the actuator. δx j = δx j (x) is the Diracmass at the pointsx j ∈ D, j = 0, 1.

The function v : (0, T ) × Ω → R represents the time variation of a voltagewhichis applied to the actuator, andwhich is assumed to be affected by some randompertur-bation. Since physical controller devices are affected of uncertainty, it is realistic todecompose the control variable into an unknown deterministic and a known stochas-tic components. Moreover, it is reasonable to assume that the stochastic part bemodulated by the deterministic one. Thus, the function v = v(t,ω), which appearsin (1.6), takes the form

v(t,ω) = u(t)(1 + u(ω)

), (1.7)

where u : (0, T ) → R is the (unknown) deterministic control and u is a (known)zero-mean random variable which accounts for uncertainty in the controller device.

The random output y(t, x,ω) represents vertical displacement at time t of theparticle on the centreline occupying position x in the equilibrium position y = 0. Inthe above equations, yt and ytt denotefirst and secondderivativesw.r.t. t ∈ (0, T ), andthe subscripts “x” and “xx” are first and second derivatives w.r.t. x ∈ D, respectively.See Fig. 1.2 for the problem configuration.

Let H = L2(D) be the space of (Lebesgue) measurable functions v : D → R

whose norm

Fig. 1.2 Problemconfiguration for theBernoulli-Euler beam

y(t, x, ω) v(t, ω)

L

u (t)

x0 (ω) x1 (ω)

[email protected]

Page 23: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

1.2 Modelling Uncertainty in the Input Data. Illustrative Examples 7

‖v‖L2(D) :=(∫

Dv2(x) dx

)1/2

is finite. Consider the Sobolev space

V = H 2(D) ∩ H 10 (D) := {

v : D → R : v, v′, v′′ ∈ L2 (D) , v(0) = v(L) = 0},

endowed with the norm

‖v‖V = ‖v‖L2(D) + ‖v′‖L2(D) + ‖v′′‖L2(D).

The topological dual space of V is denoted by V �. For each random event ω ∈ Ω

and each control u ∈ L2 (0, T ), a measure of the amplitude and velocity of the beamvibrations at a control time T > 0 is expressed by

J (u,ω) = 1

2

(‖y (T,ω) ‖2H + ‖yt (T,ω) ‖2V �

). (1.8)

In this problem, one may be interested in computing a control u for rare occur-rences of the random inputs (which therefore have a very little impact on themean andvariance of J ), but which could have catastrophic consequences. For that purpose,the following risk averse model may be considered:

Jε (u) = P {ω ∈ Ω : J (u,ω) > ε} , (1.9)

where ε > 0 is a prescribed threshold parameter. The corresponding control problemis then formulated as:

⎧⎨⎩Minimize in u : Jε (u) = P {ω ∈ Ω : J (u,ω) > ε}subject to

y = y (u) solves (1.6), u ∈ L2 (0, T ) .

(1.10)

In the remaining chapters, we will consistently use the three above examples toillustrate the main theoretical and numerical methods in optimal control of randomPDEs.

References

1. Chiba, R.: Stochastic analysis of heat conduction and thermal stresses in solids: a review. In:Kazi, S.N. (ed.) Chapter 9 in Heat Transfer Phenomena and Applications. InTech (2012)

2. de Rocquigny, E., Devictor, N., Tarantola, S.: Uncertainty in Industrial Practice: A Guide toQuantitative Uncertainty Management. John Wiley & Sons, Ltd. (2008)

[email protected]

Page 24: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

8 1 Introduction

3. Ghanem, R.G., Spanos, P.D.: Stochastic Finite Elements: A Spectral Approach. Springer-Verlag, New York (1991)

4. Gunzburger, M.D., Webster, C., Zhang, G.: Stochastic finite element methods for partial dif-ferential equations with random input data. Acta Numer. 23, 521–650 (2014)

5. Smith, R.C.: Uncertainty quantification. Theory, implementation, and applications. Comput.Sci. Eng., 12. Society for Industrial andAppliedMathematics (SIAM), Philadelphia, PA (2014)

6. Zhang,D.: StochasticMethods for Flow inPorousMedia:CopingwithUncertainties.AcademicPress (2002)

[email protected]

Page 25: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Chapter 2Mathematical Preliminaires

Everything should be made as simple as possible, but notsimpler.

Albert Einstein.

This chapter briefly collects the main theoretical and numerical tools which areneeded to the analysis and numerical resolution of control problems under uncer-tainty.Westartwith a fewbasic definitions andgeneral notations.Then, somematerialon tensor product of Hilbert spaces is reviewed. Finally, the numerical approximationof random fields by using Karhunen-Loève expansions is studied.

2.1 Basic Definitions and Notations

From now on in this text, D ⊂ Rd , d = 1, 2, 3 in applications, is a bounded and

Lipschitz domain with boundary ∂D. We denote by Lp (D), 1 ≤ p ≤ ∞, the space ofall (equivalence classes of) Lebesgue measurable functions g : D → R whose norm

‖g‖Lp(D) ={(∫

D |g(x)|p dx)1/p , p < ∞ess supx∈D |g(x)|, p = ∞

is finite. To make explicit the Lebesgue measure and hence avoid confusion, for1 ≤ p < ∞, Lp (D) will be denoted by Lp (D; dx) in some places in the next section.

Similarly, for 1 ≤ p < ∞,

Lp (∂D) ={g : ∂D → R measurable :

∫∂D

|g(x)|p ds < ∞}

,

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2018J. Martínez-Frutos and F. Periago Esparza, Optimal Control of PDEs under Uncertainty,SpringerBriefs in Mathematics, https://doi.org/10.1007/978-3-319-98210-6_2

9

[email protected]

Page 26: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

10 2 Mathematical Preliminaires

where ds represents the surface Lebesgue measure on ∂D.The following Sobolev spaces will be of special importance in the remaining

of this book. H 1 (D) denotes the space of all functions y ∈ L2 (D) having first-orderpartial derivatives (in the sense of distributions) ∂xj y in L

2 (D). This space is endowedwith the inner product

< y, z >H 1(D)=∫Dyz dx +

∫D

∇y · ∇z dx,

which induces the norm

‖y‖H 1(D)

(∫D

(y2 + |∇y|2) dx

)1/2

.

As is well-known, H 1(D) is a separable Hilbert space. The space

H 10 (D) = {

y ∈ H 1(D) : y |∂D= 0},

where y |∂D denotes the trace of y, is also considered. In H 10 (D),

‖y‖H 10 (D) =

(∫D

|∇y|2 dx)1/2

defines a norm, which is equivalent to the norm in H 1 (D).(Ω,F , P) denotes a complete probability space. By complete it is meant thatF

contains all P-null sets, i.e., if B1 ⊂ Ω is such that there exists B2 ∈ F satisfyingB1 ⊂ B2 and P (B2) = 0, then B1 ∈ F . Notice that this concept of completenessmeans that the measure P is complete [10, p. 31].

For 1 ≤ p ≤ ∞, LpP(Ω) denotes the space composed of all (equivalence classes

of) measurable functions g : Ω → R whose norm

‖g‖LpP(Ω) =

{(∫Ω

|g (ω) |p dP (ω))1/p

, p < ∞ess supω∈Ω |g (ω) |, p = ∞

is finite. For convenience, LpP(Ω) will also be denoted by Lp (Ω; dP(ω)) in the next

section.If (X , ‖ · ‖X ) is a Banach space and 1 ≤ p ≤ ∞, Lp

P(Ω;X ) denotes the Lebesgue-

Bochner space composed of all (equivalence classes of) strongly measurable func-tions g : Ω → X whose norm

[email protected]

Page 27: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

2.1 Basic Definitions and Notations 11

‖g‖LpP(Ω;X ) =

{(∫Ω

‖g (·, ω) ‖pX dP (ω))1/p

, p < ∞ess supω∈Ω‖g (·, ω) ‖X , p = ∞

is finite. Since (Ω,F , P) is complete, LpP(Ω;X ) is complete (in the sense that every

Cauchy sequence in LpP(Ω;X ) is a convergent sequence; see [10, pp. 107 and 175]).

Given the probability space (Ω,F , P), the probability measure P induces a dis-tance d inF as follows: given E,F ∈ F ,

d (E,F) = P (E�F) ,

where E�F = (E F) ∪ (F E) is the symmetric difference between E and F . Theprobability space (Ω,F , P) is said to be separable if its associated metric space(F , d) is separable.1 It is proved that if (Ω,F , P) is separable, then Lp

P(Ω;X ) is

also separable [10, p. 177].Of particular interest in this book is the case p = 2. Precisely, let (H ,< ·, · >H )

be a separable Hilbert space. Analogously to the scalar case, it may be proved thatthe vector-valued space L2

P(Ω;H ), with the inner product

< f , g >L2P(Ω;H )=

∫Ω

< f (ω), g(ω) >H dP (ω)

is also a separable Hilbert space. The induced norm is given by

‖f ‖L2P(Ω;H ) =

(∫Ω

‖f (ω)‖2H dP (ω)

)1/2

.

2.2 Tensor Product of Hilbert Spaces

In this section, we review on some definitions and results on tensor product of Hilbertspaces, which shall be very useful in the study of random PDEs. We follow theapproach in [17, Chapter 2] and refer the reader to that reference for further details.All the Hilbert spaces considered in this section are assumed to be real and separable.

Definition 2.1 Let(H1,< ·, · >H1

),(H2,< ·, · >H2

)be two Hilbert spaces. Given

h1 ∈ H1 and h2 ∈ H2, the pure tensor h1 ⊗ h2 is defined as the bilinear form

h1 ⊗ h2 : H1 × H2 → R

(ϕ1, ϕ2) �→ (h1 ⊗ h2) (ϕ1, ϕ2) =< h1, ϕ1 >H1< h2, ϕ2 >H2 .(2.1)

LetH := span {h1 ⊗ h2, h1 ∈ H1, h2 ∈ H2} be the set of all finite linear combina-tions of such pure tensors. An inner product < ·, · >H is defined onH by definingit on pure tensors as

1We recall that a metric space (F , d) is separable if it has a countable dense subset.

[email protected]

Page 28: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

12 2 Mathematical Preliminaires

∀ϕ1 ⊗ ϕ2, ψ1 ⊗ ψ2, < ϕ1 ⊗ ϕ2, ψ1 ⊗ ψ2 >H =< ϕ1, ψ1 >H1< ϕ2, ψ2 >H2 ,

and then by extending it by linearity toH .It is proved that < ·, · >H is well defined in the sense that < ϕ,ψ >H , with

ϕ,ψ ∈ H , does not depend on which finite combinations are used to express ϕ andψ .

Definition 2.2 Let(H1,< ·, · >H1

),(H2,< ·, · >H2

)be two Hilbert spaces. The

tensor product of H1 and H2, denoted by H1 ⊗ H2, is the Hilbert space which isdefined as the completion of H with respect to the inner product < ·, · >H .

Proposition 2.1 [17, p. 50] If {ϕi}i∈N and{ψj

}j∈N are orthonormal bases of H1 and

H2, respectively, then{ϕi ⊗ ψj

}i,j∈N is an orthonormal basis of H1 ⊗ H2.

Remark 2.1 To simplify notation, both the spacesH1 andH2 in the preceding propo-sition have been taken infinite dimensional. The other cases, i.e. both spaces are finitedimensional or one of them is finite dimensional, are obviously similar.

In this book we are mainly concerned with random fields, i.e., with functionsf (x, ω) : D × Ω → R which depend on a spatial variable x ∈ D and on a randomevent ω ∈ Ω . Let us assume that f (x, ω) ∈ L2 (D × Ω; dx × dP (ω)), where dx isthe Lebesgue measure on D and P is the probability measure on Ω . As is usual,dx × dP (ω) denotes the associated product measure. Since f (x, ω) intrinsically hasa different structure with respect to x andω, thinking on the numerical approximationof such an f , the use of tensor product spaces is very convenient. Indeed, let {ϕi(x)}i∈Nand

{ψj(ω)

}j∈N be orthonormal bases of L2 (D; dx) and L2 (Ω; dP(ω)), respectively.

By Fubini’s theorem, it is not hard to show that{ϕi(x)ψj(ω)

}i,j∈N is an orthonormal

basis of L2 (D × Ω; dx × dP (ω)). Now, consider the mapping

U : L2 (D; dx) ⊗ L2 (Ω; dP(ω)) → L2 (D × Ω; dx × dP (ω)) ,

which is defined on the basis{ϕi(x) ⊗ ψj(ω)

}i,j∈N by

U(ϕi(x) ⊗ ψj(ω)

) = ϕi(x)ψj(ω)

and can be uniquely extended to a unitarymapping fromL2 (D; dx) ⊗ L2 (Ω; dP(ω))

onto L2 (D × Ω; dx × dP (ω)). Hence, if ϕ(x) = ∑∞i=1 ciϕi(x) ∈ L2 (D; dx) and

ψ (ω) = ∑∞j=1 djψj(ω) ∈ L2 (Ω; dP(ω)), then one has

U (ϕ ⊗ ψ) = U

⎛⎝∑

i,j

cidjϕi(x) ⊗ ψj(ω)

⎞⎠ =

∑i,j

cidjϕi(x)ψj(ω) = ϕ(x)ψ(ω).

In this sense, it is said that U defines a natural isomorphism between the spacesL2 (D; dx) ⊗ L2 (Ω; dP(ω)) and L2 (D × Ω; dx × dP (ω)). We write

[email protected]

Page 29: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

2.2 Tensor Product of Hilbert Spaces 13

L2 (D; dx) ⊗ L2 (Ω; dP(ω)) ∼= L2 (D × Ω; dx × dP (ω)) . (2.2)

Another interesting use of tensor product spaces is that they enable a nicerepresentation of vector-valued functions. Indeed, let (H ,< ·, · >H ) be a Hilbertspace with basis {ϕi}i∈N and consider the Bochner space L2

P(Ω;H ). Every function

f ∈ L2P(Ω;H ) may be written as

f (ω) = limN→∞

N∑k=1

fk(ω)ϕk , fk(ω) =< f (ω), ϕk >H and ω ∈ Ω.

The mapping

U :N∑

k=1

fk(ω) ⊗ ϕk �→N∑

k=1

fk(ω)ϕk

is well-defined from a dense subset of L2 (Ω, dP(ω)) ⊗ H onto a dense subset ofL2P(Ω;H ) and it preserves the norms. Hence, U may be uniquely extended to a

unitary mapping from L2 (Ω, dP(ω)) ⊗ H onto L2P(Ω;H ). As a consequence, if

f = f (ω) ∈ L2 (Ω, dP(ω)) andϕ ∈ H , then U (f (ω) ⊗ ϕ) = f (ω)ϕ. Because of thisproperty, it is said that U is a natural isomorphism between L2 (Ω, dP(ω)) ⊗ H andL2P(Ω;H ) and we may write

L2 (Ω, dP(ω)) ⊗ H ∼= L2P(Ω;H ) (2.3)

Remark 2.2 In the particular case in which H = L2 (D; dx), by (2.2) and (2.3) wehave the isomorphisms

L2P

(Ω;L2 (D)

) ∼= L2 (D; dx) ⊗ L2 (Ω; dP(ω)) ∼= L2 (D × Ω; dx × dP (ω)) .

As it will be shown in the next chapter, another very interesting case in which theisomorphism (2.3) applies is when H is a Sobolev space.

All the above discussion is summarized in the following abstract result:

Theorem 2.1 [17, Theorem II.10] Let (M1, μ1) and (M2, μ2) be measure spacesso that L2 (M1, dμ1) and L2 (M2, dμ2) are separable. Then:

(a) There is a unique isomorphism from the tensor space L2 (M1, dμ1) ⊗ L2

(M2, dμ2) to L2 (M1 × M2, dμ1 × dμ2) so that ϕ ⊗ ψ �→ ϕψ .(b) If (H ,< ·, · >H ) is a separableHilbert space, then there is a unique isomorphism

from L2 (M1, dμ1) ⊗ H to L2 (M1, dμ1;H ) so that f (x) ⊗ ϕ �→ f (x)ϕ.

Three typical situations to which the abstract results of this section will be appliedhereafter are the following:

• Let V be a Sobolev space, e.g. V = H 10 (D), and consider the space L2

P(Ω). By

part (b) of Theorem 2.1 one has the isomorphism

[email protected]

Page 30: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

14 2 Mathematical Preliminaires

L2P(Ω; V ) ∼= L2

P(Ω) ⊗ V .

• In the context of random PDEs, let us assume that inputs of the PDE dependon a finite number of uncorrelated real-valued random variables ξn : Ω → R,1 ≤ n ≤ N . Let us denote by Γn = ξn (Ω), by Γ = ∏N

n=1 Γn, and by ρ : Γ → R+the joint probability density function (PDF) of the multivariate random variableξ = (ξ1, . . . , ξN ), which is assumed to factorize as

ρ (z) =N∏n=1

ρn (zn) , z = (z1, . . . , zn) ∈ Γ,

with ρn : Γn → R+. Instead of working in the abstract space L2P (Ω) onemayworkin the image space

L2ρ (Γ ) ={f : Γ → R :

∫Γ

f 2 (z) ρ (z) dz < ∞}

,

which is equipped with the inner product

< f , g >L2ρ(Γ )=∫

Γ

f (z)g(z)ρ(z) dz.

We refer the reader to Sect. 4.1 in Chap. 4 for more details on this passage.Let

{ψpn (zn)

}∞pn=0, 1 ≤ n ≤ N , be an orthonormal basis of L2ρn

(Γn) composed of

a suitable class of orthonormal polynomials. Since L2ρ (Γ ) = ⊗Nn=1 L

2ρn

(Γn), amultivariate orthonormal polynomial basis of L2ρ (Γ ) is constructed as

{ψp (z) =

N∏n=1

ψpn (zn) , z = (z1, . . . , zN )

}

p=(p1,...,pN )∈NN

.

• Let y = y (z) ∈ L2ρ(Γ ;H 1

0 (D)). As it will be illustrated in Chaps. 4 and 5,

typically y is the solution of a given elliptic random PDE and one is inter-ested in its numerical approximation. As is usual, the vector valued functionz �→ y(·, z) ∈ H 1

0 (D) is identifiedwith the real-valued functionD × Γ � (x, z) �→y(x, z). Having in mind that L2ρ

(Γ ;H 1

0 (D)) ∼= L2ρ (Γ ) ⊗ H 1

0 (D), to approxi-mate numerically y(x, z), a standard finite element space Vh (D) ⊂ H 1

0 (D) andthe space Pp (Γ ) ⊂ L2ρ (Γ ) composed of polynomials of degree less than orequal to p, are considered. Hence, we look for a numerical approximationof y ∈ L2ρ (Γ ) ⊗ H 1

0 (D) in the finite dimensional space Pp (Γ ) ⊗ Vh (D). Let{ψm(z), 0 ≤ m ≤ Mp

}be a basis ofPp (Γ ) composed of orthonormal polynomi-

als in L2ρ (Γ ), and let {φn(x), 1 ≤ n ≤ Nh} be a basis of shape functions in Vh (D).Then,

{ψm ⊗ φn, 0 ≤ m ≤ Mp, 1 ≤ n ≤ Nh

}is a basis of Pp (Γ ) ⊗ Vh (D),

[email protected]

Page 31: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

2.2 Tensor Product of Hilbert Spaces 15

which is identified with{ψm(z)φn(x), 0 ≤ m ≤ Mp, 1 ≤ n ≤ Nh

}. Thus, an

approximation yhp (x, z) ∈ Pp (Γ ) ⊗ Vh (D) of y(x, z) is expressed in the form

yhp (x, z) =Nh∑n=1

p∑m=0

yn,mφn(x)ψm(z).

2.3 Numerical Approximation of Random Fields

As the examples in the preceding chapter show, uncertainty in the random inputsof a PDE are typically represented by functions a : D × Ω → R, which depend ona spatial variable x ∈ D and on a random event ω ∈ Ω . Such functions are calledrandom fields. Associated to a random field a (x, ω) it is interesting to consider thefollowing quantities:

• Mean (or expected) function a : D → R, which is defined by

a (x) = E [a (x, ·)] =∫

Ω

a (x, ω) dP (ω) . (2.4)

• Covariance function: Cova : D × D → R, which is defined by

Cova(x, x′) = E

[(a (x, ·) − a (x))

(a(x′, ·) − a

(x′))] (2.5)

and expresses the spatial correlation of the random field, i.e., Cova(x, x′) is the

covariance of the random variables ω �→ a (x, ω) and ω �→ a(x′, ω

).

• Variance function: Vara (x) = Cova (x, x).

The random field a is said to be second order if Vara (x) < ∞ for all x ∈ D. Aparticular type of random fields, which are widely used to model uncertainty insystems, is the class of Gaussian fields. A random field a (x, ω) is said to be Gaussianif for any M ∈ N and x1, x2, . . . , xM ∈ D the R

M -valued random variable X (ω) =(a(x1, ω

), a

(x2, ω

), . . . , a

(xM , ω

))follows the multivariate Gaussian distribution

N (μ,C), where

μ = (a(x1), . . . , a

(xM

))and Cij = Cova

(xi, xj

), 1 ≤ i, j ≤ M .

For computational purposes, it is very convenient to express a randomfield a(x, ω)

by means of a finite number of random variables {ξn (ω)}1≤n≤N , with N ∈ N+ ={1, 2, 3, . . .}. Among the different possibilities, Karhunen-Loève expansion is a veryusual choice to represent a random field.

[email protected]

Page 32: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

16 2 Mathematical Preliminaires

2.3.1 Karhunen-Loève Expansion of a Random Field

As it will be illustrated along this text, it is very appealing to have a representationof a random field in the form

a (x, ω) = a (x) +∞∑n=1

bn (x) ξn (ω) , (2.6)

with pairwise uncorrelated2 random variables{ξn (ω)

}∞n=1

having, for convenience,

zero mean and σ 2n variances, and L2 (D)-orthonormal functions {bn (x)}∞n=1.

To derive how the spatial functions bn (x) and the random variables ξn (ω) looklike, we continue our discussion in this section proceeding in a formal way. If (2.6)holds, then the covariance function (2.5) admits (at least formally) the representation

Cova(x, x′) = ∫

Ω

∞∑i=1

bi (x) ξi (ω)

∞∑n=1

bn(x′) ξn (ω) dP (ω)

=∞∑i=1

∞∑n=1

bi (x) bn(x′) ∫

Ω

ξi (ω) ξn (ω) dP (ω)

=∞∑n=1

σ 2n bn (x) bn

(x′) .

By multiplying this expression by bj(x′) and integrating in D, a direct (still formal)

computation leads to

∫DCova

(x, x′) bj (x′) dx′ = σ 2

j bj (x) ,

which means that{σ 2n , bn (x)

}∞n=1 are the eigenpairs associated to the operator

ψ �→∫DCova

(x, x′)ψ

(x′) dx′, ψ ∈ L2 (D) .

Denoting by ξn = 1σn

ξn, 1 ≤ n ≤ ∞, the random variables ξn have zero mean andunit variance. Then, (2.6) rewrites as

a (x, ω) = a (x) +∞∑n=1

σnbn (x) ξn (ω) . (2.7)

2We recall that a family of random variables{ξn (ω)

}∞n=1

having zero mean is said to be pairwise

uncorrelated if∫Ω

ξi (ω) ξj (ω) dP (ω) = 0 for i �= j.

[email protected]

Page 33: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

2.3 Numerical Approximation of Random Fields 17

Hence,

∫D[a (x, ω) − a (x)] bn (x) dx =

∞∑i=1

σiξi (ω)

∫Dbi (x) bn (x) dx = σnξn (ω) ,

that is,

ξn (ω) = 1

σn

∫D[a (x, ω) − a (x)] bn (x) dx.

These formal computations can be made rigorous, as the following results state.Before that, let us recall that a function C : D × D → R is symmetric if C

(x, x′) =

C(x′, x

)for all x, x′ ∈ D and C is non-negative definite if for any xj ∈ D and αj ∈ R,

1 ≤ j ≤ N , the inequalityN∑

j,k=1

αjαkC(xj, xk

) ≥ 0

holds.

Theorem 2.2 (Mercer) [14, Theorem 1.8] Let D ⊂ Rd be a bounded domain and

let C : D × D → R be a continuous, symmetric and non-negative definite function.Consider the integral operator C : L2 (D) → L2 (D) defined by

(Cψ) (x) =∫DC(x, x′)ψ

(x′) dx′. (2.8)

Then, there exist a sequence of, continuous in D, eigenfunctions {bn}∞n=1 of C , whichcan be chosen orthonormal in L2 (D), such that the corresponding sequence of eigen-values {λn}∞n=1 is positive. Moreover,

C(x, x′) =

∞∑n=1

λnbn (x) bn(x′) , x, x′ ∈ D, (2.9)

where the series converges absolutely and uniformly on D × D.

Theorem 2.3 (Karhunen-Loève expansion) [14, Theorems 5.28, 5.29, 7.52, 7.53]Let D ⊂ R

d be a bounded domain and let a = a (x, ω) be a random field with contin-uous covariance3 function Cova : D × D → R. Then, a (x, ω) admits the Karhunen-Loève expansion

3We recall that not any function may be taken as the covariance function of a random field. To bea valid covariance function it must be, in particular, non-negative definite. See [14, Theorem 5.18]for more details on this issue. All covariance functions considered in this text are assumed to becontinuous, symmetric and non-negative definite.

[email protected]

Page 34: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

18 2 Mathematical Preliminaires

a (x, ω) = a (x) +∞∑n=1

√λnbn (x) ξn (ω) , (2.10)

where the sum converges in L2P

(Ω;L2 (D)

). {λn, bn (x)}∞n=1 are the eigenpairs of the

operator C defined in Mercer’s theorem, with C(x, x′) = Cova

(x, x′), and

ξn (ω) = 1√λn

∫D[a (x, ω) − a (x)] bn (x) dx (2.11)

are pairwise uncorrelated random variables with zero mean and unit variance. If therandomfielda(x, ω) isGaussian, then ξn ∼ N (0, 1)are independent and identicallydistributed (i.i.d.) standard Gaussian variables.

Remark 2.3 Note that, thanks to Theorem 2.3, a Gaussian field a(x, ω) is completelydetermined by its expected function a (x) and its covariance function. Indeed, fromCova

(x, x′) one computes the eigenpairs {λn, bn(x)}∞n=1 of the operator (2.8). Thus,

if a (x) is known, then the KL expansion (2.10) determines the Gaussian field a(x, ω)

since the random variables that appear in (2.10) are i.i.d. standardGaussian variables.As a consequence, if the mean and covariance functions of the Gaussian field a(x, ω)

are properly estimated from experimental data, then a is completely characterized(see also Remark 2.4 for more details on this issue).

For numerical simulation purposes, the expansion (2.10) is always truncated at somelarge enough term N , i.e., the random field a (x, ω) is approximated as

a (x, ω) ≈ aN (x, ω) = a (x) +N∑n=1

√λnbn (x) ξn (ω) . (2.12)

The convergence rate of aN (x, ω) depends on the decay of the eigenvalues λn, whichin turn depends on the smoothness of it associated covariance function. Precisely, forpiecewise Sobolev regular covariance functions, the eigenvalue decay is algebraic.For piecewise analytic covariance functions, the rate decay is quasi-exponential (see[19] and the references therein).

A measure of the relative error committed by truncation of a KL expansion isgiven by the so-called variance error [14, 18]

EN (x) = Vara−aN (x)

Vara (x)= C (x, x) − ∑N

n=1 λnb2n (x)

C (x, x)= 1 −

∑Nn=1 λnb2n (x)

C (x, x), (2.13)

where, to alleviate the notation, from now on C = Cova is the covariance functionof a. Second equality in (2.13) is easily obtained from Theorems 2.2 and 2.3. See[14, Corollary 7.54] for more details. A global measure of this relative error, whichis called mean error variance, is

[email protected]

Page 35: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

2.3 Numerical Approximation of Random Fields 19

EN =∫D Vara−aN (x) dx∫

D Vara (x) dx=

∫D C (x, x) dx − ∑N

n=1 λn∫D C (x, x) dx

= 1 −∑N

n=1 λn∫D C (x, x) dx

.

(2.14)For some specific covariance functions and simple geometries of the physical

domain D, it is possible to find analytical expressions of KL expansions, as thefollowing example shows.

Example 2.1 (Exponential Covariance) Consider a Gaussian field a = a (x, ω) withzero mean, unit variance and having the following exponential covariance function:

C(x, x′) = e−|x−x′|/Lc , x, x′ ∈ D = (0,L) , (2.15)

where Lc is the so-called correlation length. The eigenpairs {λn, bn(x)}∞n=1 associatedto C

(x, x′) are obtained by solving the spectral problem

∫ x

0e− x−x′

Lc b(x′) dx′ +

∫ L

xe− x′−x

Lc b(x′) dx′ = λb (x) . (2.16)

By differentiating twice and by evaluating b′ (x) at the end points x = 0,L, thespectral problem (2.16) takes the form

{b′′ (x) + 1

Lc

(2λ

− 1Lc

)b (x) = 0, 0 < x < L

b′ (0) = 1Lcb (0) , b′ (L) = − 1

Lcb (L) .

(2.17)

Denoting by � = L/Lc, the eigenvalues of (2.17) are given by

λn = 2�2Lc�2 + γ 2

n

, (2.18)

where γn, n ∈ N, are the positive solutions of the transcendental equation

cot (γn) = γ 2n − �2

2�γn. (2.19)

Its associated normalized eigenfunctions are

bn (x) =√

2

�2 + γ 2n + 2�

[γn cos

(γnx

L

)+ � sin

(γnx

L

)], 0 < x < L. (2.20)

The energy of a is given by

‖a‖2L2(D)⊗L2P(Ω)

=∫ L

0Vara (x) dx =

∫ L

0C (x, x) dx = L. (2.21)

[email protected]

Page 36: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

20 2 Mathematical Preliminaires

Consider theN -term truncationaN (x, ω) = ∑Nn=1

√λnbn (x) ξn (ω),where the eigen-

values are ordered in decreasing order, i.e., λ1 ≥ λ2 ≥ · · · ≥ λN . Taking into accountthat {bn (x)}∞n=1 are orthonormal in L2 (D) and that ξn ∼ N (0, 1) are independent,a direct computation shows that

‖aN‖2L2(D)⊗L2P(Ω)

=N∑n=1

λn. (2.22)

Hence, if one aims to capture a certain amount, e.g. 90%, of the energy of the fielda, then the truncation term N must be chosen as to satisfy the condition EN ≤ 0.1,with EN given by (2.14), which, in this example, corresponds to

N∑n=1

λn ≥ 0.9L.

Figure 2.1a shows the first 100 eigenvalues obtained using (2.18), with L = 1 anddifferent values of Lc. The transcendental equation (2.19) is solved by the classicalNewton method. Figure 2.1b displays the first 5 eigenfunctions for Lc = 0.5.

The rate decay of eigenvalues determines how fast the variance error and themean error variance are reduced. This fact is observed in Fig. 2.2, where these errorsare depicted as a function of N and for several values of the correlation length Lc.One can observe how the number of terms needed to reduce the relative error to adesired threshold increases as Lc → 0. This issue is of paramount importance froma computational point of view since the number of terms of the KL expansion andthus the number of simulations required for an accurate representation of the field

(a) (b)

Fig. 2.1 Example 2.1: (a) Eigenvalues λn, 1 ≤ n ≤ 100 for the correlation lengths Lc = 0.05(dash red line), Lc = 0.5 (dot-dash blue line) and Lc = 5 (solid green line). (b) Eigenfunctions for1 ≤ n ≤ 5 and Lc = 0.5

[email protected]

Page 37: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

2.3 Numerical Approximation of Random Fields 21

(a) (b)

Fig. 2.2 Example 2.1: (a) Mean error variance EN for 1 ≤ N ≤ 100 and Lc = 0.05 (dash line),Lc = 0.5 (dot-dash line) and Lc = 5 (solid line), and (b) variance error EN (x) for Lc = 0.5 andN = 9, 21, 42

Fig. 2.3 Example 2.1: (a) Covariance of a(x, ω), and (b) Covariance of a42(x, ω)

would be large in case the spectrum decays slowly. Notice also that the largest valuesof EN (x) in Fig. 2.2b are located at the end points of the interval. This is due to thefact that since the eigenvalues are ordered in decreasing order, the dominant termin EN (x) is the first one, whose associated eigenfunction b1(x) takes its minimumvalues at x = 0, 1 (see Fig. 2.1b).

In the same vein, Fig. 2.3 displays the covariance functions of the random fielda(x, ω) and of its truncation at the 42-th term.

Three realizations of aN (x, ω) are shown in Fig. 2.4.Observe that, for small valuesof the spatial correlation length, the random field tends to be rough. When Lc → 0the random field can be viewed as a infinite set of uncorrelated random variables.

[email protected]

Page 38: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

22 2 Mathematical Preliminaires

Fig. 2.4 Example 2.1:realizations of aN (x, ω) forLc = 0.05 (dash red line),Lc = 0.5 (dot-dash blue line)and Lc = 5 (solid greenline). In each case, thenumber of terms N has beenfixed to capture the 95% ofthe energy field

0 0.2 0.4 0.6 0.8 1-3

-2

-1

0

1

2

Conversely, the random field becomes smoother as the correlation length increases.In the limit, when Lc → ∞, the random field is uniform.

For more complicated geometries and/or covariance functions, the eigenpairs{λn, bn (x)} need to be approximated numerically because in these cases it is notpossible to have explicit solutions of the spectral problem: find λ ∈ R and b ∈ L2 (D),b �= 0, such that ∫

DC(x, x′) b (x′) dx′ = λb (x) , x ∈ D. (2.23)

Several numericalmethods have been proposed in the literature [3]. Since for the classof problems considered in this book, a mesh of the physical domain D is assumed tobe readily available, herewe describe a discetized version of (2.23) based on the finiteelement method (FEM). Let Vh (D) ⊂ L2 (D) be a finite element space associatedto a mesh of size h of D and let {ϕ1, ϕ2, . . . , ϕM } be a basis of shape functions ofVh (D). Then, we look for approximated eigenvalues λn and eigenfunctions bn (x) =∑M

k=1 bknϕk (x) by solving the generalized matrix eigenvalue problem

M∑k=1

bkn

∫D

∫DC(x, x′)ϕk

(x′)ϕi (x) dx

′dx = λn

M∑k=1

bkn

∫D

ϕk (x) ϕi (x) dx, 1 ≤ i ≤ M ,

(2.24)which, in matrix form, is written as

Cbn = λnBbn, (2.25)

with bn = [b1n, . . . , b

Mn

]T,

[email protected]

Page 39: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

2.3 Numerical Approximation of Random Fields 23

Fig. 2.5 Example 2.2.Relative error of eigenvaluesfor 0 ≤ n ≤ 42 and differentsizes of the finite elementmesh. Lc = 0.5 in (2.15)

[C]ij =∫D

∫DC(x, x′)ϕi

(x′)ϕj (x) dx

′dx, and [B]ij =∫D

ϕi (x) ϕj (x) dx.

For error estimates of this numerical method we refer to [4, 19]. The variance errorassociated to the finite element discretization is given by (2.13), with λn replaced byλn, and bn (x) by bn (x). The same applies for (2.14). Let us now revisit Example 2.1.

Example 2.2 (Continuation of Example 2.1) For the covariance function (2.15),the generalized matrix eigenvalue problem (2.25) is solved by using Lagrange P1finite elements on a uniform mesh of size h = 1/(M − 1). Figure 2.5 shows therelative error, w.r.t. the eigenvalues λn computed in Example 2.1, committed forM = 101, 201, 501.

Remark 2.4 (Log-normal random fields)When random fields or random variablesare used to model input data of a mathematical model which are, in nature, positive,Gaussian fields may not be the best choice because negative values have positiveprobability. In addition, it is observed [13] that experimental data for most of theseparameters are log-normal distributed, which means that the logarithm (in practice,log10 although in theory the natural logarithmic log is used) of the original data arenormally (or Gaussian) distributed. The reason for this behaviour is that many phys-ical laws involve products of physical magnitudes. Thus, the additive hypothesis ofsmall errors, which state that if random variation is the sum of many small randomeffects, then a normal distribution is the result, does not apply. The multiplicativeversion of the Central Limit Theorem (“Geometric means of (non-log-normal) vari-ables are approximately log-normally distributed”) must be used instead. We referthe reader to [1, 13] for more details on this passage. This supports the multiplica-tive hypothesis of small errors: if random variation is the product of several randomeffects, then a log-normal distribution must be the result. A log-normal random fieldis expressed in the form

[email protected]

Page 40: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

24 2 Mathematical Preliminaires

a (x, ω) = eμ(x)+σ(x)z(x,ω), (2.26)

where z(x, ω) is a standard Gaussian random field with zero mean and unit variance,the so-called geometric mean μ∗

a (x) = eμ(x) is a scale parameter, and the geometricstandard deviation σ ∗

a (x) = eσ(x) is a shape parameter. The mean a (x) and varianceVara (x) functions of the field a are given by

a (x) = eμ(x)+ 12 σ 2(x), Vara (x) = e2μ(x)+σ 2(x)

(eσ 2(x) − 1

).

Hence, if the mean a (x) and variance Vara (x) functions of the field a are known,then μ (x) and σ (x) are obtained through the transformation

μ (x) = log( a2 (x)√

a2 (x) + Vara (x)

), σ 2 (x) = log

(1 + Vara (x)

a2 (x)

). (2.27)

In practice, μ∗a (x) = eμ(x) and σ ∗

a = eσ(x) are more useful than the usual arithmeticmean and arithmetic standard deviation of a (x, ω). For instance, for a log-normal

random variable, the interval[μ∗a/

(σ ∗a

)2, μ∗

a

(σ ∗a

)2]covers a probability of 95.5%.

Due to its relevance in the problems addressed along the book, the discretizationprocess of a two-dimensional (2D), in space, log-normal random field with squaredexponential covariance is presented in the following example.

Example 2.3 (2D Log-normal random field) Let z(x, ω) be a Gaussian random fieldwith zeromean, unit variance and the following isotropic squared exponential covari-ance function:

C(x, x′) = exp

[−

2∑i=1

(xi − x′i)2

L2i

], x = (x1, x2), x

′ = (x′1, x

′2) ∈ D = (0, 1)2 ,

(2.28)

where L1 and L2 are the correlation lengths in the two spatial directions. Eigenvaluesand eigenfunctions associated to C may be numerically approximated by using theFEM described above. However, since the covariance function (2.28) is expressedas

C(x, x′) =2∏

j=1

Cj

(xj, x

′j

),

with Cj

(xj, x′

j

)= exp

[−

(xj − x′

j

)2/L2j

], xj, x′

j ∈ (0, 1), j = 1, 2, its eigenvalues

(and eigenfunctions) are expressed as the product of the eigenvalues (eigenfunctions)

associated to the one-dimensional (1D) covariance functionsCj

(xj, x′

j

), j = 1, 2. See

[14, Example 7.56] for a similar situation. From a truncated KL expansion of theGaussian field z (x, ω) one obtains an approximation of a (x, ω) in the form

[email protected]

Page 41: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

2.3 Numerical Approximation of Random Fields 25

a (x, ω) ≈ aN (x, ω) = eμ(x)+σ(x)∑N

n=1

√λnbn(x)ξn(ω). (2.29)

Even if the random field (2.29) takes positive values, it is not completely satis-factory from a mathematical point of view because since the random variablesξn (ω) are Gaussian, aN (x, ω) may approximate to zero and to ∞. Hence, whenaN (x, ω) appears in the principal part of an elliptic differential operator, the uni-form (w.r.t. ω) ellipticity condition, which is required in most of PDEs, does nothold (see next chapter). A natural way of overcoming this technical difficulty is bytruncating the Gaussians. Indeed, instead of using the probability density functionφ (s) = 1√

2πe−s2/2, s ∈ R, of the Gaussian variable ξn (ω), one considers, for some

large enough d > 0, the probability density function

φ (s) ={

φ(s)�(d)−�(−d)

, −d ≤ s ≤ d0, otherwise,

(2.30)

where �(s) = ∫ s−∞ φ (t) dt, s ∈ R, is the cumulative density function of the normal

distribution. For instance, since ξn have unit variance, for d = 2 a probability of0.9545 is covered.

The Gaussian random field z(x, ω) is discretized using a Karhunen-Loève expan-sion with 15 terms. Figure 2.6 shows the first four eigenfunctions. This random fielddiscretization is able to capture the 99.7% of the total energy of the field providinga good approximation of the variance of the field, which is depicted in Fig. 2.7.

Figure 2.8 shows some realizations of zN (x, ω) and aN (x, ω), with N = 15,for Gaussian and log-normal isotropic (L1 = L2 = 0.5) and anisotropic (L1 = 0.05,L2 = 0.5) random fields. The log-normal random field is assumed to have mean andvariance functions both equal to 1. Hence, from (2.27) it is obtained μ = −0.3466and σ = 0.8326. Figure 2.8c, d show how the directional preference of the anisotropyis evident. The realizations appears streched from left to right due to the anisotropyof the covariance function.

2.4 Notes and Related Software

In Sect. 2.1 we have introduced some basic definitions and properties on FunctionalAnalysis and Probability Theory. The reader is referred to [15] for details concerningSobolev spaces and to [7, 10, 16] for definitions, properties and proofs, concerningprobability spaces.

Apart fromKL expansion, several methods of discretization of randomfields havebeen proposed in the literature. These methods can be broadly classified into threemain groups [18]. The first group consists of methods that discretize the randomfield by a finite set of random variables averaged over local domains, such as theweighted integral method [6] and the local averagingmethod [20]. The second groupis based on point discretization methods in which the random variables represent

[email protected]

Page 42: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

26 2 Mathematical Preliminaires

(a) λ1= 0.42523, b1(x1,x2) (b) λ2= 0.17475, b2(x1,x2)

(c) λ3= 0.17475, b3(x1,x2) (d) λ 4= 0.071813, b4(x1,x2)

Fig. 2.6 Example 2.3. First four eigenfunctions of z(x, ω)

the values of the random field at some points. The midpoint method [11] or theoptimal linear estimation method (OLE) [12] are some of the methods includedin the category. The third group is composed of series expansion methods whichapproximate the random field by a finite sum of products of deterministic spatialfunctions and random variables. The polynomial chaos (PC) expansion [21] and theabove mentioned Karhunen-Loève expansion [14] belong to this category.

The KL expansion has received much attention in the literature because it isoptimal in the sense that it minimizes the mean square truncation error [9, Chap. 2].

Methods for numerically approximate eigenvalues and eigenfunctions associatedto covariance functions may be broadly classified into three categories [3]: Nys-tröm methods [2], Degenerate kernel methods and Projection methods. An in-depth

[email protected]

Page 43: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

2.4 Notes and Related Software 27

(a) Variance of z(x,ω ) (b) Variance of z15(x,ω )

Fig. 2.7 Example 2.3: (a) Variance of of z(x, ω), and (b) variance of z15(x, ω)

discussion of all these methods lies beyond the scope of this book. The reader isreferred to [3, 18] for a comprehensive overview.

The exponential and squared exponential covariance functions introduced in thischapter arewidely used in applications (see, e.g. [5,Chap. 4] and [8]). The exponentialcovariance function generates a random field which is only mean-square continuous.The random field constructed from the squared exponential covariance function ismean-square differentiable. We refer to [14, Sect. 5.5] for more details on regularityissues of random fields.

To illustrate the implementation issues of the examples presented along thechapter, several MATLAB scripts are provided accompanying this book. For theinterested reader, below are some references to open source code for UQ using ran-dom fields:

• FERUM - Finite Element Reliability Using MATLAB.http://www.ce.berkeley.edu/projects/ferum/.

• OpenTURNS - Free Software for uncertainty quantification, uncertainty propaga-tion, sensitivity analysis and meta-modelling (written in C++/Python).http://www.openturns.org/.

• UQlab - Computational tools that provide efficient, flexible and easy to use pro-grams for UQ on a variety of applications (written in MATLAB).http://www.uqlab.com/.

• Uribe, Felipe (2015). MATLAB software for stochastic finite element analysis.http://www.mathworks.com/matlabcentral/profile/authors/2912338-felipe-uribe.

[email protected]

Page 44: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

28 2 Mathematical Preliminaires

Fig. 2.8 Example 2.3: left panel shows realizations of z15(x, ω)with correlation lengths [L1,L2] =[0.5, 0.5] (a) and [L1,L2] = [0.05, 0.5] (c). Right panel shows realizations of a15(x, ω) with cor-relation lengths [L1,L2] = [0.5, 0.5] (b) and [L1,L2] = [0.05, 0.5] (d)

References

1. Aitchison, J., Brown, J. A. C.: The Lognormal Distribution, with Special Reference to its Usesin Economics. Cambridge University Press (1957)

2. Atkinson, K.: The Numerical Solution of Integral Equations of the Second Kind. CambridgeUniversity Press (1997)

3. Betz, W., Papaioannou, I., Straub, D.: Numerical methods for the discretization of randomfields by means of the Karhunen-Loève expansion. Comput. Methods Appl. Mech. Engrg. 271,109–129 (2014)

4. Chatelin, F.: Spectral Approximation of Linear Operators. Academic Press (1983)5. Cressie, N., Wikle, C.K.: Statistics for Spatio-Temporal Data. Wiley Series in Probability and

Statistics. John Wiley & Sons Inc., Hoboken, NJ (2011)6. Deodatis, G., Shinozuka,M.:Weighted integral method. II: Response variability and reliability.

J. Eng. Mech. 117(8), 1865–1877 (1991)

[email protected]

Page 45: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

References 29

7. Durrett, R.: Probability: Theory and Examples. 4 edn. Cambridge Series in Statistical andProbabilistic Mathematics, vol. 31. Cambridge University Press, Cambridge (2010)

8. Gneiting, T., Guttorp, P.: Continuous Parameter Stochastic Process Theory. Handbook of Spa-tial Statistics, 17–28, Chapman & Hall/CRC Handb. Mod. Stat. Methods, CRC Press, BocaRaton, FL (2010)

9. Ghanem, R.G., Spanos, P.D.: Stochastic Finite Elements: A Spectral Approach. Springer, NewYork (1991)

10. Halmos, P.: Measure Theory. Springer, New York (1950)11. Kiureghian, A.D., Ke, J.-B.: The stochastic finite element method in structural reliability.

Probab. Eng. Mech. 3(2), 83–91 (1988)12. Li, C.-C., Kiureghian, A.D.: Optimal discretization of random fields. J. Eng. Mech. ASCE.

119(6), 1136–1154 (1993)13. Limpert, E., Stahel, W. A., Abbt, M.: The log-normal distribution across sciences: keys and

clues. BioScience 51(5) (2001)14. Lord, G.L., Powell, C.E., Shardlow, T.: An Introduction to Computational Stochastic PDEs.

Cambridge University Press (2014)15. Necas, J.: Les Méthodes Directes en Théorie des Equations Elliptiques. Éditeurs Academia,

Prague (1967)16. Rao, M.M., Swift, R.J.: Probability Theory with Applications. Vol. 582. 2 edn. Springer (2006)17. Reed, M., Simon, B.: Methods of Modern Mathematical Physics, vol. 1. Academic Press,

Functional Analysis (1980)18. Sudret, B., Kiureghian, A. D.: Stochastic Finite Element Methods and Reliability: A State-

of-the-Art Report. Technical Report UCB/SEMM-2000/08, Department of Civil and Environ-mental Engineering, University of California, Berkeley (2000)

19. Todor, R.A.: Robust eigenvalue computation for smoothing operators. SIAM J. Numer. Anal.44(2), 865–878 (2006)

20. Vanmarcke, E., Grigoriu, M.: Stochastic finite element analysis of simple beams. J. Eng. Mech.109(5), 1203–1214 (1983)

21. Wiener, N.: The homogeneous chaos. Amer. J. Math. 60(4), 897–936 (1938)

[email protected]

Page 46: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Chapter 3Mathematical Analysis of OptimalControl Problems Under Uncertainty

... the methods of mathematical analysis have allowed torigorously identify the missing elements for the models to becomplete, facilitating the reliable use of computational methods.

Jacques Louis Lions.Translated from Spanish. Bolletín of SEMA 15, 2000.

This chapter is focussed on existence theory for the solutions of robust and riskaverse optimal control problems. As a first step, the classical variational theory ofdeterministic PDEs is extended to the case of random PDEs. This variational the-ory may be developed by using either the formalism of tensor product of Hilbertspaces or abstract functions, i.e., functions with values in Banach or Hilbert spaces.However, tensor products of Hilbert spaces have the advantage that the numericalapproximation of such random PDEs becomes very natural in such a formalism.

3.1 Variational Formulation of Random PDEs

Well-posedness of Random Partial Differential Equations (RPDEs) may be studiedby working only on the spatial domain D and thus, by considering a deterministicPDE for each realization of the random event ω ∈ Ω . Another different approachconsists of working in D × Ω . Hence, solutions of RPDEs are looked for in a Hilbertspace such as L2

P(Ω; V ), with V a suitable Sobolev space of functions defined in

D. The latter approach has the advantage that solutions of random PDEs are foundin L2

P(Ω; V ) so that their mean and variance are well-defined. For this reason, a

formulation in D × Ω is presented next.Since the goal of this section is to illustrate how to deal with the new ingredient of

working with random input parameters, our study is limited to the three toy models

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2018J. Martínez-Frutos and F. Periago Esparza, Optimal Control of PDEs under Uncertainty,SpringerBriefs in Mathematics, https://doi.org/10.1007/978-3-319-98210-6_3

31

[email protected]

Page 47: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

32 3 Mathematical Analysis of Optimal Control Problems Under Uncertainty

presented in Chap. 1, which, however, cover the usual cases of elliptic equations,and first and second order in time PDEs. Other linear RPDEs may be studied in asimilar way.

3.1.1 The Laplace-Poisson Equation Revisited I

Consider the elliptic problem

{−div (a(x, ω)∇ y(x, ω)) = f (x, ω) in D × Ω

y(x, ω) = 0 on ∂D × Ω(3.1)

The following hypotheses on the input data are assumed to hold:

(A1) a = a(x, ω) ∈ L∞P

(Ω; L∞(D)) and there exist amin , amax > 0 such that

0 < amin ≤ a(x, ω) ≤ amax < ∞ a. e. x ∈ D and a. s. ω ∈ Ω,

(A2) f ∈ L2P

(Ω; L2(D)

).

Definition 3.1 A random field y ∈ L2P

(Ω; H 1

0 (D))is said to be a weak solution to

(3.1) if it satisfies the so-called variational formulation

∫Ω

∫Da∇ y · ∇v dxdP(ω) =

∫Ω

∫Df v dxdP(ω) ∀v ∈ L2

P

(Ω; H 1

0 (D)). (3.2)

By considering the bilinear form A : L2P

(Ω; H 1

0 (D)) × L2

P

(Ω; H 1

0 (D)) → R,

A (y, v) =∫

Ω

∫Da∇ y · ∇v dxdP(ω), (3.3)

and the linear form F : L2P

(Ω; H 1

0 (D)) → R defined by

F (v) =∫

Ω

∫Df v dxdP(ω),

the variational formulation (3.2) is rewritten in the abstract form

A (y, v) = F (v) ∀v ∈ L2P

(Ω; H 1

0 (D)).

Thanks to assumption (A1), the bilinear form A (·, ·) is continuous and elliptic. By(A2), F (·) is continuous. As a straightforward application of Lax-Milgram’s lemmaone has:

Theorem 3.1 Under assumptions (A1) and (A2) above, the variational problem(3.2) has a unique solution y = y (x, ω) which satisfies the estimate

[email protected]

Page 48: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

3.1 Variational Formulation of Random PDEs 33

‖y‖L2P(Ω;H 1

0 (D)) ≤ CP

amin‖ f ‖L2

P(Ω;L2(D)), (3.4)

where CP = CP (D) is the Poincaré constant, i.e., CP is the optimal constant in theinequality ‖v‖L2(D) ≤ CP‖∇v‖L2(D) for all v ∈ H 1

0 (D).

Inequality (3.4) is eaisly obtained from asummption (A1) and the Cauchy-Schwarzinequality. Indeed,

amin∫Ω

∫D |∇ y|2 dxdP (ω) ≤ ∫

Ω

∫D a|∇ y|2 dxdP (ω)

= ∫Ω

∫D f ydxdP (ω)

≤ ‖ f ‖L2P(Ω;L2(D))‖y‖L2

P(Ω;L2(D))≤ CP‖ f ‖L2

P(Ω;L2(D))‖y‖L2P(Ω;H 1

0 (D)).

Remark 3.1 Note that due to the isomorphism L2P

(Ω; H 1

0 (D)) ∼= L2

P(Ω) ⊗ H 1

0 (D),a completely analogous analysis of (3.1) may be carried out in the tensor productspace L2

P(Ω) ⊗ H 1

0 (D) with the force term f ∈ L2P(Ω) ⊗ L2(D).

Before proceeding, the following remark on assumptions (A1) and (A2) is in order.

Remark 3.2 As it has been already commented in Sect. 2.3.1, in most applications,Gaussian fields are the model of choice to represent uncertain parameters whichshow a spatial correlation. Moreover, for numerical simulation purposes, Gaussianfields are typically approximated by truncated KL expansions of the form

f (x, ω) ≈N∑

n=1

√λnbn (x) ξn (ω) .

This choice, which is appropriate for forcing terms and initial conditions, is nolonger suitable for random fields which are positive in nature, as the coefficienta (x, ω) which appears in the principal part of elliptic differential operators. In fact,experimental data (e.g., reported in [15] for the case of steady-state, single-phasegroundwater flows) reveal that a (x, ω) is often log-normal distributed. Hence, asdescribed in Example 2.3, a (x, ω) is typically approximated as

a (x, ω) ≈ eμ(x)+σ(x)∑N

n=1

√λnbn(x)ξn(ω),

where ξn are truncated standard Gaussian variables.As a conclusion, assumptions (A1) and (A2) are suitable both from amathematical

point of view and for applications.Nevertheless, assumptions (A1) and (A2) may be relaxed and still guarantee the

well-posedness of (3.2). See [1, Lemma 1.2] .

[email protected]

Page 49: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

34 3 Mathematical Analysis of Optimal Control Problems Under Uncertainty

3.1.2 The Heat Equation Revisited I

Let us now study the well-posedness of the parabolic system,

⎧⎪⎪⎨⎪⎪⎩

yt − div (a∇ y) = 0, in (0, T ) × D × Ω

a∇ y · n = 0, on (0, T ) × ∂D0 × Ω

a∇ y · n = α (u − y) , on (0, T ) × ∂D1 × Ω,

y (0) = y0 in D × Ω,

(3.5)

which was introduced in Sect. 1.2.2. We recall that in problem (3.5), D ⊂ Rd is

decomposed into two disjoint parts ∂D = ∂D0 ∪ ∂D1, and n is the unit outwardnormal vector to ∂D. The random coefficient a = a(x, ω) satisfies hypothesis (A1)above. In addition, the following conditions are assumed to hold:

(B1) α = α(x, ω) ∈ L2P

(Ω; L2(∂D1)

), with α(x, ω) ≥ 0. There exists a positive

constant αmax < ∞ such that α(x, ω) ≤ αmax a. e. x ∈ ∂D1 and a. s. ω ∈ Ω ,and ∫

Ω

∫∂D1

α(x, ω) dsdP(ω) > 0, (3.6)

where ds is the Lebesgue measure on ∂D1.(B2) u = u(t, x, ω) ∈ L2

(0, T ; L2

P(Ω; L2(∂D1))

)(B3) y0 = y0(x, ω) ∈ L2

P(Ω; L2(D)).

Let V = H 1(D) and consider theHilbert space VP = L2P(Ω; V ) equippedwith norm

‖v‖VP=

(∫Ω

∫D

(v2 + |∇v|2) dx dP(ω)

)1/2

.

The dual space of VP is denoted by V P. Finally, consider the space

WP(0, T ) = {y ∈ L2 (0, T ; VP) : yt ∈ L2

(0, T ; V

P

)}

endowed with the norm

‖y‖WP(0,T ) =(∫ T

0

(‖y(t)‖2VP

+ ‖yt (t)‖2V P

)dt

)1/2

.

Definition 3.2 A random field y = y(t, x, ω) ∈ WP (0, T ) is a solution to (3.5) if itsatisfies ∫ T

0 (yt , v)V P,VP

dt + ∫ T0

∫Ω

∫D a∇ y · ∇v dxdP(ω)dt

+ ∫ T0

∫Ω

∫∂D1

αyv dsdP(ω)dt

= ∫ T0

∫Ω

∫∂D1

αuv dsdP(ω)dt

[email protected]

Page 50: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

3.1 Variational Formulation of Random PDEs 35

for all v ∈ L2 (0, T ; VP). Here, (·, ·)V P,VP

is the duality product1 between VP and V P.

The initial condition y(0, x, ω) = y0(x, ω) must be satisfied in L2P

(Ω; L2 (D)

).

Notice that the space WP(0, T ) is continuously embedded in C([0, T ] ; L2

P(Ω;

L2(D)))

so that the above initial condition makes sense.

Theorem 3.2 Let (Ω,F ,P) be a complete and separable probability space. Letus assume that (A1), (B1)-(B3) hold. Then, there exists a unique solution of (3.5).Moreover, there exists c > 0 such that

‖y‖WP(0,T ) ≤ c(‖y0‖L2

P(Ω;L2(D)) + ‖u‖L2(0,T ;L2

P(Ω;L2(∂D1)))

). (3.7)

Proof The proof follows the same lines as in the deterministic case so that here weonly focus on the novelty brought about by the probabilistic framework.

We shall apply the abstract result [14, Theorem 26.1]. Both existence and unique-ness are proved as in the deterministic case. Indeed, consider the Hilbert spaceHP = L2

P

(Ω; L2 (D)

)and the Gelfand triple VP ↪→ HP ↪→ V

P. Existence of solu-

tions follows from the Galerkin method. The proof is carried out in several steps:

(i) Since VP is isomorphic to L2P(Ω) ⊗ V , given the orthonormal bases {φi }i≥1 and{

ψ j}j≥1 of L

2P(Ω) andV , respectively, the set

{φi ⊗ ψ j

}i, j≥1 is anorthonormal

basis of L2P(Ω) ⊗ V . Thus, by considering the above orthonormal basis, which

for simplicity is denoted by {ζi }i≥1, the solution y is approximated by a sequence

ym (t, x, ω) =m∑i=1

gi (t) ζi (x, ω) .

(ii) It is proved that ‖ym‖L2(0,T ;VP) ≤ K , with K constant. As a consequence, thereexists z = z (t, x, ω) ∈ L2 (0, T ; VP) such that, up to a subsequence, ym weaklyconverges to z. At this point, the crucial step is to prove that the bilinear formA : VP × VP → R defined as

A (y, v) =∫

Ω

∫Da(x, ω)∇ y · ∇v dxdP(ω) +

∫Ω

∫∂D1

αyv dsdP(ω) (3.8)

is continuous and VP-elliptic, and that the linear form

F : VP → R

v �→ F(v) = ∫Ω

∫∂D1

αuv dsdP(ω)(3.9)

is continuous for a. e. t ∈ (0, T ). The continuity of A(·, ·) is an immediate con-sequence of the Cauchy-Schwarz inequality, the uniform upper bound imposed

1We recall that for t ∈ (0, T ), yt (t) : VP → R is a continuous linear form and v(t) ∈ VP. Thus,(yt (t) , v (t))V

P,VP := yt (t) (v(t)) .

[email protected]

Page 51: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

36 3 Mathematical Analysis of Optimal Control Problems Under Uncertainty

on α in Assumption (B1), and the continuity of the trace operator. VP-ellipticityof A(·, ·) is a consequence of (3.6) and may be proved by using an adaptedversion of the Poincaré inequality, which may be proved by following the samelines as in the deterministic case (see [13, p. 36]). Finally, the continuity of Ffollows from assumptions (B1)-(B2).

(iii) Finally, it is proved that z = y, solution to (3.5).

The rest of the proof is as in the deterministic case (see the proof of [14, Theorem26.1]). �Remark 3.3 Due to the tensor product structure of Bochner spaces, WP (0, T ) maybe identified with the spaces

WP (0, T ) ∼= L2P(Ω) ⊗ [(

L2 (0, T ) ⊗ V) ∩ (

H 1 (0, T ) ⊗ V )]

∼= L2P

(Ω; L2 (0, T ; V ) ∩ H 1 (0, T ; V )

).

Thus, the well-posedness of (3.5) may be studied in these tensor product spaces.

3.1.3 The Bernoulli-Euler Beam Equation Revisited I

To illustrate how the transposition method2 for proving the well-posedness of sec-ond order in time, deterministic, PDEs, extends to random PDEs, we consider thefollowing system for the Bernoulli-Euler beam equation, which was introduced inSect. 1.2.3:

⎧⎪⎪⎨⎪⎪⎩

ytt + (ayxx )xx = v[δx1(ω) − δx0(ω)

]x, in (0, T ) × D × Ω

y(t, 0, ω) = yxx (t, 0, ω) = 0, on (0, T ) × Ω

y(t, L , ω) = yxx (t, L , ω) = 0, on (0, T ) × Ω

y(0, x, ω) = y0(x, ω), yt (0, x, ω) = y1(x, ω), in D × Ω,

(3.10)

with D = (0, L).Consider again the Sobolev space V = H 2(D) ∩ H 1

0 (D) and the Hilbert spaceH = L2(D). Consider also the spaces VP = L2

P(Ω; V ) and HP = L2

P(Ω; H). The

topological dual space of VP is denoted by V P.

The random coefficient a = a(x, ω) is assumed to satisfy assumption (A1), asgiven above. In addition, the following hypotheses on the input data of system (3.10)are assumed to hold:

(C1) v ∈ L2P

(Ω; L2(0, T )

),

(C2) 0 < x0(ω) < x1(ω) < L for all random events ω ∈ Ω , and(C3)

(y0, y1

) ∈ HP × V P.

2The concept of transposition lets give a meaning to solutions of PDEs with non-regular inputdata (initial, boundary conditions and/or forcing terms). To illustrate the main ideas underlying thismethod, a very simple deterministic example is presented in Sect. 3.4.

[email protected]

Page 52: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

3.1 Variational Formulation of Random PDEs 37

Following the same lines as in the case of deterministic PDEs (see [8, Chap. 3]),the concept of a solution for system (3.10) is defined by transposition as follows:

• In a first step, for a given f ∈ L2 (0, T ; HP), consider the problem of find-ing a function ζ = ζ(t, x, ω) with ζ ∈ L2 (0, T ; VP), ζ ′ ∈ L2 (0, T ; HP), ζ ′′ ∈L2

(0, T ; V

P

)such that

(ζ ′′, ϕ

)V P,VP

+∫

Ω

∫ L

0a (x, ω) ζxxϕxx dxdP(ω) =

∫Ω

∫ L

0f ϕ dxdP(ω) ∀ϕ ∈ VP, (3.11)

and a.e. t ∈ (0, T ), where (·, ·)V P,VP

is the duality product between VP and V P. In

addition, the final conditions ζ(T ) = ζ ′(T ) = 0 must be satisfied.Thanks to hypothesis (A1), the bilinear form A : VP × VP → R, with

A (ϕ, ψ) =∫

Ω

∫ L

0a(x, ω)ϕxxψxx dxP(ω) (3.12)

is continuous and VP-elliptic. Assuming that (Ω,F ,P) is separable and followingthe same lines as in the proof of Theorem 3.2, the Galerkin method may be appliedto conclude that there exists a unique solution to (3.11).

• In a second step, the space

XP = {ζ = ζ f solution of (51) with f ∈ L2 (0, T ; HP)

},

is considered. XP, endowed with the norm ‖ζ‖XP= ‖ f ‖L2(0,T ;HP), is a Hilbert

space. In addition, the operator

ζ �→ ζ ′′ + (aζxx )xx

is an isomorphism from XP onto L2 (0, T ; HP).Next, consider the linear and continuous form L : XP → R defined by

L (ζ ) =∫ T

0

(v

[δx1 − δx0

]x , ζ

)V P,VP

dt +(y1, ζ(0)

)VP,V

P

−∫

Ω

∫ L

0y0ζ ′(0) dxdP(ω).

• By transposition [8, Theorem 9.3], there exists a unique y ∈ L2 (0, T ; HP) suchthat ∫ T

0

(y, ζ ′′ + (aζxx )xx

)HP

dt = L (ζ ) for all ζ ∈ XP. (3.13)

Definition 3.3 The function y = y (t, x, ω), solution to (3.13) is called a solutionto (3.10) in the sense of transposition.

Theorem 3.3 Let (Ω,F ,P) be a complete and separable probability space. Let usassume that (A1) and (C1)-(C3) hold. Then, there exists a unique solution y to (3.10)in the sense of transposition. Moreover, y has the regularity

y ∈ C ([0, T ] ; HP) ∩ C1 ([0, T ] ; V

P

)

[email protected]

Page 53: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

38 3 Mathematical Analysis of Optimal Control Problems Under Uncertainty

and there exists c > 0 such that

‖y‖L∞(0,T ;HP) + ‖y′‖L∞(0,T ;V P) ≤ c

(‖y0‖HP

+ ‖y1‖V P

+ ‖v [δx1 − δx0

]x ‖L2(0,T ;V

P)

).

(3.14)

Proof The proof follows exactly the same lines as in [8, Theorem 9.3] so that it isomitted here. �

3.2 Existence of Optimal Controls Under Uncertainty

In this section, we are concerned with existence of optimal controls for the classesof robust and risk averse control problems introduced in Chap.1. To illustrate howexistence theory of deterministic optimal control problems extends to the randomcase, robust and risk averse control problems for the following Poisson’s system arestudied next.

With the same notations as in Sect. 3.1.1, let us consider the system

{−div (a∇ y) = 1Ou, in D × Ω

y = 0, on ∂D × Ω,(3.15)

where y = y (x, ω) is the state variable and u = u (x) is the control variable, whichacts on the spatial region O ⊂ D, a (Lebesgue) measurable subset. As usual, 1Ostands for the characteristic function of O.

It is assumed that the random coefficient a = a (x, ω) satisfies assumption (A1)and that the control u ∈ Uad , where

Uad = {u ∈ L2 (D) : m ≤ u (x) ≤ M a.e. x ∈ D

}, (3.16)

with m, M ∈ [−∞,+∞], is the set of admissible controls.

3.2.1 Robust Optimal Control Problems

For a desired target yd ∈ L2 (D), the two following cost functionals are considered:

J1 (u) = 1

2

∫Ω

∫D

|y (x, ω) − yd (x) |2 dx dP (ω) + γ

2

∫Ou2 (x) dx, (3.17)

where γ ≥ 0 is a weighting parameter, and

J2 (u) = 1

2

∫D

(∫Ω

y(x, ω) dP(ω) − yd (x)

)2

dx + β

2

∫DVar (y(x)) dx + γ

2

∫Ou2 (x) dx,

(3.18)

[email protected]

Page 54: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

3.2 Existence of Optimal Controls Under Uncertainty 39

where

Var (y(x)) =∫

Ω

y2(x, ω) dP(ω) −(∫

Ω

y(x, ω) dP(ω)

)2

is the variance of y (x, . . .), and β, γ ≥ 0 are weighting parameters.

Remark 3.4 Note that tanks to Theorem 3.1, both J1 (u) and J2 (u) are well-definedfor each u ∈ Uad .

Eventually, consider the two following robust optimal control problems:

(P1)

⎧⎨⎩Minimize in u ∈ Uad : J1 (u)

subject toy = y (u) solves (3.15),

and

(P2)

⎧⎨⎩Minimize in u ∈ Uad : J2 (u)

subject toy = y (u) solves (3.15).

Following the same ideas as in deterministic quadratic optimal control problems [13,Theorem 2.14], the following existence results are established.

Theorem 3.4 Assume that β = 0. The following assertions hold:

(i) If −∞ < m < M < +∞, then problems (P1) and (P2) have, at least, one solu-tion. If, in addition, γ > 0, then the solution is unique.

(ii) If m = −∞ and/or M = +∞, and if γ > 0, then problems (P1) and (P2) have,respectively, a unique solution.

Proof (i) In the constrained case (−∞ < m < M < +∞), the set of admissiblecontrols Uad is bounded, closed and convex. Moreover, by Theorem 3.1, the control-to-state linear operator

S : L2 (D) → L2P

(Ω; L2 (D)

), u �→ y (u)

is continuous. Hence, J1 and J2 are also continuous. Moreover, J1 is convex, and,for β = 0, J2 is also convex. Consequently, both functionals are weakly lower semi-continuous [2, Chap. 2]. The existence of a solution for (P1) and (P2) then follows.The solution is unique if γ > 0 since, in that case, both J1 and J2 are strictly convex.

(ii) In the case m = −∞ and/or M = +∞, the set Uad is not bounded. However,if γ > 0, then for a fixed u0 ∈ Uad , the search for a solution to (P1)may be restrictedto the closed, convex and bounded set

{u ∈ L2 (D) : ‖u‖2L2(D) ≤ 2γ −1 J1

(u0

)}.

[email protected]

Page 55: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

40 3 Mathematical Analysis of Optimal Control Problems Under Uncertainty

Indeed, if u ∈ L2 (D) satisfies ‖u‖2L2(D)> 2γ −1 J1

(u0

), then

J1 (u) ≥ γ

2‖u‖2L2(D) > J1

(u0

).

The same argument applies for J2. The rest of the proof is the same as in the con-strained case. �

3.2.2 Risk Averse Optimal Control Problems

Let yd ∈ L2 (D) be a given target, and ε > 0 a threshold parameter. Consider thecost functional

Jε (u) = P

{ω ∈ Ω : I (u, ω) := ‖y(ω) − yd‖2L2(D) > ε

}.

Notice that since for each u ∈ Uad , I (u, . . .) : Ω → R is F-measurable, it definesa real-valued random variable. As a consequence, {ω ∈ Ω : I (u, ω) > ε} ∈ F andthus the cost functional Jε (u) is well-defined.

A typical risk averse, optimal control problem is formulated as

(Pε)

⎧⎨⎩Minimize in u ∈ Uad : Jε (u)

subject toy = y (u) solves (3.15).

Existence of a solution to (Pε) is established next.

Theorem 3.5 Assume that−∞ < m < M < +∞. Then, problem (Pε) has, at least,one solution.

Proof Let un be a (bounded)minimizing sequence.Up to a subsequence, still labelledby n, un ⇀ u weakly in L2 (D). For a fixed ω ∈ Ω , since I (·, ω) is continuous andconvex, it is weakly lower semicontinuous. Hence, I (u, ω) ≤ lim infn→∞ I (un, ω).Moreover, the cost functional Jε may be expressed in integral form as

Jε (u) =∫

Ω

H (I (u, ω) − ε) dP(ω),

where

H (s) ={1, s > 00, s ≤ 0,

is the Heaviside function. Hence, by Fatou’s lemma [4, Theorem 9.1],

[email protected]

Page 56: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

3.2 Existence of Optimal Controls Under Uncertainty 41

∫Ω

lim infn→∞ H (I (un, ω) − ε) dP(ω) ≤ lim inf

n→∞

∫Ω

H (I (un, ω) − ε) dP(ω).

(3.19)By the lower semicontinuity of the Heaviside function,

H (I (u, ω) − ε) ≤ lim infn→∞ H (I (un, ω) − ε)).

Thus, integrating in both sides and using (3.19),

∫ΩH (I (u) − ε) dP(ω) ≤ ∫

Ωlim infn→∞ H (I (un) − ε) dP(ω)

≤ lim infn→∞

∫Ω

H (I (un) − ε) dP(ω),

which completes the proof. �

Remark 3.5 Existence of optimal controls for (Pε) in the unconstrained case(m = −∞ and/or M = +∞) mayc be obtained by using the same arguments as inTheorem 3.4.

3.3 Differences Between Robust and Risk-Averse OptimalControl

Although in some situations the optimal control obtained using a risk-averse for-mulation exhibits an increased robustness, these two approaches show fundamentaldifferences in the control objective and the analysis type.

Regarding the control objective, the robust control approach aims at reducing thesystem variability under unexpected variations in the input data. However, the risk-averse approach is less concerned with the system variability andmore with reducinga risk function that quantifies the expected losses related with the damages caused byextrem events (see Fig. 3.1). The same ideas apply in the context of optimal designproblems [5] .

Fig. 3.1 Differencescenarios concerned inrobust and risk-averseoptimization

[email protected]

Page 57: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

42 3 Mathematical Analysis of Optimal Control Problems Under Uncertainty

Fig. 3.2 Difference betweenrobustness and risk-aversion

These conceptual differences entail different formulations of the optimal controlproblem. On the one hand, the robust control formulation incorporates measures ofthe dispersion of the objective (which for simplicity here and in Fig. 3.2 is denotedby I (u, ω)) around its mean value, such as the standard deviation. On the other hand,the risk-averse approach minimizes a risk function, such as the excess probability,that quantifies the expected losses. These differences in formulation are illustratedin Fig. 3.2. Observe that, whereas in the robust control the attention is paid to theregions around the mean value of the probability density function (PDF) of I (u, ω),the risk averse approach is focused on the tail of that PDF. As a consequence, sincethe tails are difficult to approximate numerically, the risk-averse approach involves ahigher computational burden compared to its robust counterpart, which only requiresthe estimation of low order statistical moments (mean and standard deviation).

3.4 Notes

As it has been noticed at the beginning of this chapter, the formalism of tensorproduct spaces is not necessary to establish an existence theory for solutions ofrandom PDEs. It is however very useful for the numerical approximation of suchsolutions, as noticed, for instance in [1, 3].

Existence of solutions for robust and risk averse control problems constrainedby evolution random PDEs is obtained by using the same ideas as in the ellipticsituation. See, e.g., [9–11].

The Idea of Transposition. To better illustrate the transposition method introducedin Sect. 3.1.3, let us consider the deterministic, non-homogeneous, boundary valueproblem {−Δy = 0 in D

y = y on ∂D,(3.20)

where D ⊂ Rd is a bounded domain with a smooth (of class C2) boundary ∂D, and

[email protected]

Page 58: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

3.4 Notes 43

−Δy = −d∑j=1

∂2y

∂x2j

is the (negative) of the Laplacian operator acting on y. By using the Lax-Milgramlemma, it is proved that if the boundary datum y belongs to the traces spaceH 1/2 (∂D), then there exists a unique weak solution y ∈ H 1 (D) to problem (3.20).However, if y ∈ L2 (∂D), then the Lax-Milgram lemma can not be applied as in thecasewhere y ∈ H 1/2 (∂D). To give ameaning to a solution of (3.20) for y ∈ L2 (∂D)

the transposition method is introduced as follows: let us suppose, for the moment,that y is regular and consider the problem

{−Δz = ϕ in Dz = 0 on ∂D,

(3.21)

with ϕ ∈ L2 (D). It is well-known that there is a unique solution z ∈ H 2 (D) to(3.21). Multiplying the PDE in (3.20) by z and integrating by parts twice one gets

∫Dyϕ dx = −

∫∂D

y∂z

∂nds. (3.22)

The operator � : L2 (D) → H 1/2 (∂D), defined by �ϕ = ∂z∂n , is linear and continu-

ous. When considered as an operator from L2 (D) to L2 (∂D) it is compact [12, p.22, Exercice 1.2]. Its adjoint �∗ is also compact from L2 (∂D) to L2 (D). Moreover,

−∫

∂Dy∂z

∂nds =< y,�ϕ >L2(∂D)=< �∗y, ϕ >L2(D) ∀ϕ ∈ L2 (D) .

Thus, by (3.22), y = �∗y. Notice that, up to now, to define y we have assumed thaty ∈ H 1/2 (∂D). However, �∗y is well-defined for y ∈ L2 (∂D). The transpositionmethod amounts to taking y = �∗y as the solution to (3.20) for y ∈ L2 (∂D).

Thismethod is of particular importance in the study of control problems associatedto that type of PDEs. We refer the reader to [8, Chaps. 2 and 3] for its use in theanalysis of PDEs, and to [6, Chap. 4] and [7] for its applications to control of PDEs.

References

1. Babuška, I., Nobile, F., Tempone, R.: A Stochastic collocation method for elliptic partial dif-ferential equations with random input data. SIAM Rev. 52(2), 317–355 (2010)

2. Barbu, V., Precupanu, T.: Convexity andOptimization in Banach Spaces. SpringerMonographsin Mathematics, 4 edn. (2012)

3. Gunzburger, M.D., Webster, C., Zhang, G.: Stochastic finite element methods for partial dif-ferential equations with random input data. Acta Numer. 23, 521–650 (2014)

4. Jacod, J., Protter, P.: ProbabilityEssentials, 2nd edn. Springer-Verlag,Berlin,Heidelberg (2003)

[email protected]

Page 59: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

44 3 Mathematical Analysis of Optimal Control Problems Under Uncertainty

5. Kang, Z.: Robust design of structures under uncertainties. Ph.D. Thesis, University of Stuttgart(2005)

6. Lions, J.L.: Optimal Control of Systems Governed by Partial Differential Equations. Springer,Berlin (1971)

7. Lions, J.L., Contröllabilité Exacte, Perturbation et Stabilisation des Systémes Disrtibués. tome1. Masson (1988)

8. Lions, J.L., Magenes, E.: Non-Homogeneous Boundary Value Problems and Applications Vol.I, Springer-Verlag (1972)

9. Marín, F.J., Martínez-Frutos, J., Periago, F.: Robust averaged control of vibrations for theBernoulli-Euler beam equation. J. Optim. Theory Appl. 174(2), 428–454 (2017)

10. Marín, F.J.,Martínez-Frutos, J., Periago, F.: A polynomial chaos approach to risk-averse controlof stochastic vibrations for the Bernoulli-Euler beam equation. Int. J. Numer. Methods Eng.,1–18 (2018). https://doi.org/10.1002/nme.5823

11. Martínez-Frutos, J., Kessler, M., Münch, A., Periago, F.: Robust optimal Robin boundarycontrol for the transient heat equation with random input data. Internat. J. Numer. MethodsEng. 108(2), 116–135 (2016)

12. Necas, J.: Les Méthodes Directes en Théorie des Equations Elliptiques. Éditeurs Academia,Prague (1967)

13. Tröltzsch, F.: Optimal Control of Partial Differential Equations: Theory, Methods and Appli-cations. Graduate Studies in Mathematics 112. AMS. Providence, Rhode Island (2010)

14. Wloka, J.: Partial Differential Equations. Cambridge University Press (1987)15. Zhang,D.: StochasticMethods for Flow inPorousMedia:CopingwithUncertainties.Academic

Press, San Diego, California (2002)

[email protected]

Page 60: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Chapter 4Numerical Resolution of Robust OptimalControl Problems

... a computational solution ... is quite definitely not routinewhen the number of variables is large. All this may be subsumedunder the heading “the curse of dimensionality.” Since this is acurse which has hung over the head of the physicist andastronomer for many a year, there is no need to feel discouragedabout the possibility of obtaining significant results despite it.

Richard Bellman.Dynamic Programming, 1957.

Both gradient-based methods and methods based on the resolution of first-orderoptimality conditions may be used for solving numerically the robust optimal controlproblems presented in the preceding chapters. In both cases, the main difficultyarises in the numerical approximation of statistical quantities of interest associatedto solutions of random PDEs. To handle this issue, it is assumed that the randominputs of the underlying PDEs depend on a finite number of random variables. Thisis the well-known finite-dimensional noise assumption. In this chapter, StochasticCollocation and Stochastic Galerkin methods are presented as powerful tools forsolving numerically robust optimal control problems by using gradient-based andone-shot methods.

Complete MatLab codes for all presented examples are available as part of thesupporting material accompanying this book.

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2018J. Martínez-Frutos and F. Periago Esparza, Optimal Control of PDEs under Uncertainty,SpringerBriefs in Mathematics, https://doi.org/10.1007/978-3-319-98210-6_4

45

[email protected]

Page 61: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

46 4 Numerical Resolution of Robust Optimal Control Problems

4.1 Finite-Dimensional Noise Assumption: From RandomPDEs to Deterministic PDEs with a Finite-DimensionalParameter

In many applications, uncertainty in the input data of a PDE is modelled by usinga finite number of random variables. In the case of random fields, which typicallyinvolve an infinite number of random variables, they are approximated by a finitenumber of random variables, e.g., with the help of a truncated KL expansion as in(2.12). Random fields which depend on a finite number of random variables arecalled finite-dimensional noise. All this motivates the following assumption:

Finite dimensional noise assumptionThe random inputs of the PDEs considered from now on depend on a finitenumber of uncorrelated real-valued random variables

ξ (ω) = (ξ1 (ω) , · · · , ξN (ω)) , (4.1)

where ξn : Ω → R, 1 ≤ n ≤ N .

For clarity in the exposition, from now on in this section we focus on problem(3.1). The following technical result establishes that if the diffusion coefficient a =a (x, ξ (ω)) and the forcing term f = f (x, ξ (ω)) both depend on ξ , then the solutiony = y (x, ω) to (3.1) is also finite-dimensional noise, i.e., y = y (x, ξ (ω)).

Proposition 4.1 In addition tohypotheses (A1) and (A2), as considered inSect.3.1.1,if a = a (x, ξ (ω)) and f = f (x, ξ (ω)), with ξ (ω) given by (4.1), then the solutiony to (3.1) is also finite-dimensional noise, i.e., y = y (x, ξ (ω)).

The proof of this result is based on the following abstract result, which is specializedto our particular case. For a proof, the reader is referred to [15, pp. 8–9].

Lemma 4.1 (Doob-Dynkin) Consider the space(R

N ,B(R

N)), whereB

(R

N)is

the σ -algebra of Borel sets in RN . Let ξ : Ω → RN be a multivariate random vari-

able and consider the measurable spaces (Ω, σ (ξ)), where σ (ξ) is the σ -algebragenerated by ξ , i.e., σ (ξ) = {ξ−1 (B) : B ∈ B

(R

N)}. Then, a random variable

g : Ω → R is σ (ξ)- measurable if and only if there exists a measurable functionh : RN → R such that g = h ◦ ξ .

Proof of Proposition 4.1. Consider the probability space (Ω, σ (ξ) ,P). Sincea(x, ·) and f (x, ·) depend on ξ , they are σ (ξ)-measurable. As a consequence, forevery x ∈ D, the solution y (x, ·) to (3.1) is also σ (ξ)-measurable. Indeed, sincea(x, ·) and f (x, ·) are σ (ξ)-measurable, by Lax-Milgram’s lemma, the variationalproblem: find y� ∈ L2

P

(Ω,σ (ξ) ; H 1

0 (D))such that

[email protected]

Page 62: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.1 Finite-Dimensional Noise Assumption: From Random PDEs … 47

Ω

Da∇ y · ∇v dxdP(ω) =

Ω

Df v dxdP(ω) ∀v ∈ L2

P

(Ω,σ (ξ) ; H 1

0 (D))

(4.2)has a unique solution y�, which is σ (ξ)-measurable as it belongs to the spaceL2P

(Ω,σ (ξ) ; H 1

0 (D)). We emphasize that the notation L2

P

(Ω,σ (ξ) ; H 1

0 (D))

indicates that σ (ξ) is the σ -algebra considered in Ω . Considering the same vari-ational problem in the probability space L2

P

(Ω,F ; H 1

0 (D))yields the solution y

to problem (3.1). Since σ (ξ) ⊂ F is a sub σ -algebra of F , by the uniqueness ofsolution of problem (4.2), one has y = y�. Hence, y(x) is σ (ξ)-measurable for everyx ∈ D.

The result then follows by applying Doob-Dynkin’s lemma, with g = y(x). �

Remark 4.1 As it will be illustrated hereafter, the main interest of Proposition 4.1is that it ensures that all statistical information contained in y (x, ·) can be obtainedfrom that included in the multivariate random variable ξ .

Remark 4.2 Since material properties and forces are seldom related eachother, in practice it is usual to have a (x, ξ (ω)) = a (x, ξa (ω)) and f (x, ξ (ω)) =f(x, ξ f (ω)

), with ξ = (ξa, ξ f

)and being ξa and ξ f independent vectors of real-

valued random variables.

The abstract probability space (Ω,F ,P) is not so appealing when statisticalquantities associated to solutions of random PDEs must be approximated numer-ically. Fortunately, under the finite dimensional noise assumption (4.1), one maytransform the random PDE (3.1) into a deterministic PDE with a finite dimensional

parameter. Indeed, let us denote by Γn = ξn (Ω) ⊂ R and by Γ =N∏

n=1

Γn ⊂ RN .

If g : Γ → R is a Borel measurable function, then the change of variable theorem([15, p. 21]) yields

Ω

g (ξ (ω)) dP (ω) =∫

Γ

g (z) dμ (z) ,

whereμ is the distribution probabilitymeasure of ξ , which is defined on theσ -algebraB (Γ ) of Borel sets on Γ , as

μ (B) = P(ξ−1 (B)

), B ∈ B (Γ ) .

If, in addition, μ is absolutely continuous with respect to the Lebesgue measure,then the Radon-Nikodym theorem ensures that there exists a joint probability densityfunction for ξ , which from now on is denoted by

ρ : Γ → R+, with ρ = ρ (z) ∈ L∞ (Γ ) .

[email protected]

Page 63: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

48 4 Numerical Resolution of Robust Optimal Control Problems

Thus,

Ω

g (ξ (ω)) dP (ω) =∫

Γ

g (z) dμ (z) =∫

Γ

g (z) ρ (z) dz,

with dz the Lebesgue measure. This way, the abstract probability space (Ω,F ,P)

is mapped to (Γ,B (Γ ) , ρ(z) dz), which is much more suitable for the numericalapproximation of statistical quantities of interest associated to solutions of randomPDEs.

When applied to problem (3.1), thanks to Proposition 4.1, the variational problem(3.2) rewrites as: find y = y (x, z) ∈ L2

ρ

(Γ ; H 1

0 (D))such that

Γ

ρ(z)∫

Da∇ y · ∇v dxdz =

Γ

ρ(z)∫

Df v dxdz ∀v ∈ L2

ρ

(Γ ; H 1

0 (D)),

(4.3)where L2

ρ

(Γ ; H 1

0 (D))denotes the Hilbert space composed of all (equivalence

classes of) strongly measurable functions g : Γ → H 10 (D) whose norm

‖g‖L2ρ(Γ ;H 1

0 (D)) =(∫

Γ

‖g (·, z) ‖2H 10 (D)

ρ(z) dz

)1/2

is finite. Notice that because of the results of Sect. 2.2 we have the isomorphism

L2ρ

(Γ ; H 1

0 (D)) ∼= L2

ρ (Γ ;R) ⊗ H 10 (D) ,

with L2ρ (Γ ;R) = {g : Γ → R : ∫

Γg2 (z) ρ(z) dz < ∞}. We emphasize that (4.3)

is a deterministic PDE with a finite-dimensional parameter z ∈ Γ . In its classicalform, (4.3) is written as

{−div (a(x, z)∇ y) = f, in D × Γ

y = 0, on ∂D × Γ.(4.4)

To sump up, under finite-dimensional noise assumption, a random PDE rewrites asa deterministic PDE with a finite-dimensional parameter.

From now on in this chapter, the finite dimensional noise assumption (4.1) isassumed to hold.

4.2 Gradient-Based Methods

As is well-known [9, Chap.2], a gradient-based algorithm for the minimization of ageneric functional J : U → R, with U a suitable space of admissible controls, isstructured as follows:

[email protected]

Page 64: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.2 Gradient-Based Methods 49

Gradient-based optimization algorithmInitialization: Take an initial u0 ∈ U .Construction: For k = 0, 1, · · · (until convergence) do:

2.1: Choose a descent direction uk (for which∂ J(uk)

∂u · uk < 0).2.2: Choose a step size λk such that J

(uk + λkuk

)< J

(uk).

2.3: Set uk+1 = uk + λkuk .Verification: The stopping criterion is the first k ∈ N for which

|J (uk)− J(uk−1

) |J(uk) ≤ η, (4.5)

where η is a prescribed tolerance.

In what follows, the chosen descent direction uk is minus the gradient of the costfunctional. As for the step size parameter λk , it will be chosen as the optimal one inthe selected search direction.

Remark 4.3 The stopping criterion (4.5) is often used in practise. Other stoppingcriteria may also be considered. For instance, criteria based on the size of the descentdirection and/or on the relative change in uk . For simplicity, throughout this text,stopping criteria based on the relative change of the involved cost functionals areused.

4.2.1 Computing Gradients of Functionals MeasuringRobustness

The gradients of the cost functionals J1 and J2, considered in problems (P1) and (P2)of Sect. 3.2.1, are derived by using the formal Lagrange method. The computationof the optimal step size parameters is also indicated next.

Problem (P1). The cost functional to be minimized is

J1 (u) = 1

2

Γ

D|y (x, z) − yd (x) |2 dx ρ(z)dz + γ

2

Ou2 (x) dx, (4.6)

with γ ≥ 0, and y solves

{−div (a∇ y) = 1O u, in D × Γ

y = 0, on ∂D × Γ.(4.7)

For simplicity, it is assumed that no constraints are imposed on the control variable,i.e., u ∈ Uad = L2 (D).

[email protected]

Page 65: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

50 4 Numerical Resolution of Robust Optimal Control Problems

To compute such a gradient, the formal Lagrange method is used. Although thisapproach is not mathematically rigorous, it provides a very simple way to derivethe optimality conditions of control problems constrained by PDEs. The methodis formal because it is assumed that all the functions that occur as well as theirderivatives are square integrable, without specifying the underlying spaces. We referthe reader to [19, Sects. 2.10 and 2.13] for the basis of the formal Lagrange methodfor deterministic PDEs. With obvious changes, the method extends to random PDEs,as it is illustrated next.

In a first step, a Lagrange multiplier p = p (x, z) for the random PDE constraintis introduced. Multiplying the PDE in (4.7) by p, integrating by parts, and addingthe result to the cost functional (4.6) gives the so-called Lagrangian associated toproblem (P1),

L(y, u, p

) = J1(y, u)−

Γ

D

[a∇ y · ∇ p − 1O u p

]dxρ(z)dz, (4.8)

where y, p ∈ L2ρ

(Γ ; H 1

0 (D)), and u ∈ L2 (D). Note that, since in (4.8) the variables

y, u and p are treated as independent, we have written J1 = J1(y, u)and not J1

(u).

Similarly to the case of non-linear mathematical programming problems, we lookfor a stationary point (y, u, p) of L . Since, formally y is unconstrained, equatingthe derivative of L w.r.t. y to zero gives the so-called adjoint equation

{−div (a∇ p) = y − yd , in D × Γ

p = 0, on ∂D × Γ.(4.9)

By inserting the adjoint state p as a test function in the variational formulationof (4.7), the second term in the r.h.s. of (4.8) vanishes. As a consequence, J1 (u) =L (y, u, p). Thus, J ′

1 (u) is formally calculated by differentiating the Lagrangianw.r.t. the control u, which leads to

J ′1 (u) (x) = γ u (x) +

Γ

p (x, z) ρ(z)dz, x ∈ O, (4.10)

where p solves (4.9).As indicated above, a descent direction of the gradient algorithmis given by u = −J ′

1 (u).Once a descent direction u is fixed, the optimal descent parameter λ is obtained

by minimizing, over R+, the function λ �→ J1 (u + λu). It is easy to see thatJ1 (u + λu) = J1 (u) + a1λ + b1λ2, where

a1 =∫

Γ

Dy (y − yd) dxρ(z)dz + γ

Ouu dx, (4.11)

b1 = 1

2

Γ

Dy2 dxρ(z)dz + γ

2

Ou2 dx (4.12)

[email protected]

Page 66: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.2 Gradient-Based Methods 51

and y = y (x, z) solves

{−div (a∇ y) = 1O u, in D × Γ

y = 0, on ∂D × Γ.

Hence, the optimal step size parameter is λ1 = − a12b1

.

Problem (P2). Consider the cost functional J2 (u), as given by (3.18), which now isexpressed as

J2 (u) = 1

2

D

(∫

Γ

y(x, z)ρ(z) dz − yd (x)

)2

dx + β

2

DVar (y(x)) dx + γ

2

Ou2 (x) dx,

(4.13)Proceeding as in the preceding case,

J ′2 (u) (x) = γ u (x) +

Γ

q (x, z) ρ(z) x ∈ O, (4.14)

where q = q (x, z) solves the adjoint system

{−div (a∇q) = ∫Γy ρ(z)dz − yd + β

(y − ∫

Γy ρ(z)dz

), in D × Γ

q = 0, on ∂D × Γ.

(4.15)The descent direction is u = −J ′

2 (u). Similarly to the preceding case, the optimalstep size parameter is λ2 = − a2

2b2, where

a2 = ∫D(∫

Γy ρ(z)dz − yd

) (∫Γy ρ(z)dz

)dx

+β∫D

[∫Γy y ρ(z)dz − ∫

Γy ρ(z)dz

∫Γy ρ(z)dz

]dx

+γ∫O uu dx,

(4.16)

b2 = 1

2

D

(∫

Γ

y ρ(z)dz

)2

dx + β

2

DVar

(y (x)

)dx + γ

2

Ou2 dx, (4.17)

and y (x, z) is a solution to

{−div(a∇ y

) = 1O u, in D × Γ

y = 0, on ∂D × Γ.

As is observed in both problems (P1) and (P2), themain task to implement a gradient-based optimization algorithm arises in the numerical approximation of statisticalquantities (mean and variance in the case of robust control problems) associated tosolutions of random PDEs. This crucial fact is addressed in the next section.

[email protected]

Page 67: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

52 4 Numerical Resolution of Robust Optimal Control Problems

4.2.2 Numerical Approximation of Quantities of Interestin Robust Optimal Control Problems

The following reference benchmark problem is considered again

{−div (a (x, z) ∇ y (x, z)) = 1O u (x) , (x, z) ∈ D × Γ

y (x, z) = 0, (x, z) ∈ ∂D × Γ,(4.18)

where Γ ⊂ RN is a compact set. As it has been showed above, in robust optimal

control problems one is mainly interested in computing the first and second orderstatistical moments of the solution y (x, z),

E[y j](x) =

Γ

y j (x, z) ρ(z)dz, j = 1, 2, x ∈ D. (4.19)

A number of methods (e.g., Monte Carlo (MC), Stochastic Galerkin (SG), StochasticCollocation (SC), ReducedBasis (RB), etc.)may be used to approximate numerically(4.19). Although MC sampling is a natural choice, its very slow convergence maylead to an unaffordable computational cost when used in optimization algorithmsthat require iteration. In addition, MC method does not exploit the possible regu-larity of the solution w.r.t. the random parameter. For the specific problems underconsideration in this chapter, the solution y of (4.18) is smooth w.r.t. the parameterz ∈ Γ . Indeed, it is known [2] that, under reasonable assumptions on the uncertaincoefficient a (x, z), the solution y to (4.18) is analytic w.r.t. the parameter z. Moredetails on this issue are provided in Sect. 4.5 below. In this situation, stochastic col-location methods are computationally very efficient, at least when the number N ofinvolved random variables is not so large, as is the case of truncatedKL expansions ofsmooth random fields. Accordingly, an adaptive, anisotropic, sparse grid, stochasticcollocation method is used to compute the statistical quantities involved in problems(P1) and (P2).

4.2.2.1 An Anisotropic, Sparse Grid, Stochastic Collocation Method

Throughout this section, we closely follow [2, 12] for the description of the methodand [3] for its numerical implementation. We focus on an algorithmic presentationof stochastic collocation methods and refer the readers to [2, 12] for error estimates.See also [17, Chap.11].

The basic idea to approximate (4.19) is to use a quadrature rule so that the solu-tion y (x, z) to (4.18) must be known at a set of nodes zk = (zk1, · · · , zkN

) ∈ Γ ,1 ≤ k ≤ M , as well as suitable weights wk , 1 ≤ k ≤ M . The choice of these nodesand weights is a crucial point as it determines the accuracy and efficiency of thecollocation method. Although the first idea would be to use a full tensor product ofa one dimensional set, even for a reduced number of random variables, full tensor

[email protected]

Page 68: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.2 Gradient-Based Methods 53

rules are, in most of the cases of interest, unnecessary, in the sense that the sameexactness may be achieved by using a lower number of nodes. Sparse grid-basedquadrature rules must be used instead. Both, isotropic and anisotropic quadraturerules may be considered. Having in mind that when random inputs are representedby truncated KL expansions of random fields, the different stochastic directions donot weigh equally, anisotropic rules may preserve similar accuracy as isotropic ones,but alleviating the involved computational cost.

For pedagogical reasons, in parallel to the theoretical description of the method,the numerical approximation of (4.19) is illustrated through the following specificdata: D = (0, 1)2 is the unit square, O is a circle centred at (0.5, 0.5) and of radius0.25, and u (x) = 1 for all x ∈ O . The random domain is Γ = [−3, 3]N , and a (x, z)is the truncated KL expansion considered in (2.29), with N = 4, i.e.,

a (x, z) = eμ+σ∑4

n=1

√λnbn(x)zn , x = (x1, x2) ∈ D, z = (z1, z2, z3, z4) ∈ Γ,

(4.20)where μ = −0.3466 and σ = 0.8326;

{λ j , b j (x)

}4j=1 are the first fourth eigenpairs

associated to the covariance function (2.28), with correlation lengths L1 = L2 = 0.5,and finally

ρ (z) =4∏

n=1

φ (zn) , z = (z1, z2, z3, z4) ∈ Γ, (4.21)

where φ (zn) is the truncated Gaussian (2.30). See Example 2.3 for more details.To better understand an anisotropic, sparse grid, quadrature rule in the multi-

dimensional case N > 1, let us start with the one dimensional case N = 1.

The one-dimensional case. Let f : Γ1 ⊂ R → R be a random function and ρ1 =ρ1 (z1) the PDF associated to the random parameter z1 ∈ Γ1. It is useful to arrangethe number of quadrature nodes in levels. Thus, for a positive integer � ∈ N+, fromnow on called the level of the quadrature rule, the number of nodes in the level � isdenoted by m�. Several choices are possible to define m�. For instance, one may addjust only one quadrature node when passing from one level to the next one, or onemay double the number of nodes. Another typical choice of m� is [18]

m1 = 1 and m� = 2�−1 + 1 for � > 1. (4.22)

A quadrature rule for f at the level � then takes the form

Γ1

f (z1) ρ1 (z1) dz1 ≈m�∑

j=1

f(z j1

)w

j1 := Q� f, (4.23)

where z j1 and wj1 are, respectively, the nodes and weights of the one-dimensional

quadrature rule.

[email protected]

Page 69: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

54 4 Numerical Resolution of Robust Optimal Control Problems

Selection of quadrature nodes and weights. Several choices are possible for thenodes z j1 and the weights w

j1 . These are typically classified into nested and non-

nested quadrature rules. The former are characterized by the fact that the nodes at thelevel � contain the nodes at the level � − 1. This property does not hold in non-nestedrules. Clenshaw-Curtis abscissas, in which the nodes are the extrema of Chebyshevpolynomials, which are given by

z j1 = − cos

(π ( j − 1)

m� − 1

), j = 1, · · · ,m�, (4.24)

are widely used as nested quadrature rules.Gaussian abscissas, in which the nodes are the zeros of a suitable class of orthog-

onal polynomials (e.g., Hermite polynomials for the case of Gaussian random vari-ables, and Legendre polynomials for uniformly distributed random variables) are theclassical examples of non-nested quadrature rules. Gauss-Hermite nodes are used inthe numerical experiments included in this chapter. The associated weights are thosefor which the quadrature rule (4.23) is exact for polynomials up to degree m� − 1.

For a comparison of Clenshaw-Curtis and Gaussian quadrature rules we referto [20].

The multi-dimensional case N > 1. Let y (x, ·) : Γ ⊂ RN → R be a solution to

(4.18) and let us assume that y (x, z) admits an analytical extension w.r.t. eachstochastic direction zn , 1 ≤ n ≤ N , in a region of the complex plane C. As it willbe indicated in Sect. 4.5 below, the size of these regions depends on

√λn‖bn‖L∞(D).

Recall that throughout this text, hypotheses of Mercer’s Theorem 2.2 are assumedto hold. As a consequence, the eigenfunctions bn ∈ L∞ (D). Stochastic directionsin which the region of analyticity of the solution is larger require less colloca-tion points than directions with less regularity. Accordingly, a vector of weightsg = (g1, · · · , gN ) for the different stochastic directions is introduced. Vector g willdetermine the number of collocation points to be used in each stochastic direction.For the specific problem under consideration, in which a (x, z) is modelled by atruncated KL expansions of a log-normal field, it is proved in [2] that an appropriateselection of g is

gn = 1

2√2λn‖bn‖L∞(D)

, 1 ≤ n ≤ N . (4.25)

Thus, the larger the√

λn‖bn‖L∞(D) (or equivalently the smaller the gn) the moreimportant the corresponding n-th stochastic direction and so the larger the numberof collocation points required in that stochastic direction (see [2, 12] for more detailson this passage).

For the specific case of the random field (4.20) one has

g = (0.3894, 0.5478, 0.5478, 0.7708) . (4.26)

[email protected]

Page 70: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.2 Gradient-Based Methods 55

Starting from a fixed quadrature level � ∈ N+ for the multi-dimensional grid, andfrom the vector of weights g, in order to construct an anisotropic tensor grid, it isvery convenient to use a multi-index notation. Hence, the following multi-index setis introduced:

Ig (�, N ) ={

i = (i1, · · · , iN ) ∈ NN+ , i ≥ 1 :

N∑

n=1

(in − 1) gn ≤ �g

}

, (4.27)

where g = min1≤n≤N gn . Note that the lower the weight gn , the larger the corre-sponding maximum index in the n-th direction. Also note that (4.27) is completelydetermined by the weighting vector g and the level �. This indices set together withthe rule used to increase the number of points in each direction (i.e., increasing byone point to pass from one level to the next one, doubling the number of points,using (4.22), etc.), determines the total number of collocation points in the multi-dimensional grid. Since the sparse grid generated may have points in common, inorder to save computational time during evaluation of a function on a sparse grid, itis then important to get ride of these repetitions.

To construct the anisotropic quadrature rule associated to the multi-index setIg (�, N ), let us first consider the difference operator

� j = Q j − Q j−1, for j = 1, 2, . . . (4.28)

withQ0 = 0, and whereQ j is the one-dimensional quadrature rule given by (4.23).By using (4.27) and (4.28), the so-called anisotropic Smolyak1 quadrature rule

approximating∫Γy (x, z) ρ(z)dz, is defined as

Ag (�, N ) y (x) :=∑

i∈Ig(�,N )

(�i1 ⊗ · · · ⊗ �iN

)y (x) , (4.29)

which is equivalent to

Ag (�, N ) y (x) =∑

i∈Yg(�,N )

cg(i)(Qi1 ⊗ · · · ⊗ QiN

)y (x) , (4.30)

with

cg(i) =∑

j ∈ {0, 1}Ni + j ∈ Ig (�, N )

(−1)| j | , | j | =N∑

k=1

jk (4.31)

1The sparse grid quadrature rule presented in this chapter was proposed by Smolyak [18]. The mainidea underlying this construction is described in Sect. 4.5.

[email protected]

Page 71: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

56 4 Numerical Resolution of Robust Optimal Control Problems

and

Yg (�, N ) = Ig (�, N ) \ Ig(

� − |g|g

, N

)

, |g| =N∑

k=1

gk . (4.32)

Thus, to compute Ag (�, N ) y (x), we only need to determine the coefficients cg(i)and to apply tensor product quadrature formulae.We recall that for a fixedmulti-indexi = (i1, · · · , iN ) ∈ Yg (�, N ),

(Qi1 ⊗ · · · ⊗ QiN

)y (x) :=

mi1∑

r1=1

· · ·miN∑

rN=1

y(x; zr11 , · · · , zrNN

)w

r1i1

· · ·wrNiN

, (4.33)

where(zr jj , w

r ji j

), 1 ≤ j ≤ N , 1 ≤ r j ≤ mi j , are the one-dimensional nodes and

weights associated to the level mi j . Thus, by considering the anisotropic nodal set

Θg (�, N ) =⋃

i∈Yg(�,N )

(Θ i1 × · · · × Θ iN

), (4.34)

with Θ in the set of one-dimensional nodes for the level in in the n-th stochasticdirection, (4.30) may be expressed in the form

Ag (�, N ) y (x) =∑

zk∈Θg(�,N )

y(x, zk

)wk, (4.35)

where the weights wk are obtained by multiplying the corresponding coefficientscg(i) by its associatedweights, which appear in (4.33). Notice that, as a consequence,some weights wk may take negative values.

Remark 4.4 Having in mind an efficient numerical implementation of the multivari-ate quadrature rule (4.35), it is convenient to compute firstly the coefficients (4.31)because some of those coefficients may be equal to zero. Hence, there is no need tocompute nodes and weights associated to indices i ∈ Yg (�, N ) for which cg(i) = 0.

Remark 4.5 It is proved in [13] that for a multivariate function f = f (z1, · · · , zN )

having continuous derivatives up to order α in each variable, if we denote byM the number of nodes in the sparse grid, then the quadrature error is of orderO(M−α (logM)(N−1)(α+1)

).

As an example, Table4.1 collects the maximum index in the n-th stochastic direc-tion (which is denoted by βn ,) obtained according to (4.27), for the vector of weights(4.26) and for several levels �. One-dimensional Gauss-Hermite nodes are ordered inlevels min , with min = in . The total number of nodes, after removing repeated nodesand nodes associated to coefficients cg (i) which are equal to zero, is shown in thelast column.

[email protected]

Page 72: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.2 Gradient-Based Methods 57

Table 4.1 Maximum indexset for the vector of weights(4.26) and for several levels �

� β1 β2 β3 β4 # Nodes

1 2 1 1 1 2

2 3 2 2 2 9

3 4 3 3 2 31

4 5 3 3 3 61

5 6 4 4 3 149

8 10 7 7 5 1851

Example 4.1 To better illustrate the construction of the above anisotropic, sparsegrid, and to show the differences between isotropic and anisotropic sparse grids,Fig. 4.1 displays the multi-index sets Ig (�, N ) for two stochastic directions (N = 2),and the vectors of weights g = (1, 1) (isotropic) and g = (1, 0.5) (anisotropic), bothfor a quadrature level � = 4. The associated isotropic and anisotropic sparse grids arecomposed of Gaussian nodes (zeros of Hermite polynomials). The number of one-dimensional nodesmin , n = 1, 2, in the n-th stochastic direction is taken asmin = in .Thus, the nodes corresponding to the one-dimensional level in are the zeros of theHermite polynomial of degree in .

To construct themulti-index setY(1,0.5) (4, 2) (as given by (4.32)) in the anisotropiccase (Fig. 4.1c–d), we only have to remove the indices (1, 1) and (1, 2) fromFig. 4.1c.From (4.31) it follows that c(1,4) (1, 0.5) = c(1,0.5) (2, 2) = 0. Consequently, thenodes associated to the indices (1, 4) and (2, 2) are not plotted in Fig. 4.1d.

Practical numerical implementation. Discretization in the physical domain.Once the set of nodes Θg (�, N ) and it associated weights have been computed, toapproximate the first two statistical moments, E

[y j](x), j = 1, 2, of the solution

y(x, z) to (4.18) we proceed as follows:

(i) Compute, by using finite elements, the solutions of the deterministic problems

{−div(a(x, zk

)∇ y(x, zk

)) = 1O u (x) , x ∈ Dy(x, zk

) = 0, x ∈ ∂D,(4.36)

at each node zk ∈ Θg (�, N ).(ii) Apply the quadrature rule (4.35) to y(x j ) (and to y2(x j )) at each node x j of the

finite element mesh in the physical domain D.

Figure4.2 displays the expectedvalueE [y] (x) (left) and thevarianceVar [y] (x) =E[y2](x) − (E [y] (x))2 (right) of the solution to (4.18) for the data: D = (0, 1)2,

O a circle centred at (0.5, 0.5) and of radius 0.25, and u (x) = 1 for all x ∈ O .The random domain Γ = [−3, 3]N , and a (x, z) is given by (4.20). Lagrange P1finite elements have been used for approximation in the physical domain D. Themesh used for discretization of D is plotted in Fig. 4.3. Approximation in the ran-dom domain Γ has been carried out by using the anisotropic sparse grid collocationmethod described in the preceding section. The level � of the multivariate quadraturerule has been fixed to � = 5. Gauss-Hermite nodes and weights have been used.

[email protected]

Page 73: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

58 4 Numerical Resolution of Robust Optimal Control Problems

Fig. 4.1 N = 2, � = 4. (a) Multi-index set Ig (�, N ), with g = (1, 1) and associated isotropicsparse grid (b). (c) Multi-index set Ig (�, N ) for g = (1, 0.5) and associated anisotropic sparse grid(d). Gaussian nodes corresponding to zeros of Hermite polynomials are plotted in (b) and (d).Both the indices sets and sparse grids have been generated by using the Sparse grids Matlab kit [3](http://csqi.epfl.ch)

4.2.2.2 Adaptive Algorithm to Select the Quadrature Level �

This subsection describes an adaptive algorithm to select the level � that will be used,along the descent algorithm, to approximate (in the random domain) the statisticalquantities of interest in problems (P1) and (P2).

Since the main goals in (P1) and (P2) are to minimize the cost functionals J1 (u)

and J2 (u), the level � used for numerical integration in the random domain is adap-tively chosen so as to comply with a prescribed accuracy level δ. Obviously, δ shouldbe less than or equal to the tolerance imposed in the stopping criterion (4.5) of the

[email protected]

Page 74: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.2 Gradient-Based Methods 59

Fig. 4.2 Approximation of the expected value E [y] (x) (a) and the variance Var [y] (x) (b) forproblem (4.18)

Fig. 4.3 Mesh of thephysical domain D used forfinite elementsapproximation

0 10.2 0.4 0.6 0.80

0.2

0.4

0.6

0.8

1

x1

x2

gradient algorithm, i.e., δ ≤ η. The idea is therefore quite simple: roughly speaking,firstly, a criterion to measure the approximation error of the statistical quantities ofinterest is considered. Then, the level � is increased linearly up to that criterion issatisfied.

To fix ideas, let us consider problem (P2), where the cost functional J2 (u), givenby (4.13), involves the numerical approximation of the first two statistical momentsconsidered in (4.19). Let us denote by M j,h

g,�,N (y) a numerical approximation of

D

Γ

y j (x, z) ρ (z) dzdx, j = 1, 2, (4.37)

obtained as described in the preceding subsection. Here, h refers to the size of thefinite element mesh for discretization of the physical domain D, and g, �, N are theparameters that appear in (4.29).

[email protected]

Page 75: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

60 4 Numerical Resolution of Robust Optimal Control Problems

Since the level � is going to be fixed before running the optimization algorithm, areference control function u = u (x), which appears in the right-hand side of (4.18),must be fixed in advance. For instance, one may consider the optimal control of theassociated deterministic control problem (see Sect. 4.2.3.1 for a specific example).

Then, to estimate the computational error in the level �, reference (or enriched)values for (4.37) are computed before optimization. This could be done, e.g., by usingMonte Carlo method. If the computational cost of MC method is unaffordable, then,following [12], one may use the following approach: take a large enough integer �.Denoting byβn

(g, �) := maxi∈Ig(�,N) {in} themaximum index, in the n-th stochastic

direction, used in the quadrature rule (4.29), enriched (or reference) values for (4.37)are obtained by using the quadrature rule (4.29) for the level � + 1 and for a largervector of weights g, which have to satisfy that βn

(g, � + 1

) ≥ βn(g, �)+ 1. As

indicated in [12], g may be defined component-wise as

gn =(

βn(g, �)− 1

βn(g, �)

)(gng

)

.

These enriched values are denoted by M j,hg,�+1,N

(y), j = 1, 2, and play the role ofexact computations for (4.37).

To sum up, the adaptive algorithm to select � is the following:

Adaptive algorithm to select the level � of quadrature1. Fix a tolerance 0 < δ � 1 and an enriched level of quadrature �.2. Compute enriched values, M j,h

g,�+1,N(y), for (4.37).

3. Increasing the level � linearly from � = 1 to � = �opt ≤ � until the follow-ing stopping criterion is satisfied

max

⎧⎨

|M 1,hg,�+1,N

(y) − M 1,hg,�,N (y) |

M 1,hg,�+1,N

(y),

|M 2,hg,�+1,N

(y) − M 2,hg,�,N (y) |

M 2,hg,�+1,N

(y)

⎫⎬

⎭≤ δ.

(4.38)If the stopping criterion (4.38) is not satisfied, then itmeans that the integer �

is not large enough as to approximate with enough accuracy the exact quantityof interest. In such a case, another � greater than the previous one is selectedand the whole process is repeated.

Remark 4.6 For the case of problem (P1), denoting byM hg,�,N a numerical approx-

imation of ∫

Γ

D|y (x, z) − yd (x) |2 dxρ(z)dz (4.39)

[email protected]

Page 76: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.2 Gradient-Based Methods 61

the adaptive algorithm described above applies, but with the stopping criterion (4.38)replaced by

|M hg,�+1,N

(y) − M hg,�,N (y) |

M hg,�+1,N

(y)≤ δ, (4.40)

where, similarly to problem (P2),M hg,�+1,N

denotes an enriched numerical approxi-mation of (4.39).

Remark 4.7 Convergence results for this type of adaptive algorithms, in terms of thelevel �, have been obtained in [2, Sect. 4.1] and in [12, Theorem 3.13].

Figures4.6 and 4.11 illustrate, in two specific examples, the performance of theadaptive algorithm described in this subsection.

4.2.3 Numerical Experiments

This section presents numerical simulation results for problems (P1) and (P2)obtained by applying the gradient algorithm described in Sect. 4.2, in combinationwith the above adaptive, anisotropic, sparse grid, stochastic, collocation method.

As in the preceding section, inwhat follows, it is assumed that the physical domainD = (0, 1)2 is the unit square and the control region

O = {x = (x1, x2) ∈ R2 : (x1 − 0.5)2 + (x2 − 0.5)2 < 0.252

},

is a circle centred at the point (0.5, 0.5) and of radius r = 0.25. The random coeffi-cient a (x, z) is the truncated KL expansion considered in (4.20), the random domainis Γ = [−3, 3]4, and ρ(z) is given by (4.21).

4.2.3.1 A Deterministic Test Example

In order to assess and compare numerical simulation results for optimal control prob-lems under uncertainty with their deterministic counterparts, the following determin-istic optimal control problem is introduced next.

Consider the control functionu (x) = 1, x ∈ O , and let yd = yd (x)be the solutionof the deterministic problem

{−div (∇ y) = 1O u, in D,

y = 0, on ∂D.(4.41)

[email protected]

Page 77: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

62 4 Numerical Resolution of Robust Optimal Control Problems

Fig. 4.4 (a) Target function yd . (b) Optimal exact control uexact

Then, obviously, the deterministic optimal control problem

(Pdet )

⎧⎪⎪⎨

⎪⎪⎩

Minimize in u ∈ L2 (D) : Jdet (u) = 12

∫D |y (x) − yd (x) |2 dx

subject to−div (∇ y) = 1O u, in D,

y = 0, on ∂D,

(4.42)

has the solution uexact (x) = 1, x ∈ O .Figure4.4 displays the target function yd (x) and the optimal exact control uexact .Problem (Pdet ) has been solved by using a standard gradient algorithm with opti-

mal step-size. Numerical simulation results are plotted in Fig. 4.5. Precisely, Fig. 4.5adisplays the error with respect to the target, i.e., the values of

|y (udet ) − y (uexact ) |

at each point x ∈ D, where udet is the numerical solution of (Pdet ). Here, y (uexact )and y (udet ) denote, respectively, the solutions to (4.18), with u = uexact and u =udet . Figure4.5b shows the optimal numerical control udet .

Figure4.7a displays the convergence history for problem (Pdet ).

4.2.3.2 Experiment 1: Problem (P1)

Next, we solve (P1) for the case γ = 0 in (4.6). The target function yd(x) is the sameas in the deterministic problem, i.e., the function yd is the solution to (4.41) for thecontrol u (x) = 1, x ∈ O .

The algorithm to solve (P1) is the following:

[email protected]

Page 78: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.2 Gradient-Based Methods 63

Fig. 4.5 (a) Picture of |y (udet ) − y (uexact ) | at each point x ∈ D, where y (uexact ) and y (udet )are, respectively, the solutions to (4.18), with u = uexact and u = udet . (b) Picture of udet

Algorithm for problem (P1)Initialization: Take an initial u0 ∈ L2 (D), stopping parameters η, δ for (4.5)

and (4.40), respectively, and discretization parameters h and � for approx-imation in the physical and random domains.

Construction: For k = 0, 1, · · · (until convergence) do:2.1:Computation of a descent direction uk = −J ′

1

(uk), as given

by (4.10), with u = uk . To this end, first solve the direct problem (4.7) andthen the adjoint system (4.9).

2.2: Choose an optimal step size parameter λk by using expres-sions (4.11) and (4.12).

2.3: Update the control variable uk+1 = uk + λkuk .Verification: The stopping criterion is (4.5).

We take δ = 10−3 in (4.40). The enriched level in the adaptive algorithm ofSect. 4.2.2.2 is taken as � = 9 leading to a quadrature level � = 5. Figure4.6 dis-plays the rate of convergence for the adaptive, anisotropic, sparse grid algorithm.Gauss-Hermite nodes and weights are used for numerical integration in the randomdomain. As shown in Table4.1, this corresponds to 149 nodes in the sparse grid. Themesh for discretization of the physical domain is shown in Fig. 4.3.

The algorithm is initiated with a control u0(x) = 0, x ∈ O . Figure4.7b plots thehistory convergence of the gradient algorithm for the first 100 iterates. Figure4.8plots the computed optimal control u1.

Let us denote by y(udet ) the solution to the random elliptic problem (4.18) withu = udet the optimal control for the deterministic problem (Pdet ), and by y(u1)the solution to (4.18), with u = u1 the optimal robust control for (P1) obtainedafter 100 iterates of the gradient algorithm. Figure4.9 presents numerical results for

[email protected]

Page 79: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

64 4 Numerical Resolution of Robust Optimal Control Problems

1 2 3 4 5-4

-3

-2

-1

0

Fig. 4.6 Experiment 1. Rates of convergence of the anisotropic sparse grid algorithm: level ofquadrature � versus stopping criterion (4.40). The dashed line represents the prescribed accuracylevel δ = 10−3

0 200 400 600 80010-15

10-10

10-5

100

100 101 10210-5

10-4

10-3(a) (b)

Fig. 4.7 Convergence histories (number of iterations versus values of cost functionals) for thegradient algorithm of Sect. 4.2 corresponding to Jdet (a) and J1 (b)

∫Γ

|y(x, z) − yd(x)|2ρ(z) dz for the cases: (a) y = y(udet ) and (b) y = y(u1). Asexpected, compared to the optimal deterministic control udet , the optimal control u1for problem (P1), reduces in a significant way the L2

ρ(Γ )-distance from y(x, ·) toyd(x).

Figure4.10 displays the expected value and variance of y(udet ) (left panel), andof y(u1) (right panel). It is observed that the the expected values of y(udet ) and y(u1)are quite similar. However, the variance of y(u1) is lower than the one of y(udet ).

[email protected]

Page 80: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.2 Gradient-Based Methods 65

Fig. 4.8 Experiment 1. Control u1, optimal solution of problem (P1)

Fig. 4.9 Experiment 1. (a)∫Γ

|y(udet ) − yd |2ρ(z) dz, and (b)∫Γ

|y(u1) − yd |2ρ(z) dz

4.2.3.3 Experiment 2: Problem (P2)

This section presents numerical simulation results for problem (P2), with γ = 0 in(4.13) and for different values of the weighting parameter β. The target functionyd(x) is the same as in the deterministic problem.

[email protected]

Page 81: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

66 4 Numerical Resolution of Robust Optimal Control Problems

Fig. 4.10 Experiment 1. Upper panel shows pictures of the expected values for y(udet ) (a), andy(u1) (b). Lower panel plots the variances of y(udet ) (c), and y(u1) (d)

The algorithm to solve (P2) is structured as follows:

Algorithm for problem (P2)Initialization: Take an initial u0 ∈ L2 (D), stopping parameters η, δ for (4.5)

and (4.38), respectively, and discretization parameters h and � for approx-imation in the physical and random domains.

Construction: For k = 0, 1, · · · (until convergence) do:2.1: Computation of a descent direction uk = −J ′

2

(uk), as given by

(4.14), with u = uk . To this end, first solve the direct problem (4.7) and thenthe adjoint system (4.15).

2.2: Choose an optimal step size parameter λk by using expressions(4.16) and (4.17).

2.3: Update the control variable uk+1 = uk + λkuk .Verification: The stopping criterion is (4.5).

[email protected]

Page 82: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.2 Gradient-Based Methods 67

1 2 3 4 5 6 7-6

-4

-1

-3

-5

-2

0

Fig. 4.11 Experiment 2. Rates of convergence of the anisotropic sparse grid algorithm: level ofquadrature � versus Log10 of the stopping criterion (4.38). The dashed-circle line corresponds to thefirst term in (4.38), i.e., the one associated to the first statistical moment of y, and the continuous-circle line corresponds to the second term in (4.38), i.e., the one associated to the second momentof y. The horizontal dashed line represents the prescribed accuracy level δ = 10−2

The tolerance in the stopping criterion (4.38) is δ = 10−2. The enriched level inthe adaptive algorithm of Sect. 4.2.2.2 is taken as � = 8 leading to a quadrature level� = 7. This corresponds to 555 Gauss-Hermite nodes in the sparse grid. Figure4.11displays the rate of convergence for the adaptive, anisotropic, sparse grid algorithm.The same mesh as in the preceding examples, for discretization of the physicaldomain D, is used.

The algorithm is initiated with u0(x) = 0, x ∈ O . The convergence history forthe first 100 iterations of the gradient algorithm is shown in Fig. 4.12, for β = 0and β = 2. Figure4.13 displays the optimal controls for problem (P2), after 100iterates of the gradient algorithm, corresponding to different values of the weightingparameter β.

Figure4.14 shows the expected values and variances of the optimal states forproblem (P2). In the pictures of the left panel, the target function is shown trans-parently as a reference. As expected, increasing the value of β produces a reductionin variance. The price paid for this variance reduction is that the expected value ofy moves further away from the target yd . This fact is also numerically observed inTable4.2. First row in Table4.2 presents values for:

∫Γ

∫D (y − yd(x))

2 dxρ(z)dz(first column),

∫D (E [y] − yd(x))

2 dx (second column), and∫D Var (y(x)) dx (third

column), which are obtained when the control udet is inserted in the (random) stateEq. (4.18). Second row shows the same quantities but for the optimal control ofproblem (P1). The rest of rows show the same quantities for the optimal controls of(P2) corresponding to β = 0 (third row), β = 1 (fourth row), and β = 2 (fifth row).

[email protected]

Page 83: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

68 4 Numerical Resolution of Robust Optimal Control Problems

100 101 10210-8

10-7

10-6

10-5

10-4

10-3

100 101 10210-5

10-4

10-3(a) (b)

Fig. 4.12 Experiment 2. Convergence histories (number of iterations versus values of cost func-tionals) for the gradient algorithm of Sect. 4.2 corresponding to J2 for β = 0 (a) and β = 2 (b)

(a) β = )b(0 β = 1

(c) β = 2

Fig. 4.13 Experiment 2. Optimal controls for β = 0 (a), β = 1 (b), and β = 2 (c)

[email protected]

Page 84: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.2 Gradient-Based Methods 69

(a) β = (b)0 β = 0

(c) β = )d(1 β = 1

(e) β = 2 (f) β = 2

Fig. 4.14 Experiment 2. Expected values and variances of y(u2), optimal state of (P2), for differentvalues of β. u2 denotes the solution to (P2)

[email protected]

Page 85: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

70 4 Numerical Resolution of Robust Optimal Control Problems

Table 4.2 Summary of results for the cost functional J1 (second column), and for the first twoterms of the cost functional J2 corresponding to the optimal deterministic control udet (first row),the optimal control u1 for problem (P1), and the optimal controls u2 of (P2) for different values ofβ. All integrals in the random domain have been approximated with a quadrature level � = 7 in thesparse grid

Control∫Γ

∫D (y − yd (x))2 dxρ(z)dz

∫D (E [y] − yd (x))2 dx

∫D Var (y(x)) dx

udet 4.5422e-04 1.1092e-04 3.4330e-04

u1 9.2445e-05 2.8380e-05 6.4065e-05

u2, β = 0 1.3362e-04 6.9242e-08 1.3355e-04

u2, β = 1 9.2690e-05 2.8197e-05 6.4493e-05

u2, β = 2 1.0383e-04 6.6029e-05 3.7798e-05

4.3 Benefits and Drawbacks of the Cost Functionals

In this section the costs J1 and J2 are compared in terms of their computationalefficiency and ability to provide robust solutions. Concerning the capability of thecosts to provide robust solutions, the first term of J1 implicitly penalizes the fluctua-tions of the state variable which permits to obtain robust solutions without includingexplicitly the variance. This issue has the advantage of avoiding the tunning of theparameter β, a priori unknown. However, the numerical experiments show that thecost J1 can be too conservative since it penalizes the differences with respect tothe target along the entire stochastic domain which notably compromises the matchbetween themean state and the target. Conversely, the cost J2 includes the variance ofthe state variable as a robustness criterion. This second term provides more flexibilitythan the cost J1 to balance the error with respect to the target and the robustness.However, this scalarization requires to solve a set of control problems to find thebest trade-off between robustness and accuracy. With respect to the computationalcomplexity, the cost J2 needs a larger number of stochastic collocation points tofulfill the prescribed accuracy due to the slow rate of convergence of the variance.

4.4 One-Shot Methods

In this section, the numerical resolution of problem (P1), as given by (4.6) and (4.7),is addressed by solving its associated first-order optimality conditions. From now onin this section, we essentially follow the approach in [16].

As is well-known, these optimality conditions are derived by equating to zero thedirectional derivatives of the Lagrangian (4.8) with respect to the state, adjoint andcontrol variables. A straightforward computation leads to the coupled system

[email protected]

Page 86: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.4 One-Shot Methods 71

⎧⎪⎪⎪⎪⎨

⎪⎪⎪⎪⎩

−div (a∇ y) = 1O u, in D × Γ

y = 0, on ∂D × Γ

−div (a∇ p) = y − yd , in D × Γ

p = 0, on ∂D × Γ

γ u + ∫Γp(x, z)ρ(z) dz = 0 in O.

(4.43)

Solving in u the last equation and substituting into the first equation gives the reducedoptimality system

⎧⎨

−div (a∇ y) = − 1γ1O∫Γp(x, z)ρ(z) dz, in D × Γ

−div (a∇ p) = y − yd , in D × Γ

y = p = 0, on ∂D × Γ.

(4.44)

Remark 4.8 Well-posedness of the optimality system (4.44) may be easily obtainedby applying the Brezzi theory on saddle-point problems [5]. See also [6] for a relatedproblem.

The numerical resolution of the optimality system (4.44) may be addressed byusing both a Stochastic Collocation (SC) or a Stochastic Galerkin (SG) finite elementmethod. As it has been illustrated in the preceding section, SC entails the solutionof a number of uncoupled deterministic problems. Thus, SC methods are very use-ful to solve uncoupled problems. However, system (4.44) is coupled via the term∫Γp(x, z)ρ(z) dz. As a consequence, the advantage associated to decoupled sys-

tems, which is associated to the SC method, is lost. It is therefore more convenientto use a SGmethod, which, typically, requires less degrees of freedom in the randomdomain than SC for the same accuracy (see [3] and [16, Remarks 1 and 2] for moredetails).

When applied to problem (4.44), SG finite element methods propose an approxi-mation of the space L2

ρ (Γ ) ⊗ H 10 (D), where both the state variable y(x, z) and the

adjoint state p(x, z) live in, in the formP�(�,N ) (Γ ) ⊗ Vh (D), with

P�(�,N ) (Γ ) = span

{

ψq (z) =N∏

i=1

ψqi (zi ), q = (q1, · · · , qN ) ∈ � (�, N )

}

⊂ L2ρ (Γ )

a multivariate polynomial space, and

Vh (D) = span {φi (x), 1 ≤ i ≤ Mh} ⊂ H 10 (D) ,

a standard finite element space associated to a given triangulationTh of the physicalspace D.

Here, � ∈ N+ denotes a level of approximation, and N is the number of terms inthe truncated KL expansion of a(x, z), e.g., if a(x, z) is a log-normal random field,then

a(x, z) ≈ aN (x, z) = eμ+σ∑N

n=1

√λnbn(x)zn , z = (z1, · · · , zN ) ∈ Γ. (4.45)

[email protected]

Page 87: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

72 4 Numerical Resolution of Robust Optimal Control Problems

�(�, N ) is a multi-index set, and{ψqi (zi )

}∞qi=1 is a family of orthonormal polyno-

mials in L2ρi

(Γi ).

Remark 4.9 Similarly to the stochastic collocation method presented in Sect. 4.2.2,when a(x, z) is given by (4.45) it is very convenient to use an anisotropic polynomialspace, e.g., the onewhichwill be considered in next chapter. For the sake of simplicity,we do not give more details on the polynomial spaceP�(�,N ) (Γ ) in this section.

With all these ingredients, a SG formulation of the optimality system (4.44) readsas: find yhq , phq ∈ P�(�,N ) (Γ ) ⊗ Vh (D) such that

⎧⎨

∫Γ

∫D aN∇ yhq · ∇vhqρ dxdz + 1

γ

∫Γ

∫D 1O

(∫Γ phqρ dz

)vhqρ dxdz = 0,∫

Γ

∫D aN∇ phq · ∇whqρ dxdz − ∫Γ

∫D yhqwhqρ dxdz = − ∫Γ

∫D ydwhqρ dxdz,

∀vhq , whq ∈ P�(�,N ) (Γ ) ⊗ Vh (D) .

(4.46)The matrix formulation of system (4.46) is as follows. Let M ∈ R

Mh×Mh be the massmatrix with entries

[M]i j =∫

Dφiφ j dx, 1 ≤ i, j ≤ Mh,

and let Kn ∈ RMh×Mh , 1 ≤ n ≤ N , be a set of stiffness matrices with entries

[Kn]i j =∫

D

√λnbn (x) ∇φi (x) · ∇φ j (x) dx, 1 ≤ i, j ≤ Mh .

Let us denote by Q = dim(P�(�,N ) (Γ )

)and consider the set of stochasticmatri-

ces Cn ∈ RQ×Q , 1 ≤ n ≤ N , with entries

[Cn]i j =∫

Γ

znψμ(i)(z)ψμ( j)(z)ρ(z) dz, 1 ≤ i, j ≤ Q,

whereμ : {1, 2, · · · , Q} → �(�, N )

is a bijection that assigns a unique integer j to each multi-index μ( j) ∈ �(�, N ).As noticed in Sect. 2.2, the set

{ψμ(1)φ1, · · · , ψμ(1)φMh , ψμ(2)φ1, · · · , ψμ(2)φMh , · · · , ψμ(Q)φ1, · · · , ψμ(Q)φMh

}

(4.47)is a basis ofP�(�,N ) (Γ ) ⊗ Vh (D). Let us denote by ym = (y1,m, · · · , yMh ,m

), pm =(

p1,m, · · · , pMh ,m) ∈ R

Mh , with 1 ≤ m ≤ Q, the degrees of freedom of yhq and phq ,respectively, i.e.,

yhq (x, z) =Mh∑

n=1

Q∑

m=1

yn,mφn(x)ψμ(m)(z), phq (x, z) =Mh∑

n=1

Q∑

m=1

pn,mφn(x)ψμ(m)(z).

[email protected]

Page 88: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.4 One-Shot Methods 73

Substituting these expressions in (4.46) and replacing vhq and whq by the elementsof the basis (4.47), system (4.46) takes the form

⎝N∑

n=1

[Cn 0Q0Q Cn

]⊗ Kn +

[0Q 0Q

−1Q 0Q

]⊗ M + 1

γ

[0Mh×Q T0Mh×Q 0Mh×Q

]⎞

⎜⎜⎜⎜⎜⎜⎝

y1· · ·yQp1· · ·pQ

⎟⎟⎟⎟⎟⎟⎠

=

⎜⎜⎜⎜⎜⎜⎝

0· · ·0yd1· · ·ydQ

⎟⎟⎟⎟⎟⎟⎠

(4.48)where 0Q ∈ R

Q×Q and 0Mh×Q ∈ R(Mh×Q)×(Mh×Q) are zero matrices, 1Q ∈ R

Q×Q isthe identity matrix,

ydj = −(∫

Γψμ( j)ρ dz

Dydφ1 dx, · · · ,

Γψμ( j)ρ dz

DydφMh dx

), 1 ≤ j ≤ Q,

and

T =⎡

⎢⎣

T11 · · · T1Q...

. . ....

TQ1 · · · TQQ

⎥⎦ ∈ R

(Mh×Q)×(Mh×Q)

is a block matrix where the entries of each block Tkl ∈ RMh×Mh , 1 ≤ k, l ≤ Q, are

given by

[Tkl]i j =

Γψμ(l)(z)ρ(z)

Oφi (x)φ j (x)

(∫

Γψμ(k)(z)ρ(z) dz

)dxdz, 1 ≤ i, j ≤ Mh .

Notation. We recall that given two matrices A = (ai j) ∈ R

m×n and B = (bi j) ∈

Rp×q , A ⊗ B denotes the Kronecker product of A and B, which is explicitly given

by

A ⊗ B =

⎢⎢⎢⎢⎢⎢⎢⎢⎢⎢⎢⎢⎢⎢⎢⎢⎢⎢⎣

a11b11 a11b12 · · · a11b1q · · · · · · a1nb11 a1nb12 · · · a1nb1qa11b21 a11b22 · · · a11b2q · · · · · · a1nb21 a1nb22 · · · a1nb2q

......

. . ....

......

. . ....

a11bp1 a11bp2 · · · a11bpq · · · · · · a1nbp1 a1nbp2 · · · a1nbpq...

......

. . ....

......

......

.... . .

......

...

am1b11 am1b12 · · · am1b1q · · · · · · amnb11 amnb12 · · · amnbmq

am1b21 am1b22 · · · am1b2q · · · · · · amnb21 amnb22 · · · amnb2q...

.... . .

......

.... . .

...

am1bp1 am1bp2 · · · am1bpq · · · · · · amnbp1 amnbp2 · · · amnbpq

⎥⎥⎥⎥⎥⎥⎥⎥⎥⎥⎥⎥⎥⎥⎥⎥⎥⎥⎦

.

[email protected]

Page 89: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

74 4 Numerical Resolution of Robust Optimal Control Problems

Table 4.3 Monomials up to2nd degree (in each variable)

1

z1 z2z21 z1z2 z22

z21z2 z22z1z21z

22

The numerical resolution of system (4.48) is typically addressed by using itera-tive preconditioned Krylov (sub)space solvers [16]. The computational cost, in therandom domain, of the SG method presentend in this section is, therefore, given bythe product of stochastic degrees of freedem of system (4.48) times the number ofiterations required by the iterative solver.

4.5 Notes and Related Software

The sparse grid described in Sect. 4.2.2.1 is usually referred in the literature as toanisotropic, Smolyak, sparse grid. Other choices for selecting the number m� ofnodes at the level � and the index set (4.27) are also possible. For a given numberof nodes M , the problem of computing the optimal grid, in the sense of minimizingnumerical error, is therefore a pertinent one. This problem is being object of anintensive research effort [4, 7]. Next, we briefly describe the main idea underlyingSmolyak sparse grid construction.

The idea of Smolyak. The main goal of sparse grids constructions is to keep thesame accuracy as full tensor rules but with a reduced number of nodes and functionsevaluations. The accuracy of a quadrature rule is quantified in terms of the degreeof the polynomials that can be integrated exactly. For instance, to integrate exactlypolynomials p (z1, z2) of degree less than or equal to 2, in two dimensions, one onlyneed to consider the following monomials:

{1, z1, z

21, z2, z

22, z1z2

}.

Thus, terms like z21z22, that arise in a full tensor product, are unnecessary for the

prescribed accuracy. More precisely, the monomials below the line in Table4.3 arenot needed.

In the same vein, the idea of Smolyak was to add low order grids together insuch a way that the resulting quadrature rule (as given by (4.29), or equivalently by(4.30)–(4.33)) is expressed as a linear combinations of full tensor grids, but eachone of them with a relatively low number of points. It is important to point out that,due to the use of the difference operators (4.28), negative weights may appear in thequadrature rule. All of this is schematically represented in Fig. 4.15.

[email protected]

Page 90: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.5 Notes and Related Software 75

Fig. 4.15 Scheme of the isotropic (g = 1) Smolyak quadrature rule (4.30) using Gauss-Hermitenodes in dimension N = 2 and level � = 3

In addition, a Smolyak’s quadrature rule that employs nested points requires fewerfunction evaluations than the corresponding formula with non-nested points. Any-way, the major computational savings come from the fact that most of the pointsin the full tensor grid are removed. This improvement is greater as the dimensionand/or the level increase.

Smoothness and convergence. Convergence results of the collocation method pre-sented in this chapter depend on the regularity of the solution y (x, z) of system(4.18), w.r.t. the random parameter z ∈ Γ . Indeed, with the same notations as inSect. 4.1, let us consider the compact set

Γ ∗n =

N∏

j=1, j �=n

Γ j ,

and let us denote by z∗n an element in Γ ∗

n . For continuous diffusion coefficients,2 i.e.,

a ∈ C0 (Γ ; L∞ (D)), then the solution y(x, z) to (4.18), considered as a functionof the variable zn ∈ Γn with values in the space of continuous functions from Γ ∗

n toH 1

0 (D), i.e.,y = y

(x, ·, z∗

n

) : Γn → C0(Γ ∗n ; H 1

0 (D)),

admits an analytical extension y(x, s, z∗

n

), s ∈ C, in the region of the complex plane

∑(Γn; τn) = {s ∈ C : dist (s, Γn) ≤ τn} , (4.49)

2For instance, when a(x, z) is expressed as a truncated KL expansion of a log-normal randon field(see [2, Example 3]).

[email protected]

Page 91: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

76 4 Numerical Resolution of Robust Optimal Control Problems

and, moreover, there exists c > 0, such that

‖y (s) ‖C0(Γ ∗n ;H 1

0 (D)) ≤ c.

For a proof, we refer to [2, Lemmas 3.1. and 3.2]. It is important to point out that τnin (4.49) depends on the eigenpair {λn, bn (x)}. Typically,

τn = α√λn‖bn‖L∞(D)

, withα > 0.

From this, error estimates, in suitable spatial norms, for the first two statisticalmoments (4.19) may be obtained (see [2, Lemmas 4.7 and 4.8]).

Although for the simple PDEmodel (4.18) considered in this chapter, the region ofanalyticity of y w.r.t. z ∈ Γ may be estimated [2, 4], thus providing a good estimatefor the vector g, as given by (4.25), in many other cases it might not be possible tosharply compute g using a priori information. An easily implementable method forselecting g using a posteriori information is presented in [12].

The methods presented in this chapter have a nice computational performancewhen the number of random inputs is not so large. However, a common criticalcomputational challenge which faces most of stochastic collocation methods is highdimensionality. Indeed,when the number N of input randomvariables becomes large,the number of stochastic degrees of freedom required for a prescribed accuracy levelmight grow exponentially when isotropic (or even anisotropic) sparse grids are used.This phenomenon is known as the curse of dimensionality. A number of methods toovercome, or at least to alleviate, this undesirable performance have been proposedduring the last years (see [1, 7, 8] and the references therein).

For sparse grids constructions and quadrature rules used in all experiments in thischapter we have used the Sparse grids Matlab kit [3] (http://csqi.epfl.ch). For finiteelement analysis we have used the Matlab vectorized codes provided by [14]. Weare indebted to the authors for letting us the use of these two packages.

In the line of Sect. 4.3, for a comparison study of different functionals for randomPDE optimization problems we refer to [10].

Apart from the codes accompanying this book, the interested reader may findadditional open source codes for interpolation and quadrature rules on sparse gridsin the links provided below:

• Sparse grids Matlab kit [3] provided by Scientific Computing and UncertaintyQuantification (CSQI) group. http://csqi.epfl.ch.

• The Dakota software project developed by Sandia National Laboratories. Pro-vides state-of-the-art research and robust, usable software for optimization anduncertainty quantification (UQ). https://dakota.sandia.gov/

• TheSparseGrid InterpolationToolbox is aMatlab toolbox for recovering (approxi-mating) expensive, possibly high-dimensionalmultivariate functions. It was devel-oped by Andreas Klimke at the Institute of Applied Analysis and Numerical

[email protected]

Page 92: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

4.5 Notes and Related Software 77

Simulation, High Performance Scientific Computing lab, Universität Stuttgart.http://www.ians.uni-stuttgart.de/spinterp/

• Computer C++ code in support of paper [16]. Very well-suited for Sect. 4.4.https://www.repository.cam.ac.uk/handle/1810/240740

• TheToolkit forAdaptiveStochasticModeling andNon-IntrusiveApproximation isa collection of robust libraries for high dimensional integration and interpolation aswell as parameter calibration. The project is sponsored byOakRidgeNational Lab-oratory Directed Research and Development as well as the Department of EnergyOffice for Advanced Scientific Computing Research. http://tasmanian.ornl.gov/

• Several useful codes supporting different programming language are provided byProfessor John Burkardt fromDepartment of Scientific Computing at Florida StateUniversity (FSU). https://people.sc.fsu.edu/ jburkardt/

References

1. Alexanderian, A., Petra, N., Stadler, G., Ghattas, O.:Mean-variance risk-averse optimal controlof systems governed by PDEs with random parameter fields using quadratic approximations.SIAM/ASA J. Uncertain. Quantif. 5(1), 1166–1192 (2017)

2. Babuška, I., Nobile, F., Tempone, R.: A Stochastic collocation method for elliptic partial dif-ferential equations with random input data. SIAM Rev. 52(2), 317–355 (2010)

3. Bäck, J., Nobile, F., Tamellini, L., Tempone, R.: Stochastic spectral Galerkin and collocationmethods for PDEs with random coefficients: a numerical comparison. In: Spectral and HighOrder Methods for Partial Differential Equations. Lecture Notes in Computational Science andEngineering, vol. 4362, p. 76. Springer, Heidelberg (2011)

4. Bäck, J., Tempone, R., Nobile, F., Tamellini, L.: On the optimal polynomial approximationof stochastic PDEs by Galerkin and collocation methods. Math. Models Methods Appl. Sci.22(9), 1250023 (2012)

5. Brezzi, F.: On the existence, uniqueness and approximation of saddle-point problems arisingfrom lagrangian multipliers. RAIRO Modél. Math. Anal. Numér. 8(2), 129–151 (1974)

6. Chen, P., Quarteroni, A., Rozza, G.: Stochastic optimal Robin boundary control problems ofadvection-dominated elliptic equations. SIAM J. Numer. Anal. 51(5), 2700–2722 (2013)

7. Chen, P., Quarteroni, A.: A new algorithm for high-dimensional uncertainty quantificationbased on dimension-adaptive sparse grid approximation and reduced basismethods. J. Comput.Phys. 298, 176–193 (2015)

8. Chen, P., Quarteroni, A., Rozza, G.: Comparison between reduced basis and stochastic collo-cation methods for elliptic problems. J. Sci. Comput. 59(1), 187–216 (2014)

9. Hinze, M., Pinnau, R., Ulbrich, M., Ulbrich, S.: Optimization with PDE constraints. In: Math-ematical Modelling: Theory and Applications, vol. 23. Springer, Berlin (2009)

10. Lee, HCh., Gunzburger, M.D.: Comparison of approaches for random PDE optimization prob-lems based on different matching functionals. Comput. Math. Appl. 73(8), 1657–1672 (2017)

11. Lord, G.L., Powell, C.E., Shardlow, T.: An Introduction to Computational Stochastic PDEs.Cambridge University Press (2014)

12. Nobile, F., Tempone, R., Webster, C.G.: An anisotropic sparse grid stochastic collocationmethod for partial differential equations with random input data. SIAM J. Numer. Anal. 46(5),2411–2442 (2008)

13. Novak, E., Ritter, K.: High-dimensional integration of smooth functions over cubes. Numer.Math. 75(1), 79–97 (1996)

14. Rahman, T., Valdman, J.: Fast MATLAB assembly of FEM matrices in 2D and 3D: nodalelements. Appl. Math. Comput. 219, 7151–7158 (2013)

[email protected]

Page 93: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

78 4 Numerical Resolution of Robust Optimal Control Problems

15. Rao, M.M., Swift, R.J.: Probability Theory with Applications, 2nd edn. Springer, New York(2006)

16. Rosseel, E., Wells, G.N.: Optimal control with stochastic PDE constraints and uncertain con-trols. Comput. Methods Appl. Mech. Eng. 213(216), 152–167 (2012)

17. Smith, R.C.: Uncertainty Quantification. Theory, Implementation, and Applications. Compu-tational Science & Engineering, vol. 12. SIAM, Philadelphia (2014)

18. Smolyak, S.: Quadrature and interpolation formulas for tensor product of certain classes offunctions. Dokl. Akad. Nauk SSSR 4, 240–243 (1963)

19. Tröltzsch, F.: Optimal Control of Partial Differential Equations: Theory, Methods and Appli-cations. Graduate Studies in Mathematics, vol. 112. AMS. Providence, Rhode Island (2010)

20. Trefethen, L.N.: Is Gauss quadrature better than Clenshaw Curtis ? SIAM Rev. 50, 67–87(2008)

[email protected]

Page 94: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Chapter 5Numerical Resolution of Risk AverseOptimal Control Problems

Far better an approximate answer to the right question, which isoften vague, than an exact answer to the wrong question, whichcan always be made precise.

John Wilder Tukey.The Future of Data Analysis, 1962.

In this chapter, an adaptive, gradient-based, minimization algorithm is proposed forthe numerical resolution of the risk averse optimal control problem (Pε), as definedin Sect. 3.2.2. The numerical approximation of the associated statistics combines anadaptive, anisotropic, non-intrusive, Stochastic Galerkin approach for the numericalresolution of the underlying state and adjoint equations with a standard Monte-Carlo(MC) sampling method for numerical integration in the random domain.

5.1 An Adaptive, Gradient-Based, Minimization Algorithm

Let yd ∈ L2 (D) be a given target and ε > 0 a threshold parameter. As indicated inthe proof of Theorem 3.5 and taking into account Sect. 4.1, the risk-averse optimalcontrol problem (Pε), as formulated in Sect. 3.2.2, may be written in the form

⎧⎪⎨

⎪⎩

Minimize in u : Jε (u) = ∫

ΓH

(‖y(z) − yd‖2L2(D)

− ε)

ρ(z) dz

subject toy = y (u) solves (4.18), u ∈ Uad ,

(5.1)

where H is the Heaviside function and, for simplicity, Uad = L2 (D).

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2018J. Martínez-Frutos and F. Periago Esparza, Optimal Control of PDEs under Uncertainty,SpringerBriefs in Mathematics, https://doi.org/10.1007/978-3-319-98210-6_5

79

[email protected]

Page 95: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

80 5 Numerical Resolution of Risk Averse Optimal Control Problems

Since the Heaviside step function H (s) is not differentiable at s = 0, the smoothapproximation

Hα (s) =(1 + e− 2s

α

)−1, with 0 < α < 1

is considered. Thus, the cost functional Jε (u) is replaced by

Jαε (u) =

Γ

[

1 + e− 2

α

(‖y(z)−yd‖2L2(D)

−ε)]−1

ρ(z) dz (5.2)

Eventually, problem (5.1) is approximated by

⎧⎪⎪⎨

⎪⎪⎩

Minimize in u : Jαε (u) = ∫

Γ

[

1 + e− 2

α

(‖y(z)−yd‖2L2(D)

−ε)]−1

ρ(z) dz

subject toy = y (u) solves (4.18), u ∈ Uad .

(5.3)

In addition to the non-differentiability issue mentioned above, another difficultywhich arises when trying to implement, in risk averse problems, the gradient algo-rithm proposed at the beginning of Sect. 4.2 is the following one: denoting by uk

the control at iterate k and by yk (z) the solution to (4.18) for the control uk , it isnot surprising that the PDF of ‖yk (·) − yd‖2L2(D)

(typically at the first iterate) beconcentrated in the region where (the approximation of) the Heaviside function isconstant and equal to one (i.e., ‖yk (z) − yd‖2L2(D)

> ε for all z ∈ Γ ). In such a case,(Jαε

)′ (uk

) = 0 and hence no descent direction is available.Apossibility to overcome this difficulty is to adapt, at each iterate k, the parameters

ε and α, which appear in (5.2), to the current location of the PDF of ‖yk (·) −yd‖2L2(D)

. More precisely, at iterate k, the parameter εk may be taken as the mean

value of ‖yk (·) − yd‖2L2(D), and αk proportional to the standard deviation of the same

quantity ‖yk (·) − yd‖2L2(D). Both parameters are then introduced in (5.2). Thus, in

order to reduce the probability of ‖yk (z) − yd‖2L2(D)to exceeding εk , the optimization

algorithm (for εk and αk kept fixed) moves the PDF of the current ‖y (·) − yd‖2L2(D)

towards the left of εk . This way, the new updated εk+1 would be lower than εk . Atthe same time, since ‖y(z) − yd‖2L2(D)

is non-negative, the dispersion of the new

‖yk+1 (·) − yd‖2L2(D)would also be lower than the previous one. See Fig. 5.1 for a

graphical illustration of this idea.This observation motivates the following minimization algorithm for risk averse

optimal control problems:

[email protected]

Page 96: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

5.1 An Adaptive, Gradient-Based, Minimization Algorithm 81

Fig. 5.1 Evolution of the probability density function of I(uk , z

) := ‖yk(z) − yd‖2L2(D)along the

minimization algorithm. Here u∗ denotes the optimal risk averse control

Adaptive, gradient-based, optimization algorithmInitialization: Take an initial u0 ∈ U .Construction: For k = 0, 1, . . . (while εk > ε) do:

2.1: Update parameters ε and α as

{εk = mean value of ‖yk (·) − yd‖2L2(D)

αk = 0.01 × standard deviation of ‖yk (·) − yd‖2L2(D).

2.2: Take as descent direction uk = − (Jαkεk

)′ (uk

).

2.3: Choose a step size λk such that Jαkεk

(uk + λkuk

)< Jαk

εk

(uk

).

2.4: Set uk+1 = uk + λkuk .2.5: Go to 2.1.

When εk ≤ ε, both εk and αk are kept fixed (εk = ε and αk equal to itscurrent value) and the algorithm continues with steps 2.2 to 2.4.

Verification: The stopping criterion is the first k ∈ N for which

|Jαε

(uk

) − Jαε

(uk−1

) |Jαε

(uk

) ≤ η, (5.4)

where η is a prescribed tolerance.If condition εk ≤ ε is not satisfied after a given, large enough, number ofiterates, then the algorithm stops. This is an indication that the prescribedε is too small to be satisfied with controls u ∈ Uad .

[email protected]

Page 97: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

82 5 Numerical Resolution of Risk Averse Optimal Control Problems

5.2 Computing Gradients of Functionals Measuring RiskAversion

The gradient of Jαε (u), which is required in step 2.2. of the above minimization

algorithm, may be obtained by following the same lines as in Sect. 4.2.1. Precisely,for y, p ∈ L2

ρ

(Γ ; H 1

0 (D)), and u ∈ L2 (D), the Lagrangian

L(y, u, p

) = Jαε

(y, u

) −∫

Γ

D

[a∇ y · ∇ p − 1O u p

]dxρ(z)dz, (5.5)

is considered. Let us denote by (y, u, p) a stationary point of L . Equating thederivative ofL w.r.t. y to zero gives the adjoint system

{−div (a∇ p) = C (z) (y − yd) , in D × Γ

p = 0, on ∂D × Γ,(5.6)

with

C (z) = 4

αe− 2

α

(‖y(z)−yd‖2L2(D)

−ε) [

1 + e− 2

α

(‖y(z)−yd‖2L2(D)

−ε)]−2

. (5.7)

The gradient of Jαε is obtained by differentiating the Lagrangian w.r.t. the control u,

which leads to(Jαε

)′(u) (x) =

Γ

p (x, z) ρ(z)dz, x ∈ O, (5.8)

where p solves (5.6).

5.3 Numerical Approximation of Quantities of Interestin Risk Averse Optimal Control Problems

Although the random variableC (z), which appears in (5.7), is theoretically smooth,for α small, numerically it is not. The same problem appears when evaluating thecost functional Jα

ε (u) and its sensitivity at each iteration of the descent algorithm.As a consequence, stochastic collocation methods are not the best choice to computethe involved integrals in the random domain Γ because if a few collocation nodesare taken in the random discontinuities, then the corresponding numerical approxi-mations could be very inaccurate. It is then more convenient to use a MC samplingstrategy. However, the use of a direct MC method requires the numerical resolutionof (4.18) and (5.6) at a large number of random sampling points zk ∈ Γ and at eachiteration of the descent method. This makes a direct MC method unaffordable froma computational point of view. For this reason, the MC method will be combinedwith a Stochastic Galerkin (SG) (also called Polynomial Chaos (PC)) method for

[email protected]

Page 98: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

5.3 Numerical Approximation of Quantities of Interest … 83

uncertainty propagation. More precisely, y(x, z), which is smooth w.r.t. z ∈ Γ , willbe approximated with a SG method, which is computationally cheaper than MCmethod. The same method will be applied to approximate the adjoint state p(x, z),solution to (5.6). Then, the MC method will be applied to these two approximationsin order to compute the integrals in the random domain, which appear in the costfunctional Jα

ε (u) and in its sensitivity. Next subsection provides details on all theseissues.

5.3.1 An Anisotropic, Non-intrusive, Stochastic GalerkinMethod

As in the preceding chapter, having in mind that in most of applications the uncer-tain coefficient a (x, z) is represented as a truncated KL expansion of a log-normalrandom field, in what follows it is assumed that the different stochastic directionsdo not have the same influence in the solution y (x, z) to (4.18). As a consequence,it is natural to use an anisotropic SG method to approximate numerically y (x, z)and its associated adjoint state p (x, z), solution to (5.6). Following [2], this subsec-tion presents the construction of such anisotropic polynomial space. We follow analgorithmic description of the method and refer the reader to [2] for error estimates.

With the same notations as in the preceding chapters, let{ψpn (zn)

}∞pn=0, 1 ≤ n ≤

N , be an orthonormal basis of L2ρn

(Γn) composed of a suitable class of orthonormal

polynomials.1 Since L2ρ (Γ ) = ⊗N

n=1 L2ρn

(Γn), a multivariate orthonormal polyno-mial basis of L2

ρ (Γ ) is constructed as

{

ψp (z) =N∏

n=1

ψpn (zn)

}

p=(p1,...,pN )∈NN

.

The construction of an anisotropic multivariate polynomial space, which should beable to keep accuracy and reduce computational cost, is based on a suitable selectionof the degrees pn of the orthonormal polynomials in each stochastic direction. To thisend, and similarly to the anisotropic stochastic collocation method described in thepreceding chapter, a positive integer ∈ N+ indicating the level of approximationin the random space L2

ρ (Γ ), the vector of weights g = (g1, . . . , gN ) ∈ RN+ for the

different stochastic directions (which for the specific model example considered inthis text is given by (4.25)) and the multi-index set

�g (, N ) ={

p = (p1, . . . , pN ) ∈ NN :

N∑

n=1

pngn ≤ g

}

, (5.9)

1In the numerical experiments of Sect. 5.4, we take ρn (zn) = φ (zn), where φ (zn) is given by(2.30).

[email protected]

Page 99: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

84 5 Numerical Resolution of Risk Averse Optimal Control Problems

with g = min1≤n≤N gn , are considered. Finally, the anisotropic approximation mul-tivariate polynomial space is defined as

P�g(,N ) (Γ ) = span{ψp (z) , p ∈ �g (, N )

}.

This choice of�g (, N ) is referred in the literature [2] as toAnisotropic Total Degreepolynomial space (ATD).

Thus, an approximate solution y (x, z) ∈ P�g(,N ) (Γ ) ⊗ H 10 (D) of (4.18) is

expressed in the form

y(x, z) ≈ y (x, z) =∑

p∈�g(,N )

yp (x) ψp (z) , yp ∈ H 10 (D) , (5.10)

where due to the orthonormality of{ψp (z)

}

p∈�g(,N ),

yp (x) =∫

Γ

y (x, z) ψp (z) ρ (z) dz ≈∫

Γ

y (x, z) ψp (z) ρ (z) dz. (5.11)

This latter integral is numerically approximated using the anisotropic sparse gridcollocation method described in the preceding chapter. Eventually, we get a fullydiscretized solution

yh (x, z) =∑

p∈�g(,N )

yhp (x) ψp (z) ∈ P�g(,N ) (Γ ) ⊗ Vh (D)

with Vh (D) a standard finite element space approximating H 10 (D), and

yhp (x) =∫

Γ

yh (x, z) ψp (z) ρ (z) dz ≈ Ag ( + 1, N ) yh (x) (5.12)

a quadrature rule as in (4.35). Here, yh (·, z) ∈ Vh (D) denotes the finite elementapproximation of y (·, z) at the point z ∈ Γ.

Notice that the multi-index set used in the quadrature rule (5.12) is Yg ( + 1, N ),which is suitable to integrate, by using Gauss-Hermite nodes and weights, polyno-mials in P�g(,N ) (Γ ).

Remark 5.1 We emphasize that to approximate numerically (5.12), the solution y to(4.18) has to be computed at a number of collocation nodes zk ∈ Γ . This computa-tional task may, therefore, be carried out in parallel computers. Due to the regularityof y w.r.t. z, the number of sampling points zk required by the stochastic collocationmethod is lower than the one needed by the Monte Carlo method.

[email protected]

Page 100: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

5.3 Numerical Approximation of Quantities of Interest … 85

5.3.2 Adaptive Algorithm to Select the Level ofApproximation

The level of approximation is adaptively chosen in a similar fashion as in Sect.4.2.2.2. For the sake of clarity and completeness, it is briefly described next.

Adaptive algorithm to select the level of approximation1. Computation of an enriched solution yh

(x, z) of (4.18) by following the

method described above for approximation in the randon domain Γ andby using finite elements for discretization in the physical domain D. Thisenriched solution (which plays the role of exact solution) is then used toobtain an approximation Jα,h

ε,(u) of Jα

ε (u) by using MC method, where

random samplings zk are applied to yh

(x, z). As in the preceding chapter,the reference control u = udet is taken as the solution of the associateddeterministic optimal control problem (4.42).

2. The level is linearly increased (from = 1 to = opt < ) until thestopping criterion

|Jα,hε, (u) − Jα,h

ε,(u)|

Jα,hε,

(u)≤ δ, (5.13)

where 0 < δ � 1 is a prescribed tolerance, is satisfied.In the case that the stopping criterion (5.13) does not hold for any < ,then a larger is taken to ensure that the enriched solution yh

(x, z) is good

enough. The whole process is then repeated taking the new value of asreference.

The use of a Monte Carlo sampling method to approximate the integrals in therandom domain Γ , which appear in the cost functional (5.2) and in its sensitivity(5.8), introduces an additional approximation error. Therefore, the number of MCsampling points used must be properly chosen.

5.3.3 Choosing Monte Carlo Samples for NumericalIntegration

Let F : Γ → R be a multivariate function. When applied to the numerical approxi-mation of

I := E [F (z)] =∫

Γ

F (z) ρ (z) dz, (5.14)

[email protected]

Page 101: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

86 5 Numerical Resolution of Risk Averse Optimal Control Problems

the classical Monte Carlo method amounts to drawing M random points zk ∈ Γ ,which are distributed according to the PDF ρ (z), and take

1

M

M∑

k=1

F(zk

)(5.15)

as an approximation of (5.14).This approach is supported by the Strong Law of Large Numbers. Indeed, let{

ξ j : � → Γ}

j≥1 be a sequence of i.i.d. random variables and let ρ = ρ (z) itsassociated joint probability density function (see Sect. 4.1 for details). Consider thesample average

I M := 1

M

M∑

j=1

F(ξ j

). (5.16)

Then, by the Strong Law of Large Numbers [7, Chap. 20],

limM→∞ I M = I a.s. and in L2

ρ (Γ ) .

Now, given small constants T OLMC > 0 and δMC > 0, this section addressesthe problem of choosing M = M (T OLMC , δMC ) large enough so that the errorprobability satisfies

P(|I M − I | > T OLMC

) ≤ δMC . (5.17)

Denoting by σ 2 the variance of F(ξ j

), the Central Limit Theorem [7, Chap 21]

states that √M

(I M − I

)

σ⇀ ν in distribution, (5.18)

where ν ∼ N (0, 1) follows a standard Gaussian distribution, i.e.,

limM→∞E

[

g

(√M

(I M − I

)

σ

)]

= E [g (ν)] ,

for all bounded and continuous functions g.As a consequence (see e.g. [7, Th. 18.4]), for M large enough,

P(|I M − I | > T OLMC

) = P

(|√M(I M−I)

σ| >

√MTOLMC

σ

)

= 1 − P

(|√M(I M−I)

σ| ≤

√MT OLMC

σ

)

≈ 1 − P

(|ν| ≤

√MT OLMC

σ

)

= 2(1 − �

(√MT OLMC

σ

)).

[email protected]

Page 102: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

5.3 Numerical Approximation of Quantities of Interest … 87

All these arguments motivate the following sample variance based algorithm[3, 6] for selecting M such that (5.17) holds:

Sample variance based algorithm to select MInput: T OLMC > 0, δMC > 0, initial number of samplesM0 and the cumu-

lative distribution function of the standard Gaussian variable �(x).Output: I M .

Set n = 0, generate Mn samples zk and compute the sample variance

σ 2Mn

= 1

Mn − 1

Mn∑

k=1

(F

(zk

) − I Mn

)2. (5.19)

while 2(1 − �

(√MnT OLMC/σ Mn

))> δMC do

Set n = n + 1 and Mn = 2Mn−1.Generate a batch of Mn i.i.d. samples zk .Compute the sample variance σ 2

Mnas given by (5.19).

end whileSet M = Mn , generate samples zk , 1 ≤ k ≤ M and compute the output I Mas given by (5.15).

5.4 Numerical Experiments

From now on in this Section, it is assumed that the physical domain D = (0, 1)2 isthe unit square and the control region is a circle centred at the point (0.5, 0.5) andof radius r = 0.25. The uncertain parameter a (x, z) is the truncated KL expansionconsidered in (2.29), with N = 5, which lets capture 89% of the energy field. Thecorrelation lengths are L1 = L2 = 0.5, and the mean and variance of the randomfield are constant and equal to 1. The random domain is Γ = [−3, 3]5, and

ρ (z) =5∏

n=1

φ (zn) , z = (z1, . . . , z5) ∈ Γ, (5.20)

where φ (zn) is the truncated Gaussian (2.30) . The target function yd is the same asin the preceding chapter (see Figure 4.4a).

The threshold parameter ε is taken as ε = 10−4. The vector of weights g, whichis needed to construct the multi-index set (5.9) is obtained from (4.25) and is givenby

g = (0.3894, 0.5478, 0.5478, 0.7708, 0.8219) . (5.21)

[email protected]

Page 103: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

88 5 Numerical Resolution of Risk Averse Optimal Control Problems

1 2 3 4 50.3

0.4

0.5

0.6

0.7

0.8

0.9

1

1e-5 1e-4 1e-3 1e-2 1e-10

2000

4000

6000

8000

10000

12000(a) (b)

Fig. 5.2 (a) Convergence history of the algorithm, and (b) evolution of the PDF of the randomvariable ‖y (z) − yd‖L2(D)

The enriched level used in the adaptive algorithm of Sect. 5.3.2 is taken as = 7,which, for a tolerance δ = 10−3 in (5.13), leads to = 5.

The number M of random Monte Carlo sampling points, which are used fornumerical integration of the cost functional and its sensitivity, is taken as M = 105.For the case of the cost functional (5.2) and being y = y(udet ) the state associatedto udet (solution to problem (4.42), the stopping criterion in the sample variancealgorithm of Sect. 5.3.3 is satisfied with T OLMC = 0.005 and δMC = 10−3.

The step-size parameterλk is chosen at each iterate in such away that it ensures thedecrease of the cost function. In the numerical experiments that follow the selectionofλk is done using a straightforward strategy: in a first stage, an acceptable step length(acceptable meaning that the cost functional decreases) is found after a finite numberof trials. Once an acceptable step length is achieved, a second stage is performed tosearch a value that improves this step length. In this new search, the acceptable steplength is increased by multiplying it by a scale factor while the cost functional isreduced w.r.t. the previous iteration. The reader is referred to the computer codes thataccompany the text for the specific test examples here included and to [12, Chapter3] for many other possibilities to select λk .

The algorithm is initiated with a constant control u0(x) = 0.1, x ∈ O . Since forthis initial control, 0 < Jε(u0) < 1 (as it may be observed in Fig. 5.2b), the parameterε is kept fixed during the optimization process. After 4 iterates, the stopping criterion(5.4) is satisfied for η = 10−4.

Figure 5.2a displays the convergence history of the gradient-based algorithm ofSect. 5.1. The evolution of the PDF of the random variable ‖y (z) − yd‖L2(D) isplotted in Fig. 5.2b. The optimal risk-averse optimal control, solution to (5.3), isplotted in Fig. 5.3. The value of the objective function (5.2) at this risk-averse controlis 0.3012.

[email protected]

Page 104: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

5.5 Notes and Related Software 89

Fig. 5.3 Optimal risk-aversecontrol, solution to (5.3)

5.5 Notes and Related Software

The notion of risk aversion finds its origin in Economy [13]. This concept was thenimported to Shape Optimization [5, 10] and to Control of Random PDEs (see [8, 9]and the references therein). Numerical resolution methods for this type of problemsmight be considered to be in its infancy. In this sense, the methods presented in thischapter are only a first step in this very challenging topic.

The risk functional considered in this chapter is only one among many otherpossible functionals measuring risk aversion (see e.g. [13, Chap. 2] and [5, 8, 11]).

Similarly to stochastic collocation methods, a primary drawback of stochasticGalerkin methods is the well known curse of dimensionality phenomenon, accordingto which the number of unknown coefficients highly increases as the space of randomvariables becomes higher. In this respect, optimality of ATD (and variants of it) incomparisonwith other possible choices for the index set�g (, N ), which determinesthe degrees of polynomials used for approximation in the random space, has beenstudied in [1, 2].

Breaking the curse of dimensionality in the computation of risk-averse functionalsis a very challenging problem. The reader is referred to [4] for an interesting reviewon this topic.

For the different notions of convergence of sequences of random variables andtheir relationships, the reader is referred to [7].

A drawback of the algorithm presented in Sect. 5.3.3 is that the sample varianceσ Mn is used instead of the true variance σ , which is unknown. A higher order basedalgorithm has been proposed in [3].

In addition to the codes accompanying this book, open source codes for StochasticGalerkin (or Polynomial Chaos) methods are provided in the following links:

[email protected]

Page 105: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

90 5 Numerical Resolution of Risk Averse Optimal Control Problems

• FERUM—http://www.ce.berkeley.edu/projects/ferum/ .• OpenTURNS—http://www.openturns.org/.• UQlab—http://www.uqlab.com/.• Dakota software: https://dakota.sandia.gov/.• Professor John Burkardt: https://people.sc.fsu.edu/jburkardt/.

References

1. Bäck, J., Nobile, F., Tamellini, L., Tempone, R.: Stochastic spectral Galerkin and colloca-tion methods for PDEs with random coefficients: a numerical comparison. Spectral and highorder methods for partial differential equations, vol. 4362, Lect. Notes Comput. Sci. Eng., 76,Springer, Heidelberg (2011)

2. Bäck, J., Tempone, R., Nobile, F., Tamellini, L.: On the optimal polynomial approximationof stochastic PDEs by Galerkin and collocation methods. Math. Models Methods Appl. Sci.22(9), 1250023 (2012)

3. Bayer, C., Hoel, H., Schwerin, E.V., Tempone, R.: On non-asymptotic optimal stopping criteriain Monte Carlo simulations. SIAM J. Sci. Comput. 36(2), 869–885 (2014)

4. Chen, P., Quarteroni, A.: Accurate and efficient evaluation of failure probability for partialdifferential equations with random input. Comput. Methods Appl. MEch. Engr. 267, 233–260(2013)

5. Conti, S., Held, H., Pach, M., Rumpf, M., Schultz, R.: Risk averse shape optimization. SIAMJ. Control Optim. 49(3), 927–947 (2011)

6. Fishman, G.S.: Monte Carlo: Concepts, Algorithms, and Applications. Springer, New York(1996)

7. Jacod, J., Protter, P.: Probability Essentials, 2nd edn. Springer, Berlin, Heidelberg (2003)8. Kouri, D.P., Surowiec, T.M.: Risk-averse PDE-constrained optimization using the conditional

value-at-risk. SIAM J. Optim. 26(1), 365–396 (2016)9. Marín, F.J., Martínez-Frutos, J., Periago, F.: A polynomial chaos approach to risk-averse piezo-

electric control of random vibrations of beams. Int J NumerMethods Eng. 1–18 (2018). https://doi.org/10.1002/nme.5823

10. Martínez-Frutos, J., Herrero-Pérez, D., Kessler,M., Periago, F.: Risk-averse structural topologyoptimization under randomfields using stochastic expansionmethods. Comput.Methods Appl.Mech. Engr. 330, 180–206 (2018)

11. Martínez-Frutos, J., Kessler, M., Periago, F.: Robust optimal shape design for an elliptic PDEwith uncertainty in its input data. ESAIM: COCV, 21(4), 901–923 (2015)

12. Nocedal, J., Wright, S.J.: Numerical Optimization. Springer Series in Operations Research.Springer, New York (1999)

13. Pflug, G.Ch., Römisch, W.: Modeling, Measuring, and Managing Risk. World Scientific Pub-lishing Co. Pte. Ltd. (2007)

[email protected]

Page 106: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Chapter 6Structural Optimization UnderUncertainty

All human things hang on a slender thread, the strongest fallwith a sudden crash.

Ovid.(43 B.C-18 A.C)

Motivated by its applications in Engineering, since the early’s 70, optimal designof structures has been a very active research topic. Roughly speaking, the struc-tural optimization problem amounts to finding the optimal distribution of materialwithin a design domain such that a certain objective function or mechanical criterionis minimized. To ensure robustness and reliability of the final designs, a numberof uncertainties should be accounted for in realistic models. These uncertaintiesinclude, among others, manufacturing imperfections, unknown loading conditionsand variations of material’s properties.

This chapter aims at illustrating how the ideas and techniques elaborated in thistext apply to the analysis and numerical resolution of the structural optimizationproblem under uncertainty. We would like to highlight that the goal of this chapteris not to provide a deep review of the techniques and results involved in structuraloptimization, but showing how the techniques described in the preceding chaptersmay be extended to the numerical resolution of the structural optimization problemunder uncertainty. Most of the material of this chapter has been taken from [18, 19].

6.1 Problem Formulation

Let (Ω,F ,P) be a complete probability space, and let O ⊂ Rd (d = 2 or 3 in

applications) be a bounded Lipschitz domain whose boundary ∂O is decomposedinto three disjoint parts

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2018J. Martínez-Frutos and F. Periago Esparza, Optimal Control of PDEs under Uncertainty,SpringerBriefs in Mathematics, https://doi.org/10.1007/978-3-319-98210-6_6

91

[email protected]

Page 107: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

92 6 Structural Optimization Under Uncertainty

∂O = �D ∪ �N ∪ �0, |�D| > 0, (6.1)

where | · | stands for the Lebesgue measure. Consider the system of linear elasticitywith random input data

⎧⎪⎪⎪⎨

⎪⎪⎪⎩

−div(Ae(u(x, ω))) = f in O × Ω,

u(x, ω) = 0 on �D × Ω,

(Ae(u(x, ω))) · n = g on �N × Ω,

(Ae(u(x, ω))) · n = 0 on �0 × Ω,

(6.2)

with u = (u1, . . . , ud) the displacement vector field, e (u) = (∇uT + ∇u)/2 the

strain tensor and n the unit outward normal vector to ∂O . The material Hooke’s lawA is defined for any symmetric matrix ζ by

Aζ = 2μζ + λ(Trζ )I d,

where λ = λ(x, ω) and μ = μ(x, ω) are the Lamé moduli of the material, whichmay depend on both x = (x1, . . . , xd) ∈ O and ω ∈ Ω . Thus, Ae (u) is the stresstensor. In the same vein, the volume and surface loads f and g depend on a spatialvariable x and on a random event ω ∈ Ω , i.e. f = f (x, ω) and g = g(x, ω). Thedivergence (div) and the gradient ∇ operators in (6.2) involve only derivatives withrespect to the spatial variable x .

Since the domainO will change during the optimization process, λ,μ and f mustbe known for all possible configurations ofO . Thus, a working bounded domain D isintroduced, which contains all admissible domains O and satisfies that �D ∪ �N ⊂∂D. Therefore, it is assumed that λ, μ and f are defined in D. Both �D and �N arekept fixed during optimization.

The following assumptions on the uncertain input parameters of system (6.2) areconsidered:

(A1) λ,μ ∈ L∞P

(Ω; L∞(D)) and there exist μmin , μmax , λmin , λmax such that

0 < μmin ≤ μ(x, ω) ≤ μmax < ∞ a.e. x ∈ D, a.s. ω ∈ Ω,

and

0 < λmin ≤ 2μ(x, ω) + dλ(x, ω) ≤ λmax < ∞ a.e. x ∈ D, a.s. ω ∈ Ω,

(A2) f = f (x, ω) ∈ L2P

(Ω; L2(D)d

),

(A3) g = g(x, ω) ∈ L2P

(Ω; L2(�N )d

).

Next, the Hilbert space

VO = {v ∈ H 1(O)d : v|�D = 0 in the sense of traces

},

[email protected]

Page 108: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

6.1 Problem Formulation 93

equipped with the usual H 1(O)d -norm, is introduced.Similarly to Sect. 3.1.1, a weak solution of (6.2) is a random field u ∈ L2

P(Ω; VO )

such that∫

Ω

OAe(u(x, ω)) : e(v(x, ω)) dxdP(ω) =

Ω

[∫

Of · v dx +

�N

g · v ds]

dP(ω),

(6.3)for all v ∈ L2

P(Ω; VO ).

Notation. We recall that given two matrices A = (ai j

)

1≤i, j≤d and B = (bi j

)

1≤i, j≤d ,

A : B = ∑di, j=1 ai j bi j denotes de full contraction of A and B. If u, v ∈ R

d , then

u · v = ∑di=1 uivi is the scalar product between vectors u and v.

By using Korn’s inequality and following the same lines as in the deterministiccase [1, Theorem 2.24], the well-posedness of (6.2) may be easily proved. Moreover,there exists c > 0 such that

‖u‖L2P(Ω;VO ) ≤ c

(‖ f ‖L2

P(Ω;L2(O )d) + ‖g‖L2P(Ω;L2(�N )d)

)

and

‖u(ω)‖VO ≤ c(‖ f (ω)‖L2(O )d + ‖g(ω)‖L2(�N )d

), a.s. ω ∈ Ω. (6.4)

Notice that, thanks to assumption (A1), the constant c = c(O, μmin, λmin) in (6.4)does not depend on ω ∈ Ω .

As is usual in structural optimization, a volume constraint is imposed on an admis-sible domain. Thus, the class of admissible domains is

Uad = {O ⊂ D, �D ∪ �N ⊂ ∂O, |O| = V0} , with 0 < V0 < |D|.

For a given O ∈ Uad , let uO = uO (x, ω) be the weak solution of (6.2). For eachrealization ω ∈ Ω , consider the compliance cost functional

J (O, ω) =∫

Of (x, ω) · uO (x, ω) dx +

�N

g(x, ω) · uO (y, ω) ds. (6.5)

Similarly to Chap. 4, the following robust-type cost functional is introduced:

JRα(O) = E(J (O, ω)) + αVar(J (O, ω)), (6.6)

where α ≥ 0 is a weighting parameter,

E(J (O, ω)) =∫

Ω

J (O, ω) dP(ω),

is the expectation of the compliance, and

[email protected]

Page 109: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

94 6 Structural Optimization Under Uncertainty

Var(J (O, ω)) =∫

Ω

J (O, ω)2 dP(ω) −(∫

Ω

J (O, ω) dP(ω)

)2

is its variance.

Remark 6.1 Notice that in order for the variance to be well-defined we shouldassume that f ∈ L4

P

(Ω; L2(D)d

)and g ∈ L4

P

(Ω; L2(�N )d

). Consequently, from

now on in this chapter, whenever the variance is considered, it is assumed thatf ∈ L4

P

(Ω; L2(D)d

)and g ∈ L4

P

(Ω; L2(�N )d

).

Next, the following risk-averse (also called excess probability or failure probabil-ity in the context of structural optimization) objective function is considered:

JEPη(O) = E (H (J (O, ω) − η)) (6.7)

where η > 0 is a threshold parameter, and H is the Heaviside function.With all these ingredients, the robust and risk-averse structural optimization prob-

lems are formulated as:

(Pα) min{JRα

(O) : O ∈ Uad},

and(PEPη

) min{JEPη

(O) : O ∈ Uad}.

Remark 6.2 Notice that amajor difference of these two optimization problems, com-pared to the control problems studied in the preceding chapters, is that the optimiza-tion variable is not a function but a domain.

6.2 Existence of Optimal Shapes

If an admissible domain is assumed to be only Lebesgue measurable, then (Pα)

and (PEPη) may be ill-posed due to the lack of closure of the set of admissible

domains with respect to the weak- topology in L∞(D; {0, 1}). To overcome thisdifficulty and so guarantee the existence of solution, additional conditions, such assmoothness, geometrical or topology constraints, may be imposed on the class ofadmissible designs. As an illustration of these ideas, the standard ε-cone property[5] is introduced next.

Let a point y ∈ Rd , a vector ξ ∈ R

d and ε > 0 be given. Consider the cone

C(y, ξ, ε) = {z ∈ R

d : (z − y, ξ) ≥ cos ε|z − y|, 0 < |z − y| < ε}.

An open setO ⊂ Rd is said to have the ε-cone property if for all x ∈ ∂O there exists

a unit vector ξx such that ∀y ∈ O ∩ B(x, ε), the inclusion C(y, ξx , ε) ⊂ O holds.

[email protected]

Page 110: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

6.2 Existence of Optimal Shapes 95

Consider the class of domains

Oε = {O open, O ⊂ D, O has the ε-cone property} .

Eventually, from now on in this section, the class of admissible domains Uad isreplaced by

U εad = {O ∈ Oε, �D ∪ �N ⊂ ∂O, |O| = V0} , with 0 < V0 < |D|.

Following [13, Definition 2.2.3], let {On}n≥1 and O be (Lebesgue) measurable setsof Rd . It is said that {On} converges to O in the sense of characteristic functions asn → ∞ if

1On −→ 1O in L ploc

(R

d) ∀p ∈ [1,∞[,

where 1B stands for the characteristic function of a measurable set B ⊂ Rd .

The class of domains Oε enjoys the following compactness property: let {On}n≥1

be a sequence in the class Oε. Then, up to a subsequence, still labelled by n, On

converges to some O ∈ Oε, in the sense of characteristic functions (also in thesense of Hausdorff distance and in the sense of compacts sets). We refer the readersto [13, Theorem 2.4.10] for more details on all these classes of convergence and theirrelationships.

Another very important property that the domains in the class Oε satisfy is thefollowing uniform extension property [5, Theorem II.1] : there exists a positiveconstant K , which depends only on ε, such that for all O ∈ Oε there exists a linearand continuous extension operator

PO : H 1 (O) → H 1(R

d), with ‖PO ‖ ≤ K . (6.8)

We are now in a position to prove the following existence result:

Theorem 6.1 Problems (Pα) and (PEPη) have, at least, one solution.

Proof Let On ⊂ U εad be a minimizing sequence. Due to the compactness property

of Oε mentioned above there exists O ∈ Oε and a subsequence, still labelled by n,such that On converges to O. Moreover, thanks to the convergence in the sense ofcharacteristic functions,O satisfies the volume constraint and henceO ∈ U ε

ad . Thedomain O is the candidate to be the minimizer we are looking for.

For an arbitrary but fixed random event ω ∈ Ω the following convergence holds:

J (On, ω) → J(O, ω

)as n → ∞. (6.9)

Indeed, let un (x, ω) and uO (x, ω) be the solutions to (6.2) associated to On andO, respectively. Denoting by un (·, ω) := POn (un) (·, ω) the extension of un (·.ω)

to Rd , by (6.4) and (6.8) one has

‖un(ω)‖VD ≤ ‖POn‖‖un(ω)‖VO n≤ Kc

(‖ f (ω)‖L2(D)d + ‖g(ω)‖L2(�N )d),

[email protected]

Page 111: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

96 6 Structural Optimization Under Uncertainty

a.s. ω ∈ Ω , which proves that un(ω) is bounded in VD . Thus, by extracting a sub-sequence, it converges weakly in VD , and strongly in L2(D)d to some u(ω) ∈ VD .Let us prove that

u|O (ω) = uO (ω), ω ∈ Ω.

From the definition of un , it follows that

On

Ae(un) : e(v) dx =∫

On

f · v dx +∫

�N

g · v dx ∀v ∈ VD, a.s. ω ∈ Ω,

which is equivalent to

D1On Ae(un) : e(v) dx =

D1On f · v dx +

�N

g · v dx ∀v ∈ VD, a.s. ω ∈ Ω.

The weak convergence of un in VD and the strong convergence in L2(D) of thecharacteristic functions permit to take the limit leading to

D1O Ae(u) : e(v) dx =

D1O f · v dx +

�N

g · v dx ∀v ∈ VD, a.s. ω ∈ Ω,

and also, thanks to the extension property (6.8),

O

Ae(u) : e(v) dx =∫

O

f · v dx +∫

�N

g · v dx , ∀v ∈ VO .

This proves that u|O (ω) = uO (ω) a.s. ω ∈ Ω . Since this is valid for any subse-quence, the whole sequence un converges to uO . Let us consider the compliance inthe form

J (On, ω) =∫

On

f · un dx +∫

�N

g · un ds =∫

D1On f · un dx +

�N

g · un ds.

Since 1On f → 1O f strongly in L2 (D) and un ⇀ u weakly in VD , the limit in thisexpression can be taken to obtain (6.9).

Let us first consider the cost functional JRα. The two contributions of this cost

functional are studied separately as follows:

• Expectation of the compliance: by the Cauchy-Schwartz inequality, the continuityof the trace operator and (6.4),

[email protected]

Page 112: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

6.2 Existence of Optimal Shapes 97

J (On, ω) = ∫

Onf (x, ω) · uOn (x, ω) dx + ∫

�Ng(x, ω) · uOn (x, ω) ds

≤ ‖ f (ω)‖L2(D)d‖uOn (ω)‖L2(On)d + ‖g(ω)‖L2(�N )d‖uOn (ω)‖L2(�N )d

≤ c‖ f (ω)‖L2(D)d(‖ f (ω)‖L2(D)d + ‖g(ω)‖L2(�N )d

)

+c‖g(ω)‖L2(�N )d(‖ f (ω)‖L2(D)d + ‖g(ω)‖L2(�N )d

) ∈ L1P(Ω).

(6.10)Combining (6.9) and (6.10), by dominated convergence, it follows that

E(J (On, ω)) =∫

Ω

J (On, ω) dP(ω) → E(J (O, ω)) =∫

Ω

J (O, ω) dP(ω).

• Variance of the compliance: from (6.9) it follows that

J (On, ω)2 → J(O, ω

)2a.s. ω ∈ Ω.

Hence, by Remark 6.1 and (6.10),

Ω

J (On, ω)2 dP(ω) →∫

Ω

J (O, ω)2 dP(ω)

and

Var(J (On, ω)) =∫

Ω

J (On, ω)2 dP(ω) −(∫

Ω

J (On, ω) dP(ω)

)2

→ Var(J (O, ω)).

Regarding the cost functional JEPη, since Fn(ω) := H (J (On, ω) − η) are inte-

grable and non-negative, by Fatou’s lemma,

Ω

lim infn→∞ Fn(ω) dP(ω) ≤ lim inf

n→∞

Ω

Fn(ω) dP(ω). (6.11)

By (6.9) and the lower semi-continuity of the Heaviside function,

H(J (O, ω) − η

) ≤ lim infn→∞ Fn(ω),

Integrating in both sides of this expression and using (6.11) yields

Ω

H(J (O, ω) − η

)dP(ω) ≤

Ω

lim infn→∞ Fn(ω) dP(ω)

≤ lim infn→∞

Ω

Fn(ω) dP(ω).

Since On is a minimizing sequence, the result follows. �

[email protected]

Page 113: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

98 6 Structural Optimization Under Uncertainty

6.3 Numerical Approximation via the Level-Set Method

In this section, a gradient-based method is used for the numerical approximation ofproblems (Pα) and (PEPη

). Thus, the shape derivatives of the two considered costfunctionals must be computed.

6.3.1 Computing Gradients of Shape Functionals

In his seminal paper [12], Hadamard introduced the concept of shape derivative ofa functional with respect to a domain. This notion was further exploited in [20]. Inthis section, we briefly recall a few basic facts about shape derivatives, which areneeded to compute the gradients of the shape functionals considered in this chapter.For further details on the notion of shape derivative and its properties we refer thereader to [1, 25].

We recall that the shape derivative of a functional J : Uad → R atO is the Fréchetderivative at θ = 0 in W 1,∞ (

Rd;Rd

)of the mapping θ → J ((I + θ) (O)), i.e.,

J ((I + θ) (O)) = J (O) + J ′ (O) (θ) + o (θ) , with limθ→0

|o (θ) |‖θ‖W 1,∞

= 0,

where I : Rd → Rd is the identity operator, θ ∈ W 1,∞ (

Rd;Rd

)is a vector field, and

J ′ (O) is a continuous linear form on W 1,∞ (R

d;Rd).

The following result is very useful when it comes to calculating shape derivatives.For a proof, the reader is referred to [1, Proposition 6.22].

Proposition 6.1 LetO bea smoothboundedopendomainand letφ(x) ∈ W 1,1(R

d).

The shape functional

J (O) =∫

Oφ(x) dx

is shape differentiable at O and its shape derivative is given by

J ′ (O) (θ) =∫

∂Oθ(x) · n (x) φ(x) ds, θ ∈ W 1,∞ (

Rd;Rd

). (6.12)

The volume constraint, as given in the definition of Uad , is incorporated in thecost functional through an augmented Lagrangian method. Accordingly, for the caseof problem (Pα), the following augmented Lagrangian function is considered:

J LRα

(O) = JRα(O) + L1

( ∫

Odx − V0

)+ L2

2

( ∫

Odx − V0

)2. (6.13)

For problem (PEPη), similarly to Chap. 5, the smooth approximation

[email protected]

Page 114: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

6.3 Numerical Approximation via the Level-Set Method 99

J L ,εEPη

(O) = ∫

Ω

(1 + e− 2

ε(J (O ,ω)−η)

)−1dP(ω)

+L1(∫

O dx − V0) + L2

2

(∫

O dx − V0)2

,(6.14)

with 0 < ε � 1, is considered.In both cases, L = (L1, L2), being L1 and L2 the Lagrange multiplier and the

penalty parameter, respectively, used to enforce the volumeconstraint at convergence.The Lagrange multiplier will be updated at each iteration n according to L(n+1)

1 =Ln1 + Ln

2

( ∫

O dx − V0

). The penalty multiplier L2 will be augmented, after every

five iterations, multiplying its previous value by 1.2.

Remark 6.3 Since during the optimization process �D and �N are kept fixed, thedeformation vector field θ ∈ W 1,∞ (

Rd;Rd

)satisfies θ = 0 on �D ∪ �N .

Theorem 6.2 The following assertions hold:

(i) The shape derivative of the cost functional J LRα

(O) is given by

(J LRα

)′(O)(θ) = ∫

�0θ · n [∫

Ω (2 f · u − Ae(u) : e(u)) dP(ω)]ds

+α∫

�0θ · n [∫

Ω (2J (O, ω) f · u − f · p + Ae(u) : e(p)) dP(ω)]ds

−2αE(J (O, ω))∫

�0θ · n [∫

Ω (2 f · u − Ae(u) : e(u)) dP(ω)]ds

+L1∫

�0θ · n ds + L2

( ∫

O dx − V0) ∫

�0θ · n ds

(6.15)where J (O, ω) is given by (6.5) and the adjoint state p ∈ L2

P(Ω; VO ) solves

the system

Ω

O Ae(p) : e(v) dxdP(ω)

= −2∫

Ω

(∫

O f · u dx + ∫

�Ng · u ds

) (∫

O f · v dx + ∫

�Ng · v ds

)dP(ω),

(6.16)all v ∈ L2

P(Ω; VO ).

(ii) The shape derivative of J L ,εEPη

(O) (θ) is expressed as

(J L ,εEPη

)′(O)(θ) = ∫

�0θ · n [∫

ΩC1 (ω) (2 f · u − Ae(u) : e(u)) dP(ω)

]ds

+L1∫

�0θ · n ds + L2

(∫

O dx − V0) ∫

�0θ · n ds,

(6.17)where the random variable C1 (ω) is given by

C1 (ω) = 2

ε

(1 + e− 2

ε(J (O ,ω)−η)

)−2e− 2

ε(J (O ,ω)−η). (6.18)

Proof To simplify, a brief formal proof using the Lagrange method due to Céa [3]is provided next. We consider only the functional J L

Rα. The proof for J L ,ε

EPηis similar

on all points so that it is omitted here (see [19] for details).

[email protected]

Page 115: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

100 6 Structural Optimization Under Uncertainty

The first term in the right hand side (r.h.s.) of (6.15), which is the one associated tothe expectation of the compliance, is obtained, after changing the order of integration,from Proposition 6.1.

For the second term in (6.15), the Lagrangian

L(O, u, p

) = ∫

Ω

[∫

O f · u dx + ∫

�Ng · u ds

]2dP (ω)

+ ∫

O

ΩAe

(u) : e (

p)dP (ω) dx

− ∫

O

Ωp · f dP (ω) dx

− ∫

�N

Ωp · g dP (ω) ds

(6.19)

is considered, where u, p ∈ L2P(Ω; H 1(Rd;Rd)) satisfy u = p = 0 on �D . For a

givenO , denoting by (u, p) a stationary point ofL and equating to zero the derivativeofL with respect to u in the direction v ∈ L2

P

(Ω; H 1(Rd;Rd)

), with v = 0 on �D ,

yields

0 =< ∂L∂ u (O, u, p) , v >

= 2∫

Ω

(∫

O f · u dx + ∫

�Ng · u ds

) (∫

O f · v dx + ∫

�Ng · v ds

)dP(ω)

+ ∫

O

ΩAe (v) : e (p) dP (ω) dx .

(6.20)After integrating by parts and by taking v with compact support in O × Ω gives

− div(Ae(p)) = −2

(∫

Of · u dx +

�N

g · u ds)

f in D × Ω. (6.21)

Similarly, varying the trace of v(·, ω) on �N and on �0, the following boundaryconditions

Ae(p) · n = −2

(∫

Of · u dx +

�N

g · u ds)

g on �N × Ω (6.22)

andAe(p) · n = 0 on �0 × Ω (6.23)

are obtained. The variational or weak formulation of (6.21)–(6.23) is given by (6.16).Notice that, from (6.16) it follows that

p = −2

(∫

Of · u dx +

�N

g · u ds)

u.

The third term in the r.h.s. of (6.15) is obtained by applying the chain rule theorem.The fourth term in (6.15) is obtained from Proposition 6.1. Recall that in the presentcase θ · n = 0 on �D ∪ �N . Finally, the fifth term in (6.15) also follows from thechain rule theorem. �

[email protected]

Page 116: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

6.3 Numerical Approximation via the Level-Set Method 101

6.3.2 Mise en Scène of the Level Set Method

Following the original idea by Osher and Sethian [22], which was imported to thetopology optimization framework in [2, 27], the unknown domain O is describedthrough the zero level set of a function ψ = ψ(t, x) defined on the working domainD as follows

∀x ∈ D, ∀t ∈ (0, T ) ,

⎧⎨

ψ(x, t) < 0 if x ∈ O,

ψ(x, t) = 0 if x ∈ ∂O,

ψ(x, t) > 0 if x /∈ O.

(6.24)

Here, t stands for a fictitious time that accounts for the step parameter in the descentalgorithm. Hence, the domainO = O (t) is evolving in time during the optimizationprocess. Let us assume that O (t) evolves with a normal velocity V (t, x). Since

ψ (x (t) , t) = 0 for any x (t) ∈ ∂O (t) ,

differentiating in t yields

∂ψ

∂t+ ∇ψ · x ′ (t) = ∂ψ

∂t+ Vn · ∇ψ = 0.

Thus, as n = ∇ψ/|∇ψ |,∂ψ

∂t+ V |∇ψ | = 0.

ThisHamilton-Jacobi equation, which is complemented with an initial condition andwith Neumann type boundary conditions, is solved in the working domain D sincethe velocity field is known everywhere in D. Indeed, returning to the optimizationproblems (Pα) and (PEPη

), we recall that the shape derivatives of the cost functionalsinvolved in those problems take the form

�0

vθ · n ds

wherev = ∫

Ω(2 f · u − Ae(u) : e(u)) dP(ω)

+α∫

Ω(2J (O, ω) f · u − f · p + Ae(u) : e(p)) dP(ω)

−2αE(J (O, ω))∫

Ω(2 f · u − Ae(u) : e(u)) dP(ω)

+L1 + L2

( ∫

O dx − V0

)(6.25)

for the case of J LRα, and

[email protected]

Page 117: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

102 6 Structural Optimization Under Uncertainty

v =∫

Ω

C1 (ω) (2 f · u − Ae(u) : e(u)) dP(ω)

+ L1 + L2

(∫

Odx − V0

)

, (6.26)

for J L ,εEPη

(see Theorem 6.2). Since u and p are computed in D, the integrand in v isdefined also in D and not only on the free boundary �0. As a consequence, a descentdirection in the whole domain D is defined by

θ = −vn.

Therefore, its normal component θ · n = −v is the advectionvelocity in theHamilton-Jacobi equation

∂ψ

∂t− v|∇ψ | = 0, t ∈ (0, T ) , x ∈ D. (6.27)

As indicated above, Eq. (6.27) is solved in D. To this end, the elasticity system(6.2) as well as the adjoint system (6.16) are solved in D by using the so-calledersatz material approach. Precisely, D \ O is filled by a weak phase that mimics thevoid, but at the same time avoids the singularity of the stiffness matrix. Thus, anelasticity tensor A∗ (x) which equals A in O and 10−3 · A in D \ O is introduced.The displacement field u(x, ω) then solves

⎧⎪⎪⎨

⎪⎪⎩

−div(A∗e(u(x, ω))) = f in D × Ω,

u(x, ω) = 0 on �D × Ω,

(A∗e(u(x, ω))) · n = g on �N × Ω,

(A∗e(u(x, ω))) · n = 0 on ∂D0 × Ω,

(6.28)

Equation (6.27) may be solved by using an explicit first order upwind scheme. Toavoid singularities, the level-set function is periodically reinitialized. More detailson these implementation issues of the level set are provided in Sect. 6.5.

A crucial step is the numerical approximation of the advection velocities (6.25)and (6.26). For the case of problem (Pα), this may be done by using the StochasticCollocation method presented in Chap. 4. For problem (PEPη

), the techniques ofChap. 5 may be applied. See [18, 19] for more details.

6.4 Numerical Simulation Results

This section presents numerical simulation results for problems (Pα) and (PEPη) in

the widely studied benchmark example of the beam-to-cantilever.The beam-to-cantilever problem consists in the shape optimization of a two-

dimensional cantilever under uncertain loading conditions. The left edge of thecantilever is anchored and a unit force with uncertainty in direction, centered in

[email protected]

Page 118: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

6.4 Numerical Simulation Results 103

Fig. 6.1 Thebeam-to-cantilever problem:(a) the design domain andboundary conditions, and (b)the initialization of level-setfunction. From [18],reproduced with permission

g = 1

D1 = 1

h1

h2

D2 = 2φ

(a) (b)

the horizontal line, is applied at the middle of the right edge. The design domainis a 1 × 2 rectangle which is tessellated using a 60 × 120 quadrangular mesh(h1 = h2 = 0.016). The uncertain loading, the boundary conditions and the tessella-tion of the design domain are depicted in Fig. 6.1a, whereas the initialization of thelevel-set function is shown in Fig. 6.1b.

All the experiments that followmake use of a fixed grid ofQ1 quadrangular planestress finite elements with Young’s modulus E = 1 and Poisson’s ratio ν = 0.3. Thelevel set function ψ(t, x) is discretized at the nodes of such a fixed grid. For JRα

,the Lagrange multipliers and the penalty parameters of the augmented Lagrangianfunctions are initialized heuristically according to L1 = 0.1 · JRα

(O)/V0 and L2 =0.01 · L1/V0. The target volume is set to 30% of the initial design domain. The samestrategy is followed for J L ,ε

EPη.

Problem (Pα). The direction φ of the unit-load follows a Gaussian distribution cen-tered at the horizontal line, φ = 0, with the deterministic loading state as the meanvalue μφ = 0. The influence of the level of uncertainty on the robust shape design isshown by the consideration of σφ = {π/6, π/12, π/24, π/32, π/42} standard devi-ation values.

The deterministic shape optimization design (classical beam) for the horizontal(φ = 0) unit-load is depicted in Fig. 6.2a. The robust optimized designs for theminimization of the cost functional J L

Rαare shown from Fig. 6.2b to Fig. 6.2e for

various magnitudes of the parameter α. The resulting robust designs incorporatestiffness in the vertical direction where the uncertain loading is introduced. Highervalues of the parameter α, increasing the weight of the variance in the functional,lead to more robust structural designs at the cost of decreasing their (compliance)optimality. One can observe how the increase in robustness is firstly achieved by

[email protected]

Page 119: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

104 6 Structural Optimization Under Uncertainty

(a) (b) (c) (d) (e)

Fig. 6.2 Problem (Pα). (a) Deterministic shape optimization design and (b–e) robust optimizeddesigns for the beam-to-cantilever problem with the direction of the unit-load φ ∼ N (0, π/12).From [18], reproduced with permission

(a) (b) (c) (d) (e)

Fig. 6.3 Problem (Pα). Robust optimized designs for J LRα, with α = 4 increasing the uncertainty

in the direction of the unit-load φ ∼ N (0, σφ). From [18], reproduced with permission

means of topology modifications, increasing the redundancy by adding paths totransmit the load to the support, and then by thickening the cantilever members.

The effect of the amount of uncertainty using a Gaussian probability distributionfor the direction φ of the unit-load is shown in Fig. 6.3. This experiment shows thedifferent shape optimal designs of the functional cost J L

Rαfor a given value of α. One

can observe how the increment in the uncertainty of the direction φ of the unit-loadincreases the span between supports and the thickness of members, which incrementthe vertical stiffness.

Problem (PEPη). The optimal designs are shown from Fig. 6.4a to Fig. 6.4e for vari-

ous magnitudes of the parameter η. One can observe that the resulting designs incor-porate stiffness in the vertical direction where the uncertain loading is introduced.Analogously to what was observed in problem (Pα), the risk is firstly minimized

[email protected]

Page 120: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

6.4 Numerical Simulation Results 105

(a) (b) (c) (d) (e)

Fig. 6.4 Problem (PEPη ). The beam-to-cantilever problem: Optimal designs for different η values.From [19], reproduced with permission

Table 6.1 Problem (PEPη ). Excess probability of optimal design for different values of η and forthe deterministic optimal design. From [19], reproduced with permission

Case P {J > 5} P {J > 6} P {J > 7} P {J > 9} P {J > 10} σJ

Deterministic 0.7524 0.6269 0.5419 0.4224 0.3763 9.2215

η = 5 0.7093 0.5274 0.4155 0.2772 0.2280 5.4176

η = 6 0.9999 0.4632 0.2884 0.1299 0.0872 2.3927

η = 7 1 0.4861 0.2178 0.0614 0.0311 1.3829

η = 9 1 1 0.6557 0.0581 0.0249 0.8515

η = 10 1 1 1 0.1093 0.0225 0.5725

by means of topology modifications, increasing the redundancy by adding paths totransmit the load to the support, and then by thickening the cantilever members.

Table 6.1 shows the probability of exceeding different magnitudes of compliancefor deterministic and risk-averse designs. Notice that the deterministic design is notalways the worst choice when compared to risk averse designs with η �= ηi . Thisis attributed to the shape of the PDF resulting from the minimization of the excessprobability. Figure 6.5 shows thePDF for the optimal deterministic design anddiverseoptimal risk-averse designs using different values of η. One can observe that themeanvalues of these PDFs are higher as the prescribed threshold ηi increases. Althoughthe probability of exceeding ηi is reduced, more realizations of the compliance areclose to such a threshold value, and thus the mean performance is degraded. Thiscan lead to a deterministic design showing lower levels of P {J > ηi } compared torisk-averse designs obtained with η > ηi values. This conclusion is only pertinent tothe problem discussed in this specific example. Finally, it should be noted that thereduction in the excess probability involves a reduction of the dispersion in the PDF.This fact can be observed in the last column of Table 6.1, where the standard deviationof the compliance is shown for the different cases analyzed. One can observe that thevalue of the standard deviation is reduced as the threshold η increases. In addition, to

[email protected]

Page 121: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

106 6 Structural Optimization Under Uncertainty

Fig. 6.5 Problem (PEPη ).PDFs calculated for theoptimal deterministic designand risk-averse designs withdifferent values of η. From[19], reproduced withpermission

reduce the excess probability, the risk-averse formulation provides designs which areless sensitive to fluctuations and thus more robust. This fact explains the topologicalsimilarities between the risk-averse designs (Fig. 6.4) and the robust designs obtainedin problem (Pα) above (Fig. 6.2).

6.5 Notes and Related Software

Structural optimization provides engineering designers with a powerful tool to findinnovative and high-performance conceptual designs at the early stages of the designprocess. Such a technique has been successfully applied to improve the design ofcomplex industrial problems, such as aeronautical, aerospace and naval applications[10]. The need for including uncertainty during the design process has shown tobe a key issue for solving real-world engineering problems in several fields, suchas aeronautical and aerospace [17], civil [14], automotive [26] and mechanical [23]engineering, to name but a few.

More recently, the structural optimization problem including uncertainties inloads, material properties and even in geometry has been studied by using boththe SIMP method (see [16] and the references therein) and the level set method (see,e.g. [4, 7–9, 11]). At the discrete level, it is also remarkable to mention [6].

A previous knowledge of Hadamard’s boundary variation method and of the levelset method has been assumed in this chapter. For a comprehensive treatment of thesemethodswe refer the reader to [1, 21, 24]. Section 6.3.2 lacksmany numerical detailsand implementation tricks, which are out of the scope of this text. The interestedreader may find open source codes for the numerical resolution of the deterministicstructural optimization problem by using the level set method in the two followinglinks:

[email protected]

Page 122: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

6.5 Notes and Related Software 107

• Several very useful codes written in Scilab and FreeFem++ are provided by Pro-fessor G. Allaire. http://www.cmap.polytechnique.fr/ allaire/

• A level set-based structural optimization educational code using FEniCS is pro-vided by A. Laurain in [15]. http://www.antoinelaurain.com/

References

1. Allaire, G.: Conception Optimale de Structures, 58. Mathématiques et Applications. Springer,Berlin (2007)

2. Allaire, G., Jouve, F., Toader, A.M.: Structural optimization using shape sensitivity analysisand a level-set method. J. Comput. Phys. 194, 363–393 (2004)

3. Céa, J.: Conception optimale ou identification de formes, calcul rapide de la dérivée direction-nelle de la function cöut. ESAIM Math. Model. Numer. Anal. 20, 371–402 (1986)

4. Chen, S., Chen, W., Lee, S.: Level set based robust shape and topology optimization underrandom field uncertainties. Struct. Multidisc. Optim. 44(1), 507–524 (2010)

5. Chenais, D.: On the existence of a solution in a domain identification problem. J. Math. Anal.Appl. 52, 189–219 (1975)

6. Choi, S.K., Grandhi, R., Canfield, R.A.: Reliability-Based Structural Design, Springer (2007)7. Conti, S., Held, H., Pach, M., Rumpf, M., Schultz, R.: Risk averse shape optimization. SIAM

J. Control Optim. 49, 927–947 (2011)8. Dambrine, M., Dapogny, C., Harbrecht, H.: Shape optimization for quadratic functionals and

states with random right-hand sides. SIAM J. Control Optim. 53(5), 3081–3103 (2015)9. Dambrine, M., Harbrecht, H., Puig, B.: Computing quantities of interest for random domains

with second order shape sensitivity analysis. ESAIMMath. Model. Numer. Anal. 49(5), 1285–1302 (2015)

10. Deaton, J.D., Grandhi, R.V.: A survey of structural and multidisciplinary continuum topologyoptimization: post 2000. Struct. Multidiscip. Optim. 49, 1–38 (2014)

11. Dunning, P.D., Kim, H.A.: Robust topology optimization: minimization of expected and vari-ance of compliance. AIAA J. 51(11), 2656–2664 (2013)

12. Hadamard, J.: Mémoire sur le problème d’analyse relatif l’équilibre des plaques élastiquesencastrées. Bull. Soc. Math, France (1907)

13. Henrot, A., Pierre, M.: Variation et optimisation de formes: Une Analyse Geometrique, 48.Mathématiques et Applications. Springer, Berlin (2005)

14. Lagaros, N.D., Papadrakakis,M.: Robust seismic design optimization of steel structures. Struct.Multidiscip. Optim. 33(6), 457–469 (2007)

15. Laurain, A.: A level set-based structural optimization code using FEniCS (2018).arXiv:1705.01442

16. Lazarov, B., Schevenels, M., Sigmund, O.: Topology optimization considering material andgeometric uncertainties using stochastic collocation methods. Struct. Multidisc. Optim. 46(4),597–612 (2012)

17. López, C., Baldomir, A., Hernández, S.: The relevance of reliability-based topology optimiza-tion in early design stages of aircraft structures. Struct. Multidiscip. Optim. 57(1), 417–439(2018)

18. Martínez-Frutos, J., Herrero-Pérez, D., Kessler, M., Periago, F.: Robust shape optimization ofcontinuous structures via the level set method. Comput. Methods Appl. Mech. Engrg. 305,271–291 (2016)

19. Martínez-Frutos, J., Herrero-Pérez, D., Kessler,M., Periago, F.: Risk-averse structural topologyoptimization under randomfields using stochastic expansionmethods. Comput.Methods Appl.Mech. Engrg. 330, 180–206 (2018)

20. Murat, F., Simon, J.: Sur le contrôle par un domaine géométrique, Technical report RR-76015,Laboratoire d’Analyse Numérique (1976)

[email protected]

Page 123: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

108 6 Structural Optimization Under Uncertainty

21. Osher, S., Fedkiw, R.: Level Set Methods and Dynamic Implicit Surfaces. Springer, New York(2003)

22. Osher, S., Sethian, J.A.: Fronts propagation with curvature-dependent speed: algorithms basedon Hamilton-Jacobi formulations. J. Comput. Phys. 79(1), 12–49 (1988)

23. Schevenelsa, M., Lazarov, B.S., Sigmund, O.: Robust topology optimization accounting forspatially varyingmanufacturing errors. Comput. Methods Appl. Mech. Engrg. 200, 3613–3627(2011)

24. Sethian, J.A.: Level Set Methods and Fast Marching Methods: Evolving Interfaces in Com-putational Geometry, Fluid Mechanics, Computer Vision and Material Science. CambridgeUniversity Press (1999)

25. Sokolowski, J., Zolesio, J. P.: Introduction to Shape Optimization: Shape Sensitivity Analysis.Springer Series in Computational Mathematics, vol. 16, Springer, Heidelberg (1992)

26. Youn, B.D., Choi, K.K., Yang, R.J., Gu, L.: Reliability-based design optimization for crash-worthiness of vehicle side impact. Struct. Multidiscip. Optim. 26(3–4), 272–283 (2004)

27. Wang, M.Y., Wang, X., Guo, D.: A level-set method for structural topology optimization.Comput. Methods Appl. Mech. Engrg. 192, 227–246 (2003)

[email protected]

Page 124: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Chapter 7Miscellaneous Topics and Open Problems

Surely the first and oldest problems in every branch ofmathematics spring from experience and are suggested by theworld of external phenomena.

David Hilbert.Lecture delivered before ICM, Paris, 1900.

In this final chapterwediscuss some related topics to the basicmethodologypresentedin detail in the previous chapters. These topics are related to (i) time-dependentproblems, and (ii) physical interpretation of robust and risk-averse optimal controls.Wealso list a number of challengingproblems in thefield of control under uncertainty.

7.1 The Heat Equation Revisited II

In this section it is shown how to apply the techniques developed in the precedingchapters to robust optimal controls problems constrained by a time-dependent PDE.As it will be shown hereafter, the methods described in Chap. 4 easily extend to thissituation.

As an illustration, uncertaintywill not appear in the principal part of the differentialoperator (as it was the case in the examples of the previous chapters), but in an inputparameter acting on the boundary conditions. The control function will act in a partof the boundary’s domain. To this end, the heat equation

⎧⎪⎪⎨

⎪⎪⎩

yt − �y = 0, in (0, T ) × D × Ω

∇ y · n = 0, on (0, T ) × ∂D0 × Ω

∇ y · n = α (u − y) , on (0, T ) × ∂D1 × Ω,

y (0) = y0 in D × Ω,

(7.1)

is revisited again.

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2018J. Martínez-Frutos and F. Periago Esparza, Optimal Control of PDEs under Uncertainty,SpringerBriefs in Mathematics, https://doi.org/10.1007/978-3-319-98210-6_7

109

[email protected]

Page 125: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

110 7 Miscellaneous Topics and Open Problems

It is well-known [2, 5] that the convective heat transfer coefficient α is difficult tomeasure in practice as it depends on random factors such as the physical propertiesof the boundary ∂D1 and the motion of the surrounding fluid. Thus, it is reasonableto assume that α = α(x, ω) is a random field.

For a given control time T > 0 and a target function yd ∈ L2 (D), the two fol-lowing cost functionals are introduced:

J1 (u) = 12

D

Ω|y(T, x, ω) − yd(x)|2 dP(ω)dx

+ γ12

D Var (y(T, x)) dx

+ γ22

∫ T0

∂D1

(|u(t, y)|2 + |ut (t, y)|2)ds dt

(7.2)

andJ2(u) = ∫

D

(∫

Ωy(T, x, ω) dP(ω) − yd(x)

)2dx

+ γ12

D Var (y(T, x)) dx

+ γ22

∫ T0

∂D1

(|u(t, y)|2 + |ut (t, y)|2)ds dt.

(7.3)

In both cases, γ1, γ2 ≥ 0 are two weighting parameters, and the space of admissiblecontrols is Uad = H 1

0

(0, T ; L2 (∂D1)

).

With all these ingredients, the two following optimal control problems are con-sidered:

(Pheat )

⎧⎨

Minimize in u ∈ Uad : J1(u) or J2(u)

subject toy = y(u) solves (7.1).

To fix ideas, in what follows, D = (0, 1)2 is the unit square. Its boundary isdivided into ∂D1 = {(x1, x2) ∈ R

2 : x2 = 0} ∪ {(x1, x2) ∈ R2 : x2 = 1} and ∂D0 =

∂D \ ∂D1.Since the randomfieldα is positive, it is typicallymodeled as a log-normal random

field. For the experiment that follows,α is a log-normal randomfieldwithmean equalto one and variance 0.01 at x2 = 0, and variance 0.1 at x2 = 1. Precisely, in a first step,a three-terms, z3(x, ω), KL expansion is computed using the exponential covariancefunction (2.15), with correlation length of 0.5. In a second step, formulae (2.27), witha = 1 and variances 0.01 at x2 = 0, and variance 0.1 at x2 = 1, are used to computethe parameters μ and σ that appear in (2.26) with z(x, ω) replaced by z3(x, ω). Thisleads to an explicit representation of α(x, ω) as given by (2.26) with a(x, ω) replacedby α(x, ω). Note that two three-terms KL expansions are considered; one of themfor the edge x2 = 0 and another one for x2 = 1. Thus, we end up with 6 stochasticdirections.

[email protected]

Page 126: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

7.1 The Heat Equation Revisited II 111

Fig. 7.1 Problem (Pheat ).Deterministic target functionyd . From [11], reproducedwith permission

Similarly to Chap.4, the random domain Γ = [−3, 3]6, and

ρ (z) =6∏

n=1

φ (zn) , z = (z1, · · · , z6) ∈ Γ, (7.4)

where φ (zn) is the truncated Gaussian (2.30).The initial condition is y(0, x, ω) = 0, the final control time is T = 0.5, and the

target function is (see Fig. 7.1)

yd(x1, x2) =⎧⎨

(−2x21 + 2x1)(−25x22 + 10x2) 0 ≤ x1 ≤ 1, 0 ≤ x2 ≤ 0.4,0 0 ≤ x1 ≤ 1, 0.4 ≤ x2 ≤ 0.6,(−4x21 + 4x1)(−25x22 + 40x2 − 15) 0 ≤ x1 ≤ 1.0, 0.6 ≤ x2 ≤ 1.

(7.5)The gradient-based minimization algorithm of Sect. 4.2 may be used to solve

problem (Pheat ). Explicit expressions for the descent directions and the optimal step-size parameters are provided in [11]. Here, the only novelty is that the discretizationprocess, which is needed to approximate numerically the state equation (7.1) andits associated adjoint equations, involves discretization w.r.t. the time variable. Thismay be done, for instance, by using a fully implicit backward second order Gearscheme. Indeed, starting from y(n−1) at time t (n−1) and y(n) at time t (n), the Gearscheme is based on the approximation

y(n+1)t ≡ ∂y

∂t

(t (n+1)

) � 3y(n+1) − 4y(n) + y(n−1)

2dt. (7.6)

[email protected]

Page 127: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

112 7 Miscellaneous Topics and Open Problems

Table 7.1 Problem (Pheat ). Summary of the error with respect to the target and variance ofthe state variable for γ2 = 1e − 8. Here y(T, x) denotes the mean of y(T, x, ·), i.e., y(T, x) =∫

Γy(T, x, z)ρ(z) dz. From [11], reproduced with permission

||y(T ) − yd ||2L2(D)||y(T ) − yd ||2L2(D)⊗L2

ρ (Γ )||Var(y(T ))||L2(D)

J1, γ1 = 0 8.669 × 10−3 1.086 × 10−2 2.187 × 10−3

J1, γ1 = 1 6.330 × 10−3 8.076 × 10−3 1.745 × 10−3

J1, γ1 = 2 5.945 × 10−3 7.466 × 10−3 1.519 × 10−3

J1, γ1 = 3 7.388 × 10−3 1.138 × 10−3 8.526 × 10−3

J2, γ1 = 0 1.128 × 10−3 7.649 × 10−3 6.521 × 10−3

J2, γ1 = 1 1.969 × 10−3 3.283 × 10−3 1.314 × 10−3

J2, γ1 = 2 1.867 × 10−3 8.559 × 10−4 2.723 × 10−3

J2, γ = 3 1.738 × 10−3 7.294 × 10−4 2.468 × 10−3

In the numerical experiments that follow, a time stepofdt = 2.5 × 10−3 is consideredin (7.6).

Discretization in space is carried out by using Lagrange P1 finite elements on atriangular mesh composed of 2436 triangles and 1310 degrees of freedom.

Following Sect. 4.2.2.1, an anisotropic sparse grid is constructed for discretizationin the random domain Γ . The same vector of weights as in (4.25) has been selected.See [12] for more details on this passage. The level of quadrature � in the sparsegrid is fixed by using the adaptive algorithm of Sect. 4.2.2.2. Precisely, a non-nestedquadrature rule whose collocation nodes are determined by the roots of Hermitepolynomials is used. Notice that the random variables of this problem follow theGaussian distribution (7.4) so that Hermite polynomials are a suitable choice forapproximation in the random domain. The nodes and weights for the anisotropicsparse grid are computed adaptively as described in Sect. 4.2.2 with N = 6, � = 9,δ = 1e − 6 and �opt = 5. Here, the only difference is that in the stopping criteria(4.38) and (4.40), y = y (T ) is the solution to (7.1) at time T .

The convergence history of the optimization algorithm, depicted in Fig. 7.2, con-firms that the cost J2 requires more iterations to achieve convergence.

To analyze the performance of the cost functionals J1 and J2, the optimizationproblem has been solved for values of γ1 = 0, 1, 2, 3, and γ2 = 1e − 8. The resultsfor the different contributions of the costs J1 and J2 are presented numerically inTable7.1 and graphically in Figs. 7.3, 7.4 and 7.5. As expected, the peak value of thevariance of the state variable is close to the upper edge due to the large covarianceof the random field on this part of the boundary. The results presented in Figs. 7.3and 7.4 are in line with the ones obtained in Chap. 4: on the one hand, the cost J1provides solutions less sensitive to variations in the input data. On the other hand, thecost J2 provides a better correspondence between the mean of state variable and thetarget function, but with high variance. This variance can be reduced by increasingthe value of the parameter γ1 as shown in Fig. 7.5 for γ1 = 1, 2 and 3.

[email protected]

Page 128: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

7.1 The Heat Equation Revisited II 113

0 20 40 60 80 100 120−4.6

−4.4

−4.2

−4

−3.8

−3.6

−3.4

−3.2

−3

−2.8

−2.6

Iter

Log(J

1)

0 200 400 600 800 1000−7

−6.5

−6

−5.5

−5

−4.5

−4

−3.5

−3

−2.5

Iter

Log

(J2)

Fig. 7.2 Problem (Pheat ). Convergence history of the gradient-basedminimization algorithm. From[11], reproduced with permission

Fig. 7.3 Problem (Pheat ). Results associated with the cost J1 with γ1 = 0, γ2 = 1e − 8. Thefollowing quantities are represented: the mean of the state variable (a), the variance of the statevariable (b), the control variable acting on x2 = 1 (c), and the control variable acting on x2 = 0 (d).The target yd is shown transparently in (a) as a reference. From [11], reproduced with permission

[email protected]

Page 129: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

114 7 Miscellaneous Topics and Open Problems

Fig. 7.4 Problem (Pheat ). Results associated with the cost J2 with γ1 = 0, γ2 = 1e − 8. Thefollowing quantities are represented: the mean of the state variable (a), the variance of the statevariable (b), the control variable acting on x2 = 1 (c), and the control variable acting on x2 = 0 (d).The target yd is shown transparently in (a) as a reference. From [11], reproduced with permission

7.2 The Bernoulli-Euler Beam Equation Revisited II

The goal of this section is to provide a first insight into the profiles of optimal controlsfor the class of robust optimal control problems considered in this text. To this end,the following control system for the Bernoulli-Euler beam equation is considered:

⎧⎨

ytt + (1 + α) yxxxx = 1O u(t, x), in (0, T ) × (0, L) × Ω

y(0) = yxx (0) = y(L) = yxx (L) = 0, on (0, T ) × Ω

y(0) = y0, yt (0) = y1, in (0, L) × Ω,

(7.7)

where α = α (ω) is a random variable uniformly distributed in the interval [−ε, ε],with 0 < ε < 1.

[email protected]

Page 130: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

7.2 The Bernoulli-Euler Beam Equation Revisited II 115

Fig. 7.5 Problem (Pheat ). Variance of the state variable associatedwith the cost J2 for γ2 = 1e − 8,γ1 = 0 (a), γ1 = 1 (b), γ1 = 2 (c), and γ1 = 3 (d). From [11], reproduced with permission

The cost functional considered is:

J (u) := δ12

∫ L0

(∫

Ωy (T ) dP(ω)

)2dx + δ2

2

∫ L0

(∫

Ωyt (T ) dP(ω)

)2dx

+ β1

2

∫ L0

Ωy2 (T ) dP(ω)dx + β2

2

∫ L0

Ωy2t (T ) dP(ω)dx

+ γ

2

∫ T0

O u2 dxdt,

(7.8)

where δ1, δ2 > 0 and β1, β2, γ ≥ 0 are weighting parameters.The rest of input parameters in (7.7) are: L = 1, T = 0.5, O = (0.2, 0.8) and(

y0 (x) , y1 (x)) = (sin (πx) , 0). See Fig. 7.6 for the problem configuration.

[email protected]

Page 131: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

116 7 Miscellaneous Topics and Open Problems

y(t, x)

x

y(0, x)

L = 1

u(t, x)

Fig. 7.6 Problem (Pbeam). Geometry, initial and boundary conditions. From [11], reproduced withpermission

Eventually, the robust optimal control problem is formulated as:

(Pbeam)

⎧⎨

Minimize in u ∈ L2 (O) : J (u)

subject toy = y(u) solves (7).

Remark 7.1 For the shake of clarity and simplicity, the control function in (Pbeam)

is not of piezoelectric type, but a more academic control u ∈ L2 (O).

The solution of (Pbeam) may be numerically approximated by using the methodsdescribed in Chap.4. The reader is referred to [10] for details. Here, we focus on thephysical interpretation of the results obtained.

Figure7.7 shows the computed optimal control u(t, x) in the cases: (a) determin-istic problem, where α = 0 in (7.7) and the cost functional is

Jdet (u) = 1

2‖y(T )‖2L2(0,L) + 1

2‖yt (T )‖2L2(0,L) + γ

2‖u‖2L2(]0,T [×O ),

and (b)–(c) solutions of (Pbeam) for different values of the weighting parametersand for ε = 0.5, i.e., α (ω) is uniformly distributed in the interval [−0.5, 0.5]. It isobserved that the optimal control obtained minimizing only the mean of the statevariable (Fig. 7.7b) is quite similar to the one obtained in the deterministic case(Fig. 7.7a). However, the optimal control including the second raw moment in thecost functional (Fig. 7.7c) presents a very different behavior.

To better understand the qualitative behavior of the control profiles, the way inwhich the beam equation propagates uncertainty is analyzed in Fig. 7.8. Precisely,let us denote by yunc (t, x, ω) the solution of the uncontrolled (u = 0 in (7.7)) beamequation. It is easy to see that

yunc (t, x, ω) = cos(√

1 + α (ω)π2t)sin (πx)

= 12 sin

(x + π t

√1 + α (ω)

)] + 12 sin

(x − π t

√1 + α (ω)

)].

(7.9)

[email protected]

Page 132: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

7.2 The Bernoulli-Euler Beam Equation Revisited II 117

Fig. 7.7 Problem (Pbeam). ε = 0.5 and γ = 10−6. Optimal controls for: (a) deterministic problem,(b) δ1 = δ2 = 1, β1 = β2 = 0 and (c) δ1 = δ2 = 1, β1 = β2 = 1. From [10], reproduced withpermission

From (7.9) one concludes: (i) the initial condition y0 (x) = sin (πx) propagates intime as a wave with random amplitude cos

(√1 + α (ω)π2t

), and (ii) each material

point x propagates as the superposition of twowaves travelling in opposite directionsat a random speed π

√1 + α (ω). Moreover, from (7.9) one may explicitly compute

the first two moments of (yunc, (yt )unc), which are plotted in Fig. 7.8. In particular,by the mean value theorem for integrals,

yunc (t, x) = 1

∫ ε

−ε

cos(√

1 + απ2t)sin (πx) dα = cos

(√1 + α�π2t

)sin (πx)

for some−ε ≤ α� ≤ ε. Notice that α� depends on t . Thus, for ε small, yunc is similarto the uncontrolled deterministic solution, where α = 0 in (7.9). This fact explainsthe similarity between the controls plotted in Fig. 7.7a–b. However, the second ordermoments of (yunc, (yt )unc), depicted in Fig. 7.8c–d, show a very different qualitativebehavior. Precisely, one observes an additional oscillation and there is a change inamplitude. The control displayed in Fig. 7.7c is in accordance with this behavior ofthe second moment of the uncontrolled random solution.

[email protected]

Page 133: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

118 7 Miscellaneous Topics and Open Problems

Fig. 7.8 Problem (Pbeam). ε = 0.5, δ1 = δ2 = 1, γ = 10−6. First column displays: (a) expectationof the uncontrolled solution yunc (t, x, ω), and (b) expectation of (yt )unc (t, x, ω). Second columnshows: (c) second moment of yunc(t, x, ω), and (d) second moment of (yt )unc(t, x, ω). From [10],reproduced with permission

7.3 Concluding Remarks and Some Open Problems

As it has been noticed along this introductory text, amajor computational challenge inthe field of control under uncertainty is breaking the curse of dimensionality. Indeed,when the number of random variables, which appear in the uncertain parameters, islarge, the number of degrees of freedom grows exponentially so that the problemmay be computationally unaffordable. This issue is exacerbated, for instance, in:

• multiphysics and/or multiscale problems, where the solution of the underlyingPDE is expensive, and in

• nonlinear PDEs, where numerical resolution methods typically require iteration.

A number of remedies are being proposed to overcome these difficulties. Forinstance, Analysis of Variance (ANOVA) based techniques, High-DimensionalModel Representation (HDMR) or a combination of them, aim at detecting the mostinfluential random variables (and the interaction between them) in the mathematical

[email protected]

Page 134: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

7.3 Concluding Remarks and Some Open Problems 119

model (see [6, 13] and the references therein). Other model order reduction meth-ods, which are receiving an increasing interest in the last decades, are Reduced Basis(RB) methods, Proper-Orthogonal Decomposition (POD) methods [1, 4] and ActiveSubspaces methods [3], among others.

Despite these theoretical and numerical developments, the topic of control ofrandom PDEs may be still considered to be in its infancy and further research isneeded to better understand and more efficiently solve this type of problems.

Another class of control problems, which has not been considered in this text,is exact controllability of PDEs under uncertainty. As an illustration of this type ofproblems, let us consider again the controlled heat equation

⎧⎪⎪⎨

⎪⎪⎩

yt − div (a∇ y) = 0, in (0, T ) × D × Ω

a∇ y · n = 0, on (0, T ) × ∂D0 × Ω

a∇ y · n = α (u − y) , on (0, T ) × ∂D1 × Ω,

y (0) = y0 in D × Ω,

(7.10)

where, for instance, the parameters a = a (x, ω), α = α (x, ω) and the initial condi-tion y0 = y0 (x, ω) are assumed to be uncertain, and u ∈ L2 ((0, T ) × ∂D1) is thecontrol function.

The concept of averaged exact control was introduced by Zuazua in [14]. Whenapplied to system (7.10), the problem of averaged null controllability is formulatedas: given a control time T > 0, find u ∈ L2 ((0, T ) × ∂D1) such that the mean ofthe solution y = y (u) to (7.10) satisfies

Ω

y (T, x, ω) dP (ω) = 0, x ∈ D. (7.11)

At the theoretical level, some positive averaged controllability results have beenobtained for the heat, wave and Schrödinger equations in [7–9]. In all these worksuncertainty appears in the coefficient a and is represented either as a single randomvariable or as a small randomperturbation. Thegeneral case of uncertaintiesmodelledas random fields and/or located in other input parameters of the model such asgeometry, location of controls or boundary conditions has not been addressed so far.It is then interesting to analyze if classical tools in deterministic exact controllabilityproblems such as Ingham’s inequalities or Carlerman estimates, among others, maybe adapted to deal with this class of controllability problems.

The numerical approximation of averaged exact controllability problems is alsoa challenging problem. By using an optimal control approach, a first step in thisdirection has been given in [10, 11].

Related to problem (7.10)–(7.11) is the following robust version:

(PR)

⎧⎨

Minimize in u ∈ L2 ((0, T ) × ∂D1) : ‖Var (y(T )) ‖2L2(D)

subject to(y, u) solve (7.10)–(7.11).

(7.12)

[email protected]

Page 135: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

120 7 Miscellaneous Topics and Open Problems

Up to the best knowledge of the authors, problem (PR) is open.

References

1. Chen, P., Quarteroni, A.: A new algorithm for high-dimensional uncertainty quantificationbased on dimension-adaptive sparse grid approximation and reduced basismethods. J. Comput.Phys. 298, 176–193 (2015)

2. Chiba, R.: Stochastic heat conduction of a functionally graded annular disc with spatiallyrandom heat transfer coefficients. Appl. Math. Model. 33(1), 507–523 (2009)

3. Constantine, P.G.: Active subspaces. Emerging ideas for dimension reduction in parameterstudies, vol. 2. SIAM Spotlights, Philadelphia, PA, 2015

4. Hesthaven, J.S., Rozza, G., Stamm,B.: Certified reduced basismethods for parametrized partialdifferential equations. SpringerBriefs in Mathematics. BCAM SpringerBriefs (2016)

5. Kuznetsov, A.V.: Stochastic modeling of heating of a one-dimensional porous slab by a flowof incompressible fluid. Acta Mech. 114, 39–50 (1996)

6. Labovsky, A., Gunzburger, M.: An efficient and accurate method for the identification ofthe most influential random parameters appearing in the input data for PDEs. SIAM/ASA J.Uncertain. Quantif. 2(1), 82–105 (2014)

7. Lazar, M., Zuazua, E.: Averaged control and observation of parameter-depending wave equa-tions. C. R. Acad. Sci. Paris. Ser. I 352, 497–502 (2014)

8. Lohéac, J., Zuazua, E.: Averaged controllability of parameter dependent wave equations. J.Differ. Equ. 262(3), 1540–1574 (2017)

9. Lü, Q., Zuazua, E.: Averaged controllability for random evolution partial differential equations.J. Math. Pures Appl. 105(3), 367–414 (2016)

10. Marín, F.J., Martínez-Frutos, J., Periago, F.: Robust averaged control of vibrations for theBernoulli-Euler beam equation. J. Optim. Theory Appl. 174(2), 428–454 (2017)

11. Martínez-Frutos, J., Kessler, M., Münch, A., Periago, F.: Robust optimal Robin boundarycontrol for the transient heat equation with random input data. Internat. J. Numer. MethodsEng. 108(2), 116–135 (2016)

12. Nobile, F., Tempone, R.: Analysis and implementation issues for the numerical approximationof parabolic equations with random coefficients. Internat. J. Numer. Methods Eng. 80(6–7),979–1006 (2009)

13. Smith, R. C.: Uncertainty Quantification. Theory, Implementation and Applications. Comput.Sci. Eng. 12 (2014)

14. Zuazua, E.: Averaged control. Automatica 50, 3077–3087 (2014)

[email protected]

Page 136: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Index

Symbols1O characteristic function of O , 3A : B full contraction of two matrices, 93A ⊗ B Kronecker product of two matrices,

73Ae(u) stress tensor, 92D ⊂ R

d spatial domain, 3, 9Γ ⊂ R

N random domain, 14, 47δx j Dirac mass , 6P probability measure, 4O ⊂ D, 3, 91∂D boundary of D, 9ρ(z) joint probability density function, 14,

47σ - algebra of events, 4ξ = (ξ1, . . . , ξN ) multivariate random vari-

able, 14ds surface Lebesgue measure, 10dz Lebesgue measure, 48e (u) strain tensor, 92h1 ⊗ h2 pure tensor, 11t time variable, 5u = (u1, . . . , ud ) displacement vector field,

92u · v scalar product between two vectors, 93x = (x1, . . . , xd ) ∈ D spatial variable, 3yd (x) target function, 4yt partial derivative of y w.r.t. t , 4ytt second derivative w.r.t. t , 6yxx second derivative w.r.t. x , 6yx partial derivative w.r.t. x , 6z = (z1, . . . , zn) ∈ Γ random parameter, 14

AAbscissas

Clenshaw-Curtis, 54

Gaussian, 54Averaged exact control, 119

CCompliance, 93Control u, 4Correlation length, 19Covariance

exponencial, 19squared exponential, 24

Cumulative density function, 25

DDoob-Dynkin lemma, 47Duality product, 35

EEquation

Bernoulli-Euler, 5Hamilton-Jacobi, 101heat, 4Laplace-Poisson, 3linear elasticity, 92

Errordata, 1mean variance, 18model, 1numerical, 1variance, 18

Ersatz material approach, 102Excess probability, 94

FFunction

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2018J. Martínez-Frutos and F. Periago Esparza, Optimal Control of PDEs under Uncertainty,SpringerBriefs in Mathematics, https://doi.org/10.1007/978-3-319-98210-6

121

[email protected]

Page 137: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

122 Index

covariance Cova(x, x ′) , 15Heaviside, 40level set, 101mean a(x) = E [a (x, ·)], 15variance Vara (x), 15

GGear scheme, 111Geometric

mean, 24standard deviation, 24

KKarhunen-Loève expansion, 16

LLamé moduli, 92

MMethod

formal Lagrange, 50transposition, 36

Multi-index setYg (�, N ), 56Λg (�, N ), 83Ig (�, N ), 55

NNoise

finite dimensional, 46

OOperator

difference Δ j , 55divergence (div), 3extension PO , 95gradient ∇, 3Laplacian Δ, 43

PPoincaré constant, 33Problem

averaged null controllability, 119beam-to-cantilever, 102generalized matrix eigenvalue, 22

Propertyε-cone, 94

uniform extension, 95

QQuadrature rule

level of, 53nested, 54non-nested, 54one-dimensional Q�, 53Smolyak Ag (�, N ), 55

RRandom

event ω, 12event ω, 3, 5field, 15, 16, 18, 22

Random fieldGaussian, 15, 18log-normal, 23, 24second order, 15

SSample variance σ 2

Mn, 87

SetP-null, 10nodal, 56

Shape derivative, 98Space

H1 (D), 10H10 (D), 10

H1 ⊗ H2, 12L2P

(Ω; H), 11L2

ρ

(Γ ; H1

0 (D)), 48

L2ρ (Γ ), 14, 48, 83

L p (D; dx), 9L p (D), 9L p (Ω; dP(ω)), 10L pP

(Ω; X), 10L pP

(Ω), 10Vh (D), 84VO , 92WP(0, T ), 34Pp (Γ ), 14PΛg(�,N ) (Γ ), 84H2(D) ∩ H1

0 (D), 7L2(D), 6V �, 7complete, 10probability, 3sample Ω , 3separable, 11

[email protected]

Page 138: Jesús Martínez-Frutos Francisco Periago Esparza Optimal ... · Thus, we focus on stochastic collocation (SC) and stochastic Galerkin (SG) methods [2], which, in this context, are

Index 123

State variable y, 4

TTheorem

Karhunen-Loève, 17Mercer, 17

UUncertainty

aleatoric, 1epistemic, 1quantification, 2

VVariance of y (T, x, ·), 5

[email protected]