Optimal Control of Complex Systems through Variational ... · a MDP in the domain of transportation...

9
Optimal Control of Complex Systems through Variational Inference with a Discrete Event Decision Process Fan Yang University at Buffalo Buffalo, New York fyang24@buffalo.edu Bo Liu Auburn University Auburn, Alabama [email protected] Wen Dong University at Buffalo Buffalo, New York wendong@buffalo.edu ABSTRACT Complex social systems are composed of interconnected individu- als whose interactions result in group behaviors. Optimal control of a real-world complex system has many applications, including road traffic management, epidemic prevention, and information dissemination. However, such real-world complex system control is difficult to achieve because of high-dimensional and non-linear system dynamics, and the exploding state and action spaces for the decision maker. Prior methods can be divided into two categories: simulation-based and analytical approaches. Existing simulation approaches have high-variance in Monte Carlo integration, and the analytical approaches suffer from modeling inaccuracy. We adopted simulation modeling in specifying the complex dynamics of a complex system, and developed analytical solutions for search- ing optimal strategies in a complex network with high-dimensional state-action space. To capture the complex system dynamics, we formulate the complex social network decision making problem as a discrete event decision process. To address the curse of di- mensionality and search in high-dimensional state action spaces in complex systems, we reduce control of a complex system to varia- tional inference and parameter learning, introduce Bethe entropy approximation, and develop an expectation propagation algorithm. Our proposed algorithm leads to higher system expected rewards, faster convergence, and lower variance of value function in a real- world transportation scenario than state-of-the-art analytical and sampling approaches. ACM Reference Format: Fan Yang, Bo Liu and Wen Dong. 2019. Optimal Control of Com- plex Systems through Variational Inference with a Discrete Event Decision Process. In Proc. of the 18th International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2019), May 13-17, 2019, Montréal, Canada. IFAAMAS, 9 pages. 1 INTRODUCTION A complex social system is a collective system composed of a large number of interconnected entities that as whole exhibit proper- ties and behaviors resulted from the interaction of its individual parts [27]. Achieving optimal control of a real-world complex social system is important and valuable. For example, in an urban trans- portation system, a central authority wants to reduce the overall delay and the total travel time by controlling traffic signals using traffic data collected from the Internet of Things [2]. In an epidemic Proc. of the 18th International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2019), N. Agmon, M. E. Taylor, E. Elkind, M. Veloso (eds.), May 13–17, 2019, Montreal, Canada. © 2019 International Foundation for Autonomous Agents and Multiagent Systems (www.ifaamas.org). All rights reserved. system, an optimal strategy is pursued to minimize the expected dis- counted losses resulting from the epidemic process over an infinite horizon, with specific aims such as varying the birth and death rates [18] or early detection of epidemic outbreaks [13]. In the sphere of public opinion, components such as information consumers, news media, social websites, and governments aim to maximize their own utility functions with optimal strategies. A specific objective in the public opinion environment is tackling fake news to create a clearer, more trusted information environment [1]. These scenar- ios exemplify the enormous potential applications of a versatile optimal control framework for general complex systems. However, key characteristics of complex systems make establish- ing such a framework very challenging. The system typically has a large number of interacting units and thus a high-dimensional state space. State transition dynamics in complex systems are al- ready non-linear, time-variant, and high-dimensional, and then these frameworks must account for additional control variables. Prior research in decision making for complex social systems has generally gone in one of two directions: a simulation or analyti- cal approach [21]. Simulation approaches specify system dynamics through a simulation model and develop sampling-based algorithms to reproduce the dynamic flow [29, 41, 46, 47]. These approaches can capture the microscopic dynamics of a complex system with high fidelity, but have a high variance and are time-consuming [20]. Analytical approaches instead formulate the decision-making problem as a constrained optimization problem with an analytical model through specifying the macroscopic state transitions directly, deriving analytical solutions for optimizing strategies [6, 36]. These approaches can provide a robust solution with less variance, but are applicable only to scenarios with small state spaces [16], or cases with low resolution intervention [9], due to modeling costs and errors [20]. The above research points to a new direction in research opportunities that combines the ability to capture system dynamics more precisely from simulation approaches with the ben- efit of having less variance and being more robust from analytical approaches. In this paper, we adopted simulation modeling in spec- ifying the dynamics of a complex system, and developed analytical solutions for searching optimal strategies in a complex network with high-dimensional state-action space specified through simula- tion modeling. We formulate the problem of decision making in a complex system as a discrete event decision process (DEDP), which iden- tifies the decision making process as a Markov decision process (MDP) and introduces a discrete event model — a kind of simula- tion model [17] — to specify the system dynamics. A discrete event model defines a Markov jump process with probability measure on a sequence of elementary events that specify how the system Session 1F: Agent Societies and Societal Issues 1 AAMAS 2019, May 13-17, 2019, Montréal, Canada 296

Transcript of Optimal Control of Complex Systems through Variational ... · a MDP in the domain of transportation...

Page 1: Optimal Control of Complex Systems through Variational ... · a MDP in the domain of transportation optimal control. To solve a DEDP analytically, we derived a duality theorem that

Optimal Control of Complex Systems throughVariational Inference with a Discrete Event Decision Process

Fan YangUniversity at BuffaloBuffalo, New York

[email protected]

Bo LiuAuburn UniversityAuburn, [email protected]

Wen DongUniversity at BuffaloBuffalo, New York

[email protected]

ABSTRACTComplex social systems are composed of interconnected individu-als whose interactions result in group behaviors. Optimal controlof a real-world complex system has many applications, includingroad traffic management, epidemic prevention, and informationdissemination. However, such real-world complex system controlis difficult to achieve because of high-dimensional and non-linearsystem dynamics, and the exploding state and action spaces for thedecision maker. Prior methods can be divided into two categories:simulation-based and analytical approaches. Existing simulationapproaches have high-variance in Monte Carlo integration, andthe analytical approaches suffer from modeling inaccuracy. Weadopted simulation modeling in specifying the complex dynamicsof a complex system, and developed analytical solutions for search-ing optimal strategies in a complex network with high-dimensionalstate-action space. To capture the complex system dynamics, weformulate the complex social network decision making problemas a discrete event decision process. To address the curse of di-mensionality and search in high-dimensional state action spaces incomplex systems, we reduce control of a complex system to varia-tional inference and parameter learning, introduce Bethe entropyapproximation, and develop an expectation propagation algorithm.Our proposed algorithm leads to higher system expected rewards,faster convergence, and lower variance of value function in a real-world transportation scenario than state-of-the-art analytical andsampling approaches.

ACM Reference Format:Fan Yang, Bo Liu and Wen Dong. 2019. Optimal Control of Com-plex Systems through Variational Inference with a Discrete EventDecision Process. In Proc. of the 18th International Conference onAutonomous Agents and Multiagent Systems (AAMAS 2019), May13-17, 2019, Montréal, Canada. IFAAMAS, 9 pages.

1 INTRODUCTIONA complex social system is a collective system composed of a largenumber of interconnected entities that as whole exhibit proper-ties and behaviors resulted from the interaction of its individualparts [27]. Achieving optimal control of a real-world complex socialsystem is important and valuable. For example, in an urban trans-portation system, a central authority wants to reduce the overalldelay and the total travel time by controlling traffic signals usingtraffic data collected from the Internet of Things [2]. In an epidemic

Proc. of the 18th International Conference on Autonomous Agents and Multiagent Systems(AAMAS 2019), N. Agmon, M. E. Taylor, E. Elkind, M. Veloso (eds.), May 13–17, 2019,Montreal, Canada. © 2019 International Foundation for Autonomous Agents andMultiagent Systems (www.ifaamas.org). All rights reserved.

system, an optimal strategy is pursued to minimize the expected dis-counted losses resulting from the epidemic process over an infinitehorizon, with specific aims such as varying the birth and death rates[18] or early detection of epidemic outbreaks [13]. In the sphere ofpublic opinion, components such as information consumers, newsmedia, social websites, and governments aim to maximize theirown utility functions with optimal strategies. A specific objectivein the public opinion environment is tackling fake news to create aclearer, more trusted information environment [1]. These scenar-ios exemplify the enormous potential applications of a versatileoptimal control framework for general complex systems.

However, key characteristics of complex systems make establish-ing such a framework very challenging. The system typically hasa large number of interacting units and thus a high-dimensionalstate space. State transition dynamics in complex systems are al-ready non-linear, time-variant, and high-dimensional, and thenthese frameworks must account for additional control variables.Prior research in decision making for complex social systems hasgenerally gone in one of two directions: a simulation or analyti-cal approach [21]. Simulation approaches specify system dynamicsthrough a simulationmodel and develop sampling-based algorithmsto reproduce the dynamic flow [29, 41, 46, 47]. These approachescan capture the microscopic dynamics of a complex system withhigh fidelity, but have a high variance and are time-consuming[20]. Analytical approaches instead formulate the decision-makingproblem as a constrained optimization problem with an analyticalmodel through specifying the macroscopic state transitions directly,deriving analytical solutions for optimizing strategies [6, 36]. Theseapproaches can provide a robust solution with less variance, butare applicable only to scenarios with small state spaces [16], orcases with low resolution intervention [9], due to modeling costsand errors [20]. The above research points to a new direction inresearch opportunities that combines the ability to capture systemdynamics more precisely from simulation approaches with the ben-efit of having less variance and being more robust from analyticalapproaches. In this paper, we adopted simulation modeling in spec-ifying the dynamics of a complex system, and developed analyticalsolutions for searching optimal strategies in a complex networkwith high-dimensional state-action space specified through simula-tion modeling.

We formulate the problem of decision making in a complexsystem as a discrete event decision process (DEDP), which iden-tifies the decision making process as a Markov decision process(MDP) and introduces a discrete event model — a kind of simula-tion model [17] — to specify the system dynamics. A discrete eventmodel defines a Markov jump process with probability measureon a sequence of elementary events that specify how the system

Session 1F: Agent Societies and Societal Issues 1 AAMAS 2019, May 13-17, 2019, Montréal, Canada

296

Page 2: Optimal Control of Complex Systems through Variational ... · a MDP in the domain of transportation optimal control. To solve a DEDP analytically, we derived a duality theorem that

components interact and change the system states. These elemen-tary events individually effect only minimal changes to the system,but in sequence together are powerful enough to induce non-linear,time-variant, high-dimensional behavior. Comparing with an MDPwhich specifies the system transition dynamics of a complex sys-tem analytically through a Markov model, a DEDP describes thedynamics more accurately through a discrete event model thatcaptures the dynamics using a simulation process over the micro-scopic component-interaction events. We will demonstrate thismerit through benchmarking with an analytical approach based ona MDP in the domain of transportation optimal control.

To solve a DEDP analytically, we derived a duality theoremthat recasts optimal control to variational inference and param-eter learning, which is an extension of the current equivalenceresults between optimal control and probabilistic inference [24, 38]in Markov decision process research. With this duality, we caninclude a number of existing probabilistic-inference and parameter-learning techniques, and integrate signal processing and decisionmaking into a holistic framework. When exact inference becomesintractable, which is often the case in complex systems due to theformidable state space, our duality theorem implies the possibilityof introducing recent approximate inference techniques to infercomplex dynamics. The method in this paper is an expectationpropagation algorithm, part of a family of approximate inference al-gorithms with local marginal projection. We will demonstrate thatour approach is more robust and has less variance in comparisonwith other simulation approaches in the domain of transportationoptimal control.

This research makes several important contributions. First, weformulate a DEDP — a general framework for modeling complexsystem decision-making problems — by combining MDP and simu-lation modeling. Second, we reduce the problem of optimal controlto variational inference and parameter learning, and develop anapproximate solver to find optimal control in complex systemsthrough Bethe entropy approximation and an expectation propaga-tion algorithm. Finally, we demonstrate that our proposed algorithmcan achieve higher system expected rewards, faster convergence,and lower variance of value function within a real-world transporta-tion scenario than even state-of-the-art analytical and samplingapproaches.

2 BACKGROUNDIn this section, we review the complex social system, the discreteevent model, the Markov decision process, and the variational in-ference framework for a probabilistic graphical model.

2.1 Complex Social SystemA complex social system is a collective system composed of a largenumber of interconnected entities that as whole exhibit proper-ties and behaviors resulted from the interaction of its individualparts. A complex system tends to have four attributes: Diversity, in-teractivity, interdependency, and adaptivity [27]. Diversity meansthe system contains a large number of entities with various at-tributes and characteristics. Interactivity means the diverse enti-ties interact with each other in an interaction structure, such as afixed network or an ephemeral contact. Interdependency means

the state change of one entity is dependent on others through theinteractions. Adaptivity means the entities can adapt to differentenvironments automatically.

In this paper, we temporarily exclude the attribute of adaptiv-ity, the study of which will be future work. We focus on studyingthe optimal control of a complex social system with the attributesof diversity, interactivity, and interdependency, which leads to asystem containing a large number of diverse components, the in-teractions of which lead to the states change. Examples of complexsocial systems include the transportation system where the trafficcongestions are formed and dissipated through the interaction andmovement of individual vehicles, the epidemic system where thedisease is spread through the interaction of different people, andthe public opinion system where people’s minds are influenced andshaped by the dissemination of news through social media.

2.2 Discrete Event ModelA discrete event model defines a discrete event process, also calleda Markov jump process. It is used to specify complex system dy-namics with a sequence of stochastic events that each changes thestate only minimally, but when combined in a sequence inducecomplex system evolutions. Specifically, a discrete event modeldescribes the temporal evolution of a system withM species X ={X (1),X (2), · · · ,X (M )} driven by V mutually independent eventsparameterized by rate coefficients c = (c1, . . . , cV ). At any specifictime t , the populations of the species are xt = (x

(1)t , . . . ,x

(M )t ).

A discrete event process initially in state x0 at time t = 0 can besimulated by: (1) Sampling the event v ∈ {∅, 1, · · · ,V } accord-ing to categorical distribution v ∼ (1 − h0,h1, . . . ,hV ), wherehv (x , cv ) = cv

∏Mm=1 д

(m)v (x

(m)t ) is the rate of eventv , which equals

to the rate coefficients cv times a total of∏M

m=1 д(m)v (x

(m)) differ-ent ways for the individuals to react, and h0(x , c) =

∑Vv=1 hv (x , cv )

the rate of all events. The formulation of hv (x , cv ) comes fromthe formulations of the stochastic kinetic model and stochasticpetri net [43, 44]. (2) Updating the network state deterministicallyx ← x + ∆v , where ∆v represents how an event v changes thesystem states, until the termination condition is satisfied. In a socialsystem, each event involves only a few state and action variables.This generative process thus assigns a probabilistic measure to asample path induced by a sequence of events v0, . . . ,vT happeningbetween times 0, 1, · · · ,T , where δ is an indicator function.

P(x0:T ,v0:T ) = p(x0)T∏t=0

p(vt |xt )δxt+1=xt+∆vt ,

where p(vt |xt ) =

{1 − h0(xt , c), vt = ∅

hk (xt , ck ), vt = k,

The discrete event model is widely used by social scientists tospecify social system dynamics [3] where the system state transi-tions are induced by interactions of individual components. Recentresearch [5, 26, 45] has also applied the model to infer the hiddenstate of social systems, but this approach has not been explored insocial network intervention and decision making.

Session 1F: Agent Societies and Societal Issues 1 AAMAS 2019, May 13-17, 2019, Montréal, Canada

297

Page 3: Optimal Control of Complex Systems through Variational ... · a MDP in the domain of transportation optimal control. To solve a DEDP analytically, we derived a duality theorem that

2.3 Markov Decision ProcessA Markov decision process [33] is a framework for modeling de-cision making in situations where outcomes are partly randomand partly under the control of a decision maker. Formally, anMDP is defined as a tuple MDP⟨S,A, P ,R,γ ⟩, where S representsthe state space and st ∈ S the state at time t , A the action spaceand at the action taken at time t , P the transition kernel of statessuch as P(st+1 |st ,at ), R the reward function such as R(st ,at ) [6]or R(st ) [42] that evaluates the immediate reward at each step,and γ ∈ [0, 1) the discount factor. Let us further define a policyπ as a mapping from a state st to an action at = µ(st ) or a dis-tribution of it parameterized by θ — that is, π = p(at |st ;θ ). Theprobability measure of a length − T MDP trajectory is p(ξT ) =p(s0)

∏T−1t=0 p(at |st ;θ )p(st+1 |st ,at ), where ξT = (s0:T ,a0:T ). Solv-

ing an MDP involves finding the optimal policy π or its asso-ciated parameter θ to maximize the expected future reward —arg maxθEξ (

∑t γ

tRt ;θ ).The graphical representation of an MDP is shown in Figure 1,

where we assume that the full system state st can be representedas a collection of component state variables st = (s

(1)t , ..., s

(M )t ),

so that the state space S is a Cartesian product of the domains ofcomponent state s(m)t : S = S(1)×S(2)×···×S(M ). Similarly, the actionvariable at can be represented as a collection of action variablesat = (a

(1)t , ...,a

(D)t ), and the action spaceA = A(1)×A(2)×· · ·×A(D).

HereM is not necessarily equal to D because st represents the stateof each component of the system while at represents the decisionstaken by the system as a whole. For example, in the problem ofoptimizing the traffic signals in a transportation system where strepresents the locations of each vehicle and at represents the statusof each traffic light, the number of vehicles M may not necessarilyequal to the number of traffic lights D. Usually in complex socialsystems, the number of individual componentsM is much greaterthan the system decision points D.

Prior research in solving a Markov decision process for a com-plex social system could be generally categorized into simulationor analytical approaches. A simulation approach reproduces the dy-namic flow through sampling-based method. It describes the statetransition dynamics with a high-fidelity simulation tool such asMATSIM [11], which simulates the microscopic interactions of thecomponents and how these interactions leads to macroscopic statechanges. Given current state st and action at at time t , a simulationapproach uses a simulation tool to generate the next state st+1.

An analytical approach develops analytical solutions to solvea constrained optimization problem. Instead of describing the dy-namics with a simulation tool, an analytical approach specifies thetransition kernel analytically with probability density functions thatdescribe the macroscopic state changes directly. Given current statest and action at , it computes the probability distribution of the nextstate st+1 according to the state transition kernel p(st+1 | st ,at ).However, approximations are required to make the computationtractable. For an MDP containing M binary state variables and Dbinary action variables, the state space is 2M , the action space is2D , the policy kernel is a 2M × 2D matrix, and the state transitionkernel (fixed action) is a 2M × 2M matrix. SinceM is usually muchlarger than D in complex social systems, the complexity bottleneckis usually the transition kernel with size 2M × 2M , the complexity

of which grows exponentially with the number of state variables.Certain factorizations and approximations must be applied to lowerthe dimensionality of the transition kernel.

Usually analytical approaches solve complex social systemMDPsapproximately by enforcing certain independence constraints [31].For example, Cheng [4] assumed that a state variable is only depen-dent on its neighboring variables. Sabbadin, Peyrard and Sabbadin[28, 30] exploited a mean field approximation to compute and up-date the local policies. Weiwei approximated the state transitionkernel with differential equations [22]. These assumptions intro-duces additional approximations that results in modeling errors.In the next section, we propose a discrete event decision processwhich reduces the complexity of the transition probabilities, andwhich does not introduce additional independence assumptions.

Two specific approaches of solving an MDP are optimal control[32] and reinforcement learning [34]. Optimal control problemsconsist of finding the optimal decision sequence or the time-variantstate-action mapping that maximizes the expected future reward,given the dynamics and reward function. Reinforcement-learningproblems target the optimal stationary policy that maximizes theexpected future reward while not assuming knowledge of the dy-namics or the reward function. In this paper, we address the problemof optimizing a stationary policy to maximize the expected futurereward, assuming known dynamics and reward function.

2.4 Variational InferenceA challenge in evaluating and improving a policy in a complexsystem is that the state space grows exponentially with the numberof state variables, which makes probabilistic inference and param-eter learning intractable. For example, in a system with M binarycomponents, the size of state space S will be 2M , let alone theexploding transition kernel. One way to resolve this issue is apply-ing variation inference to optimize a tractable lower bound of thelog expected future reward through conjugate duality. Variationalinference is a classical framework in the probabilistic graphicalmodel community [40]. It exploits the conjugate duality betweenlog-partition function and the entropy function for exponentialfamily distributions. Specifically, it solves the variational prob-lem log

∫exp ⟨θ ,ϕ(x)⟩ dx = supq(x )

{∫q(x) ⟨θ ,ϕ(x)⟩ dx + H (q)

},

where θ and ϕ(x) are respectively canonical parameters and suffi-cient statistics of an exponential family distribution, q is an aux-iliary distribution and H (q) = −

∫dxq(x) logq(x) is the entropy

of q. For a tree-structured graphical model (here we use the sim-plified notation of a series of xt ), the Markov property admits afactorization of q into the product and division of local marginalsq(x) =

∏t qt (xt−1,t )/

∏t qt (xt ). Substituting the factored forms

of q into the variational target, we get an equivalent constrainedoptimization problem involving local marginals and consistencyconstraints among those marginals, and a fixed-point algorithminvolving forward statistics αt and backward statistics βt . Thereare two primary forms of approximation of the original variationalproblem: the Bethe entropy problem and a structured mean field.The Bethe entropy problem is typically solved by loopy belief prop-agation or an expectation propagation algorithm.

Session 1F: Agent Societies and Societal Issues 1 AAMAS 2019, May 13-17, 2019, Montréal, Canada

298

Page 4: Optimal Control of Complex Systems through Variational ... · a MDP in the domain of transportation optimal control. To solve a DEDP analytically, we derived a duality theorem that

s1,t-1

sM,t-1

a1,t-1

t-1 tTime

t+1

s1,t

sM,t

aD,t-1

a1,t

s1,t+1

sM,t+1

aD,t

Figure 1: Graph representation of an MDP

s1,t-1

sM,t-1

a1,t-1

vt-1

t-1 tTime

t+1

s1,t

sM,t

aD,t-1

a1,t

vt

s1,t+1

sM,t+1

aD,t

Figure 2: Graph representation of a DEDP

3 METHODOLOGYIn this section, we develop a DEDP and present a duality theoremthat extends the equivalence of optimal control with probabilisticinference and parameter learning. We also develop an expectationpropagation algorithm as an approximate solver to be applied inreal-world complex systems.

3.1 Discrete Event Decision ProcessThe primary challenges in modeling a real-world complex systemare the exploding state space and complex system dynamics. Oursolution is to model the complex system decision-making processas a DEDP.

The graphical representation of a DEDP is shown in Figure 2. For-mally, a DEDP is defined as a tuple DEDP⟨S,A,V,C, P ,R,γ ⟩, whereS is the state space and st =

(s(1)t , ..., s

(M )t

)∈ S a M-dimensional

vector representing the state of each component at time t , A is theaction space and at =

(a(1)t , ...,a

(D)t

)∈ A a D-dimensional vector

representing the action taken by the system at time t . As before,Mis not necessarily equal to D, and M is usually much larger thanD in complex social systems. Both s(m)t and a(d )t could take real orcategorical values depending on the applications.V = {∅, 1, ...,V } is the set of events and vt ∈ V a scalar fol-

lowing categorical distributions indicating the event taken at timet and changing the state by ∆vt . C is the function mapping ac-tions to event rate coefficients which takes a D-dimensional vectoras input, and outputs a V -dimensional vector c = (c1, ..., cV ) =C(at ), and P is the transition kernel of states induced by eventsP(st+1,vt |st ,at ) = p(vt | st ,at )δst+1=st+△vt , where p(vt = v |st ,at ) represents the probability of an event v happened at time t :

p(vt = v | st ,at ) =

{hv (st , cv ) if v , ∅1 −

∑Vv=1 hv (st , cv ) if v = ∅

Following the definitions in the discrete eventmodel,hv (st , cv ) =cv ·

∏Mm=1 д

(m)v (s

(m)t ) is the probability for eventv to happen, which

equals to the rate coefficients cv times a total of∏M

m=1 д(m)v (s

(m)t )

ways that the components can interact to trigger an event.The immediate reward function R is a function of system states,

defined as the summation of reward evaluated at each componentR(st ) =

∑Mm=1 R

(m)t (s

(m)t ) , and γ ∈ [0, 1) is the discount factor.

We further define a policy π as a mapping from a state st to anaction at = µ(st ;θ ) or a distribution of it parameterized by θ — thatis, π = p(at |st ;θ ). The parameterized policy can take any form,such as a lookup table where θ represents values of each entriesin the table, a Gaussian distribution where θ represents the meanand variance, or a neural network where θ represents the networkweights. Solving a DEDP involves finding the optimal policy π orits associated parameter θ to maximize the expected future reward— arg maxθEξ (

∑t γ

tRt ;θ ). The probability measure of a length−TDEDP trajectory with a stochastic policy is as follows, where δ isan indicator function and ξT = (s0:T ,a0:T ,v0:T ):

p(ξT ) = p(s0)∏T−1

t=0

(p(at | st )p(vt | st ,at )δst+1=st+∆vt

)The probability measure of that with a deterministic policy is

this:

p(ξT ) = p(s0)∏T−1

t=0

(δat=µ(st )p(vt | st ,at )δst+1=st+∆vt

)= p(s0)

∏T−1t=0

(p(vt | st , µ(st ))δst+1=st+∆vt

):= p(s0)

∏T−1t=0

(p(vt | st )δst+1=st+∆vt

)ADEDPmakes a tractable representation of complex system con-

trol problem by representing the non-linear and high-dimensionalstate transition kernel with microscopic events. A vanilla MDP isan intractable representation because the state-action space growsexponentially with the number of state-action variables. In compar-ison, the description length of a DEDP grows linearly in the numberof events. As such, a DEDP greatly reduces the complexity of speci-fying a complex system control problem through introducing anauxiliary variable (event), and can potentially describe complex andhigh-fidelity dynamics of a complex system.

A DEDP could be reduced to an MDP if marginalizing out theevents∑

v0:T p(s0:T ,a0:T ,v0:T )

=∑v0:T p(s0)

∏T−1t=0

(p(at | st )p(vt | st ,at )δst+1=st+∆vt

)= p(s0)

∏T−1t=0

(p(at | st )

∑vt p(st+1,vt | st ,at )

)= p(s0)

∏T−1t=0 (p(at | st )p(st+1 | st ,at ))

Thus, the only difference between a DEDP and an MDP is that aDEDP introduces events to describe the state transition dynamics.Compared with the aforementioned models introducing indepen-dence constraints to make MDPs tractable, a DEDP does not make

Session 1F: Agent Societies and Societal Issues 1 AAMAS 2019, May 13-17, 2019, Montréal, Canada

299

Page 5: Optimal Control of Complex Systems through Variational ... · a MDP in the domain of transportation optimal control. To solve a DEDP analytically, we derived a duality theorem that

any independence assumptions. It defines a simulation process thatdescribes how components interact and trigger an event, and howthe aggregation of events leads to system state changes. In thisway, a DEDP captures the microscopic dynamics in a macroscopicsystem, leading to more accurate dynamics modeling.

3.2 Duality Theorem on Value FunctionTo solve a DEDP with a high-dimensional state and action space,we derive the convex conjugate duality between the log expectedfuture reward function and the entropy function of a distributionover finite-length DEDP trajectories, and the corresponding dual-ity in the parameter space between the log discounted trajectory-weighted reward and the distribution of finite-length DEDP trajec-tories. As a result, we can reduce the policy evaluation problem toa variational inference problem that involves the entropy functionand can be solved by various probabilistic inference techniques.

Specifically, in a complex system decision processDEDP⟨S,A,V,C, P ,R,γ ⟩, let T be a discrete time, m the compo-nent index, ξT the length-T trajectory of a DEDP starting frominitial state s0, and V π = E

(∑∞t=0 γ

tRt ;π)the expected future re-

ward (value function), where γ ∈ [0, 1) is a discount factor. Defineq(T ,m, ξT ) as a proposal joint probability distribution over finitelength−T DEDP trajectories, r (T ,m, ξT ;π ) = γT P(ξT ;π )R

(m)T (s

(m)T )

as the discounted trajectory-weighted reward component whereP(ξT ;π ) is the probability distribution of a trajectory with policyπ , and R(m)T (s

(m)T ) as the reward evaluated at componentm at time

T with state s(m)T . We thus have the following duality theorem.Theorem 1.

logV π (r ) = supq

©­«∑

T ,m,ξT

q(T ,m, ξT )log r (T ,m, ξT ; π ) + H (q)ª®¬

In the above, equality is satisfiedwhen∑T

∑m

∑ξT r (T ,m, ξT ;π ) <

∞ and q(T ,m, ξT ) ∝ γT P(ξT )R(m)T (s

(m)T ). As such, logV

π (r ) is theconvex conjugate of H (q(T ,m, ξT )). H (q(T ,m, ξT )) is a convexfunction of q(T ,m, ξT ), so by property of the convex conjugate,H (q(T ,m, ξT )) is also a conjugate of logV π (r ). The proof for thistheorem is shown in the Appendix.

Theorem 1 provides a tight lower bound of the log expectedreward and, more importantly, defines a variational problem interms analogous to well-known variational inference formulationsin the graphic model community [40], where a number of varia-tional inference methods can be introduced such as exact inferencemethods, sampling-based approximate solutions, and variationalinference. Theorem 1 extends the equivalence between optimalcontrol and probability inference in recent literatures to a generalvariational functional problem. Specifically, it gets rid of the proba-bility likelihood interpretation of the value function in [38, 39], andthe prior assumption that value function is in the multiplicationform of local-scope value functions [24].

3.3 Expectation Propagation for OptimalControl

Theorem 1 implies a generalized policy-iteration paradigm aroundthe duality form: solving the variational problem as policy evalu-ation and optimizing the target over parameter θ with a known

mixture of finite-length trajectories as policy improvement. In thefollowing, we develop the policy evaluation and improvement al-gorithm with a deterministic policy at = µ(st ;θ ), the derivationfor which is given in the Appendix. The stochastic policy caseπ = p(at |st ;θ ) will lead to a similar result, which is not presentedhere due to the limit of space.

In policy evaluation, the Markov property of a DEDP admits fac-torizations P(ξT ;π ) = p(s0)

∏Tt=1

(p(vt | st ;θ )δst+1=st+∆vt

)and

q(T ,m, ξT ) = q(T ,m)q(s0)∏T

t=1 q(st−1,t ,vt−1 |T )/∏T−1

t=1 q(st |T ),where q(T ,m) = γT (1 − γ )/M is the length prior distribution tomatch the discount factor, and q(st |T ) and q(st−1,t ,vt−1 |T ) are lo-cally consistent one-slice and two-slice marginals. To cope withthe exploding state space, we apply the Bethe entropy approxima-tion. Specifically, we relax the formidable searching state space ofst into an amenable space through the mean field approximationq(st |T ,m) =

∏Mm̂=1 q(s

(m̂)t |T ,m), whereq(s(m̂)t |T ,m) is the one-slice

marginal involving only component m̂. Applying the factorizationand approximation, we get the following Bethe entropy problem(let

∑st−1,t ,vt−1\s

(m̂)t

represent the summation over all value combi-

nations of st−1, st ,vt−1 except a fixed s(m̂)t ).

max over q(s (m̂)t |T ,m), q(st−1,t , vt |T ,m) ∀T , t ≤ T ,m∑T ,m

T−1∑t=1

∑m̂

∑s (m̂)t

∑T ,m

q(T ,m, s (m̂)t )logq(s (m̂)t |T ,m)

−∑T ,m

∑T−1t=1

∑st−1,t ,vt−1 q(T ,m, st−1,t , vt−1)log

(q(st−1,t ,vt−1 |T ,m)p(st ,vt−1 |st−1 ;θ )

)−

∑T ,m,sT−1,T ,vT−1 q(T ,m, sT−1,T , vT−1)log

(q(sT−1,T ,vT−1 |T ,m)

p(sT ,vT−1 |sT−1 ;θ )R(m)T

)(1)

subject to:

∑st−1,t ,vt−1\s

(m̂)t−1

q(st−1,t , vt−1 |T ,m) = q(s (m̂)t−1 |T ,m),∑st−1,t ,vt−1\s

(m̂)t

q(st−1,t , vt−1 |T ,m) = q(s (m̂)t |T ,m)

We solve this with the method of Lagrange multipliers, whichleads to a forward-backward algorithm that updates the forwardmessages α (m̂)t |T ,m (s

(m̂)t ) and backward messages β (m̂)t |T ,m (s

(m̂)t )

marginally according to the average effects of all other components— that is, a projected marginal kernel p(s(m̂)t ,at |s

(m̂)t−1 ;θ ). Therefore,

the algorithm achieves linear complexity over the number of com-ponents for eachT andm, and quadratic time complexity over timehorizon H to compute all messages for all T ≤ H . To further lowerdown the time complexity, we define β (m̂)t (s (m̂)t )=

∑m

∞∑T=t

q(T,m)β (m̂)t |T ,m (s (m̂)t )

by gathering together backwardmessages sharing t andα (m̂)t (s(m̂)t ) =

α(m̂)t |T ,m (s

(m̂)t ) by noting that α (m̂)t |T ,m (s

(m̂)t ) doesn’t depend on T ,m.

This leads to the following forward-backward algorithm, which islinear in time horizon:

α(m̂)t (s

(m̂)t ) ∝

∑s (m̂)t−1 ,vt−1

α(m̂)t−1(s

(m̂)t−1) · p(s

(m̂)t ,vt−1 |s

(m̂)t−1 ;θ ) (2)

β(m̂)t (s

(m̂)t ) =

∑m

q(t ,m)β(m̂)t |t,m (s

(m̂)t )

+∑

s (m̂)t+1 ,vt

p(s(m̂)t+1,vt |s

(m̂)t ;θ )β (m̂)t+1(s

(m̂)t+1) (3)

Session 1F: Agent Societies and Societal Issues 1 AAMAS 2019, May 13-17, 2019, Montréal, Canada

300

Page 6: Optimal Control of Complex Systems through Variational ... · a MDP in the domain of transportation optimal control. To solve a DEDP analytically, we derived a duality theorem that

In policy improvement, we maximize the log expected future re-ward functionL(θ ) =

∑T ,m,ξT q(T ,m, ξT ;θold)log

(γT P(ξT ;θ )R

(m)T

)over parameter θ with q(T ,m, ξT ) inferred from θold via gradientascent update θnew = θold + ϵ · ∂L

∂θ

��� θ old , or more aggressively by

setting θnew so that ∂L∂θ |θ new = 0 . This objective can be simplified

by dropping irrelevant terms and keeping only those involving pol-icy: L(θ ) =

∑T ,m,t,vt ,st

q(T ,m,vt , st ;θold)·logP(vt |st ;θ )+const. The

gradient is obtained from chain rule and messages αt (xt ), βt (xt )through dynamic programming:

∂L(θ )∂θ =

∑t,st

∏m̂ α (m̂)t (s (m̂)t ,vt =v )β

(m̂)t (s (m̂)t ,vt =v )

cv∂cv∂θ

−∑t,st

∏m̂ α (m̂)t (s (m̂)t ,vt =∅)β

(m̂)t (s (m̂)t ,vt =∅)·

∏m дmv (s

(m)t )

1−∑Vv=1 cv ·

∏m дmv (s

(m)t )

∂cv∂θ (4)

In summary, we give our optimal control algorithm of complexsystems as Algorithm 1.

Algorithm 1 Optimal control of social systems

Input: The DEDP tuple DEDP⟨S,A,V,C, P ,R,γ ⟩, initial policyparameter θOutput: Optimal policy parameter θProcedure: Iterate until convergence:• policy evaluation

Iterate until convergence:- Update forward messages according to Eq. (2)- Update backward messages according to Eq. (3)• policy improvement

Update parameter θ according to the gradient in Eq. (4).

3.4 DiscussionsIn the above we developed a DEDP for modeling complex systemdecision making problems, and derived a expectation propagationalgorithm for solving the DEDP. our algorithm is also applicableon a Markov decision process with other simulation models. Whilewe used a discounted expected total future reward in the previousderivation, our framework is also applicable to other types of ex-pected future reward, such as a finite horizon future reward, wherewe use a different probability distribution of time q(T ) = 1

T .

4 EXPERIMENTSIn this experiment, we benchmark algorithm 1 against severaldecision-making algorithms for finding the optimal policy in acomplex system.

Overview: The complex social system in this example is a trans-portation system. The goal is to optimize the policy such thateach vehicle arriving at the correct facilities at correct time (beingat work during work hours and at home during rest hours) andspending minimum time on roads. We formulate the transporta-tion optimal control problem as a discrete event decision processDEDP⟨S,A,V,C, P ,R,γ ⟩. The state variables st = (s

(1)t , · · · , s

(M−1)t , t)

represent the populations at (M − 1) locations and the currenttime t . All events are of the form p ·m1

cm1m2→ p ·m2—an individ-

ual p moving from location m1 to location m2 with rate (proba-bility per unit time) cm1m2—decreasing the population at m1 by

one and increasing the population at m2 by one. We also intro-duce auxiliary event ∅ that doesn’t change any system state, andset the rates of leaving facilities and selecting alternative down-stream links as action variables. We implement the state transitionp(st+1,vt | st ,at ) following the fundamental diagram of traffic flow[12] that simulate the movement of vehicles. The reward functionR(st ) =

∑m β(m)t,perfs

(m)t + β

(m)t, travs

(m)t emulates the Charypa-Nagel

scoring function in transportation research [12] to reward perform-ing the correct activities at facilities and penalize traveling on roads, where β (m)t, trav and β

(m)t,perf are the score coefficients. We implement

the deterministic policy as a function of states through a neuralnetwork at = µ(st ) = NN(st ;θ ) and apply Algorithm 1 to find theoptimal policy parameter θ .

Benchmark Model Description: The transition kernel basedon an MDP in this general form p(st+1 | st ,at ) in a complex systemis too complicated to be modeled exactly with an analytical form,due to the high-dimensional state-action space, and the complexsystem dynamics. As such, we benchmark with the following algo-rithms: (1) Analytical approaches based on an MDP that uses Taylorexpansion to approximate the intractable transition dynamics withdifferential equations, which leads to a guided policy search (GPS)algorithm [25] with a known dynamics that uses iterative linearquadratic regulator (iLQR) for trajectory optimization and super-vised learning for training a global policy µ(st ), implemented as afive-layer neural network. Other aforementioned approximations[4, 28, 30] are not applicable because in their settings each compo-nent takes an action, resulting in local policies for each component.Whereas in our problem the action is taken by the system as awhole, and therefore no local policies. (2) Simulation approachesthat reproduce the dynamic flow through sampling the state actiontrajectories from the current policy and the system transition dy-namics, which leads to a policy gradient (PG) algorithm, the policyof which is implemented as a four-layer neural network; and (3) anactor-critic (AC) algorithm [23] that implements the policy µ(st )as an actor network with four layers and the state-action valuefunction Q(st ,at ) as a critic network with five layers.

Performance and Efficiency: We benchmark the algorithmsusing the SynthTown scenario (Fig. 3), which has one home facility,one work facility, 23 road links, and 50 individuals going to workfacility in the morning (9 am) and home facility in the evening (5pm). A training episode is one day. This scenario is small enoughfor studying the details of different algorithms.

In Figure 4, the value-epoch curve of VI (our algorithm) dom-inates those of the other algorithms almost everywhere. Table 1indicates that VI requires the fewest training epochs to convergeto the highest total rewards per episode. Figure 5 presents the av-erage vehicle distribution of ten runs at different locations (Home(h), Work (w), and roads 1-23) using the learned policy with eachalgorithm. This figure implies that the learned policy of VI leads tothe largest amount of vehicles being at work during work hours(9 am to 5 pm), and least amount of time the vehicles spending onroads.

In the SynthTown scenario, high rewards require the individualsto perform the correct activities (home, work, and so on) at theright time, and to spend less time on roads. VI achieves the bestperformance when evaluating a policy by considering the whole

Session 1F: Agent Societies and Societal Issues 1 AAMAS 2019, May 13-17, 2019, Montréal, Canada

301

Page 7: Optimal Control of Complex Systems through Variational ... · a MDP in the domain of transportation optimal control. To solve a DEDP analytically, we derived a duality theorem that

Figure 3: SynthTown roadnetwork

−10

0

10

20

0 25 50 75 100Training Epochs

Tota

l Rew

ard

Per E

piso

de

AlgorithmsVI GPS

AC PG

Figure 4: Training process onSynthTown

0

50

h

VIGPSACPG

0

49

w

011

013

016

0112

0115

0220

0

321

0

8

22

0

223

1 3 5 7 9 11 13 15 17 19 21

Hours of day

Figure 5: Average number ofvehicles with trained policies

state-action space with VI approximation — the evolution of eachstate variables in the mean field of the other state variables. Mod-eling the complex system transition dynamics based on an MDPanalytically with Taylor approximation will introduce modelingerrors (GPS). Value estimation from Monte Carlo integration inhigh-dimensional state space has high variance (PG). A small per-turbation of policy will result in significant change to the immediatereward and value in later steps, which makes it difficult to estimatethe correct value from sampled trajectories, and as a result difficultto compute the correct value gradient (AC).

Dataset SynthTown BerlinMetrics TRPE EC TRPE ECVI 24.21 75 11.52 200GPS 14.79 100 -10.33 -AC 16.76 100 -9.69 -PG 11.77 150 -15.52 -

Table 1: Comparing algorithms in total reward per episode(TRPE) and epochs to converge (EC)

We also benchmark the performance of all algorithms usingthe Berlin scenario, which consists of a network of 1,530 locationsand the trips of 9, 178 synthesized individuals [12]. Table 1 showsthat VI outperformed in total rewards per episode, while the otheralgorithms did not even converge in a reasonable number of epochs.

In summary, VI outperformed guided policy search, policy gradi-ent, and actor-critic algorithms in all scenarios. These benchmark-ing algorithms provide comparable results in a small dataset such asthe SynthTown scenario, but became difficult to train when appliedto a larger dataset such as Berlin.

5 RELATEDWORKSA number of prior works have explored the connection betweendecision making and probabilistic inference. The earliest such re-search is the Kalman duality proposed by Rudolf Kalman [14]. Sub-sequent works have expanded the duality in stochastic optimalcontrol, MDP planning, and reinforcement learning. Todorov andKappen expanded the connection between control and inference byrestricting the immediate cost and identifying the stochastic opti-mal control problems as linearly-solvable MDPs [37] or KL control

problems [15]. However, the passive dynamics in a linearly-solvableMDPs and the uncontrolled dynamics in a KL control problem be-come intractable in a social system scenario due to the complicatedsystem dynamics. Toussaint, Hoffman, David, and colleagues broad-ened the connection by framing planning in an MDP as inference ina mixture of PGMs [38]. The exact [8] and variational inference [7]approaches in their framework encounter the transition dynamicsintractability issue in a social network setting due to the compli-cated system dynamics. Their sampling-based alternatives [10]experience the high-variance problem in a social network due tothe exploding state space. Levine, Ziebart, and colleagues widenedthe connection by establishing the equivalence between maximumentropy reinforcement learning, where the standard reward ob-jective is augmented with an entropy term [48], and probabilisticinference in a PGM [19, 48]. They optimize a different objectivefunction compared with our method. Our approach is an extensionof Toussaint’s formulation [38], but differs in that we establish theduality with a DEDP, and provide new insights into the duality fromthe perspective of convex conjugate duality. Compared with theexisting exact inference [8] and approximate inference solutions[7, 10], our algorithm is more efficient with gathering the messages,and more scalable with the use of Bethe entropy approximation.

Other formulations of the decision-making problems which aresimilar to our framework are the Option network [35] and themulti-agent Markov decision process (MMDP) [31]. An option net-work extends a traditional Markov decision process with options —closed loop policies for taking actions over a period of time, withthe goal of providing a temporal abstraction of the representation.In our framework, we introduce an auxiliary variable v with thegoal of providing a more accurate and efficient modeling of the sys-tem dynamics. An MMDP represents sequential decision-makingproblems in cooperative multi-agent settings. There are two fun-damental differences between our framework and MMDP. First,unlike MMDP where each agent takes an action at each time step,in a DEDP the actions are taken by the system and the dimensional-ity of states not necessarily equals to the dimensionality of actions.Second, unlike MMDP where each agent has its own information,a DEDP models the decision making of a large system where onlyone controller has all the information and makes decisions for theentire system.

Session 1F: Agent Societies and Societal Issues 1 AAMAS 2019, May 13-17, 2019, Montréal, Canada

302

Page 8: Optimal Control of Complex Systems through Variational ... · a MDP in the domain of transportation optimal control. To solve a DEDP analytically, we derived a duality theorem that

6 CONCLUSIONSIn this paper, we have developed DEDP for modeling complex sys-tem decision-making processes. We reduce the optimal controlof DEDP to variational inference and parameter learning arounda variational duality theorem. Our framework is capable of opti-mally controlling real-world complex systems, as demonstrated byexperiments in a transportation scenario.

7 APPENDIX7.1 Derivation of Theorem 1

logV π =log(∑

T∑m

∑ξT q(T ,m,ξT )

r (T ,m,ξT ;π )q(T ,m,ξT )

)≥∑T

∑m

∑ξT q(T ,m,ξT )log

r (T ,m,ξT ;π )q(T ,m,ξT )

=∑T

∑m

∑ξT q(T ,m,ξT )log r (T ,m,ξT ;π )+H(q(T ,m,ξT ))

7.2 Derivation of Eq. (1)Rearranging the terms and applying the approximations describedin the main text, the target becomes the following:∑

T∑m

∑ξT q(T ,m,ξT )log

(γT P (ξT ;θ )R(m)T

)+H(q(T ,m,ξT ))

=∑T

∑m

∑ξT q(T ,m,ξT )log

(γT

T−1∏t=1

p(st ,vt−1 |st−1;θ )·p(sT ,vT−1 |sT−1;θ )R(m)T

)−

∑T

∑m

∑ξT q(T ,m,ξT )log

©­­«q(T ,m)T∏t=1

q(st−1,t ,vt−1 |T ,m)

T−1∏t=1

q(st |T ,m)

ª®®¬=∑T

∑m (q(T ,m)log

(γT

q(T ,m)

)−T−1∑t=1

∑st−1,t ,vt−1 q(T ,m,st−1,t )log

(q(st−1,t ,vt−1 |T ,m)p(st ,vt−1 |st−1;θ )

)−∑t∑m

∑st−1,t ,vt−1 q(T ,m,st−1,t )log

(q(st−1,t ,vt−1 |T ,m)

p(st ,vt−1 |st−1;θ )R(m)t

)+

∑T ,m

T−1∑t=1

∑m̂

∑s (m̂)t

q(T ,m,s (m̂)t )logq(s (m̂)t |T ,m)

7.3 Derivation to Solve Eq. (1)We solve this maximization problem with the method of Lagrangemultipliers.L =

∑T ,m q(T ,m)log

(γT

q(T ,m)

)−

∑T ,m

T−1∑t=1

∑st−1,t ,vt−1 q(T ,m,st−1,t ,vt−1)log

(q(st−1,t ,vt−1 |T ,m)p(st ,vt−1 |st−1;θ )

)−

∑T ,m

∑sT−1,T q(T ,m,sT−1,T ,vT−1)log

(q(sT−1,T ,vT−1 |T ,m)

p(sT ,vT−1 |sT−1;θ )R(m)T

)+

∑T ,m

T−1∑t=1

∑m̂

∑s (m̂)t

q(T ,m,s (m̂)t )logq(s (m̂)t |T ,m)

+∑t,m̂,s (m̂)t−1 ,T ,m

αm̂t−1,s (m̂)t−1 ,T ,m

(∑st−1,t ,vt−1\s

(m̂)t−1

q(st−1,t ,vt−1 |T ,m)−q(s(m̂)t−1 |T ,m)

)+∑t,m̂,s (m̂)t ,T ,m

βm̂t,s (m̂)t ,T ,m

(∑st−1,t ,vt−1\s

(m̂)t

q(st−1,t ,vt−1 |T ,m)−q(s(m̂)t |T ,m)

)Taking derivative with respect to q(st−1,t ,at |T ,m) and q(s (m̂)t |T ,m),

and setting it to zero, we then getfor t=1, ...T−1

q(st−1,t ,vt−1 |T ,m) ∝exp©­­«

∑m̂ α (m̂)

t−1,s (m̂)t−1 ,T ,mq(T ,m)

ª®®¬·p(st ,vt−1 |st−1;θ )exp©­­«

∑m̂ β (m̂)

t,s (m̂)t ,T ,mq(T ,m)

ª®®¬for t=T

q(st−1,t ,vt−1 |T ,m) ∝exp©­­«

∑m̂ α (m̂)

t−1,s (m̂)t−1 ,T ,mq(T ,m)

ª®®¬·p(st ,vt−1 |st−1;θ )R(m)T exp

©­­«∑m̂ β (m̂)

t,s (m̂)t ,T ,mq(T ,m)

ª®®¬q(s (m̂)t |T ,m) = 1

Z (m)T

exp©­­«α (m̂)

t,s (m̂)t ,T ,m+β (m̂)

t,s (m̂)t ,T ,mq(T ,m)

ª®®¬

Marginalizing over q(st−1,t |T ,m), we have

q(s (m̂)t−1,t ,vt−1 |T ,m)

=∑s (m̂′)

t−1,t ;m̂′,m̂

1Zt

exp©­­«

∑m̂ α (m̂)

t−1,s (m̂)t−1 ,T ,mq(T ,m)

ª®®¬·p(st ,vt−1 |st−1;θ )exp©­­«

∑m̂ β (m̂)

t,s (m̂)t ,T ,mq(T ,m)

ª®®¬:= 1

Ztexp

©­­«α (m̂)

t−1,s (m̂)t ,T ,m+β (m̂)

t,s (m̂)t ,T ,mq(T ,m)

ª®®¬·p(s(m̂)t ,vt−1 |s

(m̂)t−1 ;θ )

Wedenote exp©­­«α (m̂)

t,s (m̂)t ,T ,mq(T ,m)

ª®®¬ as α(m̂)t |T ,m

(s (m̂)t

), exp

©­­«β (m̂)

t,s (m̂)t ,T ,mq(T ,m)

ª®®¬ as β(m̂)t |T ,m

(s (m̂)t

).

We can compute α , β through a forward-backward iterativeapproach:

forward: q(s (m̂)t |T ,m)=∑s (m̂)t−1 ,vt−1

q(s (m̂)t−1,t ,vt−1 |T ,m)

⇒ α (m̂)t |T ,m (s(m̂)t )=

Z (m̂)tZt

∑s (m̂)t−1 ,vt−1

α (m̂)t−1|T ,m (s(m̂)t−1 )·p(s

(m̂)t ,vt−1 |s

(m̂)t−1 ;θ )

backward: q(s (m̂)t−1 |T ,m)=∑s (m̂)t ,vt−1

q(s (m̂)t−1,t ,vt−1 |T ,m)

⇒ β (m̂)t−1|T ,m (s(m̂)t−1 )=

Z (m̂)tZt

∑s (m̂)t ,vt−1

p(s (m̂)t ,vt−1 |s(m̂)t−1 ;θ )·β

(m̂)t |T ,m (s

(m̂)t )

for t=T ,m̂=m

β (m̂)t−1|T ,m (s(m̂)t−1 )=

Z (m̂)tZt

∑s (m̂)t ,vt−1

p(s (m̂)t ,vt−1 |s(m̂)t−1 ;θ )·β

(m̂)t |T ,m (s

(m̂)t )R(m)T

7.4 Derivation of Eq. (3)We can further simplify our algorithm by gathering the messages.For notational simplicity, we can absorb the reward R(m)T (s (m)T ) intothe β (m)T |T ,m (s

(m)T ) term, so that β (m)T |T ,m (s

(m)T )=R(m)T (s (m)T ) and β (m̂)T |T ,m (s

(m̂)T )=1

for m̂,m. We then define β (m̂)t (s (m̂)t )=∑m

∞∑T=t

q(T ,m)β (m̂)t |T ,m (s(m̂)t ) and have

β (m̂)t (s (m̂)t )

=∑m

∞∑T=t

q(T ,m)β (m̂)t |T ,m (s(m̂)t )

=∑mq(t,m)β (m̂)t |t,m (s

(m̂)t )+

∑s (m̂)t+1 ,vt

p(s (m̂)t+1 ,vt |s(m̂)t ;θ )β (m̂)t+1 (s

(m̂)t+1 )

Observing that∞∑T=1

T∑t=1⇐⇒

∞∑t=1

∞∑T=t

, this summation form of mes-sages will be useful in the parameter-learning phase.

7.5 Derivation of Eq. (4)∂L∂θ =

∑T ,m,t

∑st

q(T ,m,vt =v,st )cv

∂cv∂θ

−∑

T ,m,t

∑st

q(T ,m,vt =∅,st )·дv (st )1−

∑Vv=1 cvдv (st )

∂cv∂θ

=∑t,st

∏m̂ α (m̂)t (s (m̂)t ,vt =v )β

(m̂)t (s (m̂)t ,vt =v )

cv∂cv∂θ

−∑t,st

∏m̂ α (m̂)t (s (m̂)t ,vt =∅)β

(m̂)t (s (m̂)t ,vt =∅)·

∏m дmv (s

(m)t )

1−∑Vv=1 cv ·

∏m дmv (s

(m)t )

∂cv∂θ

Session 1F: Agent Societies and Societal Issues 1 AAMAS 2019, May 13-17, 2019, Montréal, Canada

303

Page 9: Optimal Control of Complex Systems through Variational ... · a MDP in the domain of transportation optimal control. To solve a DEDP analytically, we derived a duality theorem that

REFERENCES[1] Hunt Allcott and Matthew Gentzkow. 2017. Social media and fake news in the

2016 election. Journal of Economic Perspectives 31, 2 (2017), 211–36.[2] PG Balaji, X German, and D Srinivasan. 2010. Urban traffic signal control using

reinforcement learning agents. IET Intelligent Transport Systems 4, 3 (2010),177–188.

[3] Andrei Borshchev. 2013. The big book of simulation modeling: multimethodmodeling with AnyLogic 6. AnyLogic North America Chicago.

[4] Qiang Cheng, Qiang Liu, Feng Chen, and Alexander T Ihler. 2013. Variationalplanning for graph-based MDPs. In Advances in Neural Information ProcessingSystems. 2976–2984.

[5] Le Fang, Fan Yang,WenDong, Tong Guan, and ChunmingQiao. 2017. ExpectationPropagation with Stochastic Kinetic Model in Complex Interaction Systems. InAdvances in Neural Information Processing Systems. 2026–2036.

[6] Mehrdad Farajtabar, Yichen Wang, Manuel Gomez Rodriguez, Shuang Li,Hongyuan Zha, and Le Song. 2015. Coevolve: A joint point process model for in-formation diffusion and network co-evolution. In Advances in Neural InformationProcessing Systems. 1954–1962.

[7] Thomas Furmston and David Barber. 2010. Variational methods for reinforcementlearning. In Proceedings of the Thirteenth International Conference on ArtificialIntelligence and Statistics. 241–248.

[8] Thomas Furmston and David Barber. 2012. Efficient inference in markov controlproblems. arXiv preprint arXiv:1202.3720 (2012).

[9] Miao He, Lei Zhao, and Warren B Powell. 2010. Optimal control of dosagedecisions in controlled ovarian hyperstimulation. Annals of Operations Research178, 1 (2010), 223–245.

[10] Matt Hoffman, Hendrik Kueck, Nando de Freitas, and Arnaud Doucet. 2009. Newinference strategies for solving Markov decision processes using reversible jumpMCMC. In Proceedings of the Twenty-Fifth Conference on Uncertainty in ArtificialIntelligence. AUAI Press, 223–231.

[11] Andreas Horni, Kai Nagel, and KayWAxhausen. 2016. The multi-agent transportsimulation MATSim. Ubiquity, London 9 (2016).

[12] Andreas Horni, Kai Nagel, and KayWAxhausen. 2016. The multi-agent transportsimulation MATSim. Ubiquity, London 9 (2016).

[13] Masoumeh T Izadi and D Buckeridge. 2007. Optimizing anthrax outbreak detec-tion methods using reinforcement learning. Citeseer.

[14] Rudolph Emil Kalman. 1960. A new approach to linear filtering and predictionproblems. Journal of basic Engineering 82, 1 (1960), 35–45.

[15] Hilbert J Kappen, Vicenç Gómez, and Manfred Opper. 2012. Optimal control as agraphical model inference problem. Machine learning 87, 2 (2012), 159–182.

[16] Lin Liao Dieter Fox Henry Kautz. 2004. Learning and inferring transportationroutines. In Proceedings: Nineteenth National Conference on Artificial Intelligence(AAAI-04): Sixteenth Innovative Applications of Artificial Intelligence Conference(IAAI-04). Aaai Press, 348.

[17] Averill M Law, W David Kelton, and W David Kelton. 1991. Simulation modelingand analysis. Vol. 2. McGraw-Hill New York.

[18] Claude Lefévre. 1981. Optimal control of a birth and death epidemic process.Operations Research 29, 5 (1981), 971–982.

[19] Sergey Levine. 2018. Reinforcement Learning and Control as Probabilistic Infer-ence: Tutorial and Review. arXiv preprint arXiv:1805.00909 (2018).

[20] Li Li, Yisheng Lv, and Fei-Yue Wang. 2016. Traffic signal timing via deep rein-forcement learning. IEEE/CAA Journal of Automatica Sinica 3, 3 (2016), 247–254.

[21] Li Li, Ding Wen, and Danya Yao. 2014. A survey of traffic control with vehicularcommunications. IEEE Transactions on Intelligent Transportation Systems 15, 1(2014), 425–432.

[22] Weiwei Li and Emanuel Todorov. 2004. Iterative linear quadratic regulator designfor nonlinear biological movement systems.. In ICINCO (1). 222–229.

[23] Timothy P Lillicrap, Jonathan J Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez,Yuval Tassa, David Silver, and Daan Wierstra. 2015. Continuous control withdeep reinforcement learning. arXiv preprint arXiv:1509.02971 (2015).

[24] Qiang Liu and Alexander Ihler. 2012. Belief Propagation for Structured DecisionMaking. In Proceedings of the Twenty-Eighth Conference Annual Conference onUncertainty in Artificial Intelligence (UAI-12). AUAI Press, Corvallis, Oregon,523–532.

[25] William H Montgomery and Sergey Levine. 2016. Guided policy search viaapproximate mirror descent. InAdvances in Neural Information Processing Systems.4008–4016.

[26] Manfred Opper and Guido Sanguinetti. 2008. Variational inference for Markovjump processes. InAdvances in Neural Information Processing Systems. 1105–1112.

[27] Scott E Page. 2015. What sociologists should know about complexity. AnnualReview of Sociology 41 (2015), 21–41.

[28] Nathalie Peyrard and Régis Sabbadin. 2006. Mean field approximation of thepolicy iteration algorithm for graph-based Markov decision processes. Frontiersin Artificial Intelligence and Applications 141 (2006), 595.

[29] Victor M Preciado, Michael Zargham, Chinwendu Enyioha, Ali Jadbabaie, andGeorge Pappas. 2013. Optimal vaccine allocation to control epidemic outbreaksin arbitrary networks. In Decision and Control (CDC), 2013 IEEE 52nd AnnualConference on. IEEE, 7486–7491.

[30] Régis Sabbadin, Nathalie Peyrard, and Nicklas Forsell. 2012. A framework anda mean-field algorithm for the local control of spatial processes. InternationalJournal of Approximate Reasoning 53, 1 (2012), 66–86.

[31] Olivier Sigaud and Olivier Buffet. 2013. Markov decision processes in artificialintelligence. John Wiley & Sons.

[32] Robert F Stengel. 1994. Optimal control and estimation. Courier Corporation.[33] Richard S Sutton and Andrew G Barto. 2011. Reinforcement learning: An intro-

duction. (2011).[34] Richard S Sutton, Andrew G Barto, et al. 1998. Reinforcement learning: An intro-

duction. MIT press.[35] Richard S Sutton, Doina Precup, and Satinder Singh. 1999. Between MDPs and

semi-MDPs: A framework for temporal abstraction in reinforcement learning.Artificial intelligence 112, 1-2 (1999), 181–211.

[36] Stelios Timotheou, Christos G Panayiotou, and Marios M Polycarpou. 2015. Dis-tributed traffic signal control using the cell transmission model via the alternatingdirection method of multipliers. IEEE Transactions on Intelligent TransportationSystems 16, 2 (2015), 919–933.

[37] Emanuel Todorov. 2007. Linearly-solvableMarkov decision problems. InAdvancesin neural information processing systems. 1369–1376.

[38] Marc Toussaint and Amos Storkey. 2006. Probabilistic inference for solvingdiscrete and continuous state Markov Decision Processes. In Proceedings of the23rd international conference on Machine learning. ACM, 945–952.

[39] Marc Toussaint, Amos Storkey, and Stefan Harmeling. 2010. Expectation-Maximization methods for solving (PO) MDPs and optimal control problems.Inference and Learning in Dynamic Models. Cambridge University Press: Cambridge,England (2010).

[40] Martin J Wainwright, Michael I Jordan, et al. 2008. Graphical models, exponentialfamilies, and variational inference. Foundations and Trends® in Machine Learning1, 1–2 (2008), 1–305.

[41] Fei-Yue Wang. 2010. Parallel control and management for intelligent transporta-tion systems: Concepts, architectures, and applications. IEEE Transactions onIntelligent Transportation Systems 11, 3 (2010), 630–638.

[42] MA Wiering, J van Veenen, Jilles Vreeken, and Arne Koopman. 2004. Intelligenttraffic light control. (2004).

[43] Darren J Wilkinson. 2011. Stochastic modelling for systems biology. CRC press.[44] Zhen Xu, Wen Dong, and Sargur N Srihari. 2016. Using social dynamics to make

individual predictions: variational inference with a stochastic kinetic model. InAdvances in Neural Information Processing Systems. 2783–2791.

[45] Fan Yang and Wen Dong. 2017. Integrating simulation and signal processingwith stochastic social kinetic model. In International Conference on Social Com-puting, Behavioral-Cultural Modeling and Prediction and Behavior Representationin Modeling and Simulation. Springer, 193–203.

[46] Fan Yang and Wen Dong. 2018. Integrating simulation and signal processing intracking complex social systems. Computational and Mathematical OrganizationTheory (2018), 1–22.

[47] Fan Yang, Alina Vereshchaka, and Wen Dong. 2018. Predicting and OptimizingCity-Scale Road Traffic Dynamics Using Trajectories of Individual Vehicles. In2018 IEEE International Conference on Big Data (Big Data). IEEE, 173–180.

[48] Brian D Ziebart. 2010. Modeling purposeful adaptive behavior with the principle ofmaximum causal entropy. Ph.D. Dissertation.

Session 1F: Agent Societies and Societal Issues 1 AAMAS 2019, May 13-17, 2019, Montréal, Canada

304