Balancing a Legged Robot Using State-Dependent Riccati ...

6
Balancing a Legged Robot Using State-Dependent Riccati Equation Control E. Najafi *,** G.A.D. Lopes * R. Babuˇ ska * * Delft Center for Systems and Control, Delft University of Technology, Mekelweg 2, 2628 CD Delft, The Netherlands. {e.najafi; g.a.delgadolopes; r.babuska}@tudelft.nl ** Department of Mechatronics Engineering, Faculty of New Sciences and Technologies, University of Tehran, Iran. Abstract: In this paper, we propose a nonlinear control approach for balancing underactuated legged robots. For the balancing task, the robot is modeled as a generalized version of a Segway. The control design is based on the State-Dependent Riccati Equation (SDRE) approach. The domain of attraction of the SDRE controller is compared to the domain of attraction of a linear quadratic controller. Using a simulation example of a four-legged robot balancing on its hind legs, we show that the SDRE controller gives a reasonably large domain of attraction, even with realistic level constraints on the control input, while the linear quadratic controller is unable to stabilize the system. Keywords: State-dependent Riccati equation control, Linear quadratic control, Legged robots, Domain of attraction. 1. INTRODUCTION Designing motion controllers for ground robots that need to traverse unstructured terrains is a challenging problem. For example, a legged robot should be able to execute different tasks such as walking, jumping, running and balancing. As legged robots are efficient in navigating various terrains, legged locomotion has been the subject of extensive research, see, for instance, (Saranli et al., 2001; Raibert et al., 2008; Kalakrishnan et al., 2011). In this paper, we address stabilizing of a four-legged RHex- type robot (Weingarten et al., 2004), called RQuad (Lopes et al., 2009) in its upright position. This task is relevant to traversing different terrains and climbing large obsta- cles. In addition, the robot can observe its environment standing on its hind legs. The robot is equipped with rotating legs of semicircular shape, see Fig. 1. During the balancing task, the hind legs of the RQuad robot roll over the ground in a way similar to wheels rotating around a pivot point placed off the wheel center, as schematically shown in Fig. 2. In this sense, the task can be regarded as a more complex variant of Segway balancing. The control of the Segway, wheeled inverted pendulums and ballbots, which are all similar platforms, has been in- vestigated in the literature (Grasser et al., 2002; Nagarajan et al., 2012). It has been shown that although the Segway is a nonlinear underactuated system, it can be controlled by linear controllers (Pathak et al., 2005). However, when the pivot point moves from the wheel’s center toward the wheel’s rim (Fig. 2), the nonlinearity of the system becomes more severe. As a result, the linear controller’s domain of attraction (DA) 1 shrinks quickly and so linear 1 The DA of a controller is a set of initial states such that each trajectory starting from the set, finally converges to the controller’s desired goal. Fig. 1. Four-legged robot RQuad developed at Delft Center for System and Control, TU Delft. control is not adequate for the legged robot. Therefore, one has to resort to nonlinear controllers in order to control the robot in a sufficient range of its configuration space. In this paper, we investigate the nonlinear counterpart of the linear quadratic regulator (LQR), which is called state- dependent Riccati equation (SDRE) controller (Cloutier, 1997). This approach has been successfully applied to various nonlinear systems, see, for example, (Dang and Lewis, 2005) where an inverted pendulum is controlled using SDRE controller. The paper is organized as follows. A brief review of the Segway model and the legged robot is presented in Section 2. In Section 3, a standard LQR is designed to stabilize the legged robot in the balancing position. In addition, the performance of the LQR in terms of its DA is investigated. Section 4, first, reviews the main concepts of the SDRE controller. Then, an SDRE controller is designed to stabilize the robot in its upright position. Moreover, the DA of the SDRE controller is compared Preprints of the 19th World Congress The International Federation of Automatic Control Cape Town, South Africa. August 24-29, 2014 Copyright © 2014 IFAC 2177

Transcript of Balancing a Legged Robot Using State-Dependent Riccati ...

Page 1: Balancing a Legged Robot Using State-Dependent Riccati ...

Balancing a Legged Robot UsingState-Dependent Riccati Equation Control

E. Najafi ∗,∗∗ G.A.D. Lopes ∗ R. Babuska ∗

∗Delft Center for Systems and Control, Delft University of Technology,Mekelweg 2, 2628 CD Delft, The Netherlands.e.najafi; g.a.delgadolopes; [email protected]

∗∗Department of Mechatronics Engineering, Faculty of New Sciencesand Technologies, University of Tehran, Iran.

Abstract: In this paper, we propose a nonlinear control approach for balancing underactuatedlegged robots. For the balancing task, the robot is modeled as a generalized version of a Segway.The control design is based on the State-Dependent Riccati Equation (SDRE) approach. Thedomain of attraction of the SDRE controller is compared to the domain of attraction of a linearquadratic controller. Using a simulation example of a four-legged robot balancing on its hindlegs, we show that the SDRE controller gives a reasonably large domain of attraction, even withrealistic level constraints on the control input, while the linear quadratic controller is unable tostabilize the system.

Keywords: State-dependent Riccati equation control, Linear quadratic control, Legged robots,Domain of attraction.

1. INTRODUCTION

Designing motion controllers for ground robots that needto traverse unstructured terrains is a challenging problem.For example, a legged robot should be able to executedifferent tasks such as walking, jumping, running andbalancing. As legged robots are efficient in navigatingvarious terrains, legged locomotion has been the subjectof extensive research, see, for instance, (Saranli et al.,2001; Raibert et al., 2008; Kalakrishnan et al., 2011). Inthis paper, we address stabilizing of a four-legged RHex-type robot (Weingarten et al., 2004), called RQuad (Lopeset al., 2009) in its upright position. This task is relevantto traversing different terrains and climbing large obsta-cles. In addition, the robot can observe its environmentstanding on its hind legs. The robot is equipped withrotating legs of semicircular shape, see Fig. 1. During thebalancing task, the hind legs of the RQuad robot roll overthe ground in a way similar to wheels rotating around apivot point placed off the wheel center, as schematicallyshown in Fig. 2. In this sense, the task can be regarded asa more complex variant of Segway balancing.

The control of the Segway, wheeled inverted pendulumsand ballbots, which are all similar platforms, has been in-vestigated in the literature (Grasser et al., 2002; Nagarajanet al., 2012). It has been shown that although the Segwayis a nonlinear underactuated system, it can be controlledby linear controllers (Pathak et al., 2005). However, whenthe pivot point moves from the wheel’s center towardthe wheel’s rim (Fig. 2), the nonlinearity of the systembecomes more severe. As a result, the linear controller’sdomain of attraction (DA) 1 shrinks quickly and so linear

1 The DA of a controller is a set of initial states such that eachtrajectory starting from the set, finally converges to the controller’sdesired goal.

Fig. 1. Four-legged robot RQuad developed at Delft Centerfor System and Control, TU Delft.

control is not adequate for the legged robot. Therefore, onehas to resort to nonlinear controllers in order to control therobot in a sufficient range of its configuration space.

In this paper, we investigate the nonlinear counterpart ofthe linear quadratic regulator (LQR), which is called state-dependent Riccati equation (SDRE) controller (Cloutier,1997). This approach has been successfully applied tovarious nonlinear systems, see, for example, (Dang andLewis, 2005) where an inverted pendulum is controlledusing SDRE controller.

The paper is organized as follows. A brief review ofthe Segway model and the legged robot is presented inSection 2. In Section 3, a standard LQR is designed tostabilize the legged robot in the balancing position. Inaddition, the performance of the LQR in terms of its DAis investigated. Section 4, first, reviews the main conceptsof the SDRE controller. Then, an SDRE controller isdesigned to stabilize the robot in its upright position.Moreover, the DA of the SDRE controller is compared

Preprints of the 19th World CongressThe International Federation of Automatic ControlCape Town, South Africa. August 24-29, 2014

Copyright © 2014 IFAC 2177

Page 2: Balancing a Legged Robot Using State-Dependent Riccati ...

with the DA of the LQR. Finally, Section 5 concludes thepaper.

2. PHYSICAL MODEL OF BALANCED LEGGEDROBOT

In order to design an SDRE controller, which is a model-based controller, we need to develop a mathematical modelof the balancing robot in the sagittal plane. Denote by

q = [α φ]T

the vector of the configuration variables, withα the angle between the body and the vertical axis and φthe wheel angle, as illustrated in Fig. 2. Let c be the ratioof the pivot point distance from the wheel’s center to thewheel’s radius. Parameter c can be set to a value betweenzero and one to make a gradual transition from the wheelto a semi-circular leg. As c increases, the nonlinearity ofthe system becomes more severe.

Fig. 2. The transition from balancing a wheeled robot(Segway) to balancing a RHex-type legged robot.Parameter c denotes the ratio of the pivot pointdistance from the wheel’s center to the wheel’s radius.

Using Euler-Lagrange equations, the following model isobtained:

M(q)q + C(q, q)q +G(q) = B(q)τ + Fext (1)

with the inertia matrix M(q) ∈ Rn×n, the Coriolis andcentrifugal forces matrix C(q, q) ∈ Rn×n, the gravitymatrix G(q) ∈ Rn, the input vector B(q) ∈ Rn and thevector of the external forces Fext ∈ Rn. These matricesare given by:

M(q) =

[mbl

2 + Ib mbrl cos(α) +M1

mbrl cos(α) +M1 (mw +mb)r2 + Iw +M2

]C(q, q) =

[0 C1

−mbrlα sin(α) + C2 C3

]G(q) =

[−mblg sin(α)

G1

](2)

B(q) = [−2 2]T

Fext =[2bj(φ− α) −2bg − 2bj(φ− α)

]Twith

M1 = cmbrl cos(α− φ), M2 = c2mbr2 + 2cmbr

2 cos(α),

C1 = cmbrlφ sin(α− φ), C2 = −cmbrlα sin(α− φ),

C3 = −cmbr2φ sin(φ), G1 = −cmbrg sin(φ). (3)

The symbols are explained in Table 1. It also lists thevalues of the RQuad’s physical parameters. These valueswere found partly by measuring and partly estimated usingnonlinear system identification.

While the RQuad robot stands on its hind legs, the motorsof both legs are controlled with the same control input.This fact is modeled by the matrix B(q) which doubles thetorque value that one motor exerts on a leg. The torque τdelivered by the motors is given by the following model ofthe DC motor:

τ = Ngτm

τm =Kt

Rm(V −Keθm) (4)

θm = Ng(φ− α)

V = Nsv.

The subscript m refers to the motor variables and v isthe control input normalized in the range [−1, 1]. Thematrix Fext includes the friction force in the pivot jointand the rolling friction force due to the contact betweenthe legs and the ground. We assume that the legs do notslip. Additionally, though the RQuad’s legs are made of aflexible material, in the model they are considered as rigidbodies.

Table 1. Physical parameters of the RQuadrobot.

Physical parameter Symbol Value Unit

Body inertia Ib 2.60 · 10−2 Kgm2

Body mass mb 2.55 KgWheel inertia Iw 4.54 · 10−4 Kgm2

Wheel mass mw 3 · 10−1 KgGravity g 9.81 ms−2

Body half length l 13.8 · 10−2 mWheel radius r 5.5 · 10−2 mRolling friction bg 6.5 · 10−3 NmsJoint friction bj 1 · 10−4 NmsTorque constant Kt 3.15 · 10−2 NmA−1

Back EMF Ke 3.15 · 10−2 NmA−1

Rotor resistance Rm 5.21 ΩGear ratio Ng 35 −Input gain Ns 25 V

Note that by setting c = 0 functions Mi, Ci and Gi inequations (3) become equal to zero and we obtain thestandard Segway model.

3. LQR AND DOMAIN OF ATTRACTION ANALYSIS

We first investigate the equilibrium manifolds of the Seg-way and the legged robot. Then, an LQR is designedto stabilize the RQuad robot in the balancing position.In order to study the performance of the controller, itsdomain of attraction is investigated for different valuesof parameter c. It is well known in the literature thatthe DA of a nonlinear controller is a complex set, see,e.g., (Vannelli and Vidyasagar, 1985; Chiang et al., 1988).Generally, an analytical representation is only found fortrivial systems as discussed by Genesio et al. (1985). Inthis paper, we simulate the control system with variousinitial conditions to approximate the DA of the controller.

3.1 Equilibrium Manifold

For c = 0, the pivot is in the wheel’s center and the leggedrobot is equivalent to an ordinary Segway. To stabilize theSegway in the upright position, the angle α between the

19th IFAC World CongressCape Town, South Africa. August 24-29, 2014

2178

Page 3: Balancing a Legged Robot Using State-Dependent Riccati ...

Fig. 3. The equilibrium manifold of the robot when c = 0 (Segway) and when c = 1 (legged robot).

body and the vertical axis has to be zero, but the wheelangle φ can take any value. The equilibrium manifold ofthe Segway is shown at the left side of Fig. 3. This manifoldillustrates that the equilibrium points of the Segway lie atthe zero value of the body angle.

If the pivot point moves from the wheel’s center to thewheel’s rim, the position of the equilibrium point willdepend on the robot’s configuration. The robot will bestable if the centers of mass of the body and the wheellie exactly above each other and the angular velocitiesare zero as well, see the right side of Fig. 3. The goal ofcontroller is to stabilize the robot in one of the equilibriumpoints (not necessarily at the zero angles).

3.2 Linear Quadratic Controller

We first design an LQR to stabilize the RQuad robot in thebalancing position. However, we expect the performanceof the linear controllers to be limited as parameter cincreases. The main purpose of this section is to analyzethe effect of c on the controller’s performance with respectto its DA.

Define the state vector of the system as:

x =[α φ α φ

]T(5)

where α and φ are the angular velocities of the bodyand the wheel, respectively. The equations of motion arewritten in the general input-affine nonlinear state-spaceform:

x = f(x) + g(x)u. (6)

To design a linear controller, such as LQR, functions f(x)and g(x) have to be linearized around the operating pointof the system (chosen at x = 0 here), yielding:

x = Ax+Bu. (7)

Given matrices A and B, one can now design an LQRdescribed by:

u = −(R−1BTP )x = −Kx, (8)

where R is a positive definite matrix and P is the solutionof the algebraic Riccati equation:

ATP + PA− PBR−1BTP +Q = 0 (9)

whit Q is a positive definite matrix.

3.3 LQR: Domain of Attraction Analysis

First an LQR is designed based on the linearized statespace model. The value of parameter c determines theposition of the pivot joint. In order to investigate theperformance of the controller, we compute its DA for dif-ferent values of c, using numerical simulation. To computethe DA, the controller is applied to stabilize the robotstarting at different initial states. The initial value of thebody angle ranges from −π/2 rad to +π/2 rad. Similarly,the initial value of the wheel angle is defined within therange of −30 rad to +30 rad. These ranges are discretizedat 100 samples. The discretization step of α and φ thereforeequals to 0.031 rad and 0.6 rad, respectively.

Fig. 4 illustrates how the DA of the LQR changes regardingto parameter c. Observe that the DA shrinks as the non-linearity increases. When c changes from 0 to 0.25, the DAshrinks slightly, but for c increased to 0.75 the DA becomesvery small. In conclusion, as the system’s nonlinearitybecomes more severe, the LQR controller cannot stabilizethe system properly. Therefore, we implement the SDREcontroller, which is a nonlinear counterpart of LQR.

4. SDRE CONTROLLER AND DOMAIN OFATTRACTION ANALYSIS

This section first reviews the main concepts of the SDREmethod. Then, an SDRE controller is designed to stabilizethe RQuad robot in the balancing position. In addition,the DA of the controller is plotted for different values ofparameter c. Finally, the result achieved for both the LQRand the SDRE controller are compared with respect to thenonlinearity of the system.

4.1 SDRE Controller

State-Dependent Riccati Equation controller is a nonlinearcounterpart of the LQR, see (Cimen, 2008). This methodhas recently become popular as it provides an algorithmfor designing nonlinear state feedback controllers. TheSDRE based controllers accept nonlinearity for the statespace model. They also offer the flexibility of using state-dependent design matrices Q(x) and R(x). SDRE method

19th IFAC World CongressCape Town, South Africa. August 24-29, 2014

2179

Page 4: Balancing a Legged Robot Using State-Dependent Riccati ...

Body angle [rad]

Whe

el a

ngle

[rad

]

c = 0 c = 0.25 c = 0.5 c = 0.75

Fig. 4. The domain of attraction of LQR for different values of parameter c.

has been developed for infinite horizon suboptimal prob-lems in nonlinear systems that are autonomous and ob-servable. The state-space equations are input-affine in theform of (6) and the task is to minimize the performanceindex defined as:

J =1

2

∫ ∞t0

(xTQ(x)x+ uTR(x)u

)dt, (10)

where x ∈ Rn, u ∈ Rm, f(x) ∈ Rn, g(x) ∈ Rk, Q(x) ∈Rn×n and R(x) ∈ Rm×m, with the assumptions thatf(0) = 0 and g(x) 6= 0, ∀x. To apply the SDRE method,the nonlinear system defined in (6) needs to be representedin a state-dependent linear-like structure:

x = A(x)x+B(x)u, (11)

where A(x) and B(x) are the state-dependent matrices.Under the assumption f(0) = 0 and f(x) ∈ Rn, the state-dependent matrix A(x) : Rn → Rn×n can be found bymathematical manipulation; however, A(x) is not unique.This fact offers some flexibility in the control design. Inany case, A(x) has to be selected such that the full statemeasurement for the system is available and the pair(A(x), B(x)) is point-wise controllable (Dang and Lewis,2005).

The main concept of computing the control law in SDREis solving the algebraic Riccati equation:

A(x)TP (x) + P (x)A(x)− P (x)B(x)R(x)−1

B(x)TP (x) +Q(x) = 0 (12)

to obtain P (x) ≥ 0, which is then applied in the feedbackcontrol law given as:

U = −K(x)x (13)

with K(x) = R(x)−1B(x)TP (x). The main difference be-tween an SDRE controller and LQR is the fact that, in theSDRE controller the matrices A(x) and B(x) and also theweighting matrices Q(x) and R(x) are state-dependent.The SDRE method provides flexibility in choosing theweighting matrices which can lead to optimize the controlaccuracy and the control effort. However, the obtainedcontroller is not necessarily optimal with respect to theperformance index (10) (Cimen, 2010). Additionally, theclosed-loop matrix ACL(x) which is computed as:

ACL(x) = A(x)−B(x)K(x) (14)

is point-wise Hurwitz. Thus, the origin of the closed-loop system is locally asymptotic stable, see (Erdem andAlleyne, 2004).

4.2 SDRE: Domain of Attraction Analysis

In order to design an SDRE controller, the general input-affine nonlinear form of the system, see (6), has to be writ-ten in the state-dependent linear form (11). Substitutingequations (2) in (6) yields:

x =

[0 I

−M−1Gs −M−1C

]x+

[0

M−1B

]u, (15)

where C is an identity matrix and G(q) = Gsq. Thereare different options for matrix Gs that can respect bothequation (6) and (15). Exploiting this flexibility, one canput emphasize on a specific state compared to the otherstates. For example, in the case of the RQuad balancing,the emphasize is on the body angle, due to the significantrole of the angle α in the stabilization. Matrix Gs istherefore given by:

Gs =

−mblgsin(α)

α0

−cmbrgsin(φ)

α0

. (16)

By substituting equations (2) and (16) in (15) and doingsome manipulation, the state-dependent matrices of theRQuad robot are derived. These equations for the casec = 0 (just to simplify the presentation) are described asfollows:

A(x) =

0 0 1 00 0 0 1a31 0 a33 0a41 0 a43 0

(17)

B(x) =

00

K(Iw +mbrl cos(α) + r2(Ib +mb)

)K(Ib +mbl

2 +mbrl cos(α))

(18)

with

a31 = (K/α)(mblg

(Iw + r2(mw +mb)

)sin(α)

)a33 = K

(−m2

br2l2 sin(α) cos(α)α

)a41 = (K/α)

(−m2

brl2g sin(α) cos(α)

)(19)

a43 = K(mbrl(Ib +mbl

2) sin(α)α)

K =1

(Ib +mbl2)(Iw + r2(mw +mb)

)−m2

br2l2 cos2(α)

.

When the state-dependent matrices are obtained, one cancompute the control input (13).

In order to assess the performance of the controller interms of its DA, we repeat the same simulations as for

19th IFAC World CongressCape Town, South Africa. August 24-29, 2014

2180

Page 5: Balancing a Legged Robot Using State-Dependent Riccati ...

the LQR. The value of c increases from 0 to 1 and thecontroller’s DA is computed. Fig. 5 illustrates the DA ofthe SDRE controller just for the values 0.75 and 1 of c,where there is a considerable improvement compared toLQR.

Body angle [rad]

Whe

el a

ngle

[rad

]

-1.5 -1 -0.5 0 0.5 1 1.5 -1.5 -1 -0.5 0 0.5 1 1.5-20

c = 0.75 c = 120

-10

10

0

Fig. 5. The domain of attraction of SDRE controller fortwo different values of parameter c.

Although the DA of the SDRE controller becomes smalleras the nonlinearity becomes more severe, it is larger thanthe DA of the LQR, both for low and high values ofparameter c. This is shown in Fig. 6 which plots fordifferent values of parameter c the maximum absolutevalue of the initial body angle α(0), while φ(0) = 0,for which the controller still can stabilize the robot. TheSDRE controller is considerably more effective than theLQR, especially when c approaches 1 (the pivot joint lieson the wheel’s rim and the wheeled robot changes toRQuad).

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10

0.2

0.4

0.6

0.8

1

1.2

1.4

1.6

Parameter c

Max

abs

olut

e in

itial

bod

y an

gle

[rad]

SDRE controllerLQR controller

Fig. 6. Maximum absolute initial value of the body angle,when the wheel angle is zero, such that the controllerstill can balance the legged robot.

5. DISCUSSION AND CONCLUSIONS

We have compared nonlinear state-dependent Riccatiequation control and linear LQR in the task of balancinga RHex-type legged robot on its hind legs. The LQR’sdomain of attraction is significantly smaller than that ofthe SDRE controller and the LQR cannot stabilize therobot. The most challenging part of the SDRE controller

design is the computation of the state-dependent matricesfrom the model equations. Once the matrices are obtained,an algebraic Riccati equation is solved for computing thecontrol input.

Besides using SDRE controller, one can apply methodsthat use a collection of controllers designed for specificparts of the state space. For example, sequential compo-sition (Burridge et al., 1999) is an effective approach toaddress complex dynamical systems. In sequential compo-sition, a collection of simple local controllers with possiblysmall domains of attraction are executed consecutively.LQR-tree is another method which implements the DAs ofa series of local linear quadratic regulators (LQR) to createa tree (path) from the initial state to the desired goalthrough the state space (Tedrake et al., 2010). In addition,a learning strategy has been recently introduced by Najafiet al. (2013) which is an augmentation of the traditionalsequential composition by learning new controllers on aneed basis in runtime.

In composition based methods, the performance of eachlocal controller, particularly its DA, plays an importantrole in the control synthesis. This paper showed that usingthe SDRE method can effectively enlarge the DAs of localcontrollers.

ACKNOWLEDGEMENTS

The first author gratefully acknowledges the support ofthe Government of the Islamic Republic of Iran, Ministryof Science, Research and Technology.

REFERENCES

Burridge, R.R., Rizzi, A.A., and Koditschek, D.E. (1999).Sequential composition of dynamically dexterous robotbehaviors. The International Journal of Robotics Re-search, 18(6), 534–555.

Chiang, H.D., Hirsch, M.W., and Wu, F.F. (1988). Stabil-ity regions of nonlinear autonomous dynamical systems.Automatic Control, IEEE Transactions on, 33(1), 16–27.

Cimen, T. (2008). State-dependent Riccati equation(SDRE) control: a survey. In Proceedings of the 17thWorld Congress, The International Federation of Auto-matic Control, 6–11.

Cimen, T. (2010). Systematic and effective design ofnonlinear feedback controllers via the state-dependentRiccati equation (SDRE) method. Annual Reviews inControl, 34(1), 32–51.

Cloutier, J.R. (1997). State-dependent Riccati equationtechniques: an overview. In American Control Confer-ence, 1997. Proceedings of the 1997, volume 2, 932–936.IEEE.

Dang, P. and Lewis, F.L. (2005). Controller for swing-upand balance of single inverted pendulum using SDRE-based solution. In Industrial Electronics Society, 2005.IECON 2005. 31st Annual Conference of IEEE, 6–pp.IEEE.

Erdem, E.B. and Alleyne, A.G. (2004). Design of a class ofnonlinear controllers via state dependent Riccati equa-tions. Control Systems Technology, IEEE Transactionson, 12(1), 133–137.

19th IFAC World CongressCape Town, South Africa. August 24-29, 2014

2181

Page 6: Balancing a Legged Robot Using State-Dependent Riccati ...

Genesio, R., Tartaglia, M., and Vicino, A. (1985). Onthe estimation of asymptotic stability regions: State ofthe art and new proposals. Automatic Control, IEEETransactions on, 30(8), 747–755.

Grasser, F., D’Arrigo, A., Colombi, S., and Rufer, A.C.(2002). JOE: a mobile, inverted pendulum. IndustrialElectronics, IEEE Transactions on, 49(1), 107–114.

Kalakrishnan, M., Buchli, J., Pastor, P., Mistry, M., andSchaal, S. (2011). Learning, planning, and control forquadruped locomotion over challenging terrain. TheInternational Journal of Robotics Research, 30(2), 236–258.

Lopes, G., Babuska, R., De Schutter, B., and van denBoom, A. (2009). Switching max-plus models forlegged locomotion. In Robotics and Biomimetics (RO-BIO), 2009 IEEE International Conference on, 221–226. IEEE.

Nagarajan, U., Kantor, G., and Hollis, R. (2012). In-tegrated planning and control for graceful navigationof shape-accelerated underactuated balancing mobilerobots. In Robotics and Automation (ICRA), 2012IEEE International Conference on, 136–141. IEEE.

Najafi, E., Lopes, G.A., and Babuska, R. (2013). Rein-forcement learning for sequential composition control. In52nd IEEE Conference on Decision and Control, 7265–7270. IEEE.

Pathak, K., Franch, J., and Agrawal, S.K. (2005). Velocityand position control of a wheeled inverted pendulum bypartial feedback linearization. Robotics, IEEE Transac-tions on, 21(3), 505–513.

Raibert, M., Blankespoor, K., Nelson, G., Playter, R.,et al. (2008). Bigdog, the rough-terrain quadrupedrobot. In Proceedings of the 17th World Congress,10823–10825.

Saranli, U., Buehler, M., and Koditschek, D.E. (2001).Rhex: A simple and highly mobile hexapod robot. TheInternational Journal of Robotics Research, 20(7), 616–631.

Tedrake, R., Manchester, I.R., Tobenkin, M., and Roberts,J.W. (2010). LQR-trees: Feedback motion planning viasums-of-squares verification. The International Journalof Robotics Research, 29(8), 1038–1052.

Vannelli, A. and Vidyasagar, M. (1985). Maximal Lya-punov functions and domains of attraction for au-tonomous nonlinear systems. Automatica, 21(1), 69–80.

Weingarten, J.D., Lopes, G.A., Buehler, M., Groff, R.E.,and Koditschek, D.E. (2004). Automated gait adap-tation for legged robots. In Robotics and Automation,2004. Proceedings. ICRA’04. 2004 IEEE InternationalConference on, volume 3, 2153–2158. IEEE.

19th IFAC World CongressCape Town, South Africa. August 24-29, 2014

2182