Lars Johannsmeier, Malkin Gerchow and Sami Haddadin · 2018-05-23 · Lars Johannsmeier, Malkin...

8
A Framework for Robot Manipulation: Skill Formalism, Meta Learning and Adaptive Control Lars Johannsmeier, Malkin Gerchow and Sami Haddadin 1 Abstract— In this paper we introduce a novel framework for expressing and learning force-sensitive robot manipulation skills. It is based on a formalism that extends our previous work on adaptive impedance control with meta parameter learning and compatible skill specifications. This way the system is also able to make use of abstract expert knowledge by incorporating process descriptions and quality evaluation metrics. We eval- uate various state-of-the-art schemes for the meta parameter learning and experimentally compare selected ones. Our results clearly indicate that the combination of our adaptive impedance controller with a carefully defined skill formalism significantly reduces the complexity of manipulation tasks even for learning peg-in-hole with submillimeter industrial tolerances. Overall, the considered system is able to learn variations of this skill in under 20 minutes. In fact, experimentally the system was able to perform the learned tasks faster than humans, leading to the first learning-based solution of complex assembly at such real-world performance. I. I NTRODUCTION Typically, robot manipulation skills are introduced as more or less formal representations of certain sets of prede- fined actions or movements. Already, there exist several ap- proaches to programming with skills, e.g. [1]–[3]. A common drawback is, however, the need for laborious and complex parameterization resulting in a manual tuning phase to find satisfactory parameters for a specific skill. Depending on the particular situation various parameters need to be adapted in order to account for different environment properties such as rougher surfaces or different masses of involved objects. Within given boundaries of certainty they could be chosen such that the skill is fulfilled optimally with respect to a specific cost function. This cost function and additional constraints are usually defined by human experts optimizing e.g. for low contact forces, short execution time or low power consumption. Typically, manual parameter tuning is a very laborious task, thus autonomous parameter selection without complex pre-knowledge about the task other than the task specification and the robot prior abilities is highly desirable. However, such automatic tuning of control and other task parameters in order to find feasible, ideally even optimal parameters, in the sense of a cost function is still a significant open problem in robot skill design So far, several approaches were proposed to tackle this problem. In [4], e.g. learning motor skills by demonstration is described. In [5] a Reinforcement Learning (RL) based approach for acquiring new motor skills from demonstration is introduced. [6], [7] employ RL methods to learn motor primitives that represent a skill. In [8] supervised learning by demonstration is used in combination with dynamic movement primitives (DMP) to learn bipedal walking in sim- ulation. An early approach utilizing a stochastic real-valued RL algorithm in combination with a nonlinear multilayer artificial neural network for learning robotic skills can be found in [9]. In [10], guided policy search was applied to learn various manipulation tasks based on a neural network policy. The drawbacks of many existing approaches are their 1 Authors are with Chair of Robotic Sciences and System Intel- ligence, Technische Universit¨ at unchen, 81797 unchen, Germany [email protected] Potential solution space Reduced, meaningful search space Skill framework Search space reduction All possible motor controls Adaptive impedance controller Dynamic system limits Manipulation and interaction control and interaction control Partitioned manipulation Skill design Quality metric Fig. 1: Meaningful complexity reduction of the search space by applying adaptive impedance control, a skill formalism and inherent system limits. high demand for computational power and memory, e.g. in form of GPUs and computing clusters, and with a few exceptions they require a large amount of learning time already in simulation. In order to execute complex manipulation tasks such as inserting a peg into a hole soft robotics controlled systems like Franka Emika’s Panda [23] or the LWR III [24] are the system class of choice. Early work in learning impedance control, the most well-known soft-robotics control concept, is for example introduced in [25]. More recent work can be found in [26], [27]. Here, the first one focuses on human- robot interaction, while the second one makes use of RL to learn controller parameters for simple real-world tasks. In [28] basic concepts from human motor control were transferred to impedance adaptation. a) Contribution: Our approach bases on several con- cepts and connects them in a novel way. Soft robotics [29] is considered the basis in terms of hardware and fundamental capabilities, i.e. the embodiment, enabling us in conjunction with impedance control [30] to apply the idea of learning robot manipulation to rather complex problems. We further extend this by making use of the adaptive impedance con- troller introduced in [31]. Both Cartesian stiffness and feed- forward wrench are adapted during execution, depending on the adaptive motion error and based on four interpretable meta parameters per task coordinate. From this follows the question how to choose these meta parameters with respect to the environment and the concrete problem at hand. To unify all these aspects we introduce a novel robot manipulation skill formalism that acts as a meaningful interpretable connection between problem definition and real world. Its purpose is to reduce the complexity of the solution space of a given manipulation problem by applying a well designed, however still highly flexible structure, see Fig. 1. arXiv:1805.08576v1 [cs.RO] 22 May 2018

Transcript of Lars Johannsmeier, Malkin Gerchow and Sami Haddadin · 2018-05-23 · Lars Johannsmeier, Malkin...

Page 1: Lars Johannsmeier, Malkin Gerchow and Sami Haddadin · 2018-05-23 · Lars Johannsmeier, Malkin Gerchow and Sami Haddadin1 ... However, such automatic tuning of control and other

A Framework for Robot Manipulation:Skill Formalism, Meta Learning and Adaptive Control

Lars Johannsmeier, Malkin Gerchow and Sami Haddadin1

Abstract— In this paper we introduce a novel frameworkfor expressing and learning force-sensitive robot manipulationskills. It is based on a formalism that extends our previous workon adaptive impedance control with meta parameter learningand compatible skill specifications. This way the system is alsoable to make use of abstract expert knowledge by incorporatingprocess descriptions and quality evaluation metrics. We eval-uate various state-of-the-art schemes for the meta parameterlearning and experimentally compare selected ones. Our resultsclearly indicate that the combination of our adaptive impedancecontroller with a carefully defined skill formalism significantlyreduces the complexity of manipulation tasks even for learningpeg-in-hole with submillimeter industrial tolerances. Overall,the considered system is able to learn variations of this skill inunder 20 minutes. In fact, experimentally the system was ableto perform the learned tasks faster than humans, leading tothe first learning-based solution of complex assembly at suchreal-world performance.

I. INTRODUCTION

Typically, robot manipulation skills are introduced asmore or less formal representations of certain sets of prede-fined actions or movements. Already, there exist several ap-proaches to programming with skills, e.g. [1]–[3]. A commondrawback is, however, the need for laborious and complexparameterization resulting in a manual tuning phase to findsatisfactory parameters for a specific skill. Depending on theparticular situation various parameters need to be adapted inorder to account for different environment properties suchas rougher surfaces or different masses of involved objects.Within given boundaries of certainty they could be chosensuch that the skill is fulfilled optimally with respect toa specific cost function. This cost function and additionalconstraints are usually defined by human experts optimizinge.g. for low contact forces, short execution time or low powerconsumption. Typically, manual parameter tuning is a verylaborious task, thus autonomous parameter selection withoutcomplex pre-knowledge about the task other than the taskspecification and the robot prior abilities is highly desirable.However, such automatic tuning of control and other taskparameters in order to find feasible, ideally even optimalparameters, in the sense of a cost function is still a significantopen problem in robot skill design

So far, several approaches were proposed to tackle thisproblem. In [4], e.g. learning motor skills by demonstrationis described. In [5] a Reinforcement Learning (RL) basedapproach for acquiring new motor skills from demonstrationis introduced. [6], [7] employ RL methods to learn motorprimitives that represent a skill. In [8] supervised learningby demonstration is used in combination with dynamicmovement primitives (DMP) to learn bipedal walking in sim-ulation. An early approach utilizing a stochastic real-valuedRL algorithm in combination with a nonlinear multilayerartificial neural network for learning robotic skills can befound in [9]. In [10], guided policy search was applied tolearn various manipulation tasks based on a neural networkpolicy. The drawbacks of many existing approaches are their

1Authors are with Chair of Robotic Sciences and System Intel-ligence, Technische Universitat Munchen, 81797 Munchen, [email protected]

Potential solution space

Reduced, meaningful search space

Skill framework

Sear

chsp

ace

redu

ctio

n

All possible motor controls

Adaptive impedance controller

Dynamic system limits

Manipulation andinteraction control

and interaction controlPartitioned manipulation

Skill designQuality metric

Fig. 1: Meaningful complexity reduction of the search space by applyingadaptive impedance control, a skill formalism and inherent system limits.

high demand for computational power and memory, e.g.in form of GPUs and computing clusters, and with a fewexceptions they require a large amount of learning timealready in simulation.

In order to execute complex manipulation tasks such asinserting a peg into a hole soft robotics controlled systemslike Franka Emika’s Panda [23] or the LWR III [24] are thesystem class of choice. Early work in learning impedancecontrol, the most well-known soft-robotics control concept,is for example introduced in [25]. More recent work can befound in [26], [27]. Here, the first one focuses on human-robot interaction, while the second one makes use of RLto learn controller parameters for simple real-world tasks.In [28] basic concepts from human motor control weretransferred to impedance adaptation.

a) Contribution: Our approach bases on several con-cepts and connects them in a novel way. Soft robotics [29]is considered the basis in terms of hardware and fundamentalcapabilities, i.e. the embodiment, enabling us in conjunctionwith impedance control [30] to apply the idea of learningrobot manipulation to rather complex problems. We furtherextend this by making use of the adaptive impedance con-troller introduced in [31]. Both Cartesian stiffness and feed-forward wrench are adapted during execution, depending onthe adaptive motion error and based on four interpretablemeta parameters per task coordinate. From this followsthe question how to choose these meta parameters withrespect to the environment and the concrete problem athand. To unify all these aspects we introduce a novelrobot manipulation skill formalism that acts as a meaningfulinterpretable connection between problem definition and realworld. Its purpose is to reduce the complexity of the solutionspace of a given manipulation problem by applying a welldesigned, however still highly flexible structure, see Fig. 1.

arX

iv:1

805.

0857

6v1

[cs

.RO

] 2

2 M

ay 2

018

Page 2: Lars Johannsmeier, Malkin Gerchow and Sami Haddadin · 2018-05-23 · Lars Johannsmeier, Malkin Gerchow and Sami Haddadin1 ... However, such automatic tuning of control and other

TABLE I: Overview of existing work about peg-in-hole

Reference Principle Motion strategy Underlyingcontroller

Adaptation/ Learningprinciple

Learning / Insertionspeed

Difficulty

[11]–[15] Geometric approaches,analysis of peg-in-hole andforce/moment guidance

Reference trajec-tory

Impedance,admittance,force control

none Insertion times varybetween 5-11 s

Round pegs withtolerances between0.25-1 mm.

[16] Combination of visual percep-tion and compliant manipula-tion strategies to approach adifficult peg-in-hole problem

Trajectory gener-ator

Impedancecontrol

none Insertion time wasabout 2-3 s

Multiple pieces ofdifferent forms wereused with tolerancesof about 0.1 mm.

[9], [17] Peg-in-hole skill is learned viareinforcement learning.

Policy in terms oftrajectory is rep-resented by neu-ral networks

Admittance,force control

Reinforcementlearning

Learning took about500-800 trials, inser-tion time is between20-40 s after learning

2D-problem with tol-erance of 0.127 mmin simulation and around peg with a tol-erance of 0.175 mmin experiments.)

[18], [19] DMPs are adapted based onmeasurements of the externalwrench to adapt demonstratedsolutions of a peg-in-hole task

Trajectory fromDMP

Impedance,Admittance,Force

Trajectoryadaptation

Adaptation requiresabout 5 cycles.Insertion requiresabout 8-10 sbased on humandemonstration.

Several basic shapesand and varying toler-ances of 0.1-1 mm.

[20] Several controllers (with andwithout adaptation) were com-pared for a peg-in-hole task.The initial strategy is learnedfrom human demonstration

(Initial) referencetrajectory

Impedancecontrol

Adaptationoforientationvia randomsamplingandGaussianmixturemodels

Average insertiontimes of 6-9 swere achieved,50 trials wereused for sampling.Data acquisition byhuman demonstrationinitially is necessary.

Round peg and holemade of steel anda wooden peg witha rubber hole wereused.

[10], [21],[22]

Policies represented byconvolutional neural networksare learned via reinforcementlearning based on joint statesand visual feedback as anend-to-end solution. Amongother tasks peg-in-hole hasbeen learned mostly insimulation but also in realworld experiments.

Policies directlygenerate jointtorques

linear-Gaussiancontrollers /neural networkcontrolpolicies

Reinforcementlearning

Depending of thetask learning took100-300 trials insimulation and real-world experiments.Additionalcomputationalrequirements ofabout 50 minutes arementioned

Tasks such as bottleopening and simplervariations of peg-in-hole were learned.

When applied to a real-world task this structure is supportedby an expert skill design and a given quality metric forevaluating the skill’s execution. The reduction of complexityis followed by the application of machine learning methodssuch as CMA-ES [32], Particle Swarm Optimization [33] andBayesian optimization [34] to solve the problem not directlyon motor control level but in the meta parameter space.

In summary, we put forward the hypothesis that learningmanipulation is much more efficient and versatile if we canmake use of local intelligence of system components such asthe adaptive impedance controller and a well structured skillformalism, essentially encoding abstract expert knowledge.These are the physics grounded computational elements thatreceive more abstract commands such as goal poses andtranslate them into basic joint torque behaviors. In this,we draw inspiration from the way humans manipulate theirsurroundings, i.e. not by consciously determining the muscleforce at every time step but rather making use of morecomplex computational elements [35].

As a particular real-world example to support our con-ceptual work, we address the well-known and researched,however, in general still unsolved [20] peg-in-hole prob-lem. Especially speed requirements and accuracy still posesignificant challenges even when programmed by experts.Different approaches to this problem class were devised.Table I depicts a representative selection of works acrossliterature aiming to solve peg-in-hole and categorizes them.As can be seen, insertion times greatly depend on theproblem difficulty, although, modern control methodolo-gies are clearly beneficial compared to older approaches.Learning performance has significantly increased over time,however, part of this improvement may have been bought

with the need for large computational power e.g. GPUsand computing clusters which might be disadvantageous inmore autonomous settings. The difficulty of the consideredproblem settings in terms of geometry and material variesfrom industrial standards to much more simple everydayobjects.

In summary, our contributions are as follows.• Extension of the adaptive impedance controller from

[31] to Cartesian space and full feed-forward tracking.• A novel meta parameter design for the adaptive con-

troller from [31] based on suitable real-world constraintsof impedance controlled systems.

• A novel graph-based skill formalism to describe robotmanipulation skills and bridge the gap between high-level specification and low-level adaptive interactioncontrol. Many existing approaches have a more practicaland high-level focus on the problem and lack a rigidformulation that is able to directly connect planningwith manipulation [1]–[3].

• State-of-the-art learning algorithms such as CovarianceMatrix Adaptation [32], Bayesian optimization [34]and particle swarm optimization [33] are comparedexperimentally within the proposed framework.

• The performance of the proposed framework is show-cased for three different (industrially relevant) peg-in-hole problems. We show that the used system can learntasks complying with industrial speed and accuracyrequirements1.

• The proposed system is able to learn complex manipu-lation tasks in a short amount of time of 5-20 minutes

1In the accompanying video it is demonstrated that on average the usedrobot is even able to perform the task faster than humans.

Page 3: Lars Johannsmeier, Malkin Gerchow and Sami Haddadin · 2018-05-23 · Lars Johannsmeier, Malkin Gerchow and Sami Haddadin1 ... However, such automatic tuning of control and other

depending on the problem while being extremely effi-cient in terms of raw computational power and memoryconsumption. In particular, our entire framework canrun on a small computer such as the Intel NUC whilemaintaining a real-time interface to the robot and at thesame time running a learning algorithm.

The remainder of the paper is organized as follows.Section II describes the adaptive impedance controller, whichbehavior can be changed fundamentally by adapting its metaparameters. Section III introduces our skill definition anddefines the formal problem at hand. In Sec. IV the learningalgorithms applied to the problem definition are investigated.In Section V we apply our approach to the well-known peg-in-hole problem. Finally, Sec. VI concludes the paper.

II. ADAPTIVE IMPEDANCE CONTROLLER

Consider the standard rigid robot dynamics

M(q)q + C(q, q)q + g(q) = τu + τext, (1)

where M(q) is the symmetric, positive definite mass matrix,C(q, q)q the Coriolis and centrifugal torques, g(q) thegravity vector and τext the vector of external link-side jointtorques. The adaptive impedance control law is defined as

τu(t) = J(q)T (−Fff (t)−Fd(t)−K(t)e−De)+τr, (2)

where Fff (t) denotes the adaptive feed-forward wrench.Fd(t) is an optional time dependent feed-forward wrenchtrajectory, K(t) the stiffness matrix, D the damping matrixand J(q) the Jacobian. The position and velocity error aree = x? − x and e = x? − x, respectively. ”?” denotesthe desired motion command. The dynamics compensatorτr can be defined in multiple ways, see for example [31].The adaptive tracking error [36] is

ε = e + κe, (3)

with κ > 0. The adaptive feed-forward wrench Fff andstiffness K are

Fff =

∫ t

0

Fff (t)dt+Fff,0, K(t) =

∫ t

0

K(t)dt+K0, (4)

where Fff,0 and K0 denote the initial values. The controlleradapts feed-forward wrench and stiffness by

Fff (t) =1

Tα(ε− γα(t)Fff (t)), (5)

K(t) =1

Tβ(diag(ε e)− γβ(t)K(t)). (6)

The positive definite matrices α, β, γα, γβ are the learningrates for feed-forward commands, stiffness and the respectiveforgetting factors. The learning rates α and β determinestiffness and feed-forward adaptation speed. γα and γβ areresponsible for slowing down the adaptation process, whichis the main dynamical process for low errors. Cartesiandamping D is designed according to [37] and T denotesthe sample time of the controller. Reordering and inserting(5) and (6) into (2) leads to the overall control policy

τu = fc(xd,Fd, α, β, γα, γβ ,Ω). (7)

Ω denotes the percept vector containing the current pose,velocity, forces etc.

A. Meta Parameter ConstraintsIn order to constrain the subsequent meta learning problem

we make use of the following reasonable constraints thatcharacterize essentially every physical real-world system.For better readability we discuss the scalar case, whichgeneralizes trivially to the multi-dimensional one. The firstconstraint of an adaptive impedance controller is the upperbound of stiffness adaptation speed

Kmax = maxt>0

1

T[β(ε(t)e(t)− γβK(t))] . (8)

If we now assume that K(t = 0) = 0 and e = 0 we maydefine emax as the error at which Kmax =

βe2max

T holds.Then, the maximum value for β can be written as

βmax =KmaxT

e2max

. (9)

Furthermore, the maximum decrease of stiffness occurs fore = 0 and K(t) = Kmax, where Kmax denotes the maximumstiffness, also being an important constraint for any real-world impedance controlled robot. Thus, we may calculatean upper bound for γβ as

γβ,max =Kmax

βmaxKmax. (10)

Finding the constraints of the adaptive feed-forwardwrench may be done analogously. In conclusion, we relatethe upper limits for α, β, γα and γβ to the inherent systemconstraints Kmax, Fmax, Kmax and Fmax.

III. MANIPULATION SKILL

In this section we introduce a mathematical formalism todescribe robot manipulation skills. A manipulation skill isdefined as the directed graph G consisting of nodes n ∈ Nand edges e ∈ E. A node n ∈ N is also called a manipulationprimitive (MP), while an edge e ∈ E is also called transition.The transitions are activated by conditions that mark thesuccess of the preceding MP. A single MP consists of aparameterized twist and feed forward wrench trajectory

xd = ft(Pt,Ω),

Fd = fw(Pw,Ω),

where xd is the desired twist and Fd the feed forwardwrench skill command. These commands are described withrespect to a task frame TF . The percept vector Ω is requiredsince information about the current pose or external forcesmay in general be needed by the MPs. Pt is the set ofparameters used by node n to generate twist commands whilePw is used to generate the wrench commands. Moreover,Pt ⊂ P and Pw ⊂ P , where P is the set of all parameters.Furthermore, we divide P into two different subsets Pc andPl. The parameters in Pc are entirely determined by thecontext of the task e.g. geometric properties of involvedobjects. Pl is the subset of all parameters that are to belearned. They are chosen from a domain D which is alsodetermined by the task context or system capabilities.

Figure 2 shows a principle depiction of the graph G withan additional initial and terminal node. The transition comingfrom the initial node is triggered by the precondition whilethe transition leading to the terminal node is activated by thesuccess condition. Furthermore, every MP has an additionaltransition that is activated by an error condition and leadsto a recovery node. When a transition is activated and the

Page 4: Lars Johannsmeier, Malkin Gerchow and Sami Haddadin · 2018-05-23 · Lars Johannsmeier, Malkin Gerchow and Sami Haddadin1 ... However, such automatic tuning of control and other

active node switches from ni to ni+1 the switching behaviorshould be regulated.

Note that, although it is possible, we do not considerbackwards directed transitions in this work since this wouldintroduce another layer of complexity to the subsequentlearning problem that is out of the scope of this paper.Clearly, this would rather become a high-level planning prob-lem that requires more sophisticated and abstract knowledge.

Cpre

n1 : (xd,Fd)

ni : (xd,Fd)

nm : (xd,Fd)

Csucei−1e1 em−1ei

Initial node Terminal node

Cerr

Recovery

Fig. 2: Skill graph G.

Conditions: There exist three condition types involved inthe execution of skills: preconditions Cpre, error conditionsCerr and success conditions Csuc. They all share the same basicdefinition, yet their application is substantially different. Inparticular, their purpose is to define the limits of the skillfrom start to end. The precondition states the conditionsunder which the skill can be initialized. The error conditionstops the execution when fulfilled and returns a negativeresult. The success condition also stops the skill and returnspositive. Note, that we also make use of the success conditiondefinition to describe the transitions between MPs.

Definition 1 (Condition): Let C ⊂ S be a closed set andc(X(t)) a function c : S → B where B = 0, 1. A conditionholds iff c(X(t)) = 1. Note that the mapping itself obviouslydepends on the specific definition type.

Definition 2 (Precondition): Cpre denotes the chosen setfor which the precondition defined by cpre(X(t)) holds. Thecondition holds, i.e. cpre(X(t0)) = 1, iff ∀ x ∈ X : x(t0) ∈Cpre. t0 denotes the time at start of the skill execution.

Definition 3 (Error Condition): Cerr denotes the chosenset for which the error condition cerr(X(t)) holds, i.e.cerr(X(t)) = 1. This follows from ∃ x ∈ X : x(t) ∈ Cerr.

Definition 4 (Success Condition): Csuc denotes the cho-sen set for which the success condition defined by csuc(X(t))holds, i.e. csuc(X(t)) = 1 iff ∀x ∈ X : x(t) ∈ Csuc.

Evaluation: Lastly, a learning metric is used to evaluatethe skill execution in terms of a success indicator and apredefined cost function.

Definition 5 (Learning Metric): Q denotes the set of all2−tuples (w, fq(X(t)) with 0 < w < 1 and the resultindicator r = 0, 1 where 0 denotes failure and 1 success.Let q =

∑i wifq,i(X(t)) ∀ (wi, fq,i(X(t))) ∈ Q be the cost

function of the skill.Note that the learning metric is designed according to the

task context and potential process requirements. Examplesfor the peg-in-hole skill would be the insertion time or theaverage contact forces during insertion.

There exist other relevant works that make use of manip-ulation primitives such as [38], [39] in which MPs are usedto switch between specific situation dependent control andmotion schemes. They are composed to MP nets to enabletask planning. Similar approaches can be found in [40], [41].

A. Peg-in-HoleIn the following, the well-known, however, still challeng-

ing peg-in-hole skill is described with the help of the aboveformalism. Figure 3 shows the graph of the skill including

the manipulation primitives. The parameters p ∈ P are theestimated hole pose Th, the region of interest around the holeROI, the depth of the hole d, the maximum allowed velocityand acceleration for translation and rotation xmax and xmax,the initial tilt of the object relative to the hole ϕinit, the forcefc, the speed factor s, the amplitude of translational androtational oscillations at, ar and their frequencies ωt, ωr.

Manipulation Primitives: The MPs for peg-in-hole aredefined as follows:• n1: The robot moves to the approach pose.

xd = fp2p(Ta, sxmax, xmax, ), Fd = 0.

fp2p generates a trapezoidal velocity profile to movefrom the current pose to a goal pose while consideringa given velocity and acceleration.

• n2: The robot moves towards the surface with the holeand establishes contact.

xd =[0 0 sxmax 0 0 0

]T, Fd = 0.

• n3: The object is moved laterally in x-direction until itis constrained.

xd =[sxmax 0 0 0 0 0

]T,

Fd =[0 0 fc 0 0 0

]T.

• n4: The object is rotated into the estimated orientationof the hole while keeping contact with the hole.

xd = fp2p(Th, sxmax, xmax),

Fd =[fc 0 fc 0 0 0

]T.

• n5: The object is inserted into the hole.

xd =

at sin (2πωt)

at sin(2π 3

4ωt)

sxmax

ar sin (2πωr)

ar sin(2π 3

4ωr)

0

T

, Fd = 0

IV. PARAMETER LEARNING

Figure 4 shows how the controller and the skill formalismintroduced in Sec. II and III are connected to a given learningmethod to approach the problem of meta learning, i.e.,finding the right (optimal) parameters for solving a giventask. The execution of a particular learning algorithm untilit terminates is named an experiment throughout the paper.A single evaluation of parameters is called a trial.

A. RequirementsA potential learning algorithm will be applied to the

system defined in Sections II and III. In particular, it has tobe able to evaluate sets of parameters per trial in a continuousexperiment. Since we apply it to a real-world physical ma-nipulation problem the algorithm will face various challengesthat result in specific requirements.

1) Generally, no feasible analytic solution2) Gradients are usually not available3) Real world problems are inherently stochastic4) No assumptions possible on minima or cost function

convexity5) Violation of safety, task or quality constraints6) Significant process noise and many repetitions7) Low desired total learning time

Page 5: Lars Johannsmeier, Malkin Gerchow and Sami Haddadin · 2018-05-23 · Lars Johannsmeier, Malkin Gerchow and Sami Haddadin1 ... However, such automatic tuning of control and other

Init n1: Approach n3: Fit n4: Align n5: Insertion

e1 = X|tt ≥ ta e3 = X|Fext,x > fcCpre = X|T ∈ U(Ta)

Csuc = X|T ∈ U(Tg)

n2: Contact

Terminal nodee2 = X|Fext,z > fc e4 = X|tt ≥ th

Cerr = X|T /∈ ROI, τext > τmax, t > tmaxError

Fig. 3: Visualization of the peg-in-hole skill graph. Dashed arrows denote velocity commands, solid ones feed forward force commands. t1 and t4 are thetrajectory durations in states n1 and n4.

minP Q

Robot

ControllerFff (t) =

1T α(ε− γα(t)Fff (t))

(S,O, Cpre, Cerr, Csuc,Xg,Xcmd,X,P,Q)

K(t) = 1T β(diag(ε e)− γβ(t)K(t))

τu(t) = J(q)T (−Fff (t)− Fd −K(t)e−D(t)e) + τr(t)

xd = fX(X,P), Fd = fF (X,P)

τu

xd,Fd

PlQ

Skill

PD Ptα, β, γα, γβ

BO, CMA-ES, PSO

Task Knowledge

X(t)

Fig. 4: Implementation of skill learning framework. Solid lines (—) indicate a continuous flow of information, dotted lines (· · ·) relate to informationexchange only once per trial and dashed lines (−−−) to information relayed only once per experiment.

Thus, suitable learning algorithms will have to fulfill sub-sequent requirements. They must provide a numerical black-box optimization and cannot rely on gradients. Stochasticitymust be regarded and the method has to be able to optimizeglobally. Furthermore, it should handle unknown and noisyconstraints and must provide fast convergence rates.

B. ComparisonTable II lists several groups of state-of-the-art optimization

methods and compares them with respect to above require-ments. In this, we follow and extend the reasoning introducedin [42]. It also shows that most algorithms do not fulfill thenecessary requirements. Note that for all algorithms thereexist extensions for the stochastic case. However, comparingall of them is certainly out of the scope of the paper.Therefore, we focus on the most classical representatives ofthe mentioned classes.

TABLE II: Suitability of existing learning algorithms with regard to theproperties no gradient (NG), stochasticity assumption (SA), global optimizer(GO) and unknown constraints (UC).

Method NG SA GO UCGrid Search + − + −

Pure Random Search + − + −Gradient-descent family − − − −Evolutionary Algorithms + − + −

Particle Swarm + + + −Bayesian Optimization + + + +

Generally, gradient-descent based algorithms require agradient to be available, which obviously makes this class un-suitable. Grid search, pure random search and evolutionaryalgorithms typically do not assume stochasticity and cannothandle unknown constraints very well without extensiveknowledge about the problem they optimize, i.e. make use of

well-informed barrier functions. The latter aspect applies alsoto particle swarm algorithms. Only Bayesian optimization(BO) in accordance to [43] is capable of explicitly handlingunknown noisy constraints during optimization.

Although it seems that Bayesian optimization is the mostsuited method to cope with our problem definition some ofthe other algorithm classes might also be capable of findingsolutions, maybe even faster. Thus, in addition, we selectCovariance Matrix Adaptation Evolutionary Strategy (CMA-ES) [32] and Particle Swarm Optimization (PSO) [33] forcomparison. Furthermore, we utilize Latin Hypercube Sam-pling (LHS) [44] (an uninformed sampler) to gain insightsinto the difficulty of the problems at hand.

1) Bayesian Optimization: For Bayesian optimization wemade use of the spearmint software package [43], [45]–[47].In general, BO finds the minimum of an unknown objectivefunction f(p) on some bounded set X by developing astatistical model of f(p). Apart from the cost function, ithas two major components, which are the prior and theacquisition function.• Prior: We use a Gaussian process as prior to derive

assumptions about the function being optimized. TheGaussian process has a mean function m : X → R anda covariance function B : X × X → R. As a kernelwe use the automatic relevance determination (ARD)Matern 5/2 kernel

BM52(p,p′) =θ0(1 +

√5r2(p,p′)

+5

3r2(p,p′)e−

√5r2(p,p′),

with

r2(p,p′) =

d∑i=1

(pi − p′i)2

θ2i.

Page 6: Lars Johannsmeier, Malkin Gerchow and Sami Haddadin · 2018-05-23 · Lars Johannsmeier, Malkin Gerchow and Sami Haddadin1 ... However, such automatic tuning of control and other

This kernels results in twice-differentiable sample func-tions which makes it suitable for practical optimizationproblems as stated in [45]. It has d+3 hyperparametersin d dimensions, i.e. one characteristic length scale perdimension, the covariance amplitude θ0, the observationnoise ν and a constant mean m. These kernel hyperpa-rameters are integrated out by applying Markov chainMonte Carlo (MCMC) via slice sampling [48].

• Acquisition function: We use predictive entropy searchwith constraints (PESC) as a means to select the nextparameters p to explore, as described in [49].

2) Latin Hypercube Sampling: Latin hypercube sampling(LHS) [44] is a method to sample a given parameter spacein a nearly random way. In contrast to pure random sam-pling LHS generates equally distributed random points inthe parameter space. It might indicate whether complexityreduction of the manipulation task was successful when it ispossible to achieve feasible solutions by sampling.

3) Covariance Matrix Adaptation: The Covariance Ma-trix Adaptation Evolutionary Strategy (CMA-ES) is an opti-mization algorithm from the class of evolutionary algorithmsfor continuous, non-linear, non-convex black-box optimiza-tion problems [32], [50].

The algorithm starts with an initial centroid m ∈ Rn,a population size λ, an initial step-size σ > 0, an initialcovariance matrix C = I and isotropic and anisotropicevolution paths pσ = 0 and pc = 0. m, λ and σ are chosenby the user. Then the following steps are executed until thealgorithm terminates.

1) Evaluation of λ individuals sampled from a normaldistribution with mean m and covariance matrix σC.

2) Update of centroid m, evolution paths pσ and pc,covariance matrix C and step-size σ based on theevaluated fitness.

4) Particle Swarm Optimization: Particle swarm opti-mization usually starts by initializing all particle’s positionsxi(0) and velocities vi(0) with a uniformly distributed ran-dom vector, i.e. xi(0) ∼ U(blb, bub) and vi(0) ∼ U(−|bub−blb|, |bub − blb|) with U being the uniform distribution. Theparticles are evaluated at their initial positions and theirpersonal best pi and the global best g are set. Then, until atermination criterion is met, following steps are executed:

1) Update particle velocity:

vi(t+ 1) =vi(t) + c1(pi − xi(t))R1

+ vi(t) + c2(g − xi(t))R2

where R1 and R2 are diagonal matrices with randomnumbers generated from a uniform distribution in [0, 1]and c1, c2 are acceleration constants usually in therange of [0, 4].

2) Update the particle position:

xi(t+ 1) = xi(t) + vi(t+ 1) (11)

3) Evaluate the fitness of the particle f(xi(t+ 1)).4) Update each pi and global best g if necessary.

V. EXPERIMENTS

In our experiments we investigate the learning methodsselected in Sec. IV and compare them for three differentpeg-in-hole variations for the introduced skill formalism andcontroller, see Sections II and III. The experimental setupconsists of a Franka Emika Panda robot [23] that executesthe following routine:

1) The robot grasps the object to be inserted.

Fig. 5: Experimental setup, puzzle (top right), key (bottom left) and peg(bottom right).

TABLE III: Parameter domain.

Parameter Min Maxα

[0 0

] [0.4 1.1765

[0 0

] [200000 173010

]γα

[0 0

] [5e− 4 3.4e− 4

]γβ

[0 0

] [0 0

]Fc 5 N 15 Ns 0 1

at 0 m 0.005 mωt 0 Hz 3.2 Hzωr 0 Hz 4.5 Hzϕinit 0 rad 0.349 rad

2) A human supervisor teaches the hole position which isfixed with respect to the robot. The teaching accuracywas below 1 mm.

3) A learning algorithm is selected and executed until itterminates after a specified number of trials.

4) For every trial the robot evaluates the chosen param-eters with four slightly different (depending on theactual problem) initial positions in order to achievea more robust result.

We investigated three variations of peg-in-hole as shownin Fig. 5, a key, a puzzle piece and an aluminum peg. Themeta-parameters and skill parameters were learned with thefour methods introduced in Sec. IV.

The parameter domain (see Tab. III) is the same formost of the parameters that are learned in the ex-periments and can be derived from system and safetylimits which are the maximum stiffness Kmax =[2000 N/m 200 Nm/rad

], the maximum stiffness adap-

tation speed Kmax =[5000 N/ms 500 Nm/rads

], the

maximum allowed feed forward wrench and wrench adap-tation of the controller Fff =

[10 N 5 Nm

]and

Fff =[1 N/s 0.5 Nm/s

], the maximum error emax =[

0.005 m 0.017 rad]

and the maximum velocity xmax =[0.1 m/s 1 rad/s

].

The domain of the learned parameters Pl is derived fromthese limits and shown in Tab. III.

In the following the specifics of the three tasks shown inFig. 5 are explained.• Puzzle: The puzzle piece is an equilateral triangle with

a side length of 0.075 m. The maximum rotationalamplitude of the oscillatory motion is given by ar =0.09 rad. The hole has a depth of d = 0.005 m. Thetolerances between puzzle piece and hole are < 0.1 mmand there are no chamfers.

• Key: The key has a depth of d = 0.0023 m. Since the

Page 7: Lars Johannsmeier, Malkin Gerchow and Sami Haddadin · 2018-05-23 · Lars Johannsmeier, Malkin Gerchow and Sami Haddadin1 ... However, such automatic tuning of control and other

hole and the key itself have chamfers to make the initialinsertion very easy we omit the learning of the initialalignment ϕinit and set it to ϕinit = 0 rad. The maximumrotational amplitude of the oscillatory motion is givenby ar = 0.0175 rad. Due to the very small hole nodeliberate initial pose deviation was applied.

• Peg: The aluminum peg has a diameter of 0.02 m andthe hole has a depth of 0.035 m. The tolerances are<< 0.1 mm and there is a 0.5 mm chamfer on the peg.The maximum rotational amplitude of the oscillatorymotion is given by ar = 0.035 rad. The hole has nowalls which results in a higher chance of getting stuckduring insertion further increasing the difficulty.

The learning algorithms are configured as follows.• LHS: The parameter space is sampled at 75 points.• CMA-ES: The algorithm ran for 15 generations with a

population of 5 individuals and σ0 = 0.1. The initialcentroid was set in the middle of the parameter space.

• PSO: We used 25 particles and let the algorithm runfor 3 episodes. The acceleration constants were set toc1 = 2 and c2 = 2.

• BO: The algorithm is initialized with 5 equally dis-tributed random samples from the parameter space.

As cost function for all three problems we used theexecution time measured from first contact to full insertion.A maximum execution time of 15 s was part of the skill’serror condition. In case the error condition was triggered weadded the achieved distance towards the goal pose in the holeto the maximum time multiplied by a factor.

A. Results

The results can be seen in Fig. 6. The blue line is the meanof the execution time per trial averaged over all experiments.The grey area indicates the 90 % confidence interval.

The results indicate that all four algorithms are suitedto a certain degree to learn easier variations of the peg-in-hole task such as the puzzle and the key. However, itmust be noted that Bayesian optimization on average findssolutions not as good as the other methods. Furthermore, theconfidence interval is notably larger. It also terminates earlyinto the experiment since the model was at some point notable to find further suitable candidates. This might indicatea solution space with high noise and discontinuities that isdifficult to model.

The comparison with LHS also indicates that the complex-ity reduction of our formal approach to manipulation skillsmakes it possible to find solutions with practically relevantexecution times by sampling rather then explicit learning.

At the bottom of Fig. 6 the results for the most difficultpeg-in-hole variation are shown. We do not include theresults of BO since it was not able to find a solution inthe given time frame. The plot showing the LHS methodindicates a very hard problem. Random sampling leadsto feasible solutions, however, the confidence interval istoo large to conclude any assurance. PSO achieves bettersolutions yet also has a very high confidence interval. CMA-ES outperforms both methods and is able to find a solutionthat is better in terms of absolute cost and confidence.

Considering the best performing algorithm CMA-ES afeasible solution for any of the tasks was already foundafter 2 − 4 minutes and optimized after 5 − 20 minutesdepending on the task, significantly outperforming existingapproaches for learning peg-in-hole, see Tab. I. Note alsothat with the exception of BO no noteworthy computationtime was necessary.

VI. CONCLUSION

In this paper we introduced an approach to learning roboticmanipulation that is based on a novel skill formalism andmeta learning of adaptive controls. Overall, the proposedframework is able to solve highly complex problems such aslearning peg-in-hole with sub-millimeter tolerances in a lowamount of time, making it even applicable for industrial useand outperforming existing approaches in terms of learningspeed and resource demands. Remarkably, the used robotwas even able to outperform humans in execution time.

Summarizing the results we conclude that the applicationof complexity reduction via adaptive impedance control andsimple manipulation strategies (skills) in combination withthe right learning method to complex manipulation problemsis feasible. The results might further indicate that methodssupported by a certain degree of random sampling arepossibly better suited for learning this type of manipulationskills than those more relying on learning a model of thecost function. These conclusions will be investigated morethoroughly in future work.

Overall, our results show that, in contrast to purely data-driven learning/control approaches that do not make useof task structures nor the concept of (adaptive) impedance,our approach of combining sophisticated adaptive controltechniques with state-of-the-art machine learning makes evensuch complex problems tractable and applicable to the realphysical world and its highly demanding requirements. Typ-ically, other existing manipulation learning schemes requireorders of magnitude more iterations already in simulationwhile it is known that nowadays simulations cannot realisti-cally (if at all) capture real-world contacts.

Clearly, the next steps we intend to take are the applicationof our framework to other manipulation problem classes andthe thorough analysis of the generalization capabilities of thesystem to similar, however, yet unknown problems.

REFERENCES

[1] M. R. Pedersen, L. Nalpantidis, R. S. Andersen, C. Schou, S. Bøgh,V. Kruger, and O. Madsen, “Robot skills for manufacturing: Fromconcept to industrial deployment,” Robotics and Computer-IntegratedManufacturing, 2015.

[2] U. Thomas, G. Hirzinger, B. Rumpe, C. Schulze, and A. Wortmann,“A new skill based robot programming language using uml/p state-charts,” in Robotics and Automation (ICRA), 2013 IEEE InternationalConference on. IEEE, 2013, pp. 461–466.

[3] R. H. Andersen, T. Solund, and J. Hallam, “Definition and initial case-based evaluation of hardware-independent robot skills for industrialrobotic co-workers,” in ISR/Robotik 2014; 41st International Sympo-sium on Robotics; Proceedings of. VDE, 2014, pp. 1–7.

[4] P. Pastor, H. Hoffmann, T. Asfour, and S. Schaal, “Learning and gener-alization of motor skills by learning from demonstration,” in Roboticsand Automation, 2009. ICRA’09. IEEE International Conference on.IEEE, 2009, pp. 763–768.

[5] P. Pastor, M. Kalakrishnan, S. Chitta, E. Theodorou, and S. Schaal,“Skill learning and task outcome prediction for manipulation,” inRobotics and Automation (ICRA), 2011 IEEE International Conferenceon. IEEE, 2011, pp. 3828–3834.

[6] J. Kober and J. Peters, “Learning motor primitives for robotics,”in Robotics and Automation, 2009. ICRA’09. IEEE InternationalConference on. IEEE, 2009, pp. 2112–2118.

[7] J. Kober and J. R. Peters, “Policy search for motor primitives inrobotics,” in Advances in neural information processing systems, 2009,pp. 849–856.

[8] S. Schaal, J. Peters, J. Nakanishi, and A. Ijspeert, “Learning move-ment primitives,” in Robotics Research. The Eleventh InternationalSymposium. Springer, 2005, pp. 561–572.

[9] V. Gullapalli, J. A. Franklin, and H. Benbrahim, “Acquiring robot skillsvia reinforcement learning,” IEEE Control Systems, vol. 14, no. 1, pp.13–24, 1994.

[10] S. Levine, N. Wagener, and P. Abbeel, “Learning contact-rich manip-ulation skills with guided policy search,” in Robotics and Automation(ICRA), 2015 IEEE International Conference on. IEEE, 2015, pp.156–163.

[11] S. R. Chhatpar and M. S. Branicky, “Search strategies for peg-in-hole assemblies with position uncertainty,” in Intelligent Robots andSystems, 2001. Proceedings. 2001 IEEE/RSJ International Conferenceon, vol. 3. IEEE, 2001, pp. 1465–1470.

Page 8: Lars Johannsmeier, Malkin Gerchow and Sami Haddadin · 2018-05-23 · Lars Johannsmeier, Malkin Gerchow and Sami Haddadin1 ... However, such automatic tuning of control and other

Puzzle

Key

Peg

Trial [1] Trial [1] Trial [1]

Cos

t[s

]C

ost

[s]

CMA-ESPSOBO

Cos

t[s

]

LHS

No solution

Trial [1]Fig. 6: Experimental results. Columns show the plots for the learning algorithms (LHS,BO,PSO,CMA-ES) and rows for the tasks (Puzzle,Key,Peg).

[12] J. F. Broenink and M. L. Tiernego, “Peg-in-hole assembly usingimpedance control with a 6 dof robot,” 1996.

[13] T. Lozano-Perez, M. T. Mason, and R. H. Taylor, “Automatic synthesisof fine-motion strategies for robots,” The International Journal ofRobotics Research, vol. 3, no. 1, pp. 3–24, 1984.

[14] H. Bruyninckx, S. Dutre, and J. De Schutter, “Peg-on-hole: a modelbased solution to peg and hole alignment,” in Robotics and Automa-tion, 1995. Proceedings., 1995 IEEE International Conference on,vol. 2. IEEE, 1995, pp. 1919–1924.

[15] W. S. Newman, Y. Zhao, and Y.-H. Pao, “Interpretation of forceand moment signals for compliant peg-in-hole assembly,” in Roboticsand Automation, 2001. Proceedings 2001 ICRA. IEEE InternationalConference on, vol. 1. IEEE, 2001, pp. 571–576.

[16] A. Stemmer, G. Schreiber, K. Arbter, and A. Albu-Schaffer, “Robustassembly of complex shaped planar parts using vision and force,” inMultisensor Fusion and Integration for Intelligent Systems, 2006 IEEEInternational Conference on. IEEE, 2006, pp. 493–500.

[17] V. Gullapalli, R. A. Grupen, and A. G. Barto, “Learning reactiveadmittance control,” in Robotics and Automation, 1992. Proceedings.,1992 IEEE International Conference on. IEEE, 1992, pp. 1475–1480.

[18] B. Nemec, F. J. Abu-Dakka, B. Ridge, A. Ude, J. A. Jorgensen, T. R.Savarimuthu, J. Jouffroy, H. G. Petersen, and N. Kruger, “Transferof assembly operations to new workpiece poses by adaptation tothe desired force profile,” in Advanced Robotics (ICAR), 2013 16thInternational Conference on. IEEE, 2013, pp. 1–7.

[19] A. Kramberger, A. Gams, B. Nemec, C. Schou, D. Chrysostomou,O. Madsen, and A. Ude, “Transfer of contact skills to new environ-mental conditions,” in Humanoid Robots (Humanoids), 2016 IEEE-RAS 16th International Conference on. IEEE, 2016, pp. 668–675.

[20] K. Kronander, E. Burdet, and A. Billard, “Task transfer via collabo-rative manipulation for insertion assembly,” in Workshop on Human-Robot Interaction for Industrial Manufacturing, Robotics, Science andSystems, 2014.

[21] S. Levine, C. Finn, T. Darrell, and P. Abbeel, “End-to-end trainingof deep visuomotor policies,” Journal of Machine Learning Research,vol. 17, no. 39, pp. 1–40, 2016.

[22] C. Devin, A. Gupta, T. Darrell, P. Abbeel, and S. Levine, “Learningmodular neural network policies for multi-task and multi-robot trans-fer,” in Robotics and Automation (ICRA), 2017 IEEE InternationalConference on. IEEE, 2017, pp. 2169–2176.

[23] S. Haddadin, S. Haddadin, and S. Parusel. (2017) Franka emikagmbh. [Online]. Available: www.franka.de

[24] G. Hirzinger, N. Sporer, A. Albu-Schaffer, M. Hahnle, R. Krenn,A. Pascucci, and M. Schedl, “Dlr’s torque-controlled light weightrobot iii-are we reaching the technological limits now?” in Roboticsand Automation, 2002. Proceedings. ICRA’02. IEEE InternationalConference on, vol. 2. IEEE, 2002, pp. 1710–1716.

[25] C.-C. Cheah and D. Wang, “Learning impedance control for roboticmanipulators,” IEEE Transactions on Robotics and Automation,vol. 14, no. 3, pp. 452–465, Jun 1998.

[26] E. Gribovskaya, A. Kheddar, and A. Billard, “Motion learning andadaptive impedance for robot control during physical interaction withhumans,” in 2011 IEEE International Conference on Robotics andAutomation, May 2011, pp. 4326–4332.

[27] J. Buchli, F. Stulp, E. Theodorou, and S. Schaal, “Learning variableimpedance control,” The International Journal of Robotics Research,vol. 30, no. 7, pp. 820–833, 2011.

[28] G. Ganesh, A. Albu-Schaffer, M. Haruno, M. Kawato, and E. Burdet,“Biomimetic motor behavior for simultaneous adaptation of force,impedance and trajectory in interaction tasks,” in Robotics and Au-tomation (ICRA), 2010 IEEE International Conference on. IEEE,2010, pp. 2705–2711.

[29] A. Albu-Schaffer, O. Eiberger, M. Grebenstein, S. Haddadin, C. Ott,T. Wimbock, S. Wolf, and G. Hirzinger, “Soft robotics,” IEEE Robotics& Automation Magazine, vol. 15, no. 3, 2008.

[30] N. Hogan, “Impedance control: An approach to manipulation,” inAmerican Control Conference, 1984. IEEE, 1984, pp. 304–313.

[31] C. Yang, G. Ganesh, S. Haddadin, S. Parusel, A. Albu-Schaeffer, andE. Burdet, “Human-like adaptation of force and impedance in stableand unstable interactions,” Robotics, IEEE Transactions on, vol. 27,no. 5, pp. 918–930, 2011.

[32] N. Hansen and A. Ostermeier, “Completely derandomized self-adaptation in evolution strategies,” Evolutionary computation, vol. 9,no. 2, pp. 159–195, 2001.

[33] R. Poli, J. Kennedy, and T. Blackwell, “Particle swarm optimization,”Swarm intelligence, vol. 1, no. 1, pp. 33–57, 2007.

[34] B. Shahriari, K. Swersky, Z. Wang, R. P. Adams, and N. de Freitas,“Taking the human out of the loop: A review of bayesian optimiza-tion,” Proceedings of the IEEE, vol. 104, no. 1, pp. 148–175, 2016.

[35] R. S. Johansson and J. R. Flanagan, “Coding and use of tactile signalsfrom the fingertips in object manipulation tasks,” Nature ReviewsNeuroscience, vol. 10, no. 5, p. 345, 2009.

[36] J.-J. E. Slotine, W. Li, et al., Applied nonlinear control. Prentice-hallEnglewood Cliffs, NJ, 1991, vol. 199, no. 1.

[37] A. Albu-Schaffer, C. Ott, U. Frese, and G. Hirzinger, “Cartesianimpedance control of redundant robots: Recent results with the DLR-light-weight-arms,” in IEEE Int. Conf. on Robotics and Automation,vol. 3, 2003, pp. 3704–3709.

[38] B. Finkemeyer, T. Kroger, and F. M. Wahl, “Executing assembly tasksspecified by manipulation primitive nets,” Advanced Robotics, vol. 19,no. 5, pp. 591–611, 2005.

[39] T. Kroger, B. Finkemeyer, and F. M. Wahl, “Manipulation primitivesauniversal interface between sensor-based motion control and robot pro-gramming,” in Robotic Systems for Handling and Assembly. Springer,2010, pp. 293–313.

[40] I. Weidauer, D. Kubus, and F. M. Wahl, “A hierarchical extension ofmanipulation primitives and its integration into a robot control archi-tecture,” in Robotics and Automation (ICRA), 2014 IEEE InternationalConference on. IEEE, 2014, pp. 5401–5407.

[41] C. Pek, A. Muxfeldt, and D. Kubus, “Simplifying synchronization incooperative robot tasks-an enhancement of the manipulation primitiveparadigm,” in Emerging Technologies and Factory Automation (ETFA),2016 IEEE 21st International Conference on. IEEE, 2016, pp. 1–8.

[42] R. Calandra, A. Seyfarth, J. Peters, and M. P. Deisenroth, “Anexperimental comparison of bayesian optimization for bipedal locomo-tion,” in Robotics and Automation (ICRA), 2014 IEEE InternationalConference on. IEEE, 2014, pp. 1951–1958.

[43] J. Snoek, “Bayesian optimization and semiparametric models withapplications to assistive technology,” Ph.D. dissertation, University ofToronto, 2013.

[44] M. D. McKay, R. J. Beckman, and W. J. Conover, “Comparison ofthree methods for selecting values of input variables in the analysisof output from a computer code,” Technometrics, vol. 21, no. 2, pp.239–245, 1979.

[45] J. Snoek, H. Larochelle, and R. P. Adams, “Practical bayesian op-timization of machine learning algorithms,” in Advances in neuralinformation processing systems, 2012, pp. 2951–2959.

[46] E. Brochu, V. M. Cora, and N. De Freitas, “A tutorial on bayesianoptimization of expensive cost functions, with application to activeuser modeling and hierarchical reinforcement learning,” arXiv preprintarXiv:1012.2599, 2010.

[47] K. Swersky, J. Snoek, and R. P. Adams, “Multi-task bayesian optimiza-tion,” in Advances in neural information processing systems, 2013, pp.2004–2012.

[48] R. M. Neal, “Slice sampling,” Annals of statistics, pp. 705–741, 2003.[49] J. M. Hernandez-Lobato, M. A. Gelbart, M. W. Hoffman, R. P.

Adams, and Z. Ghahramani, “Predictive entropy search for bayesianoptimization with unknown constraints.” in ICML, 2015, pp. 1699–1707.

[50] N. Hansen, “The cma evolution strategy: a comparing review,” To-wards a new evolutionary computation, pp. 75–102, 2006.