Graphical Models - and Protein Signalling Networks

71
Graphical Models and Protein Signalling Networks Marco Scutari [email protected] Genetics Institute University College London November 5, 2012 Marco Scutari University College London

Transcript of Graphical Models - and Protein Signalling Networks

Page 1: Graphical Models - and Protein Signalling Networks

Graphical Modelsand Protein Signalling Networks

Marco Scutari

[email protected] Institute

University College London

November 5, 2012

Marco Scutari University College London

Page 2: Graphical Models - and Protein Signalling Networks

Graphical Models

Marco Scutari University College London

Page 3: Graphical Models - and Protein Signalling Networks

Graphical Models

Graphical Models

Graphical models are defined by:

• a network structure, G = (V, E), either an undirected graph(Markov networks, gene association networks, correlationnetworks, etc.) or a directed graph (Bayesian networks). Eachnode vi ∈ V corresponds to a random variable Xi;

• a global probability distribution, X, which can be factorisedinto a small set of local probability distributions according tothe edges eij ∈ E present in the graph.

This combination allows a compact representation of the jointdistribution of large numbers of random variables and simplifiesinference on the resulting parameter space.

Marco Scutari University College London

Page 4: Graphical Models - and Protein Signalling Networks

Graphical Models

A Simple Bayesian Network: Watson’s Lawn

TRUE FALSE

SPRINKLER

0.4 0.6

TRUE FALSE

RAIN

0.2 0.8

SPRINKLERFALSE

GRASS WET

0.0 1.0

TRUERAIN

FALSEFALSE

0.8 0.2TRUEFALSE

0.9 0.1FALSETRUE

0.99 0.01TRUETRUE

RAIN

FALSE

0.01 0.99TRUE

SPRINKLERSPRINKLERSPRINKLER RAIN

GRASS WET

Marco Scutari University College London

Page 5: Graphical Models - and Protein Signalling Networks

Graphical Models

Graphical Separation and Independence

The main role of the graph structure is to express the conditionalindependence relationships among the variables in the model, thusspecifying the factorisation of the global distribution. Differentclasses of graphs express these relationships with differentsemantics, which have in common the principle that graphicalseparation of two (sets of) nodes implies the conditionalindependence of the corresponding (sets of) random variables.

For networks considered here, separation is defined as:

• (u-)separation in Markov networks;

• d-separation in Bayesian networks.

Marco Scutari University College London

Page 6: Graphical Models - and Protein Signalling Networks

Graphical Models

Graphical Separation

separation (undirected graphs)

d-separation (directed acyclic graphs)

CA B

CA B

CA B

CA B

Marco Scutari University College London

Page 7: Graphical Models - and Protein Signalling Networks

Graphical Models

Maps and Independence

A graph G is a dependency map (or D-map) of the probabilisticdependence structure P of X if there is a one-to-one correspondencebetween the random variables in X and the nodes V of G, such that forall disjoint subsets A, B, C of X we have

A ⊥⊥P B |C =⇒ A ⊥⊥G B |C.

Similarly, G is an independency map (or I-map) of P if

A ⊥⊥P B |C⇐= A ⊥⊥G B |C.

G is said to be a perfect map of P if it is both a D-map and an I-map,that is

A ⊥⊥P B |C⇐⇒ A ⊥⊥G B |C,

and in this case P is said to be isomorphic to G.

Graphical models are formally defined as I-maps under the respectivedefinitions of graphical separation.

Marco Scutari University College London

Page 8: Graphical Models - and Protein Signalling Networks

Graphical Models

Bayesian Networks, Equivalence Classes and Moral Graphs

Following the definitions given in the previous couple of slides, the graphassociated with a Bayesian network has three useful transforms:

• the skeleton: the undirected graph underlying a Bayesian network,i.e. the graph we get if we disregard edges’ direction.

• the equivalence class: the graph (CPDAG) in which only edgeswhich are part of a v-structure (i.e. A→ C ← B) and/or mightresult in one are directed. All valid combinations of the other edges’directions result in networks representing the same dependencestructure P .

• the moral graph: the graph obtained by disregarding edges’direction and joining the two parents in each v-structure with anedge. This is essentially a way to transform a Bayesian network intoa Markov network.

Marco Scutari University College London

Page 9: Graphical Models - and Protein Signalling Networks

Graphical Models

Skeletons and Equivalence Classes

DAG

X1

X10

X2 X3

X4

X5

X6

X7

X8X9

Skeleton

X1

X10

X2 X3

X4

X5

X6

X7

X8X9

CPDAG

X1

X10

X2 X3

X4

X5

X6

X7

X8X9

An Equivalent DAG

X1

X10

X2 X3

X4

X5

X6

X7

X8X9

Marco Scutari University College London

Page 10: Graphical Models - and Protein Signalling Networks

Graphical Models

Factorisation into Local Distributions

The most important consequence of defining graphical models as I-mapsis the factorisation of the global distribution into local distributions:

• in Markov networks, local distributions are associated with thecliques Ci (maximal subsets of nodes in which each element isadjacent to all the others) in the graph,

P(X) =

k∏i=1

ψi(Ci),

and the ψk functions are called potentials.

• in Bayesian networks, each local distribution is associated with asingle node Xi and depends only on the joint distribution of itsparents ΠXi

:

P(X) =

p∏i=1

P(Xi |ΠXi)

Marco Scutari University College London

Page 11: Graphical Models - and Protein Signalling Networks

Graphical Models

Neighbourhoods and Markov Blankets

Furthermore, for each node Xi two sets are defined:

• the neighbourhood, the set of nodes that are adjacent to Xi. Thesenodes cannot be made independent from Xi.

• the Markov blanket, the set of nodes that completely separates Xi

from the rest of the graph. Generally speaking, it is the set of nodesthat includes all the knowledge needed to do inference on Xi, fromestimation to hypothesis testing to prediction, because all the othernodes are conditionally independent from Xi given its Markovblanket.

These sets are related in Markov and Bayesian networks; in particular,Markov blankets can be shown to be the same using a moral graph.

Marco Scutari University College London

Page 12: Graphical Models - and Protein Signalling Networks

Graphical Models

Neighbourhoods and Markov Blankets

G

F

C

KB

A

H

E

D

L G

F

C

KB

A

H

E

D

L

Bayesian network Markov network

Markov blanket

Parents Children

Children's otherparents

Neighbours

Marco Scutari University College London

Page 13: Graphical Models - and Protein Signalling Networks

Graphical Models

Probability Distributions: Discrete and Continuous

Data used in graphical modelling should respect the followingassumptions:

• if all the variables Xi are discrete, both the global and thelocal distributions are assumed to be multinomial. Localdistributions are described using conditional probability tables;

• if all the variables Xi are continuous, the global distribution isassumed to be a multivariate Gaussian distribution, and thelocal distributions are univariate or multivariate Gaussiandistributions. Local distributions are described using partialcorrelation coefficients;

• if both continuous and discrete variables are present, we canassume a mixture or conditional Gaussian distribution,discretise continuous attributes or use a nonparametricapproach.

Marco Scutari University College London

Page 14: Graphical Models - and Protein Signalling Networks

Graphical Models

Other Distributional Assumptions

Other fundamental distributional assumptions are:

• observations must be independent. If some form of temporalor spatial dependence is present, it must be specificallyaccounted for in the definition of the network (as in dynamicBayesian networks);

• if the model will be used as a causal graphical model, that is,to infer cause-effect relationship from experimental or (morefrequently) observational data, there must be no latent orhidden variables that influence the dependence structure ofthe model;

• all the relationships between the variables in the network mustbe conditional independencies, because they are by definitionthe only ones that can be expressed by graphical models.

Marco Scutari University College London

Page 15: Graphical Models - and Protein Signalling Networks

Graphical Models

A Gaussian Markov Network (MARKS)

mechanics analysis

vectors statistics

algebra

mechanics

vectors

algebra

analysis

statistics

algebra

Marco Scutari University College London

Page 16: Graphical Models - and Protein Signalling Networks

Graphical Models

A Discrete Bayesian Network (ASIA)

visit to Asia? smoking?

tuberculosis? lung cancer? bronchitis?

either tuberculosisor lung cancer?

positive X-ray?dyspnoea?

Marco Scutari University College London

Page 17: Graphical Models - and Protein Signalling Networks

Graphical Models

A Discrete Bayesian Network (ASIA)

visit to Asia?

tuberculosis?

smoking?

lung cancer?

smoking?

bronchitis?

tuberculosis? lung cancer?

either tuberculosisor lung cancer?

either tuberculosisor lung cancer?

positive X-ray?

bronchitis?either tuberculosisor lung cancer?

dyspnoea?

visit to Asia? smoking?

Marco Scutari University College London

Page 18: Graphical Models - and Protein Signalling Networks

Graphical Models

Limitations of These Probability Distribution

• no real-world, multivariate data set follows a multivariate Gaussiandistribution; even if the marginal distributions are normal, not alldependence relationships are linear.

• computing partial correlations is problematic in most large data sets(and in a lot of small ones, too).

• parametric assumptions for mixed data have strong limitations, asthey impose constraints on which edges may be present in the graph(e.g. a continuous node cannot be the parent of a discrete node).

• discretisation is a common solution to the above problems, but itdiscards useful information and it is tricky to get right (i.e. choosinga set of intervals such that the dependence relationships involvingthe original variable are preserved).

• ordered categorical variables are treated as unordered, again losinginformation.

Marco Scutari University College London

Page 19: Graphical Models - and Protein Signalling Networks

Graphical Model Learning

Marco Scutari University College London

Page 20: Graphical Models - and Protein Signalling Networks

Graphical Model Learning

Learning a Graphical Model

Model selection and estimation are collectively known as learning, andare usually performed as a two-step process:

1. structure learning, learning the graph structure from the data.

2. parameter learning, learning the local distributions implied by thegraph structure learned in the previous step.

This workflow is implicitly Bayesian; given a data set D and if we denotethe parameters of the global distribution as X with Θ, we have

P(M|D) = P(G,Θ | D)︸ ︷︷ ︸learning

= P(G |D)︸ ︷︷ ︸structure learning

· P(Θ | G,D)︸ ︷︷ ︸parameter learning

and structure learning is done in practice as

P(G |D) ∝ P(G) P(D |G) = P(G)

∫P(D |G,Θ) P(Θ | G)dΘ.

Marco Scutari University College London

Page 21: Graphical Models - and Protein Signalling Networks

Graphical Model Learning

Local Distributions: Divide and Conquer

Most tasks related to both learning and inference are NP-hard (theycannot be solved in polynomial time in the number of variables). Theyare still feasible thanks to the decomposition of X into the localdistributions; under some assumptions (parameter independence) there isnever the need to manipulate more than one of them at a time.

In Bayesian networks, for example, structure learning boils down to

P(D |G) =

∫ ∏[P(Xi |ΠXi

,ΘXi) P(ΘXi

|ΠXi)] dΘ

=∏[∫

P(Xi |ΠXi,ΘXi

) P(ΘXi|ΠXi

)dΘXi

]and parameter learning boils down to

P(Θ | G,D) =∏

P(ΘXi|ΠXi

,D).

Marco Scutari University College London

Page 22: Graphical Models - and Protein Signalling Networks

Structure Learning

Marco Scutari University College London

Page 23: Graphical Models - and Protein Signalling Networks

Structure Learning

The Big Three: Constraint-based, Score-based and Hybrid

Despite the (sometimes confusing) variety of theoretical backgrounds andterminology they can all be traced to only three approaches:

• constraint-based algorithms: they use statistical tests to learnconditional independence relationships (called constraints in thissetting) from the data and assume that the graph underlying theprobability distribution is a perfect map to determine the correctnetwork structure.

• score-based algorithms: each candidate network is assigned a scorereflecting its goodness of fit, which is then taken as an objectivefunction to maximise.

• hybrid algorithms: conditional independence tests are used to learnat least part of the conditional independence relationships from thedata, thus restricting the search space for a subsequent score-basedsearch. The latter determines which edges are actually present inthe graph and, in the case of Bayesian networks, their direction.

Marco Scutari University College London

Page 24: Graphical Models - and Protein Signalling Networks

Structure Learning

Constraint-based Structure Learning Algorithms

The mapping between edges and conditional independence relationshipslies at the core of graphical modelling; therefore, one way to learn thestructure of a graphical model is to check which ones of suchrelationships hold according to a suitable conditional independence test.

Such an approach results in a set of conditional independence constraintsthat identify a single graph (for a Markov network) or a single equivalenceclass (for a Bayesian network). In the latter case, the relevant edgedirections are determined using more conditional independence tests toidentify which v-structures are present in the graph.

The first constraint-based algorithm was pioneered by Verma & Pearl,and is names Inductive Causation. It’s not usable in practice, but itprovided a theoretical framework for later algorithms.

Marco Scutari University College London

Page 25: Graphical Models - and Protein Signalling Networks

Structure Learning

Conditional Independence Tests

Classic tests are used because they are fast but are not particularly good.

• asymptotic discrete tests: mutual information/log-likelihood ratioand Pearson’s X2 with a χ2 distribution.

• asymptotic continuous tests: Fisher’s Z, with a N(0, 1) distribution,and mutual information/log-likelihood ratio, with a χ2 distribution.

• exact continuous tests: t test with a Student’s t distribution.

Better alternatives are:

• permutation tests: all of the above, evaluated using thepermutation distribution as the null distribution. The resultingstructure is better for goodness-of-fit and prediction.

• shrinkage tests: log-likelihood ratio tests can be reworked asshrinkage tests whose behaviour is determined by a regularisationparameter λ. The resulting structure is closer to the “real” one andis therefore better for causal reasoning.

Marco Scutari University College London

Page 26: Graphical Models - and Protein Signalling Networks

Structure Learning

Other Constraint-based algorithms

• Peter & Clark (PC): a true-to-form implementation of the InductiveCausation algorithm, specifying only the order of the conditionalindependence tests. Starts from a saturated network and performstests gradually increasing the number of conditioning nodes.

• Grow-Shrink (GS) and Incremental Association (IAMB) variants:these algorithms learn the Markov blanket of each node to reducethe number of tests required by the Inductive Causation algorithm.Markov blankets are learned using different forward and step-wiseapproaches; the initial network is assumed to be empty (i.e. not tohave any edge).

• Max-Min Parents & Children (MMPC): uses a minimax approach toavoid conditional independence tests known a priori to accept thenull hypothesis of independence.

Marco Scutari University College London

Page 27: Graphical Models - and Protein Signalling Networks

Structure Learning

Pros & Cons of Constraint-based Algorithms

• They depend heavily on the quality of the conditionalindependence tests they use; all proofs of correctness assumetests are always right. That’s why asymptotic tests are bad,and non-regularised parametric tests are not ideal.

• They are consistent, but converge is slower than score-basedand hybrid algorithms.

• At any single time they evaluate a small subset of variables,which makes them very memory efficient.

• They do not require multiple testing adjustment in most cases.

• They are embarrassingly parallel, so they scale extremely well.

Marco Scutari University College London

Page 28: Graphical Models - and Protein Signalling Networks

Structure Learning

Score-based Structure Learning Algorithms

The dimensionality of the space of graph structures makes an exhaustivesearch unfeasible in practice, regardless of the goodness-of-fit measure(called network score) used in the process. However, heuristics can stillbe used in conjunction with decomposable scores, i.e.

Score(G) =∑

Score(Xi |ΠXi)

such as

BIC(G) =∑

log P(Xi |ΠXi)− |ΘXi |

2log n

BDe(G),BGe(G) =∑

log

[∫P(Xi |ΠXi

,ΘXi) P(ΘXi

|ΠXi)dΘXi

]

if each comparison involves structures differing in only one localdistribution at a time.

Marco Scutari University College London

Page 29: Graphical Models - and Protein Signalling Networks

Structure Learning

The Hill-Climbing Algorithm

Initial BIC score: −1807.528

MECH

VECT

ALG

ANLSTAT

Current BIC score: −1778.804

MECH

VECT

ALG

ANLSTAT

Current BIC score: −1755.383

MECH

VECT

ALG

ANLSTAT

Current BIC score: −1737.176

MECH

VECT

ALG

ANLSTAT

Current BIC score: −1723.325

MECH

VECT

ALG

ANLSTAT

Current BIC score: −1720.901

MECH

VECT

ALG

ANLSTAT

Current BIC score: −1720.150

MECH

VECT

ALG

ANLSTAT

Final BIC score: −1720.150

MECH

VECT

ALG

ANLSTAT

Marco Scutari University College London

Page 30: Graphical Models - and Protein Signalling Networks

Structure Learning

Other Score-based Algorithms

• Hill-Climbing + Random Restart: performs several hill-climbingruns, perturbing the result of each one as the initial network for thenext. It does get stuck in local maxima as often as plainhill-climbing.

• Greedy Equivalent Search: hill-climbing over equivalence classesrather than graph structures; the search space is much smaller.

• Tabu Search: a modified hill-climbing that keeps a list of the last kstructures visited, and returns only if they are all worse than thecurrent one.

• Genetic Algorithms: they perturb (mutation) and combine crossoverfeatures through several generations of structures, and keep theones leading to better scores. Inspired by Darwinian evolution.

• Simulated Annealing: again similar to hill-climbing, but not lookingat the maximum score improvement at each step. Very difficult touse in practice because of its tuning parameters.

Marco Scutari University College London

Page 31: Graphical Models - and Protein Signalling Networks

Structure Learning

Pros & Cons of Score-based Algorithms

• Convergence to the global maximum (i.e. the best structure)is not guaranteed for finite samples, the search may get stuckin a local maximum.

• They are consistent, and they converge faster thanconstraint-based algorithms, but this is more due to theproperties of the BDe and BGe scores than the algorithmsthemselves.

• They require a definition of both the global and the localdensities, and a matching decomposable network score.

• Most scores have tuning parameters, whereas conditionalindependence tests do not.

Marco Scutari University College London

Page 32: Graphical Models - and Protein Signalling Networks

Structure Learning

Hybrid Structure Learning Algorithms

Hybrid algorithms combine constraint-based and score-based algorithmsto complement the respective strengths and weaknesses; they areconsidered the state of the art in current literature.

They work by alternating the following two steps:

• restrict: learn some conditional independence constraints to reducethe number of candidate networks;

• maximise: find the best network that satisfies those constraints anddefine a new set of constraints to improve on.

These steps can be repeated several times (until convergence), but one ortwo times is usually enough.

The algorithm that pioneered this approach is the Sparse Candidate byFriedman et al., and more recently Max-Min Hill-Climbing (MMHC).

Marco Scutari University College London

Page 33: Graphical Models - and Protein Signalling Networks

Structure Learning

Pros & Cons of Structure Learning Algorithms

• Since only the general framework is defined, it is easy tomodify them to use newer constraint-based and score-basedalgorithms.

• You can mix and match conditional independence tests andnetwork scores to create a learning algorithm ranging fromfrequentist to Bayesian to information-theoretic and anythingin between (within reason).

• They are usually faster than the alternatives, and more stable.

• Tuning parameters can be difficult to tune for someconfigurations of algorithms, tests and scores.

Marco Scutari University College London

Page 34: Graphical Models - and Protein Signalling Networks

Parameter Learning

Marco Scutari University College London

Page 35: Graphical Models - and Protein Signalling Networks

Parameter Learning

The Big Three: Likelihood, Bayesian and Shrinkage

Once the structure of the model is known, the problem of estimating theparameters of the global distribution can be solved by estimating theparameters of the local distributions, one at a time.

Three common choices are:

• maximum likelihood estimators: just the usual empirical estimators.Often described as either maximum entropy or minimum divergenceestimators in information-theoretic literature.

• Bayesian posterior estimators: posterior estimators, based onconjugate priors to keep computations fast, simple and in closedform.

• shrinkage estimators: regularised estimators based either onJames-Stein or Bayesian shrinkage results.

Marco Scutari University College London

Page 36: Graphical Models - and Protein Signalling Networks

Parameter Learning

Maximum Likelihood and Maximum Entropy Estimation

The classic estimators for (conditional) probabilities and (partial)correlations are a bad choice for almost all real-world problems.They are still around because:

• they are used in benchmark simulations;

• computer scientists do not care much about parameter estimation.

However:

• maximum likelihood estimates are unstable in most multivariateproblems, both discrete and continuous;

• for the multivariate Gaussian distribution, James & Stein proved inthe 1950s that the maximum likelihood estimator for the mean isnot admissible in 3+ dimensions;

• partial correlations are often ill-behaved because of that, even withMoore-Penrose pseudo-inverses;

• maximum likelihood estimates are non-smooth and create problemswhen using the graphical model for inference.

Marco Scutari University College London

Page 37: Graphical Models - and Protein Signalling Networks

Parameter Learning

Maximum a Posteriori Bayesian Estimation

Bayesian posterior estimates are the sensible choice for parameterestimation according to Koller’s & Friedman’s tome on graphical models.Choices for the priors are limited (for computational reasons) toconjugate distributions, namely:

• the Dirichlet for discrete models, i.e.

Dir(αk |ΠXi=π)

data−→ Dir(αk |ΠXi=π + nk |ΠXi

=π)

meaning that pk |ΠXi=π = αk |ΠXi

=π/∑π αk |ΠXi

=π.

• the Inverse Wishart for Gaussian models, i.e.

IW (Ψ,m)data−→ IW (Ψ + nΣ,m+ n).

In both cases (when a non-informative prior is used) the only freeparameter is the equivalent or imaginary sample size, which gives therelative weight of the prior compared to the observed sample.

Marco Scutari University College London

Page 38: Graphical Models - and Protein Signalling Networks

Model Averaging

Marco Scutari University College London

Page 39: Graphical Models - and Protein Signalling Networks

Model Averaging

The Big Three: Frequentist, Bayesian and Hybrid

The results of both structure and parameter learning are noisy in mostreal-world settings, due to limitations in the data and in our knowledge ofthe processes that control them. Since parameters are learned conditionalon the results of structure learning, using model averaging to obtain astable network structure from the data is an essential step in graphicalmodelling.

• frequentist: generating network structures using bootstrap andmodel averaging (aka bagging).

• Bayesian: generating network structures from the posterior P(G |D)using exhaustive enumeration or Markov Chain Mote Carloapproximations.

• hybrid: generating network structures again using bootstrap, butweighting them with their posterior probabilities when performingmodel averaging.

Marco Scutari University College London

Page 40: Graphical Models - and Protein Signalling Networks

Model Averaging

A Frequentist Approach: Friedman’s Confidence

Friedman et al. proposed an approach to model validation basedon bootstrap resampling and model averaging:

1. For b = 1, 2, . . . ,m:

1.1 sample a new data set X∗b from the original data X usingeither parametric or nonparametric bootstrap;

1.2 learn the structure of the graphical model Gb = (V, Eb) fromX∗b .

2. Estimate the confidence that each possible edge ei is presentin the true network structure G0 = (V, E0) as

pi = P(ei) =1

m

m∑b=1

1l{ei∈Eb},

where 1l{ei∈Eb} is equal to 1 if ei ∈ Eb and 0 otherwise.

Marco Scutari University College London

Page 41: Graphical Models - and Protein Signalling Networks

Model Averaging

A Frequentist Approach: Friedman’s Confidence

Marco Scutari University College London

Page 42: Graphical Models - and Protein Signalling Networks

Model Averaging

Identifying Significant Edges

• The confidence values p = {pi} do not sum to one and aredependent on one another in a nontrivial way; the value of theconfidence threshold (i.e. the minimum confidence for an edge to beaccepted as an edge of G0) is an unknown function of both the dataand the structure learning algorithm.

• The ideal/asymptotic configuration p of confidence values would be

pi =

{1 if ei ∈ E0

0 otherwise,

i.e. all the networks Gb have exactly the same structure.

• Therefore, identifying the configuration p “closest” to p provides aprincipled way of identifying significant edges and the confidencethreshold.

Marco Scutari University College London

Page 43: Graphical Models - and Protein Signalling Networks

Model Averaging

The Confidence Threshold

Consider the order statistics p(·) and p(·) and the cumulativedistribution functions (CDFs) of their elements:

Fp(·)(x) =1

k

k∑i=1

1l{p(i)<x}

and

Fp(·)(x; t) =

0 if x ∈ (−∞, 0)

t if x ∈ [0, 1)

1 if x ∈ [1,+∞)

.

t corresponds to the fraction of elements of p(·) equal to zero andis a measure of the fraction of non-significant edges, and providesa threshold for separating the elements of p(·):

e(i) ∈ E0 ⇐⇒ p(i) > F−1p(·)(t).

Marco Scutari University College London

Page 44: Graphical Models - and Protein Signalling Networks

Model Averaging

The CDFs Fp(·)(x) and Fp(·)(x; t)

0.0 0.4 0.8

0.0

0.2

0.4

0.6

0.8

1.0

0.0 0.4 0.8

0.0

0.2

0.4

0.6

0.8

1.0

0.0 0.4 0.8

0.0

0.2

0.4

0.6

0.8

1.0

One possible estimate of t is the value t that minimises somedistance between Fp(·)(x) and Fp(·)(x; t); an intuitive choice isusing the L1 norm of their difference (i.e. the shaded area in thepicture on the right).

Marco Scutari University College London

Page 45: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling

Networks

Marco Scutari University College London

Page 46: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Source

What follows reproduces (to the best of my ability, and KarenSachs’ recollections about the implementation details that did notend up in the Methods section) the statistical analysis in thefollowing paper:

DOI: 10.1126/science.1105809, 523 (2005);308Science, et al.Karen Sachs

Causal Protein-Signaling Networks Derived fromMultiparameter Single-Cell Data

That’s a landmark paper in applying Bayesian Networks because:

• it highlights the use of observational vs interventional data;

• results are validated using existing literature.

Marco Scutari University College London

Page 47: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

An Overview of the Data

The data consist in the simultaneous measurements of 11phosphorylated proteins and phospholypids derived from thousandsof individual primary immune system cells:

• 1800 data subject only to general stimolatory cues, so thatthe protein signalling paths are active;

• 600 data with with specific stimolatory/inhibitory cues foreach of the following 4 proteins: pmek, PIP2, pakts473, PKA;

• 1200 data with specific cues for PKA.

Overall, the data set contains 5400 observations with no missingvalue.

Marco Scutari University College London

Page 48: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Network Reconstructed from Literature

P38

p44.42

pakts473

PIP2

PIP3

pjnk

PKA

PKC plcg

pmek

praf

Marco Scutari University College London

Page 49: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Using Only Observational Data

As a first, exploratory analysis, we can try to learn a network fromthe data that were subject only to general stimolatory cues. Sincethese cues only ensure the pathways are active, but do not tamperwith them in any way, such data are observational (as opposed tointerventional).

> library(bnlearn)

> hc(sachs, score = "bge", iss = 5)

Classic algorithms in literature are not designed to handleinterventional data, but work out-of-the box with observationalones.

Marco Scutari University College London

Page 50: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Network Reconstructed from the Observational Data

praf

pmek

plcg

PIP2

PIP3

p44.42

pakts473

PKA

PKC

P38

pjnk

Arc highlighted in red are also present in the network reconstructed from literature.

Marco Scutari University College London

Page 51: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Expression Data are not Symmetric

expression levels

dens

ity

0 200 400 600 800

PIP2

0 200 400 600 800

PIP30 200 400 600 800

pmek

0 200 400 600 800

P38

Therefore, assuming a Gaussian distribution is problematic.

Marco Scutari University College London

Page 52: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Expression Data are not Linked by Linear Relationships

PKC

PK

A

010

0020

0030

0040

00

0 20 40 60 80 100

●● ●

● ●

●●

●●

●●

●●

●●

● ●

●●

●●●

●●

●●

●●●

●●

●●●

●●

● ●● ●

●●

●●

●●

●●

● ●●● ● ●

●●●

●●

●●●

● ●

● ● ●

●● ●

●●

●●

●●

● ●

●●

●●●●

● ●

●●

●●

●●

●●

●●

● ●●

● ●

●●

● ● ●

● ●●

●●

● ●●

●●

●●

● ●

●●

● ●●

●● ●●●

●●

●●

●●

● ●

●●●

●●

● ●

● ●●

●●

● ●

●●

●●

●●

●●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●● ●

●●

●●

●●

● ●

●●

● ●●

●●

●●

●●

●●●●●

●●

●●

● ●

●●

● ●

●●

●●

●●

● ●

●●

●●

●●

●●

● ●

●● ●

● ●

●●●

●●

● ●

●●

●●

●●

●●

● ●●

● ●

●● ●

● ●●●

●●

● ●

●●

●●

●●

● ●

●●●

●●

●●

●●

●●

●●

●●

●●

●●● ●●

●●

●●

●●

●●

●●

●●●●

●● ●

● ●●

●● ●●● ●●●

●●

●●

●●●

●●

● ●●

●●

●●

●●

●●

●●●● ●●

●●

●●

●●

●●●

●●

● ●

●●

● ●

●● ●

●● ●

●●●

●●● ●●

●●

●●

●●

●●

●●

●●

● ●●

●●

●●

●● ●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●● ●

Therefore, tests for correlation are biased and have extremely low power.

Marco Scutari University College London

Page 53: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Discretize!

Since we cannot use Gaussian Bayesian networks, we can discretizethem instead. Hartemink’s method is designed to preserve as muchas possible all pairwise dependencies, as opposed to marginaldiscretization methods.

> dsachs = discretize(sachs, method = "hartemink",

+ breaks = 3, ibreaks = 60,

+ idisc = "quantile")

Data are first marginalised in 60 intervals, which are subsequentlycollapsed while reducing the mutual information between thevariables as little as possible. The process stops when each variablehas 3 levels (i.e. low, average and high expression).

Marco Scutari University College London

Page 54: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Network Reconstructed from the Discretized Data

praf

pmek

plcg

PIP2

PIP3

p44.42

pakts473

PKA

PKC

P38

pjnk

Two more arcs are correctly identified, but most are still missing.Marco Scutari University College London

Page 55: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Considering Interventional Data

It is apparent from the previous networks that most signalling pathsare not statistically recognisable unless we inhibit or stimulate theexpression of at least some of the proteins in the network.Therefore, we include the interventional data in the analysis.

> INT = sapply(1:11, function(x)

+ { which(isachs$INT == x) })

> names(INT) = names(isachs)[1:11]

> hc(isachs[, 1:11], score = "mbde",

+ exp = INT, iss = 5)

Since the standard BDe score does not take interventions intoaccount, we use a modified BDe score that disregards any causalinfluence for the proteins that have been inhibited or stimulated.

Marco Scutari University College London

Page 56: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Network Reconstructed from the Interventional Data

P38

p44.42

pakts473

PIP2

PIP3pjnk

PKA

PKC

plcg

pmek

praf

More arcs are included, but there are many false positives.

Marco Scutari University College London

Page 57: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Removing Noisy Arcs with Model Averaging

Two simple steps can be taken to remove noisy arcs:

• average multiple networks learned using different startingpoints for the structure learning algorithm;

• use TABU search instead of Hill-Climbing.

> start = random.graph(nodes = nodes,

+ method = "melancon", num = 500, burn.in = 10^5,

+ every = 100)

> netlist = lapply(start, function(net) {

+ tabu(isachs[, 1:11], score = "mbde", exp = INT,

+ iss = 10, start = net, tabu = 50) })

> arcs = custom.strength(netlist, nodes = nodes)

A similar approach was chosen as the best performing in Sachs etal. [27], with minor differences in results.

Marco Scutari University College London

Page 58: Graphical Models - and Protein Signalling Networks

Causal Protein-Signalling Networks

Interventional Data with Model Averaging

P38

p44.42

pakts473

PIP2

PIP3

pjnk

PKA

PKC plcg

pmek

praf

All the arcs supported by literature are present in the network.

Marco Scutari University College London

Page 59: Graphical Models - and Protein Signalling Networks

Conclusions

Marco Scutari University College London

Page 60: Graphical Models - and Protein Signalling Networks

Conclusions

Conclusions

• Graphical models combine many ideas from different fields toallow an intuitive manipulation of high-dimensional problemsand the corresponding multivariate probability distributions.

• A sensible use of Bayesian and shrinkage techniques instructure and parameter learning allows a great deal offlexibility and results in good models.

• Properly validated graphical models can capture thedependence structure of the data even with very small samplesizes.

• The use of interventions and model averaging improves thequality of the learned networks dramatically.

Marco Scutari University College London

Page 61: Graphical Models - and Protein Signalling Networks

Thanks!

Marco Scutari University College London

Page 62: Graphical Models - and Protein Signalling Networks

References

Marco Scutari University College London

Page 63: Graphical Models - and Protein Signalling Networks

References

References I

R. R. Bouckaert.Bayesian Belief Networks: from Construction to Inference.PhD thesis, Utrecht University, The Netherlands, 1995.

D. M. Chickering.Optimal Structure Identification with Greedy Search.Journal of Machine Learning Resesearch, 3:507–554, 2002.

D. I. Edwards.Introduction to Graphical Modelling.Springer, 2nd edition, 2000.

J. Friedman, T. Hastie, and R. Tibshirani.Sparse Inverse Covariance Estimation With the Graphical Lasso.Biostatistics, 9:432–441, 2007.

N. Friedman, M. Goldszmidt, and A. Wyner.Data Analysis with Bayesian Networks: A Bootstrap Approach.In Proceedings of the 15th Annual Conference on Uncertainty in ArtificialIntelligence (UAI-99), pages 206 – 215. Morgan Kaufmann, 1999.

Marco Scutari University College London

Page 64: Graphical Models - and Protein Signalling Networks

References

References II

N. Friedman and D. Koller.Being Bayesian about Bayesian Network Structure: A Bayesian Approach toStructure Discovery in Bayesian Networks.Machine Learning, 50(1–2):95–126, 2003.

N. Friedman, D. Pe’er, and I. Nachman.Learning Bayesian Network Structure from Massive Datasets: The “SparseCandidate” Algorithm.In Proceedings of 15th Conference on Uncertainty in Artificial Intelligence (UAI),pages 206–221. Morgan Kaufmann, 1999.

D. Geiger and D. Heckerman.Learning Gaussian Networks.Technical report, Microsoft Research, Redmond, Washington, 1994.Available as Technical Report MSR-TR-94-10.

F. Harary and E. M. Palmer.Graphical Enumeration.Academic Press, 1973.

Marco Scutari University College London

Page 65: Graphical Models - and Protein Signalling Networks

References

References III

T. Hastie, R. Tibshirani, and J. Friedman.The Elements of Statistical Learning: Data Mining, Inference, and Prediction.Springer, 2nd edition, 2009.

J. Hausser and K. Strimmer.Entropy Inference and the James-Stein Estimator, with Application to NonlinearGene Association Networks.Journal of Machine Learning Resesearch, 10:1469–1484, 2009.

D. Heckerman, D. Geiger, and D. M. Chickering.Learning Bayesian Networks: The Combination of Knowledge and StatisticalData.Machine Learning, 20(3):197–243, September 1995.Available as Technical Report MSR-TR-94-09.

C. J. Hoggart, J. C. Whittaker, M. De Iorio, and D. J. Balding.Simultaneous Analysis of All SNPs in Genome-Wide and Re-SequencingAssociation Studies.PLoS Genetics, 4(7), 2008.

Marco Scutari University College London

Page 66: Graphical Models - and Protein Signalling Networks

References

References IV

J. S. Ide and F. G. Cozman.Random Generation of Bayesian Networks.In Proceedings of the 16th Brazilian Symposium on Artificial Intelligence, pages366–375. Springer-Verlag, 2002.

S. Imoto, S. Y. Kim, H. Shimodaira, S. Aburatani, K. Tashiro, S. Kuhara, andS. Miyano.Bootstrap Analysis of Gene Networks Based on Bayesian Networks andNonparametric Regression.Genome Informatics, 13:369–370, 2002.

W. James and C. Stein.Estimation with Quadratic Loss.In J. Neyman, editor, Proceedings of the 4th Berkeley Symposium onMathematical Statistics and Probability, pages 361–379, 1961.

D. Koller and N. Friedman.Probabilistic Graphical Models: Principles and Techniques.MIT Press, 2009.

Marco Scutari University College London

Page 67: Graphical Models - and Protein Signalling Networks

References

References V

K. Korb and A. Nicholson.Bayesian Artificial Intelligence.Chapman and Hall, 2nd edition, 2009.

S. Kullback.Information Theory and Statistics.Dover Publications, 1968.

O. Ledoit and M. Wolf.Improved Estimation of the Covariance Matrix of Stock Returns with anApplication to Portfolio Selection.Journal of Empirical Finance, 10:603–621, 2003.

P. Legendre.Comparison of Permutation Methods for the Partial Correlation and PartialMantel Tests.Journal of Statistical Computation and Simulation, 67:37–73, 2000.

Marco Scutari University College London

Page 68: Graphical Models - and Protein Signalling Networks

References

References VI

D. Margaritis.Learning Bayesian Network Model Structure from Data.PhD thesis, School of Computer Science, Carnegie-Mellon University, Pittsburgh,PA, May 2003.Available as Technical Report CMU-CS-03-153.

G. Melancon, I. Dutour, and M. Bousquet-Melou.Random Generation of DAGs for Graph Drawing.Technical Report INS-R0005, Centre for Mathematics and Computer Sciences,Amsterdam, 2000.

J. Pearl.Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference.Morgan Kaufmann, 1988.

F. Pesarin and L. Salmaso.Permutation Tests for Complex Data: Theory, Applications and Software.Wiley, 2010.

S. J. Russell and P. Norvig.Artificial Intelligence: A Modern Approach.Prentice Hall, 3rd edition, 2009.

Marco Scutari University College London

Page 69: Graphical Models - and Protein Signalling Networks

References

References VII

K. Sachs, O. Perez, D. Pe’er, D. A. Lauffenburger, and G. P. Nolan.Causal Protein-Signaling Networks Derived from Multiparameter Single-CellData.Science, 308(5721):523–529, 2005.

J. Schafer and K. Strimmer.A Shrinkage Approach to Large-Scale Covariance Matrix Estimation andImplications for Functional Genomics.Statistical Applications in Genetics and Molecular Biology, 4:32, 2005.

G. E. Schwarz.Estimating the Dimension of a Model.Annals of Statistics, 6(2):461 – 464, 1978.

M. Scutari and K. Strimmer.Introduction to Graphical Modelling.In D. J. Balding, M. Stumpf, and M. Girolami, editors, Handbook of StatisticalSystems Biology. Wiley, 2011.In print.

Marco Scutari University College London

Page 70: Graphical Models - and Protein Signalling Networks

References

References VIII

P. Spirtes, C. Glymour, and R. Scheines.Causation, Prediction, and Search.MIT Press, 2000.

C. Stein.Inadmissibility of the Usual Estimator for the Mean of a MultivariateDistribution.In J. Neyman, editor, Proceedings of the 3rd Berkeley Symposium onMathematical Statistics and Probability, pages 197–206, 1956.

I. Tsamardinos, C. F. Aliferis, and A. Statnikov.Algorithms for Large Scale Markov Blanket Discovery.In Proceedings of the 16th International Florida Artificial Intelligence ResearchSociety Conference, pages 376–381. AAAI Press, 2003.

I. Tsamardinos, L. E. Brown, and C. F. Aliferis.The Max-Min Hill-Climbing Bayesian Network Structure Learning Algorithm.Machine Learning, 65(1):31–78, 2006.

T. S. Verma and J. Pearl.Equivalence and Synthesis of Causal Models.Uncertainty in Artificial Intelligence, 6:255–268, 1991.

Marco Scutari University College London

Page 71: Graphical Models - and Protein Signalling Networks

References

References IX

J. Whittaker.Graphical Models in Applied Multivariate Statistics.Wiley, 1990.

S. Yaramakala and D. Margaritis.Speculative Markov Blanket Discovery for Optimal Feature Selection.In Proceedings of the 5th IEEE International Conference on Data Mining, pages809–812. IEEE Computer Society, 2005.

Marco Scutari University College London