Weighted Lasso for Network inference
-
Upload
laboratoire-statistique-et-genome -
Category
Documents
-
view
795 -
download
2
Transcript of Weighted Lasso for Network inference
![Page 1: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/1.jpg)
Inferring Gene Regulatory Networks fromTime-Course Gene Expression Data
Camille Charbonnier and Julien Chiquet and Christophe Ambroise
Laboratoire Statistique et Genome,La genopole - Universite d’Evry
JOBIM – 7-9 septembre 2010
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 1
![Page 2: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/2.jpg)
Problem
t0t1
tn
≈ 10s microarrays over time
≈ 1000s probes (“genes”)
Inference
Which interactions?
The main statistical issue is the high dimensional setting.
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 2
![Page 3: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/3.jpg)
Handling the scarcity of the dataBy introducing some prior
Priors should be biologically grounded
1. few genes effectively interact (sparsity),2. networks are organized (latent clustering),
G1
G2
G3
G4 G5
G6
G7
G8
G9
G10
G11
G12
G13
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 3
![Page 4: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/4.jpg)
Handling the scarcity of the dataBy introducing some prior
Priors should be biologically grounded
1. few genes effectively interact (sparsity),2. networks are organized (latent clustering),
G1
G2
G3
G4 G5
G6
G7
G8
G9
G10
G11
G12
G13
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 3
![Page 5: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/5.jpg)
Handling the scarcity of the dataBy introducing some prior
Priors should be biologically grounded
1. few genes effectively interact (sparsity),2. networks are organized (latent clustering),
A1
A2
A3
A4 A
B1
B2
B3
B4
B5
B
C1
C
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 3
![Page 6: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/6.jpg)
Outline
Statistical modelsGaussian Graphical Model for Time-course dataStructured Regularization
Algorithms and methodsOverall viewModel selection
Numerical experimentsInference methodsPerformance on simulated dataE. coli S.O.S DNA repair network
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 4
![Page 7: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/7.jpg)
Gaussian Graphical Model for Time-course data
Collecting gene expression
1. Follow-up of one single experiment/individual;2. Close enough time-points to ensure
I dependency between consecutive measurements;I homogeneity of the Markov process.
X1
X2X3
X4
X5
stands for
X1t
X2t X2
t+1
X3t+1
X4t+1G
X5t+1
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 5
![Page 8: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/8.jpg)
Gaussian Graphical Model for Time-course data
Collecting gene expression
1. Follow-up of one single experiment/individual;2. Close enough time-points to ensure
I dependency between consecutive measurements;I homogeneity of the Markov process.
X1
X2X3
X4
X5
stands for
X1t
X2t X2
t+1
X3t+1
X4t+1G
X5t+1
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 5
![Page 9: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/9.jpg)
Gaussian Graphical Model for Time-course data
Collecting gene expression
1. Follow-up of one single experiment/individual;2. Close enough time-points to ensure
I dependency between consecutive measurements;I homogeneity of the Markov process.
X1t X1
2. . . X1
n
X21 X2
2. . . X2
n
X31 X3
2. . . X3
n
X41 X4
2. . . X4
n
G G GX5
1 X52
. . . X5n
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 5
![Page 10: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/10.jpg)
Gaussian Graphical Model for Time-Course data
AssumptionA microarray can be represented as a multivariate Gaussianvector X = (X(1), . . . , X(p)) ∈ Rp, following a first order vectorautoregressive process V AR(1):
Xt = ΘXt−1 + b + εt, t ∈ [1, n]
where we are looking for Θ = (θij)i,j∈P .
Graphical interpretationconditional dependency between Xt−1(j) and Xt(i)
lnon null partial correlation between Xt−1(j) and Xt(i)
lθij 6= 0
if and only ifi
j
k
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 6
![Page 11: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/11.jpg)
Gaussian Graphical Model for Time-Course data
AssumptionA microarray can be represented as a multivariate Gaussianvector X = (X(1), . . . , X(p)) ∈ Rp, following a first order vectorautoregressive process V AR(1):
Xt = ΘXt−1 + b + εt, t ∈ [1, n]
where we are looking for Θ = (θij)i,j∈P .
Graphical interpretationconditional dependency between Xt−1(j) and Xt(i)
lnon null partial correlation between Xt−1(j) and Xt(i)
lθij 6= 0
if and only ifi
j
?
k
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 6
![Page 12: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/12.jpg)
Gaussian Graphical Model for Time-Course data
AssumptionA microarray can be represented as a multivariate Gaussianvector X = (X(1), . . . , X(p)) ∈ Rp, following a first order vectorautoregressive process V AR(1):
Xt = ΘXt−1 + b + εt, t ∈ [1, n]
where we are looking for Θ = (θij)i,j∈P .
Graphical interpretationconditional dependency between Xt−1(j) and Xt(i)
lnon null partial correlation between Xt−1(j) and Xt(i)
lθij 6= 0
if and only ifi
j
?
k
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 6
![Page 13: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/13.jpg)
Gaussian Graphical Model for Time-course data
LetI X be the n× p matrix whose kth row is Xk,I S = n−1Xᵀ\nX\n be the within time covariance matrix,
I V = n−1Xᵀ\nX\0 be the across time covariance matrix.
The log-likelihood
Ltime(Θ;S,V) = n Trace (VΘ)− n
2Trace (ΘᵀSΘ) + c.
Maximum Likelihood Estimator ΘMLE = S−1V
I not defined for n < p;I even if n > p, requires multiple testing.
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 7
![Page 14: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/14.jpg)
Structured regularization`1 penalized log-likelihood
Charbonnier, Chiquet, Ambroise, SAGMB 2010
Θλ = arg maxΘ
Ltime(Θ;S,V)− λ ·∑i,j∈P
PZij |Θij |
where λ is an overall tuning parameter and PZ is a(non-symmetric) matrix of weights depending on the underlyingclustering Z.
It performs1. regularization (needed when n� p),2. selection (specificity of the `1-norm),3. cluster-driven inference (penalty adapted to Z).
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 8
![Page 15: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/15.jpg)
Structured regularization“Bayesian” interpretation of `1 regularization
Laplacian prior on Θ depends on the clustering Z
P(Θ|Z) ∝∏i,j
exp{−λ ·PZ
ij · |Θij |}.
PZ summarizes prior information on the position of edges
-1.0 -0.5 0.0 0.5 1.0
0.2
0.4
0.6
0.8
1.0
Θij
pri
ord
istr
ibu
tion
λPZij = 1
λPZij = 2
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 9
![Page 16: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/16.jpg)
How to come up with a latent clustering?
Biological expertise
I Build Z from prior biological informationI transcription factors vs. regulatees,I number of potential binding sites,I KEGG pathways, . . .
I Build the weight matrix from Z.
Inference: Erdos-Renyi Mixture for Networks(Daudin et al., 2008)
I Spread the nodes into Q classes;I Connexion probabilities depends upon node classes:
P(i→ j|i ∈ class q, j ∈ class `) = πq`.
I Build PZ ∝ 1− πq`.
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 10
![Page 17: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/17.jpg)
Algorithm
SIMoNESuppose you want to recover a clustered network:
Target Adjacency Matrix
Graph
Target Network
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 11
![Page 18: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/18.jpg)
Algorithm
SIMoNEStart with microarray data
Data
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 11
![Page 19: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/19.jpg)
Algorithm
SIMoNE
DataAdjacency Matrix
corresponding to G?
Inference without prior
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 11
![Page 20: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/20.jpg)
Algorithm
SIMoNE
DataAdjacency Matrix
corresponding to G?
Inference without prior
πZ
Connectivity matrix
Mixer
Penalty matrix PZ
Decreasing transformation
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 11
![Page 21: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/21.jpg)
Algorithm
SIMoNE
DataAdjacency Matrix
corresponding to G?
Inference without prior
πZ
Connectivity matrix
Mixer
Penalty matrix PZ
Decreasing transformation
Adjacency Matrixcorresponding to G?Z
+
Inference with clustered prior
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 11
![Page 22: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/22.jpg)
Tuning the penalty parameterBIC / AIC
Degrees of freedom of the Lasso (Zou et al. 2008)
df(βλ) =∑k
1(βλk 6= 0)
Straightforward extensions to the graphical framework
BIC(λ) = L(Θλ;X)− df(Θλ)log n
2
AIC(λ) = L(Θλ;X)− df(Θλ)
I Rely on asymptotic approximations,but still relevant on simulated small samples.
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 12
![Page 23: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/23.jpg)
Inference methods
1. Lasso (Tibshirani)
2. Adaptive Lasso (Zou et al.)Weights inversely proportional to an initial Lasso estimate.
3. KnwClWeights structured according to true clustering.
4. InfClWeights structured according to inferred clustering.
5. Renet-VAR (Shimamura et al.)Edge estimation based on a recursive elastic net.
6. G1DBN (Lebre et al.)Edge estimation based on dynamic Bayesian networks followed by statistical
testing of edges.
7. Shrinkage (Opgen-Rhein et al.)Edge estimation based on shrinkage followed by multiple testing local false
discovery rate correction.
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 13
![Page 24: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/24.jpg)
Simulations: time-course data with star-pattern
Simulation settings
I 2 classes, hubs and leaves, with proportions α = (0.1, 0.9),I K = 2p edges, among which:
I 85% from hubs to leaves,I 15% between hubs.
p genes n arrays samples20 40 50020 20 50020 10 500
100 100 200100 50 200800 20 100
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 14
![Page 25: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/25.jpg)
Simulations: time-course data with star-pattern
0.00.51.01.5
a) p=20, n=40
0.00.51.01.5
b) p=20, n=20
0.00.51.01.5
c) p=20, n=10
0.00.51.01.5
d) p=100, n=100
0.00.51.01.5
e) p=100, n=50
LASSO−A
IC
LASSO−B
IC
Adapt
ive−A
IC
Adapt
ive−B
IC
KnwCl−A
IC
KnwCl−B
IC
InfC
l−AIC
InfC
l−BIC
Renet
−VAR
G1DBDN
Shrink
age
0.00.51.01.5
f) p=800, n=20
PrecisionRecall
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 15
![Page 26: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/26.jpg)
Reasonnable computing time
2.5 3.0 3.5 4.0 4.5 5.0
02
46
810
CPU time (log scale)
log of p/n
log
of C
PU ti
me
(sec
)
1 se
c.10
sec
.1
min
.10
min
.1
hr.
4 hr
.
G1DBNRenet−VARInfCl
Figure: Computing times on the log-log scale for Renet-VAR, G1DBNand InfCl (including inference of classes). Intel Dual Core 3.40 GHzprocessor.
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 16
![Page 27: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/27.jpg)
E. coli S.O.S DNA repair network
E.coli S.O.S data
SOS.pdf
I 8 major genes involved inthe S.O.S. response
I 50 time points
figures/fig_sos_pathway.pdf
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 17
![Page 28: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/28.jpg)
E. coli S.O.S DNA repair networkPrecision and Recall rates
0.0
0.5
1.0
1.5Exp. 1
0.0
0.5
1.0
1.5Exp. 2
0.0
0.5
1.0
1.5Exp. 3
Lass
o
Adapt
ive
G1DBN
Renet
KnownC
l
Infer
Cl0.0
0.5
1.0
1.5Exp. 4
PrecisionRecall
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 18
![Page 29: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/29.jpg)
E. coli S.O.S DNA repair networkInferred networks
Lasso Adaptive-Lasso
uvrD
lexA
umuDC
recA
uvrA
uvrY
ruvA
polB
uvrDlexA
umuDC
recA
uvrA
uvrY
ruvA
polB
Known Classification Inferred Classification
uvrD
lexA
umuDC
recA
uvrAuvrY
ruvA
polB
uvrDlexA
umuDC
recA
uvrA
uvrY
ruvA
polB
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 19
![Page 30: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/30.jpg)
Conclusions
To sum-up
I cluster-driven inference of gene regulatory networks fromtime-course data,
I expert-based or inferred latent structure,I embedded in the SIMoNe R package along with similar
algorithms dealing with steady-state or multitask data.
Perspectives
I inference of truely dynamic networks,I use of additive biological information to refine the
inference,I comparison of inferred networks.
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 20
![Page 31: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/31.jpg)
Publications
Ambroise, Chiquet, Matias, 2009.Inferring sparse Gaussian graphical models with latent structureElectronic Journal of Statistics, 3, 205-238.
Chiquet, Smith, Grasseau, Matias, Ambroise, 2009.SIMoNe: Statistical Inference for MOdular NEtworks Bioinformatics,25(3), 417-418.
Charbonnier, Chiquet, Ambroise, 2010.Weighted-Lasso for Structured Network Inference from TimeCourse Data., SAGMB, 9.
Chiquet, Grandvalet, Ambroise, 2010. Statistics and Computing.Inferring multiple Gaussian graphical models.
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 21
![Page 32: Weighted Lasso for Network inference](https://reader034.fdocuments.in/reader034/viewer/2022051314/559ae8621a28ab784e8b45ea/html5/thumbnails/32.jpg)
Publications
Ambroise, Chiquet, Matias, 2009.Inferring sparse Gaussian graphical models with latent structureElectronic Journal of Statistics, 3, 205-238.
Chiquet, Smith, Grasseau, Matias, Ambroise, 2009.SIMoNe: Statistical Inference for MOdular NEtworks Bioinformatics,25(3), 417-418.
Charbonnier, Chiquet, Ambroise, 2010.Weighted-Lasso for Structured Network Inference from TimeCourse Data., SAGMB, 9.
Chiquet, Grandvalet, Ambroise, 2010. Statistics and Computing.Inferring multiple Gaussian graphical models.
Working paper: Chiquet, Charbonnier, Ambroise, Grasseau.SIMoNe: An R package for inferring Gausssian networks withlatent structure, Journal of Statistical Softwares.
Working paper: Chiquet, Grandvalet, Ambroise, Jeanmougin.Biological analysis of breast cancer by multitasks learning.
Camille Charbonnier and Julien Chiquet and Christophe Ambroise 21