A Discriminative Topic Model using Document Network...

65
A Discriminative Topic Model using Document Network Structure Weiwei Yang 1 , Jordan Boyd-Graber 2 , and Philip Resnik 1 1 University of Maryland College Park and 2 University of Colorado Boulder August 8, 2016 1/22

Transcript of A Discriminative Topic Model using Document Network...

Page 1: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

A Discriminative Topic Modelusing Document Network Structure

Weiwei Yang1, Jordan Boyd-Graber2, and Philip Resnik1

1University of Maryland College Park and 2University of Colorado Boulder

August 8, 2016

1/22

Page 2: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Paper and Slides

Paper

http://ter.ps/bv5

Slides

http://ter.ps/bv7

2/22

Page 3: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Documents are Linked

A Discriminative Topic Model using Document Network Structure

Weiwei YangComputer Science

University of MarylandCollege Park, MD

[email protected]

Jordan Boyd-GraberComputer Science

University of ColoradoBoulder, CO

[email protected]

Philip ResnikLinguistics and UMIACSUniversity of Maryland

College Park, [email protected]

Abstract

Document collections often have links be-tween documents—citations, hyperlinks,or revisions—and which links are added isoften based on topical similarity. To modelthese intuitions, we introduce a new topicmodel for documents situated within a net-work structure, integrating latent blocksof documents with a max-margin learningcriterion for link prediction using topic-and word-level features. Experiments ona scientific paper dataset and collectionof webpages show that, by more robustlyexploiting the rich link structure within adocument network, our model improveslink prediction, topic quality, and blockdistributions.

1 Introduction

Documents often appear within a network struc-ture: social media mentions, retweets, and fol-lower relationships; Web pages by hyperlinks; sci-entific papers by citations. Network structure in-teracts with the topics in the text, in that docu-ments linked in a network are more likely to havesimilar topic distributions. For instance, a cita-tion link between two papers suggests that theyare about a similar field, and a mentioning linkbetween two social media users often indicatescommon interests. Conversely, documents’ sim-ilar topic distributions can suggest links betweenthem. For example, topic model (Blei et al., 2003,LDA) and block detection papers (Holland et al.,1983) are relevant to our research, so we cite them.Similarly, if a social media user A finds anotheruser B with shared interests, then A is more likelyto follow B.

Our approach is part of a natural progressionof network modeling in which models integrate

more information in more sophisticated ways.Some past methods only consider the network it-self (Kim and Leskovec, 2012; Liben-Nowell andKleinberg, 2007), which loses the rich informationin text. In other cases, methods take both links andtext into account (Chaturvedi et al., 2012), but theyare modeled separately, not jointly, limiting themodel’s ability to capture interactions between thetwo. The relational topic model (Chang and Blei,2010, RTM) goes further, jointly modeling topicsand links, but it considers only pairwise documentrelationships, failing to capture network structureat the level of groups or blocks of documents.

We propose a new joint model that makes fulleruse of the rich link structure within a documentnetwork. Specifically, our model embeds theweighted stochastic block model (Aicher et al.,2014, WSBM) to identify blocks in which docu-ments are densely connected. WSBM basically cat-egorizes each item in a network probabilisticallyas belonging to one of L blocks, by reviewingits connections with each block. Our model canbe viewed as a principled probabilistic extensionof Yang et al. (2015), who identify blocks in a doc-ument network deterministically as strongly con-nected components (SCC). Like them, we assign adistinct Dirichlet prior to each block to capture itstopical commonalities. Jointly, a linear regressionmodel with a discriminative, max-margin objec-tive function (Zhu et al., 2012; Zhu et al., 2014) istrained to reconstruct the links, taking into accountthe features of documents’ topic and word distri-butions (Nguyen et al., 2013), block assignments,and inter-block link rates.

We validate our approach on a scientific pa-per abstract dataset and collection of webpages,with citation links and hyperlinks respectively, topredict links among previously unseen documentsand from those new documents to training docu-ments. Embedding the WSBM in a network/topic

Our Paper

: Topic Model,

Document Network, Block Detection

Journal of Machine Learning Research 3 (2003) 993-1022 Submitted 2/02; Published 1/03

Latent Dirichlet Allocation

David M. Blei [email protected]

Computer Science DivisionUniversity of CaliforniaBerkeley, CA 94720, USA

Andrew Y. Ng [email protected]

Computer Science DepartmentStanford UniversityStanford, CA 94305, USA

Michael I. Jordan [email protected]

Computer Science Division and Department of StatisticsUniversity of CaliforniaBerkeley, CA 94720, USA

Editor: John Lafferty

Abstract

We describelatent Dirichlet allocation(LDA), a generative probabilistic model for collections ofdiscrete data such as text corpora. LDA is a three-level hierarchical Bayesian model, in which eachitem of a collection is modeled as a finite mixture over an underlying set of topics. Each topic is, inturn, modeled as an infinite mixture over an underlying set of topic probabilities. In the context oftext modeling, the topic probabilities provide an explicit representation of a document. We presentefficient approximate inference techniques based on variational methods and an EM algorithm forempirical Bayes parameter estimation. We report results in document modeling, text classification,and collaborative filtering, comparing to a mixture of unigrams model and the probabilistic LSImodel.

1. Introduction

In this paper we consider the problem of modeling text corpora and other collections of discretedata. The goal is to find short descriptions of the members of a collection that enable efficientprocessing of large collections while preserving the essential statistical relationships that are usefulfor basic tasks such as classification, novelty detection, summarization, and similarity and relevancejudgments.

Significant progress has been made on this problem by researchers in the field of informa-tion retrieval (IR) (Baeza-Yates and Ribeiro-Neto, 1999). The basic methodology proposed byIR researchers for text corpora—a methodology successfully deployed in modern Internet searchengines—reduces each document in the corpus to a vector of real numbers, each of which repre-sents ratios of counts. In the populartf-idf scheme (Salton and McGill, 1983), a basic vocabularyof “words” or “terms” is chosen, and, for each document in the corpus, a count is formed of thenumber of occurrences of each word. After suitable normalization, this term frequency count iscompared to an inverse document frequency count, which measures the number of occurrences of a

c©2003 David M. Blei, Andrew Y. Ng and Michael I. Jordan.

LDA [Blei et al., 2003]

: Topic Model

The Annals of Applied Statistics2010, Vol. 4, No. 1, 124–150DOI: 10.1214/09-AOAS309© Institute of Mathematical Statistics, 2010

HIERARCHICAL RELATIONAL MODELSFOR DOCUMENT NETWORKS

BY JONATHAN CHANG1 AND DAVID M. BLEI2

Facebook and Princeton University

We develop the relational topic model (RTM), a hierarchical model ofboth network structure and node attributes. We focus on document networks,where the attributes of each document are its words, that is, discrete obser-vations taken from a fixed vocabulary. For each pair of documents, the RTMmodels their link as a binary random variable that is conditioned on theircontents. The model can be used to summarize a network of documents, pre-dict links between them, and predict words within them. We derive efficientinference and estimation algorithms based on variational methods that takeadvantage of sparsity and scale with the number of links. We evaluate thepredictive performance of the RTM for large networks of scientific abstracts,web documents, and geographically tagged news.

1. Introduction. Network data, such as citation networks of documents, hy-perlinked networks of web pages, and social networks of friends, are pervasive inapplied statistics and machine learning. The statistical analysis of network data canprovide both useful predictive models and descriptive statistics. Predictive modelscan point social network members toward new friends, scientific papers towardrelevant citations, and web pages toward other related pages. Descriptive statisticscan uncover the hidden community structure underlying a network data set.

Recent research in this field has focused on latent variable models of link struc-ture, models that decompose a network according to hidden patterns of connectionsbetween its nodes [Kemp, Griffiths and Tenenbaum (2004); Hofman and Wiggins(2007); Airoldi et al. (2008)]. These models represent a significant departure fromstatistical models of networks, which explain network data in terms of observedsufficient statistics [Fienberg, Meyer and Wasserman (1985); Wasserman and Pat-tison (1996); Getoor et al. (2001); Newman (2002); Taskar et al. (2004)].

While powerful, current latent variable models account only for the structureof the network, ignoring additional attributes of the nodes that might be available.For example, a citation network of articles also contains text and abstracts of thedocuments, a linked set of web-pages also contains the text for those pages, andan on-line social network also contains profile descriptions and other information

Received January 2009; revised November 2009.1Work done while at Princeton University.2Supported by ONR 175-6343, NSF CAREER 0745520 and grants from Google and Microsoft.Key words and phrases. Mixed-membership models, variational methods, text analysis, network

models.

124

Relational Topic Model

[Chang and Blei, 2010]

:

Topic Model, Document Network

Learning Latent Block Structure in Weighted Networks

Christopher AicherDepartment of Applied Mathematics, University of Colorado, Boulder, CO, 80309

[email protected]

Abigail Z. JacobsDepartment of Computer Science, University of Colorado, Boulder, CO, 80309

Aaron ClausetDepartment of Computer Science, University of Colorado, Boulder, CO, 80309

BioFrontiers Institute, University of Colorado, Boulder, CO 80303

Santa Fe Institute, Santa Fe, NM 87501

Abstract

Community detection is an important task in network analysis, in which we aim to learn anetwork partition that groups together vertices with similar community-level connectivity pat-terns. By finding such groups of vertices with similar structural roles, we extract a compactrepresentation of the network’s large-scale structure, which can facilitate its scientific interpre-tation and the prediction of unknown or future interactions. Popular approaches, including thestochastic block model, assume edges are unweighted, which limits their utility by discarding po-tentially useful information. We introduce the weighted stochastic block model (WSBM), whichgeneralizes the stochastic block model to networks with edge weights drawn from any exponen-tial family distribution. This model learns from both the presence and weight of edges, allowingit to discover structure that would otherwise be hidden when weights are discarded or thresh-olded. We describe a Bayesian variational algorithm for efficiently approximating this model’sposterior distribution over latent block structures. We then evaluate the WSBM’s performanceon both edge-existence and edge-weight prediction tasks for a set of real-world weighted net-works. In all cases, the WSBM performs as well or better than the best alternatives on thesetasks. community detection, weighted relational data, block models, exponential family,variational Bayes.

1 Introduction

Networks are an increasingly important form of structured data consisting of interactions betweenpairs of individuals in large social and biological data sets. Unlike attribute data where each obser-vation is associated with an individual, network data is represented by graphs, where individuals arevertices and interactions are edges. Because vertices are pairwise related, network data violates tra-ditional assumptions of attribute data, such as independence. This intrinsic difference in structureprompts the development of new tools for handling network data.

In social and biological networks, vertices often play distinct structural roles in generating thenetwork’s large-scale structure. To identify such latent structural roles, we aim to identify a net-work partition that groups together vertices with similar group-level connectivity patterns. We callthese groups “communities,” and their inference produces a compact description of the large-scale

1

arX

iv:1

404.

0431

v2 [

stat

.ML

] 3

Jun

201

4

Weighted Stochastic Block Model

[Aicher et al., 2014]

: Block Detection

3/22

Page 4: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Documents are Linked

A Discriminative Topic Model using Document Network Structure

Weiwei YangComputer Science

University of MarylandCollege Park, MD

[email protected]

Jordan Boyd-GraberComputer Science

University of ColoradoBoulder, CO

[email protected]

Philip ResnikLinguistics and UMIACSUniversity of Maryland

College Park, [email protected]

Abstract

Document collections often have links be-tween documents—citations, hyperlinks,or revisions—and which links are added isoften based on topical similarity. To modelthese intuitions, we introduce a new topicmodel for documents situated within a net-work structure, integrating latent blocksof documents with a max-margin learningcriterion for link prediction using topic-and word-level features. Experiments ona scientific paper dataset and collectionof webpages show that, by more robustlyexploiting the rich link structure within adocument network, our model improveslink prediction, topic quality, and blockdistributions.

1 Introduction

Documents often appear within a network struc-ture: social media mentions, retweets, and fol-lower relationships; Web pages by hyperlinks; sci-entific papers by citations. Network structure in-teracts with the topics in the text, in that docu-ments linked in a network are more likely to havesimilar topic distributions. For instance, a cita-tion link between two papers suggests that theyare about a similar field, and a mentioning linkbetween two social media users often indicatescommon interests. Conversely, documents’ sim-ilar topic distributions can suggest links betweenthem. For example, topic model (Blei et al., 2003,LDA) and block detection papers (Holland et al.,1983) are relevant to our research, so we cite them.Similarly, if a social media user A finds anotheruser B with shared interests, then A is more likelyto follow B.

Our approach is part of a natural progressionof network modeling in which models integrate

more information in more sophisticated ways.Some past methods only consider the network it-self (Kim and Leskovec, 2012; Liben-Nowell andKleinberg, 2007), which loses the rich informationin text. In other cases, methods take both links andtext into account (Chaturvedi et al., 2012), but theyare modeled separately, not jointly, limiting themodel’s ability to capture interactions between thetwo. The relational topic model (Chang and Blei,2010, RTM) goes further, jointly modeling topicsand links, but it considers only pairwise documentrelationships, failing to capture network structureat the level of groups or blocks of documents.

We propose a new joint model that makes fulleruse of the rich link structure within a documentnetwork. Specifically, our model embeds theweighted stochastic block model (Aicher et al.,2014, WSBM) to identify blocks in which docu-ments are densely connected. WSBM basically cat-egorizes each item in a network probabilisticallyas belonging to one of L blocks, by reviewingits connections with each block. Our model canbe viewed as a principled probabilistic extensionof Yang et al. (2015), who identify blocks in a doc-ument network deterministically as strongly con-nected components (SCC). Like them, we assign adistinct Dirichlet prior to each block to capture itstopical commonalities. Jointly, a linear regressionmodel with a discriminative, max-margin objec-tive function (Zhu et al., 2012; Zhu et al., 2014) istrained to reconstruct the links, taking into accountthe features of documents’ topic and word distri-butions (Nguyen et al., 2013), block assignments,and inter-block link rates.

We validate our approach on a scientific pa-per abstract dataset and collection of webpages,with citation links and hyperlinks respectively, topredict links among previously unseen documentsand from those new documents to training docu-ments. Embedding the WSBM in a network/topic

Our Paper

: Topic Model,

Document Network, Block Detection

Journal of Machine Learning Research 3 (2003) 993-1022 Submitted 2/02; Published 1/03

Latent Dirichlet Allocation

David M. Blei [email protected]

Computer Science DivisionUniversity of CaliforniaBerkeley, CA 94720, USA

Andrew Y. Ng [email protected]

Computer Science DepartmentStanford UniversityStanford, CA 94305, USA

Michael I. Jordan [email protected]

Computer Science Division and Department of StatisticsUniversity of CaliforniaBerkeley, CA 94720, USA

Editor: John Lafferty

Abstract

We describelatent Dirichlet allocation(LDA), a generative probabilistic model for collections ofdiscrete data such as text corpora. LDA is a three-level hierarchical Bayesian model, in which eachitem of a collection is modeled as a finite mixture over an underlying set of topics. Each topic is, inturn, modeled as an infinite mixture over an underlying set of topic probabilities. In the context oftext modeling, the topic probabilities provide an explicit representation of a document. We presentefficient approximate inference techniques based on variational methods and an EM algorithm forempirical Bayes parameter estimation. We report results in document modeling, text classification,and collaborative filtering, comparing to a mixture of unigrams model and the probabilistic LSImodel.

1. Introduction

In this paper we consider the problem of modeling text corpora and other collections of discretedata. The goal is to find short descriptions of the members of a collection that enable efficientprocessing of large collections while preserving the essential statistical relationships that are usefulfor basic tasks such as classification, novelty detection, summarization, and similarity and relevancejudgments.

Significant progress has been made on this problem by researchers in the field of informa-tion retrieval (IR) (Baeza-Yates and Ribeiro-Neto, 1999). The basic methodology proposed byIR researchers for text corpora—a methodology successfully deployed in modern Internet searchengines—reduces each document in the corpus to a vector of real numbers, each of which repre-sents ratios of counts. In the populartf-idf scheme (Salton and McGill, 1983), a basic vocabularyof “words” or “terms” is chosen, and, for each document in the corpus, a count is formed of thenumber of occurrences of each word. After suitable normalization, this term frequency count iscompared to an inverse document frequency count, which measures the number of occurrences of a

c©2003 David M. Blei, Andrew Y. Ng and Michael I. Jordan.

LDA [Blei et al., 2003]

: Topic Model

The Annals of Applied Statistics2010, Vol. 4, No. 1, 124–150DOI: 10.1214/09-AOAS309© Institute of Mathematical Statistics, 2010

HIERARCHICAL RELATIONAL MODELSFOR DOCUMENT NETWORKS

BY JONATHAN CHANG1 AND DAVID M. BLEI2

Facebook and Princeton University

We develop the relational topic model (RTM), a hierarchical model ofboth network structure and node attributes. We focus on document networks,where the attributes of each document are its words, that is, discrete obser-vations taken from a fixed vocabulary. For each pair of documents, the RTMmodels their link as a binary random variable that is conditioned on theircontents. The model can be used to summarize a network of documents, pre-dict links between them, and predict words within them. We derive efficientinference and estimation algorithms based on variational methods that takeadvantage of sparsity and scale with the number of links. We evaluate thepredictive performance of the RTM for large networks of scientific abstracts,web documents, and geographically tagged news.

1. Introduction. Network data, such as citation networks of documents, hy-perlinked networks of web pages, and social networks of friends, are pervasive inapplied statistics and machine learning. The statistical analysis of network data canprovide both useful predictive models and descriptive statistics. Predictive modelscan point social network members toward new friends, scientific papers towardrelevant citations, and web pages toward other related pages. Descriptive statisticscan uncover the hidden community structure underlying a network data set.

Recent research in this field has focused on latent variable models of link struc-ture, models that decompose a network according to hidden patterns of connectionsbetween its nodes [Kemp, Griffiths and Tenenbaum (2004); Hofman and Wiggins(2007); Airoldi et al. (2008)]. These models represent a significant departure fromstatistical models of networks, which explain network data in terms of observedsufficient statistics [Fienberg, Meyer and Wasserman (1985); Wasserman and Pat-tison (1996); Getoor et al. (2001); Newman (2002); Taskar et al. (2004)].

While powerful, current latent variable models account only for the structureof the network, ignoring additional attributes of the nodes that might be available.For example, a citation network of articles also contains text and abstracts of thedocuments, a linked set of web-pages also contains the text for those pages, andan on-line social network also contains profile descriptions and other information

Received January 2009; revised November 2009.1Work done while at Princeton University.2Supported by ONR 175-6343, NSF CAREER 0745520 and grants from Google and Microsoft.Key words and phrases. Mixed-membership models, variational methods, text analysis, network

models.

124

Relational Topic Model

[Chang and Blei, 2010]

:

Topic Model, Document Network

Learning Latent Block Structure in Weighted Networks

Christopher AicherDepartment of Applied Mathematics, University of Colorado, Boulder, CO, 80309

[email protected]

Abigail Z. JacobsDepartment of Computer Science, University of Colorado, Boulder, CO, 80309

Aaron ClausetDepartment of Computer Science, University of Colorado, Boulder, CO, 80309

BioFrontiers Institute, University of Colorado, Boulder, CO 80303

Santa Fe Institute, Santa Fe, NM 87501

Abstract

Community detection is an important task in network analysis, in which we aim to learn anetwork partition that groups together vertices with similar community-level connectivity pat-terns. By finding such groups of vertices with similar structural roles, we extract a compactrepresentation of the network’s large-scale structure, which can facilitate its scientific interpre-tation and the prediction of unknown or future interactions. Popular approaches, including thestochastic block model, assume edges are unweighted, which limits their utility by discarding po-tentially useful information. We introduce the weighted stochastic block model (WSBM), whichgeneralizes the stochastic block model to networks with edge weights drawn from any exponen-tial family distribution. This model learns from both the presence and weight of edges, allowingit to discover structure that would otherwise be hidden when weights are discarded or thresh-olded. We describe a Bayesian variational algorithm for efficiently approximating this model’sposterior distribution over latent block structures. We then evaluate the WSBM’s performanceon both edge-existence and edge-weight prediction tasks for a set of real-world weighted net-works. In all cases, the WSBM performs as well or better than the best alternatives on thesetasks. community detection, weighted relational data, block models, exponential family,variational Bayes.

1 Introduction

Networks are an increasingly important form of structured data consisting of interactions betweenpairs of individuals in large social and biological data sets. Unlike attribute data where each obser-vation is associated with an individual, network data is represented by graphs, where individuals arevertices and interactions are edges. Because vertices are pairwise related, network data violates tra-ditional assumptions of attribute data, such as independence. This intrinsic difference in structureprompts the development of new tools for handling network data.

In social and biological networks, vertices often play distinct structural roles in generating thenetwork’s large-scale structure. To identify such latent structural roles, we aim to identify a net-work partition that groups together vertices with similar group-level connectivity patterns. We callthese groups “communities,” and their inference produces a compact description of the large-scale

1

arX

iv:1

404.

0431

v2 [

stat

.ML

] 3

Jun

201

4

Weighted Stochastic Block Model

[Aicher et al., 2014]

: Block Detection

3/22

Page 5: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Links Indicate Topic Similarity

A Discriminative Topic Model using Document Network Structure

Weiwei YangComputer Science

University of MarylandCollege Park, MD

[email protected]

Jordan Boyd-GraberComputer Science

University of ColoradoBoulder, CO

[email protected]

Philip ResnikLinguistics and UMIACSUniversity of Maryland

College Park, [email protected]

Abstract

Document collections often have links be-tween documents—citations, hyperlinks,or revisions—and which links are added isoften based on topical similarity. To modelthese intuitions, we introduce a new topicmodel for documents situated within a net-work structure, integrating latent blocksof documents with a max-margin learningcriterion for link prediction using topic-and word-level features. Experiments ona scientific paper dataset and collectionof webpages show that, by more robustlyexploiting the rich link structure within adocument network, our model improveslink prediction, topic quality, and blockdistributions.

1 Introduction

Documents often appear within a network struc-ture: social media mentions, retweets, and fol-lower relationships; Web pages by hyperlinks; sci-entific papers by citations. Network structure in-teracts with the topics in the text, in that docu-ments linked in a network are more likely to havesimilar topic distributions. For instance, a cita-tion link between two papers suggests that theyare about a similar field, and a mentioning linkbetween two social media users often indicatescommon interests. Conversely, documents’ sim-ilar topic distributions can suggest links betweenthem. For example, topic model (Blei et al., 2003,LDA) and block detection papers (Holland et al.,1983) are relevant to our research, so we cite them.Similarly, if a social media user A finds anotheruser B with shared interests, then A is more likelyto follow B.

Our approach is part of a natural progressionof network modeling in which models integrate

more information in more sophisticated ways.Some past methods only consider the network it-self (Kim and Leskovec, 2012; Liben-Nowell andKleinberg, 2007), which loses the rich informationin text. In other cases, methods take both links andtext into account (Chaturvedi et al., 2012), but theyare modeled separately, not jointly, limiting themodel’s ability to capture interactions between thetwo. The relational topic model (Chang and Blei,2010, RTM) goes further, jointly modeling topicsand links, but it considers only pairwise documentrelationships, failing to capture network structureat the level of groups or blocks of documents.

We propose a new joint model that makes fulleruse of the rich link structure within a documentnetwork. Specifically, our model embeds theweighted stochastic block model (Aicher et al.,2014, WSBM) to identify blocks in which docu-ments are densely connected. WSBM basically cat-egorizes each item in a network probabilisticallyas belonging to one of L blocks, by reviewingits connections with each block. Our model canbe viewed as a principled probabilistic extensionof Yang et al. (2015), who identify blocks in a doc-ument network deterministically as strongly con-nected components (SCC). Like them, we assign adistinct Dirichlet prior to each block to capture itstopical commonalities. Jointly, a linear regressionmodel with a discriminative, max-margin objec-tive function (Zhu et al., 2012; Zhu et al., 2014) istrained to reconstruct the links, taking into accountthe features of documents’ topic and word distri-butions (Nguyen et al., 2013), block assignments,and inter-block link rates.

We validate our approach on a scientific pa-per abstract dataset and collection of webpages,with citation links and hyperlinks respectively, topredict links among previously unseen documentsand from those new documents to training docu-ments. Embedding the WSBM in a network/topic

Our Paper: Topic Model

,

Document Network, Block Detection

Journal of Machine Learning Research 3 (2003) 993-1022 Submitted 2/02; Published 1/03

Latent Dirichlet Allocation

David M. Blei [email protected]

Computer Science DivisionUniversity of CaliforniaBerkeley, CA 94720, USA

Andrew Y. Ng [email protected]

Computer Science DepartmentStanford UniversityStanford, CA 94305, USA

Michael I. Jordan [email protected]

Computer Science Division and Department of StatisticsUniversity of CaliforniaBerkeley, CA 94720, USA

Editor: John Lafferty

Abstract

We describelatent Dirichlet allocation(LDA), a generative probabilistic model for collections ofdiscrete data such as text corpora. LDA is a three-level hierarchical Bayesian model, in which eachitem of a collection is modeled as a finite mixture over an underlying set of topics. Each topic is, inturn, modeled as an infinite mixture over an underlying set of topic probabilities. In the context oftext modeling, the topic probabilities provide an explicit representation of a document. We presentefficient approximate inference techniques based on variational methods and an EM algorithm forempirical Bayes parameter estimation. We report results in document modeling, text classification,and collaborative filtering, comparing to a mixture of unigrams model and the probabilistic LSImodel.

1. Introduction

In this paper we consider the problem of modeling text corpora and other collections of discretedata. The goal is to find short descriptions of the members of a collection that enable efficientprocessing of large collections while preserving the essential statistical relationships that are usefulfor basic tasks such as classification, novelty detection, summarization, and similarity and relevancejudgments.

Significant progress has been made on this problem by researchers in the field of informa-tion retrieval (IR) (Baeza-Yates and Ribeiro-Neto, 1999). The basic methodology proposed byIR researchers for text corpora—a methodology successfully deployed in modern Internet searchengines—reduces each document in the corpus to a vector of real numbers, each of which repre-sents ratios of counts. In the populartf-idf scheme (Salton and McGill, 1983), a basic vocabularyof “words” or “terms” is chosen, and, for each document in the corpus, a count is formed of thenumber of occurrences of each word. After suitable normalization, this term frequency count iscompared to an inverse document frequency count, which measures the number of occurrences of a

c©2003 David M. Blei, Andrew Y. Ng and Michael I. Jordan.

LDA [Blei et al., 2003] : Topic Model

The Annals of Applied Statistics2010, Vol. 4, No. 1, 124–150DOI: 10.1214/09-AOAS309© Institute of Mathematical Statistics, 2010

HIERARCHICAL RELATIONAL MODELSFOR DOCUMENT NETWORKS

BY JONATHAN CHANG1 AND DAVID M. BLEI2

Facebook and Princeton University

We develop the relational topic model (RTM), a hierarchical model ofboth network structure and node attributes. We focus on document networks,where the attributes of each document are its words, that is, discrete obser-vations taken from a fixed vocabulary. For each pair of documents, the RTMmodels their link as a binary random variable that is conditioned on theircontents. The model can be used to summarize a network of documents, pre-dict links between them, and predict words within them. We derive efficientinference and estimation algorithms based on variational methods that takeadvantage of sparsity and scale with the number of links. We evaluate thepredictive performance of the RTM for large networks of scientific abstracts,web documents, and geographically tagged news.

1. Introduction. Network data, such as citation networks of documents, hy-perlinked networks of web pages, and social networks of friends, are pervasive inapplied statistics and machine learning. The statistical analysis of network data canprovide both useful predictive models and descriptive statistics. Predictive modelscan point social network members toward new friends, scientific papers towardrelevant citations, and web pages toward other related pages. Descriptive statisticscan uncover the hidden community structure underlying a network data set.

Recent research in this field has focused on latent variable models of link struc-ture, models that decompose a network according to hidden patterns of connectionsbetween its nodes [Kemp, Griffiths and Tenenbaum (2004); Hofman and Wiggins(2007); Airoldi et al. (2008)]. These models represent a significant departure fromstatistical models of networks, which explain network data in terms of observedsufficient statistics [Fienberg, Meyer and Wasserman (1985); Wasserman and Pat-tison (1996); Getoor et al. (2001); Newman (2002); Taskar et al. (2004)].

While powerful, current latent variable models account only for the structureof the network, ignoring additional attributes of the nodes that might be available.For example, a citation network of articles also contains text and abstracts of thedocuments, a linked set of web-pages also contains the text for those pages, andan on-line social network also contains profile descriptions and other information

Received January 2009; revised November 2009.1Work done while at Princeton University.2Supported by ONR 175-6343, NSF CAREER 0745520 and grants from Google and Microsoft.Key words and phrases. Mixed-membership models, variational methods, text analysis, network

models.

124

Relational Topic Model

[Chang and Blei, 2010]:

Topic Model

, Document Network

Learning Latent Block Structure in Weighted Networks

Christopher AicherDepartment of Applied Mathematics, University of Colorado, Boulder, CO, 80309

[email protected]

Abigail Z. JacobsDepartment of Computer Science, University of Colorado, Boulder, CO, 80309

Aaron ClausetDepartment of Computer Science, University of Colorado, Boulder, CO, 80309

BioFrontiers Institute, University of Colorado, Boulder, CO 80303

Santa Fe Institute, Santa Fe, NM 87501

Abstract

Community detection is an important task in network analysis, in which we aim to learn anetwork partition that groups together vertices with similar community-level connectivity pat-terns. By finding such groups of vertices with similar structural roles, we extract a compactrepresentation of the network’s large-scale structure, which can facilitate its scientific interpre-tation and the prediction of unknown or future interactions. Popular approaches, including thestochastic block model, assume edges are unweighted, which limits their utility by discarding po-tentially useful information. We introduce the weighted stochastic block model (WSBM), whichgeneralizes the stochastic block model to networks with edge weights drawn from any exponen-tial family distribution. This model learns from both the presence and weight of edges, allowingit to discover structure that would otherwise be hidden when weights are discarded or thresh-olded. We describe a Bayesian variational algorithm for efficiently approximating this model’sposterior distribution over latent block structures. We then evaluate the WSBM’s performanceon both edge-existence and edge-weight prediction tasks for a set of real-world weighted net-works. In all cases, the WSBM performs as well or better than the best alternatives on thesetasks. community detection, weighted relational data, block models, exponential family,variational Bayes.

1 Introduction

Networks are an increasingly important form of structured data consisting of interactions betweenpairs of individuals in large social and biological data sets. Unlike attribute data where each obser-vation is associated with an individual, network data is represented by graphs, where individuals arevertices and interactions are edges. Because vertices are pairwise related, network data violates tra-ditional assumptions of attribute data, such as independence. This intrinsic difference in structureprompts the development of new tools for handling network data.

In social and biological networks, vertices often play distinct structural roles in generating thenetwork’s large-scale structure. To identify such latent structural roles, we aim to identify a net-work partition that groups together vertices with similar group-level connectivity patterns. We callthese groups “communities,” and their inference produces a compact description of the large-scale

1

arX

iv:1

404.

0431

v2 [

stat

.ML

] 3

Jun

201

4

Weighted Stochastic Block Model

[Aicher et al., 2014]

: Block Detection

3/22

Page 6: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Links Indicate Topic Similarity

A Discriminative Topic Model using Document Network Structure

Weiwei YangComputer Science

University of MarylandCollege Park, MD

[email protected]

Jordan Boyd-GraberComputer Science

University of ColoradoBoulder, CO

[email protected]

Philip ResnikLinguistics and UMIACSUniversity of Maryland

College Park, [email protected]

Abstract

Document collections often have links be-tween documents—citations, hyperlinks,or revisions—and which links are added isoften based on topical similarity. To modelthese intuitions, we introduce a new topicmodel for documents situated within a net-work structure, integrating latent blocksof documents with a max-margin learningcriterion for link prediction using topic-and word-level features. Experiments ona scientific paper dataset and collectionof webpages show that, by more robustlyexploiting the rich link structure within adocument network, our model improveslink prediction, topic quality, and blockdistributions.

1 Introduction

Documents often appear within a network struc-ture: social media mentions, retweets, and fol-lower relationships; Web pages by hyperlinks; sci-entific papers by citations. Network structure in-teracts with the topics in the text, in that docu-ments linked in a network are more likely to havesimilar topic distributions. For instance, a cita-tion link between two papers suggests that theyare about a similar field, and a mentioning linkbetween two social media users often indicatescommon interests. Conversely, documents’ sim-ilar topic distributions can suggest links betweenthem. For example, topic model (Blei et al., 2003,LDA) and block detection papers (Holland et al.,1983) are relevant to our research, so we cite them.Similarly, if a social media user A finds anotheruser B with shared interests, then A is more likelyto follow B.

Our approach is part of a natural progressionof network modeling in which models integrate

more information in more sophisticated ways.Some past methods only consider the network it-self (Kim and Leskovec, 2012; Liben-Nowell andKleinberg, 2007), which loses the rich informationin text. In other cases, methods take both links andtext into account (Chaturvedi et al., 2012), but theyare modeled separately, not jointly, limiting themodel’s ability to capture interactions between thetwo. The relational topic model (Chang and Blei,2010, RTM) goes further, jointly modeling topicsand links, but it considers only pairwise documentrelationships, failing to capture network structureat the level of groups or blocks of documents.

We propose a new joint model that makes fulleruse of the rich link structure within a documentnetwork. Specifically, our model embeds theweighted stochastic block model (Aicher et al.,2014, WSBM) to identify blocks in which docu-ments are densely connected. WSBM basically cat-egorizes each item in a network probabilisticallyas belonging to one of L blocks, by reviewingits connections with each block. Our model canbe viewed as a principled probabilistic extensionof Yang et al. (2015), who identify blocks in a doc-ument network deterministically as strongly con-nected components (SCC). Like them, we assign adistinct Dirichlet prior to each block to capture itstopical commonalities. Jointly, a linear regressionmodel with a discriminative, max-margin objec-tive function (Zhu et al., 2012; Zhu et al., 2014) istrained to reconstruct the links, taking into accountthe features of documents’ topic and word distri-butions (Nguyen et al., 2013), block assignments,and inter-block link rates.

We validate our approach on a scientific pa-per abstract dataset and collection of webpages,with citation links and hyperlinks respectively, topredict links among previously unseen documentsand from those new documents to training docu-ments. Embedding the WSBM in a network/topic

Our Paper: Topic Model,

Document Network

, Block Detection

Journal of Machine Learning Research 3 (2003) 993-1022 Submitted 2/02; Published 1/03

Latent Dirichlet Allocation

David M. Blei [email protected]

Computer Science DivisionUniversity of CaliforniaBerkeley, CA 94720, USA

Andrew Y. Ng [email protected]

Computer Science DepartmentStanford UniversityStanford, CA 94305, USA

Michael I. Jordan [email protected]

Computer Science Division and Department of StatisticsUniversity of CaliforniaBerkeley, CA 94720, USA

Editor: John Lafferty

Abstract

We describelatent Dirichlet allocation(LDA), a generative probabilistic model for collections ofdiscrete data such as text corpora. LDA is a three-level hierarchical Bayesian model, in which eachitem of a collection is modeled as a finite mixture over an underlying set of topics. Each topic is, inturn, modeled as an infinite mixture over an underlying set of topic probabilities. In the context oftext modeling, the topic probabilities provide an explicit representation of a document. We presentefficient approximate inference techniques based on variational methods and an EM algorithm forempirical Bayes parameter estimation. We report results in document modeling, text classification,and collaborative filtering, comparing to a mixture of unigrams model and the probabilistic LSImodel.

1. Introduction

In this paper we consider the problem of modeling text corpora and other collections of discretedata. The goal is to find short descriptions of the members of a collection that enable efficientprocessing of large collections while preserving the essential statistical relationships that are usefulfor basic tasks such as classification, novelty detection, summarization, and similarity and relevancejudgments.

Significant progress has been made on this problem by researchers in the field of informa-tion retrieval (IR) (Baeza-Yates and Ribeiro-Neto, 1999). The basic methodology proposed byIR researchers for text corpora—a methodology successfully deployed in modern Internet searchengines—reduces each document in the corpus to a vector of real numbers, each of which repre-sents ratios of counts. In the populartf-idf scheme (Salton and McGill, 1983), a basic vocabularyof “words” or “terms” is chosen, and, for each document in the corpus, a count is formed of thenumber of occurrences of each word. After suitable normalization, this term frequency count iscompared to an inverse document frequency count, which measures the number of occurrences of a

c©2003 David M. Blei, Andrew Y. Ng and Michael I. Jordan.

LDA [Blei et al., 2003] : Topic Model

The Annals of Applied Statistics2010, Vol. 4, No. 1, 124–150DOI: 10.1214/09-AOAS309© Institute of Mathematical Statistics, 2010

HIERARCHICAL RELATIONAL MODELSFOR DOCUMENT NETWORKS

BY JONATHAN CHANG1 AND DAVID M. BLEI2

Facebook and Princeton University

We develop the relational topic model (RTM), a hierarchical model ofboth network structure and node attributes. We focus on document networks,where the attributes of each document are its words, that is, discrete obser-vations taken from a fixed vocabulary. For each pair of documents, the RTMmodels their link as a binary random variable that is conditioned on theircontents. The model can be used to summarize a network of documents, pre-dict links between them, and predict words within them. We derive efficientinference and estimation algorithms based on variational methods that takeadvantage of sparsity and scale with the number of links. We evaluate thepredictive performance of the RTM for large networks of scientific abstracts,web documents, and geographically tagged news.

1. Introduction. Network data, such as citation networks of documents, hy-perlinked networks of web pages, and social networks of friends, are pervasive inapplied statistics and machine learning. The statistical analysis of network data canprovide both useful predictive models and descriptive statistics. Predictive modelscan point social network members toward new friends, scientific papers towardrelevant citations, and web pages toward other related pages. Descriptive statisticscan uncover the hidden community structure underlying a network data set.

Recent research in this field has focused on latent variable models of link struc-ture, models that decompose a network according to hidden patterns of connectionsbetween its nodes [Kemp, Griffiths and Tenenbaum (2004); Hofman and Wiggins(2007); Airoldi et al. (2008)]. These models represent a significant departure fromstatistical models of networks, which explain network data in terms of observedsufficient statistics [Fienberg, Meyer and Wasserman (1985); Wasserman and Pat-tison (1996); Getoor et al. (2001); Newman (2002); Taskar et al. (2004)].

While powerful, current latent variable models account only for the structureof the network, ignoring additional attributes of the nodes that might be available.For example, a citation network of articles also contains text and abstracts of thedocuments, a linked set of web-pages also contains the text for those pages, andan on-line social network also contains profile descriptions and other information

Received January 2009; revised November 2009.1Work done while at Princeton University.2Supported by ONR 175-6343, NSF CAREER 0745520 and grants from Google and Microsoft.Key words and phrases. Mixed-membership models, variational methods, text analysis, network

models.

124

Relational Topic Model

[Chang and Blei, 2010]:

Topic Model, Document Network

Learning Latent Block Structure in Weighted Networks

Christopher AicherDepartment of Applied Mathematics, University of Colorado, Boulder, CO, 80309

[email protected]

Abigail Z. JacobsDepartment of Computer Science, University of Colorado, Boulder, CO, 80309

Aaron ClausetDepartment of Computer Science, University of Colorado, Boulder, CO, 80309

BioFrontiers Institute, University of Colorado, Boulder, CO 80303

Santa Fe Institute, Santa Fe, NM 87501

Abstract

Community detection is an important task in network analysis, in which we aim to learn anetwork partition that groups together vertices with similar community-level connectivity pat-terns. By finding such groups of vertices with similar structural roles, we extract a compactrepresentation of the network’s large-scale structure, which can facilitate its scientific interpre-tation and the prediction of unknown or future interactions. Popular approaches, including thestochastic block model, assume edges are unweighted, which limits their utility by discarding po-tentially useful information. We introduce the weighted stochastic block model (WSBM), whichgeneralizes the stochastic block model to networks with edge weights drawn from any exponen-tial family distribution. This model learns from both the presence and weight of edges, allowingit to discover structure that would otherwise be hidden when weights are discarded or thresh-olded. We describe a Bayesian variational algorithm for efficiently approximating this model’sposterior distribution over latent block structures. We then evaluate the WSBM’s performanceon both edge-existence and edge-weight prediction tasks for a set of real-world weighted net-works. In all cases, the WSBM performs as well or better than the best alternatives on thesetasks. community detection, weighted relational data, block models, exponential family,variational Bayes.

1 Introduction

Networks are an increasingly important form of structured data consisting of interactions betweenpairs of individuals in large social and biological data sets. Unlike attribute data where each obser-vation is associated with an individual, network data is represented by graphs, where individuals arevertices and interactions are edges. Because vertices are pairwise related, network data violates tra-ditional assumptions of attribute data, such as independence. This intrinsic difference in structureprompts the development of new tools for handling network data.

In social and biological networks, vertices often play distinct structural roles in generating thenetwork’s large-scale structure. To identify such latent structural roles, we aim to identify a net-work partition that groups together vertices with similar group-level connectivity patterns. We callthese groups “communities,” and their inference produces a compact description of the large-scale

1

arX

iv:1

404.

0431

v2 [

stat

.ML

] 3

Jun

201

4

Weighted Stochastic Block Model

[Aicher et al., 2014]

: Block Detection

3/22

Page 7: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Links Indicate Topic Similarity

A Discriminative Topic Model using Document Network Structure

Weiwei YangComputer Science

University of MarylandCollege Park, MD

[email protected]

Jordan Boyd-GraberComputer Science

University of ColoradoBoulder, CO

[email protected]

Philip ResnikLinguistics and UMIACSUniversity of Maryland

College Park, [email protected]

Abstract

Document collections often have links be-tween documents—citations, hyperlinks,or revisions—and which links are added isoften based on topical similarity. To modelthese intuitions, we introduce a new topicmodel for documents situated within a net-work structure, integrating latent blocksof documents with a max-margin learningcriterion for link prediction using topic-and word-level features. Experiments ona scientific paper dataset and collectionof webpages show that, by more robustlyexploiting the rich link structure within adocument network, our model improveslink prediction, topic quality, and blockdistributions.

1 Introduction

Documents often appear within a network struc-ture: social media mentions, retweets, and fol-lower relationships; Web pages by hyperlinks; sci-entific papers by citations. Network structure in-teracts with the topics in the text, in that docu-ments linked in a network are more likely to havesimilar topic distributions. For instance, a cita-tion link between two papers suggests that theyare about a similar field, and a mentioning linkbetween two social media users often indicatescommon interests. Conversely, documents’ sim-ilar topic distributions can suggest links betweenthem. For example, topic model (Blei et al., 2003,LDA) and block detection papers (Holland et al.,1983) are relevant to our research, so we cite them.Similarly, if a social media user A finds anotheruser B with shared interests, then A is more likelyto follow B.

Our approach is part of a natural progressionof network modeling in which models integrate

more information in more sophisticated ways.Some past methods only consider the network it-self (Kim and Leskovec, 2012; Liben-Nowell andKleinberg, 2007), which loses the rich informationin text. In other cases, methods take both links andtext into account (Chaturvedi et al., 2012), but theyare modeled separately, not jointly, limiting themodel’s ability to capture interactions between thetwo. The relational topic model (Chang and Blei,2010, RTM) goes further, jointly modeling topicsand links, but it considers only pairwise documentrelationships, failing to capture network structureat the level of groups or blocks of documents.

We propose a new joint model that makes fulleruse of the rich link structure within a documentnetwork. Specifically, our model embeds theweighted stochastic block model (Aicher et al.,2014, WSBM) to identify blocks in which docu-ments are densely connected. WSBM basically cat-egorizes each item in a network probabilisticallyas belonging to one of L blocks, by reviewingits connections with each block. Our model canbe viewed as a principled probabilistic extensionof Yang et al. (2015), who identify blocks in a doc-ument network deterministically as strongly con-nected components (SCC). Like them, we assign adistinct Dirichlet prior to each block to capture itstopical commonalities. Jointly, a linear regressionmodel with a discriminative, max-margin objec-tive function (Zhu et al., 2012; Zhu et al., 2014) istrained to reconstruct the links, taking into accountthe features of documents’ topic and word distri-butions (Nguyen et al., 2013), block assignments,and inter-block link rates.

We validate our approach on a scientific pa-per abstract dataset and collection of webpages,with citation links and hyperlinks respectively, topredict links among previously unseen documentsand from those new documents to training docu-ments. Embedding the WSBM in a network/topic

Our Paper: Topic Model,

Document Network, Block Detection

Journal of Machine Learning Research 3 (2003) 993-1022 Submitted 2/02; Published 1/03

Latent Dirichlet Allocation

David M. Blei [email protected]

Computer Science DivisionUniversity of CaliforniaBerkeley, CA 94720, USA

Andrew Y. Ng [email protected]

Computer Science DepartmentStanford UniversityStanford, CA 94305, USA

Michael I. Jordan [email protected]

Computer Science Division and Department of StatisticsUniversity of CaliforniaBerkeley, CA 94720, USA

Editor: John Lafferty

Abstract

We describelatent Dirichlet allocation(LDA), a generative probabilistic model for collections ofdiscrete data such as text corpora. LDA is a three-level hierarchical Bayesian model, in which eachitem of a collection is modeled as a finite mixture over an underlying set of topics. Each topic is, inturn, modeled as an infinite mixture over an underlying set of topic probabilities. In the context oftext modeling, the topic probabilities provide an explicit representation of a document. We presentefficient approximate inference techniques based on variational methods and an EM algorithm forempirical Bayes parameter estimation. We report results in document modeling, text classification,and collaborative filtering, comparing to a mixture of unigrams model and the probabilistic LSImodel.

1. Introduction

In this paper we consider the problem of modeling text corpora and other collections of discretedata. The goal is to find short descriptions of the members of a collection that enable efficientprocessing of large collections while preserving the essential statistical relationships that are usefulfor basic tasks such as classification, novelty detection, summarization, and similarity and relevancejudgments.

Significant progress has been made on this problem by researchers in the field of informa-tion retrieval (IR) (Baeza-Yates and Ribeiro-Neto, 1999). The basic methodology proposed byIR researchers for text corpora—a methodology successfully deployed in modern Internet searchengines—reduces each document in the corpus to a vector of real numbers, each of which repre-sents ratios of counts. In the populartf-idf scheme (Salton and McGill, 1983), a basic vocabularyof “words” or “terms” is chosen, and, for each document in the corpus, a count is formed of thenumber of occurrences of each word. After suitable normalization, this term frequency count iscompared to an inverse document frequency count, which measures the number of occurrences of a

c©2003 David M. Blei, Andrew Y. Ng and Michael I. Jordan.

LDA [Blei et al., 2003] : Topic Model

The Annals of Applied Statistics2010, Vol. 4, No. 1, 124–150DOI: 10.1214/09-AOAS309© Institute of Mathematical Statistics, 2010

HIERARCHICAL RELATIONAL MODELSFOR DOCUMENT NETWORKS

BY JONATHAN CHANG1 AND DAVID M. BLEI2

Facebook and Princeton University

We develop the relational topic model (RTM), a hierarchical model ofboth network structure and node attributes. We focus on document networks,where the attributes of each document are its words, that is, discrete obser-vations taken from a fixed vocabulary. For each pair of documents, the RTMmodels their link as a binary random variable that is conditioned on theircontents. The model can be used to summarize a network of documents, pre-dict links between them, and predict words within them. We derive efficientinference and estimation algorithms based on variational methods that takeadvantage of sparsity and scale with the number of links. We evaluate thepredictive performance of the RTM for large networks of scientific abstracts,web documents, and geographically tagged news.

1. Introduction. Network data, such as citation networks of documents, hy-perlinked networks of web pages, and social networks of friends, are pervasive inapplied statistics and machine learning. The statistical analysis of network data canprovide both useful predictive models and descriptive statistics. Predictive modelscan point social network members toward new friends, scientific papers towardrelevant citations, and web pages toward other related pages. Descriptive statisticscan uncover the hidden community structure underlying a network data set.

Recent research in this field has focused on latent variable models of link struc-ture, models that decompose a network according to hidden patterns of connectionsbetween its nodes [Kemp, Griffiths and Tenenbaum (2004); Hofman and Wiggins(2007); Airoldi et al. (2008)]. These models represent a significant departure fromstatistical models of networks, which explain network data in terms of observedsufficient statistics [Fienberg, Meyer and Wasserman (1985); Wasserman and Pat-tison (1996); Getoor et al. (2001); Newman (2002); Taskar et al. (2004)].

While powerful, current latent variable models account only for the structureof the network, ignoring additional attributes of the nodes that might be available.For example, a citation network of articles also contains text and abstracts of thedocuments, a linked set of web-pages also contains the text for those pages, andan on-line social network also contains profile descriptions and other information

Received January 2009; revised November 2009.1Work done while at Princeton University.2Supported by ONR 175-6343, NSF CAREER 0745520 and grants from Google and Microsoft.Key words and phrases. Mixed-membership models, variational methods, text analysis, network

models.

124

Relational Topic Model

[Chang and Blei, 2010]:

Topic Model, Document Network

Learning Latent Block Structure in Weighted Networks

Christopher AicherDepartment of Applied Mathematics, University of Colorado, Boulder, CO, 80309

[email protected]

Abigail Z. JacobsDepartment of Computer Science, University of Colorado, Boulder, CO, 80309

Aaron ClausetDepartment of Computer Science, University of Colorado, Boulder, CO, 80309

BioFrontiers Institute, University of Colorado, Boulder, CO 80303

Santa Fe Institute, Santa Fe, NM 87501

Abstract

Community detection is an important task in network analysis, in which we aim to learn anetwork partition that groups together vertices with similar community-level connectivity pat-terns. By finding such groups of vertices with similar structural roles, we extract a compactrepresentation of the network’s large-scale structure, which can facilitate its scientific interpre-tation and the prediction of unknown or future interactions. Popular approaches, including thestochastic block model, assume edges are unweighted, which limits their utility by discarding po-tentially useful information. We introduce the weighted stochastic block model (WSBM), whichgeneralizes the stochastic block model to networks with edge weights drawn from any exponen-tial family distribution. This model learns from both the presence and weight of edges, allowingit to discover structure that would otherwise be hidden when weights are discarded or thresh-olded. We describe a Bayesian variational algorithm for efficiently approximating this model’sposterior distribution over latent block structures. We then evaluate the WSBM’s performanceon both edge-existence and edge-weight prediction tasks for a set of real-world weighted net-works. In all cases, the WSBM performs as well or better than the best alternatives on thesetasks. community detection, weighted relational data, block models, exponential family,variational Bayes.

1 Introduction

Networks are an increasingly important form of structured data consisting of interactions betweenpairs of individuals in large social and biological data sets. Unlike attribute data where each obser-vation is associated with an individual, network data is represented by graphs, where individuals arevertices and interactions are edges. Because vertices are pairwise related, network data violates tra-ditional assumptions of attribute data, such as independence. This intrinsic difference in structureprompts the development of new tools for handling network data.

In social and biological networks, vertices often play distinct structural roles in generating thenetwork’s large-scale structure. To identify such latent structural roles, we aim to identify a net-work partition that groups together vertices with similar group-level connectivity patterns. We callthese groups “communities,” and their inference produces a compact description of the large-scale

1

arX

iv:1

404.

0431

v2 [

stat

.ML

] 3

Jun

201

4

Weighted Stochastic Block Model

[Aicher et al., 2014]: Block Detection

3/22

Page 8: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Goal

Make use of the rich information in document links

Improve topic modeling

Replicate existing links

Predict held-out links

4/22

Page 9: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Outline

1 Block Detection and RTM

2 LBH-RTM

3 Link Prediction Results

4 ConclusionsPaper

http://ter.ps/bv5Slides

http://ter.ps/bv7

5/22

Page 10: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Block Detection

Find densely-connected blocks in a graph

Deterministic: Strongly connected components (SCC)

Puts any linked nodes into the same componentDoes not consider link density

Probabilistic: Weighted stochastic block model (WSBM)

6/22

Page 11: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Block Detection

Find densely-connected blocks in a graph

Deterministic: Strongly connected components (SCC)

Puts any linked nodes into the same componentDoes not consider link density

Probabilistic: Weighted stochastic block model (WSBM)

6/22

Page 12: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Block Detection

Find densely-connected blocks in a graph

Deterministic: Strongly connected components (SCC)

Puts any linked nodes into the same componentDoes not consider link density

Probabilistic: Weighted stochastic block model (WSBM)

6/22

Page 13: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Block Detection

Find densely-connected blocks in a graph

Deterministic: Strongly connected components (SCC)

Puts any linked nodes into the same componentDoes not consider link density

Probabilistic: Weighted stochastic block model (WSBM)

6/22

Page 14: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model [Chang and Blei, 2010]

A topic model for link prediction

Jointly models topics and links

Each topic is assigned a weight

Indicate the correlationsbetween topics and links

Composes a regression value Rd,d′

Pr (Bd,d′ = 1) = σ (Rd,d′)

Kd

d

d

z

z

z

,

,

,

2

1

Kd

d

d

z

z

z

,'

,'

,'

2

1

K

'dN

dNd

'd

dz dw

'dz 'dw

7/22

Page 15: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model [Chang and Blei, 2010]

A topic model for link prediction

Jointly models topics and links

Each topic is assigned a weight

Indicate the correlationsbetween topics and links

Composes a regression value Rd,d′

Pr (Bd,d′ = 1) = σ (Rd,d′)

Kd

d

d

z

z

z

,

,

,

2

1

Kd

d

d

z

z

z

,'

,'

,'

2

1

T

K

'dN

dNd

'd

dz dw

'dz 'dw

7/22

Page 16: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model [Chang and Blei, 2010]

A topic model for link prediction

Jointly models topics and links

Each topic is assigned a weight

Indicate the correlationsbetween topics and links

Composes a regression value Rd,d′

Pr (Bd,d′ = 1) = σ (Rd,d′)

Kd

d

d

z

z

z

,

,

,

2

1

Kd

d

d

z

z

z

,'

,'

,'

2

1

T',ddR

K

'dN

dNd

'd

dz dw

'dz 'dw

7/22

Page 17: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model [Chang and Blei, 2010]

A topic model for link prediction

Jointly models topics and links

Each topic is assigned a weight

Indicate the correlationsbetween topics and links

Composes a regression value Rd,d′

Pr (Bd,d′ = 1) = σ (Rd,d′)

Kd

d

d

z

z

z

,

,

,

2

1

Kd

d

d

z

z

z

,'

,'

,'

2

1

T',ddR

K

'dN

dNd

'd

',ddB

dz dw

'dz 'dw

7/22

Page 18: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Outline

1 Block Detection and RTM

2 LBH-RTM

3 Link Prediction Results

4 ConclusionsPaper

http://ter.ps/bv5Slides

http://ter.ps/bv7

8/22

Page 19: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model

K

'dN

dNd

'd

',ddB

dz dw

'dz 'dw

9/22

Page 20: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model with Weighted Stochastic BlockModel

K

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

9/22

Page 21: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model with Weighted Stochastic BlockModel, Block Priors

1

2

3

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

9/22

Page 22: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model with Weighted Stochastic BlockModel, Block Priors

1

2

3

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

9/22

Page 23: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model with Weighted Stochastic BlockModel, Block Priors, Various Features

Rd,d′ = ηT

zd,1zd,2

...zd,K

zd′,1zd′,2

...zd′,K

︸ ︷︷ ︸

Topical Feature

+ τT

wd,1

wd,2

...wd,V

wd′,1wd′,2

...wd′,V

︸ ︷︷ ︸

Lexical Feature

+ ρl,l′Ωl,l′︸ ︷︷ ︸Block Feat.

Pr(Bd,d′ = 1

)= σ(Rd,d′ ) (Sigmoid Loss)

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

9/22

Page 24: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model with Weighted Stochastic BlockModel, Block Priors, Various Features

Rd,d′ = ηT

zd,1zd,2

...zd,K

zd′,1zd′,2

...zd′,K

︸ ︷︷ ︸

Topical Feature

+ τT

wd,1

wd,2

...wd,V

wd′,1wd′,2

...wd′,V

︸ ︷︷ ︸

Lexical Feature

+ ρl,l′Ωl,l′︸ ︷︷ ︸Block Feat.

Pr(Bd,d′ = 1

)= σ(Rd,d′ ) (Sigmoid Loss)

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

9/22

Page 25: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model with Weighted Stochastic BlockModel, Block Priors, Various Features

Rd,d′ = ηT

zd,1zd,2

...zd,K

zd′,1zd′,2

...zd′,K

︸ ︷︷ ︸

Topical Feature

+ τT

wd,1

wd,2

...wd,V

wd′,1wd′,2

...wd′,V

︸ ︷︷ ︸

Lexical Feature

+ ρl,l′Ωl,l′︸ ︷︷ ︸Block Feat.

Pr(Bd,d′ = 1

)= σ(Rd,d′ ) (Sigmoid Loss)

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

9/22

Page 26: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model with Weighted Stochastic BlockModel, Block Priors, Various Features

Rd,d′ = ηT

zd,1zd,2

...zd,K

zd′,1zd′,2

...zd′,K

︸ ︷︷ ︸

Topical Feature

+ τT

wd,1

wd,2

...wd,V

wd′,1wd′,2

...wd′,V

︸ ︷︷ ︸

Lexical Feature

+ ρl,l′Ωl,l′︸ ︷︷ ︸Block Feat.

Pr(Bd,d′ = 1

)= σ(Rd,d′ ) (Sigmoid Loss)

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

9/22

Page 27: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model with Weighted Stochastic BlockModel, Block Priors, Various Features, and Hinge Loss

Pr (Bd,d′ = 1) = σ(Rd,d′) (Sigmoid Loss)

→ Pr (Bd,d′) = exp (−2 max (0, 1− Bd,d′Rd,d′)) (Hinge Loss)

Make more effective use of side information when inferring topics.

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

9/22

Page 28: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model with Weighted Stochastic BlockModel, Block Priors, Various Features, and Hinge Loss

Pr (Bd,d′ = 1) = σ(Rd,d′) (Sigmoid Loss)

→ Pr (Bd,d′) = exp (−2 max (0, 1− Bd,d′Rd,d′)) (Hinge Loss)

Make more effective use of side information when inferring topics.

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

9/22

Page 29: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Relational Topic Model with Weighted Stochastic BlockModel, Block Priors, Various Features, and Hinge Loss

Pr (Bd,d′ = 1) = σ(Rd,d′) (Sigmoid Loss)

→ Pr (Bd,d′) = exp (−2 max (0, 1− Bd,d′Rd,d′)) (Hinge Loss)

Make more effective use of side information when inferring topics.

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

9/22

Page 30: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

LBH-RTM

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

Relational Topic Model with

Lexical weights

Block priors

Hinge loss

10/22

Page 31: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Baseline Models

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

Vanilla LDA: Infers topics based on words.

Relational Topic Model with various features:

Encourages linked docs to have similar topic distributions.Links indicate topic similarity.

Weighted Stochastic Block Model:

Does not understand the content at all.Finds blocks and provides informative priors.

11/22

Page 32: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Baseline Models

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

Vanilla LDA: Infers topics based on words.

Relational Topic Model with various features:

Encourages linked docs to have similar topic distributions.Links indicate topic similarity.

Weighted Stochastic Block Model:

Does not understand the content at all.Finds blocks and provides informative priors.

11/22

Page 33: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Baseline Models

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

Vanilla LDA: Infers topics based on words.

Relational Topic Model with various features:

Encourages linked docs to have similar topic distributions.Links indicate topic similarity.

Weighted Stochastic Block Model:

Does not understand the content at all.Finds blocks and provides informative priors.

11/22

Page 34: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Baseline Models

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

Vanilla LDA: Infers topics based on words.

Relational Topic Model with various features:

Encourages linked docs to have similar topic distributions.Links indicate topic similarity.

Weighted Stochastic Block Model:

Does not understand the content at all.

Finds blocks and provides informative priors.

11/22

Page 35: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Baseline Models

'

K

L

LL

'dN

dNd

'd

a

b

dy

'dy

',ddA ',ddB

dz dw

'dz 'dw

Vanilla LDA: Infers topics based on words.

Relational Topic Model with various features:

Encourages linked docs to have similar topic distributions.Links indicate topic similarity.

Weighted Stochastic Block Model:

Does not understand the content at all.Finds blocks and provides informative priors.

11/22

Page 36: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Outline

1 Block Detection and RTM

2 LBH-RTM

3 Link Prediction Results

4 ConclusionsPaper

http://ter.ps/bv5Slides

http://ter.ps/bv7

12/22

Page 37: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Datasets

Cora: Scientific papers and citation links

WebKB: Web pages and hyperlinks

Corpus #Docs #Links #VocabularyCora 2,362 4,231 1,240

WebKB 877 1,608 1,703

13/22

Page 38: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Task

Training input: Training documents with links

Test input: Test documents only

Predict links within test documents

Predict links from test documents to training documents

14/22

Page 39: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Task

Training input: Training documents with links

Test input: Test documents only

Predict links within test documents

Predict links from test documents to training documents

Training Corpus

14/22

Page 40: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Task

Training input: Training documents with links

Test input: Test documents only

Predict links within test documents

Predict links from test documents to training documents

Training Corpus Test Corpus

14/22

Page 41: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Task

Training input: Training documents with links

Test input: Test documents only

Predict links within test documents

Predict links from test documents to training documents

Training Corpus Test Corpus

14/22

Page 42: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Task

Training input: Training documents with links

Test input: Test documents only

Predict links within test documents

Predict links from test documents to training documents

Training Corpus Test Corpus

14/22

Page 43: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Evaluation Metric

Predictive link rank (PLR)

For a document d , we compute and sort all other documentsby their link probabilities to dThen compute the average rank of actually linked documentsPLR=(2+3+5)/3=3.33

15/22

Page 44: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Evaluation Metric

Predictive link rank (PLR)

For a document d , we compute and sort all other documentsby their link probabilities to d

Then compute the average rank of actually linked documentsPLR=(2+3+5)/3=3.33

Rank Doc ID Link Probability1 5 0.902 3 0.853 2 0.824 4 0.705 6 0.636 1 0.50

15/22

Page 45: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Evaluation Metric

Predictive link rank (PLR)

For a document d , we compute and sort all other documentsby their link probabilities to dThen compute the average rank of actually linked documents

PLR=(2+3+5)/3=3.33

Rank Doc ID Link Probability1 5 0.902 3 0.853 2 0.824 4 0.705 6 0.636 1 0.50

True Link?

YesYes

Yes

15/22

Page 46: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Evaluation Metric

Predictive link rank (PLR)

For a document d , we compute and sort all other documentsby their link probabilities to dThen compute the average rank of actually linked documentsPLR=(2+3+5)/3=3.33

Rank Doc ID Link Probability1 5 0.902 3 0.853 2 0.824 4 0.705 6 0.636 1 0.50

True Link?

YesYes

Yes

15/22

Page 47: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Predictive Link Rank Results

300 350 400 450 500

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

100 110 120 130 140 150 160

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

Incrementally add components onRTM framework

Block priorsLexical weightsHinge loss

Each component contributes tolink prediction improvement

Strongly connected componentsruin the link prediction

16/22

Page 48: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Predictive Link Rank Results

300 350 400 450 500

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

100 110 120 130 140 150 160

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

Incrementally add components onRTM framework

Block priors

Lexical weightsHinge loss

Each component contributes tolink prediction improvement

Strongly connected componentsruin the link prediction

16/22

Page 49: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Predictive Link Rank Results

300 350 400 450 500

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

100 110 120 130 140 150 160

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

Incrementally add components onRTM framework

Block priorsLexical weights

Hinge loss

Each component contributes tolink prediction improvement

Strongly connected componentsruin the link prediction

16/22

Page 50: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Predictive Link Rank Results

300 350 400 450 500

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

100 110 120 130 140 150 160

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

Incrementally add components onRTM framework

Block priorsLexical weightsHinge loss

Each component contributes tolink prediction improvement

Strongly connected componentsruin the link prediction

16/22

Page 51: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Predictive Link Rank Results

300 350 400 450 500

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

100 110 120 130 140 150 160

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

Incrementally add components onRTM framework

Block priorsLexical weightsHinge loss

Each component contributes tolink prediction improvement

Strongly connected componentsruin the link prediction

16/22

Page 52: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Predictive Link Rank Results

300 350 400 450 500

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

100 110 120 130 140 150 160

LCH-RTM

LBH-RTM

LBS-RTM

BS-RTM

RTM

Incrementally add components onRTM framework

Block priorsLexical weightsHinge loss

Each component contributes tolink prediction improvement

Strongly connected componentsruin the link prediction

16/22

Page 53: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Link Example

Using Fourier-Neural Recurrent Networks

to Fit Sequential Input/Output Data

Renee KoplonDept. of Mathematics and Statistics

Wright State UniversityDayton, Ohio 45435

Eduardo D. SontagDept. of Mathematics

Rutgers UniversityNew Brunswick, New Jersey 08903

March 1, 1995Revised: September 15, 1996

Abstract

This paper suggests the use of Fourier-type activation functions in fully recurrentneural networks. The main theoretical advantage is that, in principle, the problem ofrecovering internal coefficients from input/output data is solvable in closed form.

Keywords: recurrent neural networks, identification, nonlinear dynamics

1 Introduction

Neural networks provide a useful approach to parallel computation. The subclass of recur-rent architectures is characterized by the inclusion of feedback loops in the information flowamong processing units. With feedback, one may exploit context-sensitivity and memory,characteristics essential in sequence processing as well as in the modeling and control ofprocesses involving dynamical elements. Recent theoretical results about neural networkshave established their universality as models for systems approximation as well as analogcomputing devices (see e.g. [16, 13]).

The use of recurrent networks has been proposed in areas as varied as the design ofcontrol laws for robotic manipulators, in speech recognition, speaker identification, formallanguage inference, and sequence extrapolation for time series prediction. In spite of theirattractive features, recurrent networks have not yet attained as much popularity as onemight expect, compared to the feedforward nets so ubiquitous in other applications. Oneimportant reason for this is that training (“learning”) algorithms for recurrent nets sufferfrom serious potential limitations. The learning problem is that of finding parameters

1

Paper 1 [Koplon and Sontag, 1997]

Report SYCON-92-03

FOR NEURAL NETWORKS, FUNCTION DETERMINES FORM

Francesca Albertini

Eduardo D. Sontag

Department of Mathematics Rutgers University, New Brunswick, NJ 08903

E-mail: [email protected], [email protected]

ABSTRACT

This paper shows that the weights of continuous-time feedback neural networks are uniquely

identiable from input/output measurements. Under very weak genericity assumptions, the

following is true: Assume given two nets, whose neurons all have the same nonlinear activation

function ; if the two nets have equal behaviors as \black boxes" then necessarily they must

have the same number of neurons and |except at most for sign reversals at each node| the

same weights.

Key words: Neural networks, realization from input/output data, control systems

Research supported in part by US Air Force Grant AFOSR-91-0343.

Rutgers Center for Systems and Control

May 1992

Paper 2 [Albertini and Sontag, 1992]

17/22

Page 54: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Link Example — RTM

0.0 0.2 0.4 0.6 0.8 1.0

NN-1

NN-2

2 1

Topic Proportions by RTM

Model Link RankRTM 1,265

NN-1: network, neural, compute, activation, pattern, modelNN-2: network, neural, learn, train, algorithm, local, weight

18/22

Page 55: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Link Example — RTM

0.0 0.2 0.4 0.6 0.8 1.0

NN-1

NN-2

2 1

Topic Proportions by RTM

Model Link RankRTM 1,265

NN-1: network, neural, compute, activation, pattern, modelNN-2: network, neural, learn, train, algorithm, local, weight

18/22

Page 56: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Link Example — BS-RTM

0.0 0.2 0.4 0.6 0.8 1.0

NN-1

2 1

Topic Proportions by BS-RTM

Model Link RankRTM 1,265

BS-RTM 635

NN: network, neural, learn, train, algorithm, weight, input

18/22

Page 57: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Link Example — LBS-RTM

0.0 0.2 0.4 0.6 0.8 1.0

NN-1

NN-2

2 1

Topic Proportions by LBS-RTM

Model Link RankRTM 1,265

BS-RTM 635

LBS-RTM 132

NN-1: network, neural, learn, train, weight, input, architectureNN-2: learn, model, agent, reinforce, action, generate, strategy

18/22

Page 58: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Link Example — LBS-RTM

0.0 0.2 0.4 0.6 0.8 1.0

NN-1

NN-2

2 1

Topic Proportions by LBS-RTM

Model Link RankRTM 1,265

BS-RTM 635

LBS-RTM 132

NN-1: network, neural, learn, train, weight, input, architectureNN-2: learn, model, agent, reinforce, action, generate, strategy

18/22

Page 59: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Link Example — LBH-RTM

0.0 0.2 0.4 0.6 0.8 1.0

NN-1

2 1

Topic Proportions by LBH-RTM

Model Link RankRTM 1,265

BS-RTM 635

LBS-RTM 132

LBH-RTM 106

NN: network, neural, train, learn, function, generate, weight

18/22

Page 60: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Link Example — LCH-RTM

0.0 0.2 0.4 0.6 0.8 1.0

NN-1

NN-2

2 1

Topic Proportions by LCH-RTM

Model Link RankRTM 1,265

BS-RTM 635

LBS-RTM 132

LBH-RTM 106

LCH-RTM 1,385

NN-1: network, model, belief, algorithm, function, approximationNN-2: network, neural, train, learn, algorithm, weight, result

18/22

Page 61: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Link Example — LCH-RTM

0.0 0.2 0.4 0.6 0.8 1.0

NN-1

NN-2

2 1

Topic Proportions by LCH-RTM

Model Link RankRTM 1,265

BS-RTM 635

LBS-RTM 132

LBH-RTM 106

LCH-RTM 1,385

NN-1: network, model, belief, algorithm, function, approximationNN-2: network, neural, train, learn, algorithm, weight, result

18/22

Page 62: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Outline

1 Block Detection and RTM

2 LBH-RTM

3 Link Prediction Results

4 ConclusionsPaper

http://ter.ps/bv5Slides

http://ter.ps/bv7

19/22

Page 63: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Summary

LBH-RTM

Topic model for link predictionIncorporate block priors from linksInclude lexical weights and hinge loss

Future directions

Directed/undirected linksBinary/nonnegative real weight linksLink suggestion

20/22

Page 64: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

Thanks

Collaborators

Jordan Boyd-Graber (UC Boulder)

Philip Resnik (UMD)

Funders

21/22

Page 65: A Discriminative Topic Model using Document Network Structurewwyang/files/2016_acl_docblock_slides.pdf · models. 124 Relational Topic Model [Chang and Blei, 2010]: Topic Model, Document

References

Aicher, C., Jacobs, A. Z., and Clauset, A. (2014).Learning latent block structure in weighted networks.Journal of Complex Networks.

Albertini, F. and Sontag, E. D. (1992).For neural networks, function determines form.In Proceedings of IEEE Conference on Decision and Control.

Blei, D. M., Ng, A. Y., and Jordan, M. I. (2003).Latent Dirichlet allocation.Journal of Machine Learning Research.

Chang, J. and Blei, D. M. (2010).Hierarchical relational models for document networks.The Annals of Applied Statistics.

Koplon, R. and Sontag, E. D. (1997).Using Fourier-neural recurrent networks to fit sequential input/output data.Neurocomputing.

Yang, W., Boyd-Graber, J., and Resnik, P. (2015).Birds of a feather linked together: A discriminative topic model using link-basedpriors.In Proceedings of Empirical Methods in Natural Language Processing.

22/22