Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative...

52
Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University of Toronto Collaborative Filtering, Missing Data, and Ranking CSC2535 Richard Zemel March 13, 2013

Transcript of Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative...

Page 1: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 1

Collaborative Filtering, Missing Data, and Ranking

CSC2535

Richard Zemel

March 13, 2013

Page 2: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 2

Contents:

Introduction

Notation Theory Of Missing Data

Factorizations MCAR MAR NMAR Inference and Learning

Multinomial Models Multinomial Mixture Multinomial Mixture/CPT-v

Collaborative Filtering Expts. Yahoo! Data

Jester Data Results

Collaborative Ranking

Page 3: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 3

Introduction

Page 4: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 4

Introduction: Collaborative Filtering

Collaborative filtering – users assign ratings to items à system uses information from all users to recommend previously unseen items that a user might like

One approach to recommendation: predict ratings for all unrated items, recommend highest predicted ratings

Page 5: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 5

Collaborative Filtering: Collaborative Prediction Problem

??

??

? ?

Page 6: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 6

Introduction: Missing Data Critical assumption: missing ratings are missing at random One way to violate: value of variable affects probability value will be missing – bias in observed ratings, and hence learned parameters Also complementary bias in standard testing procedure – distribution of observed data different from distribution of complete data, so estimated error on observed test data poor estimate of complete data error

Page 7: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 7

Introduction: Survey Sampling Example

Data Variables

Response Variables

R

X Answers to questions.

Did the respondent answer the question?

Page 8: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 8

Introduction: Medical Diagnosis Example

Latent Variables

Data Variables

Response Variables

R

X

Z Diseases

Symptoms

Was a test performed to check for the symptom?

Page 9: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 9

Introduction: Recommender Systems Example

Latent Variables

Data Variables

Response Variables

R

X

Z Preferences and Tastes

Ratings or Purchase History

Did the user rate or buy the item?

Page 10: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 10

Introduction: Basic Notation

Number of data cases.

Number of clusters or hidden units.

Number of multinomial values.

Number of classes.

Number of data dimensions.

Page 11: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 11

Missing Data

Observed Data

Missing Dimensions

Observed Dimensions

Response Vector

Data Vector

Introduction: Notation for Missing Data 0.3 0.7 0.2 0.9 0.1

1 1 0 0 1

5 4 1

3 2

0.3 0.7 0.1

0.2 0.9

Page 12: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 12

Contents:

Introduction

Notation Theory Of Missing Data

Factorizations MCAR MAR NMAR Inference and Learning

Multinomial Models Multinomial Mixture Multinomial Mixture/CPT-v

Collaborative Filtering Expts. Yahoo! Data

Jester Data Results

Collaborative Ranking

Page 13: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 13

Theory of Missing Data: Generative Process

Latent Variables

Data Variables

Response Variables

Data Model

R

X

Z

Missing Data Model

R

X

Z

µ

θ

Page 14: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 14

Theory of Missing Data: Factorizations Data/Selection Model Factorization:

•  The probability of selection depends on the true values of the data variables and latent variables.

Pattern Mixture Model Factorization:

•  Each response vector defines a different pattern, and each pattern has a different distribution over the data.

Page 15: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 15

Theory of Missing Data: Classification

MCAR:

R

X

Z

µ

θ

Missing Completely at Random:

•  Response probability is independent of data variables and latent variables.

MCAR Examples:

Send questionnaires to a random subset of the population or use random digit dialing.

Page 16: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 16

Theory of Missing Data: Classification Missing at Random:

•  Typically written in a short-hand form that looks like a statement of probabilistic independence:

•  MAR is actually a different type of condition that requires a particular set of symmetries hold in P(r|x,z,µ):

Page 17: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 17

Theory of Missing Data: Classification Missing at Random Examples:

Respondents are not required to provide information about their employer if they are not currently employed.

Doctor only orders test B if the result of test A was negative. If result of test A is positive, result for test B is missing.

Page 18: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 18

Theory of Missing Data: Classification What Does it mean to be Missing at Random?

•  MAR is not a statement of independence between random variables. MAR requires that particular symmetries hold so that P(R=r|X=x) can be determined from observed data only.

Page 19: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 19

Theory of Missing Data: Classification

MCAR:

R

X

Z

µ

θ

Not Missing at Random:

•  Allows for arbitrary dependence of response probabilities on missing data values and latent variables:

No Simplifications

An Easy Way to Violate MAR:

•  Let the probability that a data variable is observed depend on the value of that data variable.

Page 20: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 20

Theory of Missing Data: Classification Not Missing at Random Examples:

Snowfall reading is likely to be missing if weather station is covered with snow.

Users are more likely to rate or buy items they like than items they don’t like.

Participants in a longitudinal health study for a heart medication may die of a heart attack during the study.

Page 21: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 21

Theory of Missing Data: Inference MCAR/MAR Posterior:

•  When MCAR or MAR holds, the posterior can be greatly simplified. Inference for θ does not depend on r, µ, or η. The missing data can be ignored.

Page 22: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 22

Theory of Missing Data: Inference NMAR Posterior:

•  When MAR fails to hold, the posterior does not simplify.

•  Basing inference on the observed data posterior and ignoring the missing data model leads to provably biased inference for data model parameters.

Page 23: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 23

Contents:

Introduction

Notation Theory Of Missing Data

Factorizations MCAR MAR NMAR Inference and Learning

Multinomial Models Multinomial Mixture Multinomial Mixture/CPT-v

Collaborative Filtering Expts. Yahoo! Data

Jester Data Results

Collaborative Ranking

Page 24: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 24

Multinomial Models: Mixture

Probability Model:

Properties: •  Allows for a fixed, finite number of clusters.

•  In the multinomial mixture, P(xn|βk) is a product of discrete distributions. The prior on β and θ is Dirichlet.

Page 25: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 25

Multinomial Models: Mixture Dirichlet Distribution:

Bayesian mixture modeling becomes much easier when conjugate priors are used for the model parameters. The conjugate prior for the mixture proportions θ is the Dirichlet distribution.

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10

0.5

1

1.5

2

2.5

3

3.5

4

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10

0.5

1

1.5

2

2.5

3

3.5

4

Page 26: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 26

Multinomial Models: Mixture MAP EM Algorithm:

Page 27: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 27

Multinomial Models: Mixture/CPT-v Probability Model:

Page 28: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 28

Multinomial Models: Mixture/CPT-v Probability Model:

Page 29: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 29

Multinomial Models: Mixture/CPT-v MAP EM Algorithm (E-Step):

Page 30: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 30

Multinomial Models: Mixture/CPT-v MAP EM Algorithm (M-Step):

Page 31: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 31

DEMO Multinomial Mixture Learning With

Random and Non-Random Missing Data

Page 32: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 32

Other Models for Missing Data: •  K-Nearest Neighbors •  Probabilistic Principal Components Analysis •  Factor Analysis •  Mixtures of Gaussians •  Mixtures of PPCA/FA •  Probabilistic Matrix Factorization •  Maximum Margin Matrix Factorization •  Conditional Restricted Boltzmann Machines

Page 33: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 33

Contents:

Introduction

Notation Theory Of Missing Data

Factorizations MCAR MAR NMAR Inference and Learning

Multinomial Models Multinomial Mixture Multinomial Mixture/CPT-v

Collaborative Filtering Expts. Yahoo! Data

Jester Data Results

Collaborative Ranking

Page 34: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 34

Collaborative Filtering: Collaborative Prediction Problem

??

??

? ?

Page 35: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 35

Collaborative Filtering : Yahoo! Data was collected through an online survey of Yahoo! Music LaunchCast radio users.

•  1000 songs selected at random.

•  Users rate 10 songs selected at random from 1000 songs.

•  Answer 16 questions.

•  Collected data from 35,000+ users.

Image copyright Yahoo! Inc. 2006. Used with permission.

Page 36: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 36

Collaborative Filtering: Yahoo!

User Selected Randomly Selected

Page 37: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 37

More Empirical Distributions

Page 38: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 38

Collaborative Filtering: Jester

Jester gauge set of 10 jokes used as complete data. Synthetic missing data was added.

•  15,000 users randomly selected

•  Missing data model: µv(s) = s(v-3)+0.5

Page 39: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 39

Experimental Protocol Randomly partition users into 5 blocks of 1080 users Three sets of ratings:

1.  Observed ratings – all but one of original ratings 2.  Test ratings for user-selected – remaining one 3.  Test ratings for randomly-selected – ten survey

responses

User-selected items – same distribution as observed Randomly selected test items -- MCAR

Page 40: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 40

Experimental Protocol

Weak Generalization •  Learn on training user observed ratings •  Evaluate on training user test ratings

Strong Generalization •  Learn on training user observed ratings •  Evaluate on test user test ratings

Page 41: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 41

Data Sets: User Splits

Page 42: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 42

Data Sets: User Splits

Page 43: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 43

Collaborative Filtering: Results Jester Results: MM vs MM/CPT-v

Page 44: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 44

Collaborative Filtering: Baselines

1.  Item-based K-nearest neighbor (iKNN) 2.  cRBM 3.  Matrix factorization

Standard CF methods implicitly assume MAR Here we compare to three other CF methods:

Page 45: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 45

Conditional RBM for CF Restricted Boltzmann Machines for Collaborative Filtering

Missing

Missing

Missing

Missing

W

Binary hidden

......

features

V

......

h

r

ratingsmovieVisible

D

Figure 2. Conditional RBM. The binary vector r, indi-cating rated/unrated movies, a!ects binary states of thehidden units.

missing ratings and it does not perform any computa-tions that scale with the number of missing ratings.

There is a more subtle source of information in theNetflix database that cannot be captured by the “stan-dard” multinomial RBM. Netflix tells us in advancewhich user/movie pairs occur in the test set, so wehave a third category: movies that were viewed butfor which the rating is unknown. This is a valuablesource of information about users who occur severaltimes in the test set, especially if they only gave asmall number of ratings in the training set. If, for ex-ample, a user is known to have rated “Rocky 5”, wealready have a good bet about the kinds of movies helikes.

The conditional RBM model takes this extra informa-tion into account. Let r ! {0, 1}M be a binary vec-tor of length M (total number of movies), indicatingwhich movies the user rated (even if these ratings areunknown). The idea is to define a joint distributionover (V,h) conditional on r. In the proposed condi-tional model, a vector r will a!ect the states of thehidden units (see Fig. 2):

p(vki = 1|h) =

exp (bki +

!Fj=1 hjW k

ij)!K

l=1 exp"bli +

!Fj=1 hjW l

ij

#

p(hj = 1|V, r) = !

$bj +

m%

i=1

K%

k=1

vki W k

ij +M%

i=1

riDij

&'

where Dij is an element of a learned matrix that mod-els the e!ect of r on h. Learning D using CD is similar

to learning biases and takes the form:

"Dij = "

$<hj>data " <hj>T

'ri (12)

We could instead define an arbitrary nonlinear func-tion f(r|#). Provided f is di!erentiable with respectto #, we could use backpropagation to learn #:

"# = "

$<hj>data " <hj>T

'$f(r|#)

$#(13)

In particular, f(r|#) can be parameterized as a multi-layer neural network.

Conditional RBM models have been successfully usedfor modeling temporal data, such as motion cap-ture data (Taylor et al., 2006), or video sequences(Sutskever & Hinton, 2006). For the Netflix task, con-ditioning on a vector of rated/unrated movies provesto be quite helpful – it significantly improves perfor-mance.

Instead of using a conditional RBM, we can imputethe missing ratings from the ordinary RBM model.Suppose a user rated a movie t, but his/her rating ismissing (i.e. it was provided as a part of the test set).We can initialize vt to the base rate of movie t, andcompute the gradient of the log-probability of the datawith respect to this input (Eq. 3). The CD learningtakes form:

"vkt = "

$<

%

j

W kt hj>data " <

%

j

W kt hj>T

'

After updating vkt , for k = 1, ..,K, vk

t are renormalizedto obtain probability distribution over K values. Theimputed values vt will now contribute to the energyterm of Eq. 4 and will a!ect the states of the hiddenunits. Imputing missing values by following an ap-proximate gradient of CD works quite well on a smallsubset of the Netflix data set, but is slow for the com-plete data set. Alternatively, we can use a set of meanfield equations Eqs. 9, 10 to impute the missing val-ues. The imputed values will be quite noisy, especiallyat the early stages of training. Nevertheless, in ourexperiments, the model performance was significantlyimproved by using imputations and was comparable tothe performance of the conditional RBM.

5. Conditional Factored RBM’s

One disadvantage of the RBM models we have de-scribed so far is that their current parameterization ofW ! RM!K!F results in a large number of free param-eters. In our current implementation, with F = 100

Page 46: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 46

Probabilistic Matrix Factorization Probabilistic Matrix Factorization

1234567...

1 2 3 4 5 6 7 ...5 3 ? 1 ... 3 ? 4 ? 3 2 ...

~~R U

V

UserFeatures

FeaturesMovie

• Let Rij represent the rating of user i for movie j. The row andcolumn vectors Ui and Vj represent user-specific and movie-specificlatent feature vectors respectively.

• The model:

p(Rij|Ui, Vj,!2) = N (Rij|U

Ti Vj,!

2)

6

Page 47: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 47

Collaborative Filtering: Results Comparison of Results on Yahoo! Data

Page 48: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 48

Collaborative Filtering: Results Comparison of Results on Yahoo! Data

!"## $% $$ &'()* +,-!.)*/012

013

014

015

016

017

8#$9:

;.<,=->?@=/>#$9:

(a) NMAE!" !# !$ !% !& !' !( !) !* !"+

+,%&

+,&

+,&&

+,'

+,'&

+,(

+,(&

+,)

+,)&

+,*

-

./01!-

23456./01!-6786-

9:990;<=7->?@A=75

@B..

(b) NDCG@L

Figure 4: Figure (a) presents the test set rating prediction error on randomly selected items. Figure (b) presents the test setNDCG@L results on randomly selected items. The methods are iKNN, MF, MM/MAR, MM/CPT-v, and MM/Logit-vd.

Figure 4(b) shows the test NDCG@L performance for eachmodel estimated on lists of ratings for the 10 randomly se-lected items. We select the optimal complexity for eachmodel based on cross validation NDCG@L scores. The re-sults again show that the MM/Logit-vd and MM/CPT-v mod-els, which do not assume the MAR condition, outperform theMM, iKNN and MF models, which do assume random miss-ing data.

7 ConclusionsIn this paper, we have explored properties of the missing dataprocess in recommender systems, discussed their impact onthe validity of standard statistical model estimation and eval-uation procedures, and described and tested extended mod-eling and evaluation frameworks that seek to overcome theproblems caused by non-random missing data. The develop-ment of more sophisticated models within the extended mod-eling framework is of great interest, as is the design of bet-ter test sets for ranking. The question of how non-randommissing data affects methods for learning to rank is also ofgreat interest. The continued convergence of recommendersystems, content-based search and social networks raises thequestion of the extent to which the detrimental effects of non-randommissing ratings can be mitigated by incorporating ad-ditional sources of information including content-based fea-tures for items (including social tags) and information aboutboth individual users and the relationships between users.This is a very interesting direction for future research.

AcknowledgmentsWe are indebted to Sam Roweis for his contributions to thisresearch. He was a wonderful colleague, mentor and friend.We thank the Natural Sciences and Engineering ResearchCouncil of Canada, the Killam Trusts, the Pacific Institutefor the Mathematical Sciences and the Canadian Institute forAdvanced Research for funding this work. We thank Yahoo!for making the data set used in this work, Yahoo! Music Rat-ings for User Selected and Randomly Selected songs, V 1.0,available through the Webscope Program.

References[Breese et al., 1998] J. S. Breese, D. Heckerman, andC. Kadie. Empirical Analysis of Predictive Algorithms forCollaborative Filtering. In Proceedings of the 14th Con-ference on Uncertainty in Artificial Intelligence, pages 43–52, 1998.

[Brin and Page, 1998] S. Brin and L. Page. The anatomy ofa large-scale hypertextual web search engine. In Proceed-ings of the seventh international conference on World WideWeb 7, pages 107–117, 1998.

[Dempster et al., 1977] A.P. Dempster, N.M. Laird, andD.B. Rubin. Maximum likelihood from incomplete datavia the EM algorithm. Journal of the Royal Statistical So-ciety, Series B, 39(1):1–38, 1977.

[Goldberg et al., 1992] D. Goldberg, D. Nichols, B. M. Oki,and D. Terry. Using collaborative filtering to weave aninformation tapestry. Communications of the ACM, 35:61–70, December 1992.

[Little and Rubin, 1987] R. J. A. Little and D. B. Rubin. Sta-tistical Analysis with Missing Data. John Wiley and Sons,Inc., 1987.

[Marlin and Zemel, 2009] B. M. Marlin and R. S. Zemel.Collaborative prediction and ranking with non-randommissing data. In Proceedings of the third ACM conferenceon Recommender systems, pages 5–12, 2009.

[Marlin et al., 2007] B. Marlin, R. Zemel, S. Roweis, andM. Slaney. Collaborative filtering and the missing at ran-dom assumption. In Uncertainty in Artificial Intelligence23, 2007.

[Salakhutdinov and Mnih, 2008] R. Salakhutdinov andA. Mnih. Probabilistic matrix factorization. In Advancesin Neural Information Processing Systems, volume 20,2008.

[Sarwar et al., 2001] B. Sarwar, G. Karypis, J. Konstan, andJ. Reidl. Item-based collaborative filtering recommenda-tion algorithms. In Proceedings of the 10th internationalconference on World Wide Web, pages 285–295, 2001.

Page 49: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 49

Application to Ranking

imization (EM) algorithm [Dempster et al., 1977]. The per-iteration complexity of the algorithm scales linearly with thenumber of observed ratings. In the case of the multinomialmixture model combined with the CPT-v and Logit-vd miss-ing data models, efficient EM algorithms can be also derivedwhere the computational complexity per iteration is domi-nated by the number of observed ratings. This is the main ad-vantage of using simplified missing data models. We use theEM algorithm to simultaneously learn the parameters of boththe data and missing data models in all of the experimentsdescribed in the following sections. Further details regard-ing the estimation and prediction algorithms for these mod-els can be found in our previous work [Marlin et al., 2007;Marlin and Zemel, 2009].

5 Evaluation ProtocolsAs described in Section 3, standard empirical protocols forrating prediction and ranking evaluation can lead to biasedperformance estimates in the presence of non-random miss-ing data, necessitating modified empirical protocols. In thecase of rating prediction, we require a test set that is as closeas possible to a random selection of unrated items. The rat-ings for randomly selected items collected during the Yahoo!Music user study described in Section 2 provide just such atest set since the expected overlap between randomly-selecteditems and previously-rated items is low.We also propose the use of ratings for randomly selected

items for the evaluation of ranking accuracy, although thischoice presents some issues. In particular, since we only haveten items per user and most of the items in the test set havelow ratings, the ranking evaluation may unduly reflect themodel’s ability to discriminate between items with low rat-ing values. However, we feel this is preferable to measuringthe ranking performance on a subset of the observed data sub-ject to a completely unknown observation process. Whether itis possible to construct better test sets for ranking evaluationgiven both sources of ratings is an open question.The full empirical protocol uses a training set containing

the 5400 users who participated our study (described in Sec-tion 2), plus an additional 10000 LaunchCast users selectedat random from those with at least 10 ratings on the 1000songs used in the study data set. All of the training set rat-ings are ratings for items selected by the users during normalinteraction with the LaunchCast music recommender system(Figure 1(e)). The validation and test sets contain ratings sub-sampled from the ratings for randomly-selected items col-lected during the user study for each of the 5400 study users(Figure 1(f)).The models we evaluate include the multinomial mix-

ture model under the MAR assumption (MM/MAR), as wellas the multinomial mixture model combined with the CPT-v (MM/CPT-v) and Logit-vd (MM/Logit-vd) missing datamodels. We also evaluate two very common collaborativefiltering models that implicitly assume random missing data:a matrix factorization (MF) model [Salakhutdinov and Mnih,2008] and an item-based K-nearest neighbor method (iKNN)[Sarwar et al., 2001].We train each mixture-based model using 1, 5, 10 and 20

mixture components and select the best setting using crossvalidation. The prior parameters for all of the mixture-basedmodels were set to yield broad priors. For the matrix fac-torization model, we considered ranks K = 1, 5, 10, 20 andregularization parameters 0.1, 1, 5, 10 and selected the bestvalues by cross validation. For the item-based KNN method,we use an adjusted cosine similarity metric [Sarwar et al.,2001], combined with the standard weighted nearest neigh-bor prediction rule.Once the models are trained, we condition on the training

set ratings for each user and predict the ratings for each ofthat user’s test items. We form a ranked list of test items foreach user by sorting that user’s test items according to theirpredicted ratings.

6 ResultsRating Prediction: We evaluate rating prediction perfor-mance in terms of normalized mean absolute error (NMAE).This error measure is proportional to the average absolute dif-ference between actual and predicted ratings. NMAE is com-puted as seen below. We assume there are T test items peruser with indices i(1, n) to i(T, n). The normalizing constant(equal to 1.6) is the expected MAE assuming uniformly dis-tributed predictions and true ratings. Note that lower NMAEindicates better prediction performance.

NMAE =N!

n=1

T!

t=1

|xni(t,n) ! x̂ni(t,n)|

1.6NT

Figure 4(a) shows the test NMAE score for each of the fivemodels. We select the optimal complexity for each modelbased on cross validation NMAE scores. Standard errorsare represented on the plots using error bars. We see thatthe MM/Logit-vd and MM/CPT-v models, which do not as-sume the MAR condition, drastically outperform the MM,iKNN and MF models, which do assume random missingdata, when measuring performance on ratings for randomlyselected items.Ranking: We evaluate the ranked lists of test items

produced by each method using a standard ranking accu-racy measure, the normalized discounted cumulative gain(NDCG). NDCG@L measures how well the predicted rank-ing matches the true ranking (obtained by sorting the itemsby their actual ratings) for a ranked list of length L. NDCGplaces more emphasis on ranking errors at the top of the or-dering and is normalized so that the true ranking yields anaccuracy of 1. The NDCG@L score is computed as seen be-low where !(l, n) is the index of the item with rank l whentest items are sorted in descending order by true rating xnd,!̂(l, n) is the index of the item with rank l when items aresorted in descending order according to their predicted rat-ings x̂nd. When sorting by true and predicted ratings, tiescan be broken arbitrarily without affecting the NDCG@Lscore. Note that higher NDCG indicates better ranking per-formance.

NDCG@L =N!

n=1

"Ll=1(2

xn!̂(l,n) ! 1)/ log(1 + l)

N"L

l=1(2xn!(l,n) ! 1)/ log(1 + l)

Page 50: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 50

Ranking Results

Page 51: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 51

Conclusions

In real recommender system data, the standard missing-at-random assumption is typically violated Methods that include explicit non-random missing data model out-perform methods that assume MAR

In practice, the important task is collaborative *ranking*, not rating prediction Our recent results show that combinations of neighbor- and model-based approaches to collaborative ranking permits scaling to large datasets

Page 52: Collaborative Filtering, Missing Data, and Rankinghinton/csc2535/notes/cf-2535.pdf · Collaborative Filtering, Missing Data, and Ranking CSC2535, Department of Computer Science, University

Collaborative Filtering, Missing Data, and Ranking

CSC2535, Department of Computer Science, University of Toronto 52

Collaborative Ranking Results