transfer learning in heterogeneous collaborative filtering domains

28
Transfer learning in heterogeneous collaborative filtering domains Authors/ Weike Pan and Qiang Yang Affiliation/ Dept. of CSE, Hong Kong University of Science and Technology Source/ Journal of Artificial Intelligence (2013) Presenter/ Allen Wu 2013/3/27 1

Upload: allen-wu

Post on 14-Jul-2015

2.170 views

Category:

Education


1 download

TRANSCRIPT

Transfer learning in

heterogeneous collaborative

filtering domainsAuthors/ Weike Pan and Qiang Yang

Affiliation/ Dept. of CSE, Hong Kong University of Science and Technology

Source/ Journal of Artificial Intelligence (2013)

Presenter/ Allen Wu

20

13

/3/2

7

1

Outline

• Introduction

• Heterogeneous collaborative filtering problems

• Transfer by collective factorization

• Experimental results

• Conclusion

20

13

/3/2

7

2

Introduction

• Data sparsity is a major challenge in collaborative filtering (CF).

• Overfitting can easily happen for prediction.

• Some auxiliary data of the form “like” or “dislike” may be more

easily obtained.

• It’s more convenient for users to express preference.

• How do we take advantage of auxiliary knowledge to alleviate the

sparsity problem?

• Most existing transfer learning methods in CF consider auxiliary data from

several perspectives.

• User-side transfer, item-side transfer, knowledge-transfer.

20

13

/3/2

7

3

Probabilistic Matrix Factorization

(NIPS’08)

20

13

/3/2

7

4

Social Recommendation (CIKM’08)

20

13

/3/2

7

5

Collective Matrix Factorization (KDD’08)

20

13

/3/2

7

6

CodeBook Transfer (IJCAI’09)

20

13

/3/2

7

7

Rating-matrix generative model (ICML’09)

• RMGM is derived and extended from FMM generative model,

which can be formulated as

• The difference:

• It learns (U, V) and (U3, V3) alternatively.

• A soft indicator matrix is used. E.g., U [0, 1]n d.

20

13

/3/2

7

8

Heterogeneous collaborative filtering

problems

• •

20

13

/3/2

7

9

Challenges

20

13

/3/2

7

10

Overview of solution

20

13

/3/2

7

11

Model formulation

• Assume a user u’s rating on an item i in the target data, rui, is

generated from

• user-specific latent feature vector Uu1 d, where u=1,…,n.

• item-specific latent feature vector Vi1 d, where i=1,…,m.

• some data-dependent effect denoted as B d d.

20

13

/3/2

7

12

Model formulation (Cont.)

• Likelihood:

• Prior:

• Posterior Likelihood Prior (Bayesian inference)

• Log(Posterior)= Log(Likelihood Prior)

20

13

/3/2

7

13

Model formulation

20

13

/3/2

7

14

Learning the TCF

20

13

/3/2

7

15

Learning U and V in CMTF• Theorem 1. Given B and V, we can obtain the user-specific

latent matrix U in a closed form.

20

13

/3/2

7

16

Learning U and V in CSVD

20

13

/3/2

7

17

Learning U and V in CSVD

(Cont.)

20

13

/3/2

7

18

20

13

/3/2

7

19

Algorithm of TCF

20

13

/3/2

7

20

Data sets

20

13

/3/2

7

21

Evaluation metrics

• Summary of Data sets

• Evaluation metrics

20

13

/3/2

7

22

Baselines and parameter settings

20

13

/3/2

7

23

Performance of Moviepilot data

20

13

/3/2

7

24

Performance of Netfliex data

20

13

/3/2

7

25

Performance on Netflix at different

sparsity levels

20

13

/3/2

7

26

• SCVD performs

better than CMTF in

all cases.

Conclusion

• This paper investigate how to address the sparsity problem in

CF via a transfer learning solution.

• The TCP framework is proposed to transfer knowledge from

auxiliary data to target data to alleviates the data sparsity.

• Experimental results show that TCP performs significantly

better than several state-of-the-art baseline algorithms.

• In the future, the “pure” cold-start problem for users without

any rating is needed to be addressed via transfer learning.

20

13

/3/2

7

27

Thank you for listening.Q & A

20

13

/3/2

7

28