geometric matrix completion with recurrent multi-graph neural … · 2017. 4. 25. · geometric...

9
Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks Federico Monti 1 Michael M. Bronstein 1234 Xavier Bresson 5 Abstract Matrix completion models are among the most common formulations of recommender systems. Recent works have showed a boost of perfor- mance of these techniques when introducing the pairwise relationships between users/items in the form of graphs, and imposing smoothness pri- ors on these graphs. However, such techniques do not fully exploit the local stationarity struc- tures of user/item graphs, and the number of pa- rameters to learn is linear w.r.t. the number of users and items. We propose a novel approach to overcome these limitations by using geomet- ric deep learning on graphs. Our matrix comple- tion architecture combines graph convolutional neural networks and recurrent neural networks to learn meaningful statistical graph-structured patterns and the non-linear diffusion process that generates the known ratings. This neural network system requires a constant number of parame- ters independent of the matrix size. We apply our method on both synthetic and real datasets, showing that it outperforms state-of-the-art tech- niques. 1. Introduction Recommender systems have become a central part of mod- ern intelligent systems. Recommending movies on Net- flix, friends on Facebook, furniture on Amazon, jobs on LinkedIn are a few examples of the main purpose of these systems. Two major approach to recommender systems are collaborative (Breese et al., 1998) and content (Pazzani & Billsus, 2007) filtering techniques. Systems based on col- laborative filtering use collected ratings of products by cus- tomers and offer new recommendations by finding similar rating patterns. Systems based on content filtering make 1 ICS USI Lugano, Switzerland 2 Tel Aviv University, Is- rael 3 Intel Perceptual Computing, Israel 4 TUM IAS, Germany 5 NTU, Singapore. Correspondence to: Federico Monti <fed- [email protected]>. Proceedings of the 34 th International Conference on Machine Learning, Sydney, Australia, 2017. JMLR: W&CP. Copyright 2017 by the author(s). use of similarities between products and customers to rec- ommend new products. Hybrid systems combine collabo- rative and content techniques. Matrix completion. Mathematically, a recommendation method can be posed as a matrix completion problem (Can- des & Recht, 2012), where columns and rows represent users and items, respectively, and matrix values represent a score determining whether a user would like an item or not. Given a small subset of known elements of the ma- trix, the goal is to fill in the rest. A famous example is the Netflix challenge (Koren et al., 2009) offered in 2009 and carrying a 1M$ prize for the algorithm that can best predict user ratings for movies based on previous ratings. The size of the Netflix is 480k movies × 18k users (8.5B entries), with only 0.011% known entries. Recently, there have been several attempts to incorporate geometric structure into matrix completion problems (Ma et al., 2011; Kalofolias et al., 2014; Rao et al., 2015; Kuang et al., 2016), e.g. in the form of column and row graphs rep- resenting similarity of users and items, respectively. Such additional information makes well-defined e.g. the notion of smoothness of data and was shown beneficial for the performance of recommender systems. These approaches can be generally related to the field of signal processing on graphs (Shuman et al., 2013), extending classical harmonic analysis methods to non-Euclidean domains. Geometric deep learning. Of key interest to the design of recommender systems are deep learning approaches. In the recent years, deep neural networks and, in particular, convolutional neural networks (CNNs) (LeCun et al., 1998) have been applied with great success to numerous com- puter vision-related applications. However, original CNN models cannot be directly applied to the recommendation problem to extract meaningful patterns in users, items and ratings because these data are not Euclidean structured, i.e. they do not lie on regular lattices like images but irregu- lar domains like graphs or manifolds. This strongly moti- vates the development of geometric deep learning (Bron- stein et al., 2016) techniques that can mathematically deal with graph-structured data, which arises in numerous appli- cations, ranging from computer graphics and vision (Masci et al., 2015; Boscaini et al., 2015; 2016b;a; Monti et al., 2017) to chemistry (Duvenaud et al., 2015). arXiv:1704.06803v1 [cs.LG] 22 Apr 2017

Upload: others

Post on 01-Apr-2021

5 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Geometric Matrix Completion with Recurrent Multi-Graph Neural … · 2017. 4. 25. · Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks The earliest attempts

Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks

Federico Monti 1 Michael M. Bronstein 1 2 3 4 Xavier Bresson 5

AbstractMatrix completion models are among the mostcommon formulations of recommender systems.Recent works have showed a boost of perfor-mance of these techniques when introducing thepairwise relationships between users/items in theform of graphs, and imposing smoothness pri-ors on these graphs. However, such techniquesdo not fully exploit the local stationarity struc-tures of user/item graphs, and the number of pa-rameters to learn is linear w.r.t. the number ofusers and items. We propose a novel approachto overcome these limitations by using geomet-ric deep learning on graphs. Our matrix comple-tion architecture combines graph convolutionalneural networks and recurrent neural networksto learn meaningful statistical graph-structuredpatterns and the non-linear diffusion process thatgenerates the known ratings. This neural networksystem requires a constant number of parame-ters independent of the matrix size. We applyour method on both synthetic and real datasets,showing that it outperforms state-of-the-art tech-niques.

1. IntroductionRecommender systems have become a central part of mod-ern intelligent systems. Recommending movies on Net-flix, friends on Facebook, furniture on Amazon, jobs onLinkedIn are a few examples of the main purpose of thesesystems. Two major approach to recommender systems arecollaborative (Breese et al., 1998) and content (Pazzani &Billsus, 2007) filtering techniques. Systems based on col-laborative filtering use collected ratings of products by cus-tomers and offer new recommendations by finding similarrating patterns. Systems based on content filtering make

1ICS USI Lugano, Switzerland 2Tel Aviv University, Is-rael 3Intel Perceptual Computing, Israel 4TUM IAS, Germany5NTU, Singapore. Correspondence to: Federico Monti <[email protected]>.

Proceedings of the 34 th International Conference on MachineLearning, Sydney, Australia, 2017. JMLR: W&CP. Copyright2017 by the author(s).

use of similarities between products and customers to rec-ommend new products. Hybrid systems combine collabo-rative and content techniques.

Matrix completion. Mathematically, a recommendationmethod can be posed as a matrix completion problem (Can-des & Recht, 2012), where columns and rows representusers and items, respectively, and matrix values representa score determining whether a user would like an item ornot. Given a small subset of known elements of the ma-trix, the goal is to fill in the rest. A famous example is theNetflix challenge (Koren et al., 2009) offered in 2009 andcarrying a 1M$ prize for the algorithm that can best predictuser ratings for movies based on previous ratings. The sizeof the Netflix is 480k movies × 18k users (8.5B entries),with only 0.011% known entries.

Recently, there have been several attempts to incorporategeometric structure into matrix completion problems (Maet al., 2011; Kalofolias et al., 2014; Rao et al., 2015; Kuanget al., 2016), e.g. in the form of column and row graphs rep-resenting similarity of users and items, respectively. Suchadditional information makes well-defined e.g. the notionof smoothness of data and was shown beneficial for theperformance of recommender systems. These approachescan be generally related to the field of signal processing ongraphs (Shuman et al., 2013), extending classical harmonicanalysis methods to non-Euclidean domains.

Geometric deep learning. Of key interest to the designof recommender systems are deep learning approaches. Inthe recent years, deep neural networks and, in particular,convolutional neural networks (CNNs) (LeCun et al., 1998)have been applied with great success to numerous com-puter vision-related applications. However, original CNNmodels cannot be directly applied to the recommendationproblem to extract meaningful patterns in users, items andratings because these data are not Euclidean structured, i.e.they do not lie on regular lattices like images but irregu-lar domains like graphs or manifolds. This strongly moti-vates the development of geometric deep learning (Bron-stein et al., 2016) techniques that can mathematically dealwith graph-structured data, which arises in numerous appli-cations, ranging from computer graphics and vision (Masciet al., 2015; Boscaini et al., 2015; 2016b;a; Monti et al.,2017) to chemistry (Duvenaud et al., 2015).

arX

iv:1

704.

0680

3v1

[cs

.LG

] 2

2 A

pr 2

017

Page 2: Geometric Matrix Completion with Recurrent Multi-Graph Neural … · 2017. 4. 25. · Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks The earliest attempts

Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks

The earliest attempts to apply neural networks to graphsare due to Scarselli et al. 2005; 2009 (see more recent for-mulation (Li et al., 2015; Sukhbaatar et al., 2016)). Brunaet al. 2013; 2015 formulated CNN-like deep neural archi-tectures on graphs in the spectral domain, employing theanalogy between the classical Fourier transforms and pro-jections onto the eigenbasis of the graph Laplacian operator(Shuman et al., 2013). In a follow-up work, Defferrard etal. 2016 proposed an efficient filtering scheme using recur-rent Chebyshev polynomials, which reduces the complex-ity of CNNs on graphs to the same complexity of standardCNNs (on grids). This model was later extended to dealwith dynamic data (Seo et al., 2016). Kipf and Welling2016 proposed a simplification of Chebychev networks us-ing simple filters operating on 1-hop neighborhoods of thegraph. Monti et al. 2017 introduced a spatial-domain gen-eralization of CNNs to graphs local patch operators repre-sented as Gaussian mixture models, showing a significantadvantage of such models in generalizing across differentgraphs.

Main contribution. In this work, we treat matrix com-pletion problem as deep learning on graph-structured data.We introduce a novel neural network architecture that isable to extract local stationary patterns from the high-dimensional spaces of users and items, and use these mean-ingful representations to infer the non-linear temporal dif-fusion mechanism of ratings. The spatial patterns are ex-tracted by a new CNN architecture designed to work onmultiple graphs. The temporal dynamics of the rating dif-fusion is produced by a Long-Short Term Memory (LSTM)recurrent neural network (RNN) (Hochreiter & Schmidhu-ber, 1997). To our knowledge, our work is the first appli-cation of graph-based deep learning to matrix completionproblem.

The rest of the paper is organized as follows. Section 2reviews the matrix completion models. Section 3 presentsthe proposed approach. Section 4 presents experimentalresults demonstrating the efficiency of our techniques onsynthetic and real-world datasets, and Section 5 concludesthe paper.

2. Background2.1. Matrix Completion

Matrix completion problem. Recovering the missingvalues of a matrix given a small fraction of its entries isan ill-posed problem without additional mathematical con-straints on the space of solutions. A well-posed problem isto assume that the variables lie in a smaller subspace, i.e.,that the matrix is of low rank,

minX

rank(X) s.t. xij = yij , ∀ij ∈ Ω, (1)

where X denotes the matrix to recover, Ω is the set of theknown entries and yij are their values. To make (1) robustagainst noise and perturbation, the equality constraint canbe replaced with a penalty

minX

rank(X) +µ

2‖Ω (X−Y)‖2F, (2)

where Ω is the indicator matrix of the known entries Ω and denotes the Hadamard pointwise product.

Unfortunately, rank minimization turns out an NP-hardcombinatorial problem that is computationally intractablein practical cases. The tightest possible convex relaxationof the previous problem is

minX‖X‖? +

µ

2‖Ω (X−Y)‖2F, (3)

where ‖·‖? is the nuclear norm of a matrix equal to the sumof its singular values (Candes & Recht, 2009). Candes andRecht 2009 proved that the `1 relaxation of the SVD lead tosolutions that recover almost exactly the original low-rankmatrix.

Geometric matrix completion An alternative relaxationof the rank operator in (1) is to constraint the space of so-lutions to be smooth w.r.t. some geometric structure of thematrix rows and columns (Ma et al., 2011; Kalofolias et al.,2014; Rao et al., 2015; Benzi et al., 2016). The simplestmodel is proximity structure represented as an undirectedweighted column graph Gc = (1, . . . , n, Ec,Wc) withadjacency matrix Wc = (wc

ij), where wcij = wc

ji, wcij = 0

if (i, j) /∈ Ec and wcij > 0 if (i, j) ∈ Ec. In our notation,

the column graph could be thought of as a social networkcapturing relations between users and the similarity of theirtastes. The row graph Gr = (1, . . . ,m, Er,Wr) repre-senting the similarities of the items is defined in a similarmanner.

On each of these graphs one can construct the (unnor-malized) graph Laplacian, an n × n symmetric positive-semidefinite matrix ∆ = I − D−1/2WD−1/2, whereD = diag

(∑j 6=i wij

)is the degree matrix. We de-

note the Laplacian associated with row and column graphsby ∆r and ∆c, respectively. Considering the columns(respectively, rows) of matrix X as vector-valued func-tions on the column graph Gc (respectively, row graphGr), their smoothness can be expressed as the Dirichletnorm ‖X‖2Gr = trace(X>∆rX) (respecitvely, ‖X‖2Gc =

trace(X∆cX>)). The geometric matrix completion prob-

lem thus boils down to minimizing

minX‖X‖2Gr + ‖X‖2Gc +

µ

2‖Ω (X−Y)‖2F, (4)

Page 3: Geometric Matrix Completion with Recurrent Multi-Graph Neural … · 2017. 4. 25. · Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks The earliest attempts

Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks

n users

mite

ms

X

W

H>

j1 j2 . . . j3

i2

...

i1

n users

mite

ms

j1 j2 . . . j3

Figure 1. Full (top) and factorized (bottom) geometric matrixcompletion representations. The column and row graphs repre-sent the relationships between users and items, respectively.

Factorized models. Matrix completion algorithms intro-duced in the previous section are well-posed as convexoptimization problems, guaranteeing existence, uniquenessand robustness of solutions. Besides, fast algorithms havebeen developed in the context of compressed sensing tosolve the non-differential nuclear norm problem. However,the variables in this formulation are the full m × n matrixX, making such methods hard to scale up to large matricessuch as the notorious Netflix challenge.

A solution is to use a factorized representation (Srebroet al., 2004; Koren et al., 2009; Ma et al., 2011; Yanez& Bach, 2012; Rao et al., 2015; Benzi et al., 2016) X =WH>, where W,H are m × r and n × r matrices, re-spectively, with r min(m,n). The use of factors W,Hreduce the number of degrees of freedom from O(mn) toO(m + n); this representation is also attractive as solvingthe matrix completion problem often assumes the originalmatrix to be low-rank, and rank(WH>) ≤ r by construc-tion. Figure 1 shows the full and factorized settings of thematrix completion problem.

The nuclear norm minimization problem in the previoussection can be equivalently rewritten in a factorized formas (Srebro et al., 2004):

minW,H

1

2‖W‖2F +

1

2‖H‖2F +

µ

2‖Ω (WH> −Y)‖2F. (5)

and the factorized formulation of the graph-based mini-mization problem (4) as

minW,H

1

2‖W‖2Gr +

1

2‖H‖2Gc +

µ

2‖Ω (WH> −Y)‖2F. (6)

The limitation of model (6) is to decouple the regularizationprocess applied simultaneously on the rows and columnsof X in (4), but the advantage is linear instead of quadraticcomplexity.

2.2. Deep learning on graphs

The key idea to our work is geometric deep learning, anextension of the popular CNNs to graphs. A graph Lapla-cian admits a spectral eigendecomposition of the form∆ = ΦΛΦ>, where Φ = (φ1, . . .φn) denotes the matrixof orthonormal eigenvectors and Λ = diag(λ1, . . . , λn) isthe diagonal matrix of the corresponding eigenvalues. Theeigenvectors play the role of Fourier atoms in classical har-monic analysis and the eigenvalues can be interpreted asfrequencies. Given a function x = (x1, . . . , xn)> on thevertices of the graph, its graph Fourier transform is givenby x = Φ>x. The spectral convolution of two functionsx,y can be defined as the element-wise product of the re-spective Fourier transforms,

x ? y = Φ(Φ>x) (Φ>y) = Φ diag(y1, . . . , yn) x. (7)

Bruna et al. 2013 used the spectral definition of convolu-tion (7) to generalize CNNs on graphs. A spectral convolu-tional layer has the form

xl = ξ

q′∑l′=1

ΦYll′Φ>xl′

, l = 1, . . . , q, (8)

where q′, q denote the number of input and output chan-nels, respectively, Yll′ = diag(yll′,1, . . . , yll′,n) is a diag-onal matrix of spectral multipliers representing a learnablefilter in the spectral domain, and ξ is a nonlinearity (e.g.ReLU) applied on the vertex-wise function values. Unlikeclassical convolutions carried out efficiently in the spec-tral domain using FFT, the computations of the forwardand inverse graph Fourier transform incur expensiveO(n2)multiplication by the matrices Φ,Φ>, as there are no FFT-like algorithms on general graphs. Furthermore, there is noguarantee that the filters represented in the spectral domainare localized in the spatial domain, which is an importantproperty of classical CNNs.

To address these issues, Defferrard et al. 2016 proposedusing an explicit expansion in the Chebyshev polynomialbasis to represent the spectral filters

τθ(∆) =

p−1∑j=0

θjTj(∆) =

p−1∑j=0

θjΦTj(Λ)Φ>, (9)

Page 4: Geometric Matrix Completion with Recurrent Multi-Graph Neural … · 2017. 4. 25. · Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks The earliest attempts

Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks

XX(t) X(t)

MGCNN RNN

dX(t)

X(t+1) = X(t) + dX(t)

row+column filtering

Figure 2. Recurrent GCNN (RGCNN) architecture using the full matrix completion model and operating simultaneously on the rows andcolumns of the matrix X. The output of the Multi-Graph CNN (MGCNN) module is a q-dimensional feature vector for each element ofthe input matrix. The number of parameters to learn is O(1) and the learning complexity is O(mn).

where ∆ = 2λ−1n ∆− I is the rescaled Laplacian such thatits eigenvalues Λ = 2λ−1n Λ− I are in the interval [−1, 1],θ is the p-dimensional vector of polynomial coefficientsparametrizing the filter, and Tj(λ) = 2λTj−1(λ)−Tj−2(λ)denotes the Chebyshev polynomial of degree j defined ina recursive manner with T1(λ) = λ and T0(λ) = 1.1 This approach benefits from several advantages. First,it does not require an explicit computation of the Lapla-cian eigenvectors, and due to the recursive definition ofthe Chebyshev polynomials, the computation of the filterincurs applying the Laplacian p times. Multiplication byLaplacian has the cost of O(|E|), and assuming the graphhas |E| = O(n) edges (which is the case for k-nearestneighbors graphs and most real-world networks), the over-all complexity isO(n) rather thanO(n2) operations, whichis the same complexity than standard CNNs. Moreover,since the Laplacian is a local operator affecting only 1-hopneighbors of a vertex and accordingly its (p − 1)st poweraffects the p-hop neighborhood, the resulting filters are spa-tially localized.

3. Our approachIn this paper, we propose formulating matrix completionas a learnable diffusion process applied to the score values.The deep learning architecture considered for this purposeconsists of a spatial part extracting spatial features from thematrix (we consider two different approaches working onthe full and factorized matrix models), and a temporal partusing a recurrent LSTM network. The two architectures are

1Tj(λ) = cos(j cos−1(λ)) is an oscillating function on[−1, 1] with j roots, j + 1 equally spaced extrema, and a fre-quency linearly dependent on j. Chebyshev polynomials form anorthogonal basis for the space of smooth functions on [−1, 1] andare thus convenient to compactly represent spectral filters.

summarized in Figures 2 and 3 and described in details inthe following.

3.1. Multi-Graph CNNs

Multi-graph convolution. Our first goal is to extend thenotion of the aforementioned graph Fourier transform tomatrices whose rows and columns are defined on row- andcolumn-graphs. We recall that a classical two-dimensionalFourier transform of an image (matrix) can be thoughtof as applying a one-dimensional Fourier transform to itsrows and columns. In our setting, the analogy of the two-dimensional Fourier transform has the form

X = Φ>r XΦc (10)

where Φc,Φr and Λc,Λr denote the n × n and m × meigenvector- and eigenvalue matrices of the column- androw-graph Laplacians ∆c,∆r, respectively. The multi-graph version of the spectral convolution (7) is given by

X ?Y = Φr(X Y)Φ>c . (11)

Representing the filters as their spectral multipliers Ywould yield O(mn) parameters, prohibitive in any prac-tical application. To overcome this limitation, we resort tothe representation of the filters in Chebychev polynomialbases of degree p,

τΘ(λc, λr) =

p∑j,j′=0

θjj′Tj(λc)Tj′(λr), (12)

where Θ = (θjj′) is the (p + 1) × (p + 1) matrix of co-efficients, i.e., O(1) parameters. The application of suchfilters to the matrix X

X =

p∑j,j′=0

θjj′Tj(∆r)XTj′(∆c) (13)

Page 5: Geometric Matrix Completion with Recurrent Multi-Graph Neural … · 2017. 4. 25. · Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks The earliest attempts

Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks

W

H>H(t) H(t)

W(t) W(t)

GCNN RNN

GCNN RNN

dH(t)

dW(t)

W(t+1) = W(t) + dW(t)

H(t+1) = H(t) + dH(t)

row filtering

column filtering

Figure 3. Separable Recurrent GCNN (sRGCNN) architecture using the factorized matrix completion model and operating separatelyon the rows and columns of the factors W, H>. The output of the GCNN module is a q-dimensional feature vector for each inputrow/column, respectively. The number of parameters to learn is O(1) and the learning complexity is O(m+ n).

results inO(mn) complexity. Here, as previously, ∆r, ∆c

denote the scaled Laplacians with eigenvalues in the inter-val [−1, 1].

A Multi-Graph CNN (MGCNN) using this parametrizationof filters (13) in the convolutional layer is applied to them×nmatrix X (single input channel), producing q outputs(i.e., a tensor of size m× n× q).

Separable convolution. A simplification of the multi-graph convolution is obtained considering the factorizedform of the matrix X = WH> and applying one-dimensional convolution on the respective graph to eachfactor,

wl =

q′∑l′=1

ΦrYll′

r Φ>r wl′ , l = 1, . . . , q (14)

hl =

q′∑l′=1

ΦcYll′

c Φ>c hl′ , l = 1, . . . , q, (15)

where Yll′

r = diag(yrll′,1, . . . , yrll′,m) and Yll′

c =diag(ycll′,1, . . . , y

cll′,n) are the row- and column-filters re-

sulting in a total of O(m + n) parameters. Similarly tothe previous case, we can express the filters resorting toChebyshev polynomials,

wl =

q′∑l′=1

p∑j=0

θrll′,jTj(∆r)wl′ (16)

hl =

q′∑l′=1

p∑j′=0

θcll′,j′Tj′(∆c)hl′ (17)

with 2(p+ 1)qq′ parameters in total.

3.2. Matrix diffusion with RNN

The next step of our approach is to feed the features ex-tracted from the matrix by the MGCNN (or alternatively,the row- and column-GCNNs) to a Recurrent Neural Net-work (RNN) implementing the score diffusion process. Weuse the classical Long-Short Term Memory (LSTM) RNNarchitecture (Hochreiter & Schmidhuber, 1997), which hasdemonstrated to be highly efficient to learn the dynami-cal property of data sequences as LSTM is able to keeplong-term internal states (in particular, avoiding the van-ishing gradient issue). The input of the LSTM gate isgiven by the static features extracted from the MGCNN,which can be seen as a projection or dimensionality reduc-tion of the original matrix in the space of the most mean-ingful and representative information (the disentanglementeffect). This representation coupled with LSTM appearsparticularly well-suited to keep a long term internal state,which allows to predict accurate small changes dX of thematrix X (or dW, dH of the factors W, H) that can prop-agate through the full temporal steps.

Figures 2 and 3 provides an illustration of the proposed ma-trix completion model. We also give a precise descriptionof the two settings of our model in Algorithms 1 and 2.We refer to the whole architecture combining the MGCNNand RNN in the full matrix completion setting as Recur-rent Graph CNN (RGCNN). The factorized version withtwo GCNNs and RNN is referred to as separable RecurrentGraph CNN (sRGCNN).

The complexity of Algorithm 1 scales quadratically asO(mn) due to the use of MGCNN. For large matrices,we can opt for Algorithm 2 that processes the rows andcolumns separately with standard GCNNs and scales lin-early as O(m+ n).

Page 6: Geometric Matrix Completion with Recurrent Multi-Graph Neural … · 2017. 4. 25. · Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks The earliest attempts

Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks

Algorithm 1 Full matrix completion model using RGCNN

input m× n matrix X(0) containing initial values1: for t = 0 : T do2: Apply the Multi-Graph CNN (13) on X(t) producing

anm×n×q output X(t) containing a q-dimensionalfeature vector for each matrix element.

3: for all elements (i, j) do4: Apply RNN to feature vector x

(t)ij =

(x(t)ij1, . . . , x

(t)ijq) producing the predicted in-

cremental value dx(t)ij

5: end for6: Update X(t+1) = X(t) + dX(t)

7: end for

Algorithm 2 Factorized matrix completion model usingsRGCNNinput m×r factor H(0) and n×r factor W(0) representing

the matrix X(0)

1: for t = 0 : T do2: Apply the Graph CNN on H(t) producing an n × q

output H(t).3: for j = 1 : n do4: Apply RNN to feature vector h

(t)j =

(h(t)j1 , . . . , h

(t)jq ) producing the predicted in-

cremental value dh(t)j

5: end for6: Update H(t+1) = H(t) + dH(t)

7: Apply the Graph CNN on W(t) producing an m× qoutput W(t).

8: for i = 1 : m do9: Apply RNN to feature vector w

(t)i =

(w(t)i1 , . . . , w

(t)iq ) producing the predicted in-

cremental value dw(t)i

10: end for11: Update W(t+1) = W(t) + dW(t)

12: end for

3.3. Training

Training of the networks is performed by minimizing theloss

`(Θ,σ) = ‖X(T )Θ,σ‖

2Gr+‖X(T )

Θ,σ‖2Gc+

µ

2‖Ω(X(T )

Θ,σ−Y)‖2F.(18)

Here, T denotes the number of diffusion iterations (appli-cations of the RNN), and we use the notation X

(T )Θ,σ to

emphasize that the matrix depends on the parameters ofthe MGCNN (Chebyshev polynomial coefficients Θ) andthose of the LSTM (denoted by σ).

In the factorized setting, we use the loss

`(θr,θc,σ) = ‖W(T )θr,σ‖2Gr + ‖H(T )

θc,σ‖2Gc (19)

2‖Ω (W

(T )θr,σ

(H(T )θc,σ

)> −Y)‖2F.

where θc,θr are the parameters of the two GCNNs.

4. ResultsExperimental settings. We closely followed the exper-imental setup of (Rao et al., 2015), using five standarddatasets: Synthetic dataset from (Kalofolias et al., 2014),MovieLens (Miller et al., 2003), Flixster (Jamali & Ester,2010), Douban (Ma et al., 2011), and YahooMusic (Droret al., 2012). Classical Matrix Completion (MC) (Candes& Recht, 2012), Inductive Matrix Completion (IMC) (Jain& Dhillon, 2013; Xu et al., 2013), Geometric Matrix Com-pletion (GMC) (Kalofolias et al., 2014), and Graph Reg-ularized Alternating Least Squares (GRALS) (Rao et al.,2015) were used as baseline methods.

In all the experiments, we used the following settings forour RGCNNs: Chebyshev polynomials of order p = 5,outputting k = 32-dimensional features, LSTM cells with32 features and T = 10 diffusion steps. All the modelswere implemented in Google TensorFlow and trained us-ing the Adam stochastic optimization algorithm (Kingma& Ba, 2014) with learning rate 10−3. In factorized mod-els, rank r = 15 and 10 was used for the synthetic and realdatasets, respectively. For all methods, hyperparameterswere chosen by cross-validation.

4.1. Synthetic data

We start our experimental evaluation showing the perfor-mance of our approach on a small synthetic dataset, inwhich the user and item graphs have strong communitiesstructure. Though rather simple, such a dataset allowsto study the behavior of different algorithms in controlledsettings. The performance of different matrix completionmethods is reported in Table 1, along with their theoreticalcomplexity. Our RGCNN model achieves the best accu-racy, followed by the separable RGCNN. Different diffu-sion time steps of these two models are visualized in Figure4. Figure 5 shows the convergence rates of different meth-ods. Figures 6 and 7 depict the spectral filters learnt by theMGCNN and row- and column-GCNNs.

We repeated the same experiment considering only the col-umn (users) graph to be given. In this setting, the RGCNNcannot be applied, while the sRGCNN has only one GCNNapplied on the factor H, and the other factor W is free.Table 2 summarizes the results of this experiment, again,showing that our approach performs the best.

Page 7: Geometric Matrix Completion with Recurrent Multi-Graph Neural … · 2017. 4. 25. · Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks The earliest attempts

Geometric Matrix Completion with Recurrent Multi-Graph Neural Networkst = 0 1 2 3 4 5 6 7 8 9 10

2.26 1.89 1.60 1.78 1.31 0.52 0.48 0.63 0.38 0.07 0.01

1.15 1.04 0.94 0.89 0.84 0.76 0.69 0.49 0.27 0.11 0.01

Figure 4. Evolution of the matrix X(t) with our architecture using full matrix completion model RGCNN (top) and factorized matrixcompletion model sRGCNN (bottom). Numbers indicate the RMS error.

Table 1. Comparison of different matrix completion methods us-ing users+items graphs in terms of number of parameters (opti-mization variables) and computational complexity order (opera-tions per iteration). Rightmost column shows the RMS error onSynthetic dataset.

METHOD PARAMETERS COMPLEXITY RMSE

GMC mn mn 0.3693GRALS m+ n m+ n 0.0114RGCNN 1 mn 0.0053sRGCNN 1 m+ n 0.0106

Table 2. Comparison of different matrix completion methods us-ing users graph only in terms of number of parameters (optimiza-tion variables) and computational complexity order (operationsper iteration). Rightmost column shows the RMS error on Syn-thetic dataset.

METHOD PARAMETERS COMPLEXITY RMSE

GRALS m+ n m+ n 0.0452sRGCNN m m+ n 0.0362

0 1 2 3 4 5

·104

10−4

10−3

10−2

Training iterations

RM

SE

GMCGRALSsRGCNNRGCNN

Figure 5. Convergence rates of the tested algorithms over the Syn-thetic Netflix dataset.

0 1 2

2

1

0

λr

λc

Figure 6. Absolute value of the first 8 spectral filters learnt by ourbidimensional convolution. On the left the first filter with the ref-erence axes associated to the row and column graph eigenvalues.

0 0.5 1 1.5 20

0.2

0.4

0.6

0.8

λr, λc

Filte

rRes

pons

e

Figure 7. Absolute value of the first four spectral filters learned bythe user (solid) and items (dashed) GCNNs.

4.2. Real data

Following (Rao et al., 2015), we evaluated the proposed ap-proach on the MovieLens, Flixster, Douban and YahooMu-sic datasets. For the MovieLens dataset we constructedthe user and item (movie) graphs as unweighted 10-nearestneighbor graphs in the space of user and movie features,respectively. For Flixster, the user and item graphs wereconstructed from the scores of the original matrix. On thisdataset, we also performed an experiment using only theusers graph. For the Douban dataset, we used only the usergraph (the provided social network of the user). For the Ya-hooMusic dataset, we used only the item graph, constructedwith unweighted 10-nearest neighbors in the space of itemfeatures (artists, albums, and genres). For the latter threedatasets, we used a sub-matrix of 3000 × 3000 entries forevaluating the performance.

Page 8: Geometric Matrix Completion with Recurrent Multi-Graph Neural … · 2017. 4. 25. · Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks The earliest attempts

Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks

Tables 3 and 4 summarize the performance of differentmethods. RGCNN outperforms the competitors in all theexperiments.

Table 3. Performance (RMS error) of different matrix completionmethods on the MovieLens dataset.

METHOD RMSE

GLOBAL MEAN 1.154USER MEAN 1.063MOVIE MEAN 1.033MC (CANDES & RECHT, 2012) 0.973IMC (JAIN & DHILLON, 2013; XU ET AL., 2013) 1.653GMC (KALOFOLIAS ET AL., 2014) 0.996GRALS (RAO ET AL., 2015) 0.945sRGCNN 0.929

Table 4. Matrix completion results on several datasets (RMS er-ror). For Douban and YahooMusic, a single graph (of users anditems, respectively) was used. For Flixter, two settings are shown:users+items graphs / only users graph.

METHOD FLIXSTER DOUBAN YAHOOMUSIC

GRALS 1.3126 / 1.2447 0.8326 38.0423sRGCNN 1.1788 / 0.9258 0.8012 22.4149

5. ConclusionIn this paper, we presented a new deep learning approachfor matrix completion based on a specially designed multi-graph convolutional neural network architecture. Amongthe key advantages of our approach compared to traditionalmethods is its low computational complexity and constantnumber of degrees of freedom independent of the matrixsize. We showed that the use of deep learning for matrixcompletion allows to beat current state-of-the-art recom-mender system methods. To our knowledge, our work isthe first application of deep learning on graphs to this classof problems. We believe that it shows the potential of thenascent field of geometric deep learning on non-Euclideandomains, and will encourage future works in this direction.

6. AcknowledgmentsFM and MB are supported in part by ERC StartingGrant No. 307047 (COMET), ERC Consolidator GrantNo. 724228 (LEMAN), Google Faculty Research Award,Nvidia equipment grant, Radcliffe fellowship from Har-vard Institute for Advanced Study, and TU Munich Insti-tute for Advanced Study, funded by the German Excellence

Initiative and the European Union Seventh Framework Pro-gramme under grant agreement No. 291763. XB is sup-ported in part by NRF Fellowship NRFF2017-10.

ReferencesBenzi, K., Kalofolias, V., Bresson, X., and Vandergheynst,

P. Song recommendation with non-negative matrix fac-torization and graph total variation. In Proc. ICASSP,2016.

Boscaini, D., Masci, J., Melzi, S., Bronstein, M. M.,Castellani, U., and Vandergheynst, P. Learning class-specific descriptors for deformable shapes using local-ized spectral convolutional networks. Computer Graph-ics Forum, 34(5):13–23, 2015. ISSN 1467-8659.

Boscaini, D., Masci, J., Rodola, E., and Bronstein, M. M.Learning shape correspondence with anisotropic convo-lutional neural networks. In Proc. NIPS, 2016a.

Boscaini, D., Masci, J., Rodola, E., Bronstein, M. M., andCremers, D. Anisotropic diffusion descriptors. Com-puter Graphics Forum, 35(2):431–441, 2016b.

Breese, J., Heckerman, D., and Kadie, C. Empirical Analy-sis of Predictive Algorithms for Collaborative Filtering.In Proc. Uncertainty in Artificial Intelligence, 1998.

Bronstein, M. M., Bruna, J., LeCun, Y., Szlam, A., andVandergheynst, P. Geometric deep learning: going be-yond euclidean data. arXiv:1611.08097, 2016.

Bruna, J., Zaremba, W., Szlam, A., and LeCun, Y. Spec-tral networks and locally connected networks on graphs.Proc. ICLR, 2013.

Candes, E. and Recht, B. Exact Matrix Completion viaConvex Optimization. Foundations of ComputationalMathematics, 9(6):717–772, 2009.

Candes, E. and Recht, B. Exact matrix completion via con-vex optimization. Comm. ACM, 55(6):111–119, 2012.

Defferrard, M., Bresson, X., and Vandergheynst, P. Con-volutional neural networks on graphs with fast localizedspectral filtering. In Proc. NIPS, 2016.

Dror, G., Koenigstein, N., Koren, Y., and Weimer, M. TheYahoo! music dataset and KDD-Cup’11. In KDD Cup,2012.

Duvenaud, D. K. et al. Convolutional networks on graphsfor learning molecular fingerprints. In Proc. NIPS, 2015.

Gori, M., Monfardini, G., and Scarselli, F. A new modelfor learning in graph domains. In Proc. IJCNN, 2005.

Page 9: Geometric Matrix Completion with Recurrent Multi-Graph Neural … · 2017. 4. 25. · Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks The earliest attempts

Geometric Matrix Completion with Recurrent Multi-Graph Neural Networks

Henaff, M., Bruna, J., and LeCun, Y. Deep convolutionalnetworks on graph-structured data. arXiv:1506.05163,2015.

Hochreiter, S. and Schmidhuber, J. Long short-term mem-ory. Neural Computation, 9(8):1735–1780, 1997.

Jain, P. and Dhillon, I. S. Provable inductive matrix com-pletion. arXiv:1306.0626, 2013.

Jamali, M. and Ester, M. A matrix factorization techniquewith trust propagation for recommendation in social net-works. In Proc. Recommender Systems, 2010.

Kalofolias, V., Bresson, X., Bronstein, M. M., andVandergheynst, P. Matrix completion on graphs.arXiv:1408.1717, 2014.

Kingma, D. P. and Ba, J. Adam: A method for stochasticoptimization. arXiv:1412.6980, 2014.

Kipf, T. N. and Welling, M. Semi-supervised classificationwith graph convolutional networks. arXiv:1609.02907,2016.

Koren, Y., Bell, R., and Volinsky, C. Matrix factorizationtechniques for recommender systems. Computer, 42(8):30–37, 2009.

Kuang, D., Shi, Z., Osher, S., and Bertozzi, A. Aharmonic extension approach for collaborative ranking.arXiv:1602.05127, 2016.

LeCun, Y., Bottou, L., Bengio, Y., and Haffner, P. Gradient-based learning applied to document recognition. Proc.IEEE, 86(11):2278–2324, 1998.

Li, Y., Tarlow, D., Brockschmidt, M., and Zemel, R. Gatedgraph sequence neural networks. arXiv:1511.05493,2015.

Ma, H., Zhou, D., Liu, C., Lyu, M., and King, I. Recom-mender systems with social regularization. In Proc. WebSearch and Data Mining, 2011.

Masci, J., Boscaini, D., Bronstein, M. M., and Van-dergheynst, P. Geodesic convolutional neural networkson Riemannian manifolds. In Proc. 3DRR, 2015.

Miller, B. N. et al. MovieLens unplugged: experienceswith an occasionally connected recommender system. InProc. Intelligent User Interfaces, 2003.

Monti, F., Boscaini, D., Masci, J., Rodola, E., Svoboda,J., and Bronstein, M. M. Geometric deep learning ongraphs and manifolds using mixture model CNNs. InProc. CVPR, 2017.

Pazzani, M. and Billsus, D. Content-based Recommenda-tion Systems. The Adaptive Web, pp. 325–341, 2007.

Rao, N., Yu, H.-F., Ravikumar, P. K., and Dhillon, I. S. Col-laborative filtering with graph information: Consistencyand scalable methods. In Proc. NIPS, 2015.

Scarselli, F., Gori, M., Tsoi, A. C., Hagenbuchner, M., andMonfardini, G. The graph neural network model. IEEETrans. Neural Networks, 20(1):61–80, 2009.

Seo, Y., Defferrard, M., Vandergheynst, P., and Bresson, X.Structured sequence modeling with graph convolutionalrecurrent networks. arXiv:1612.07659, 2016.

Shuman, D. I., Narang, S. K., Frossard, P., Ortega, A., andVandergheynst, P. The emerging field of signal process-ing on graphs: Extending high-dimensional data analysisto networks and other irregular domains. IEEE Sig. Proc.Magazine, 30(3):83–98, 2013.

Srebro, N., Rennie, J., and Jaakkola, T. Maximum-MarginMatrix Factorization. In Proc. NIPS, 2004.

Sukhbaatar, S., Szlam, A., and Fergus, R. Learn-ing multiagent communication with backpropagation.arXiv:1605.07736, 2016.

Xu, M., Jin, R, and Zhou, Z.-H. Speedup matrix com-pletion with side information: Application to multi-labellearning. In Proc. NIPS, 2013.

Yanez, F. and Bach, F. Primal-dual algorithms for non-negative matrix factorization with the kullback-leiblerdivergence. arXiv:1412.1788, 2012.