deep multi-graph clustering via attentive cross-graph...

1
Deep Multi-Graph Clustering via Attentive Cross-Graph Association Dongsheng Luo 1* , Jingchao Ni 2 , Suhang Wang 1 , Yuchen Bian 3 , Xiong Yu 4 , Xiang Zhang 1 1. The Pennsylvania State University, 2. NEC Lab, 3. Baidu Research, 4. Case Western Reserve University *contact: [email protected] Motivation Multi-graph clustering aims to improve clustering accuracy by leveraging information from different domains, which has been shown to be extremely effective for achieving better clustering results than single graph based clustering algorithms. Despite the previous success, existing multi-graph clustering methods mostly use shallow models, which are incapable to capture the highly non-linear structures and the complex cluster associations in multigraph, thus result in sub-optimal results. we propose an end to-end deep learning model to simultaneously infer cluster assignments and cluster associations in multi graph. Problem Input: multi-network & # clusters =1 Output: Cluster assignment () with () indicating the probability that assign node to cluster in graph () . Cluster association () , with () indicating the association weight of cluster in () and cluster in () Framework Losses: Clustering loss: sharpen probability distributions. Proximity loss: capture within-network relations. Cross-network loss: capture cross-network relations. Parameters: Θ: parameters in Deep AutoEncoders : cluster centroids in the unified space : parameters in the attention network Objective Function Deep AutoEncoder and Clustering The original graph is in high-dimensional space, where node are hard to separate. We adopt deep neural networks to map nodes to a low dimensional space, which is cluster-friendly. Deep Neural Network Inside Graph Regularization Both 1 st and 2 nd order proximity are utilized to capture the local and global structure of each graph. Cross-Graph Regularization Cauthy distribution is used to compute probability of assign a node to a cluster. Minimum-entropy based Clustering is proposed to sharpen the probability distribution. To avoid the trivial solution, we also add the constraint to balance the cluster sizes. Cluster-level relations are considered in our cross-graph regularization. Intuitively, if node in the first graph is linked to node in the second graph, then the clusters they belong to should related. This observation leads to our cross-graph regularization. Experiments Codes are available at: https://github.com/flyingdoog/DMGC

Upload: others

Post on 12-Aug-2020

17 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Deep Multi-Graph Clustering via Attentive Cross-Graph ...personal.psu.edu/dul262/DMGC/dmgc_poster.pdf · Deep Multi-Graph Clustering via Attentive Cross-Graph Association Dongsheng

Deep Multi-Graph Clustering via Attentive Cross-Graph AssociationDongsheng Luo1*, Jingchao Ni2, Suhang Wang1, Yuchen Bian3, Xiong Yu4 , Xiang Zhang1

1. The Pennsylvania State University, 2. NEC Lab, 3. Baidu Research, 4. Case Western Reserve University

*contact: [email protected]

Motivation

Multi-graph clustering aims to improve clustering accuracy by leveraging

information from different domains, which has been shown to be

extremely effective for achieving better clustering results than single

graph based clustering algorithms. Despite the previous success,

existing multi-graph clustering methods mostly use shallow models,

which are incapable to capture the highly non-linear structures and the

complex cluster associations in multigraph, thus result in sub-optimal

results. we propose an end to-end deep learning model to

simultaneously infer cluster assignments and cluster associations inmulti graph.

Problem

Input: multi-network & # clusters 𝐾𝑖 𝑖=1𝑔

Output:

• Cluster assignment 𝑸(𝑖) with 𝑞𝑥𝑘(𝑖)

indicating the probability that

assign node 𝑥 to cluster 𝑘 in graph 𝐺(𝑖).

• Cluster association 𝐶(𝑖𝑗), with 𝑐𝑘𝑙(𝑖𝑗)

indicating the association

weight of cluster 𝑘 in 𝐺(𝑖) and cluster 𝑙 in 𝐺(𝑗)

Framework

Losses:

Clustering loss: sharpen probability distributions.

Proximity loss: capture within-network relations.

Cross-network loss: capture cross-network relations.

Parameters:

Θ: parameters in Deep AutoEncoders

𝑍: cluster centroids in the unified space

𝑊: parameters in the attention network

Objective Function

Deep AutoEncoder and Clustering

The original graph is in high-dimensional space, where node are

hard to separate. We adopt deep neural networks to map nodes to a

low dimensional space, which is cluster-friendly.

Deep Neural Network

Inside Graph RegularizationBoth 1st and 2nd order proximity are utilized to capture the local and

global structure of each graph.

Cross-Graph Regularization

Cauthy distribution is used to compute probability of assign a node to

a cluster. Minimum-entropy based Clustering is proposed to sharpen

the probability distribution. To avoid the trivial solution, we also add the

constraint to balance the cluster sizes.

Cluster-level relations are considered in our cross-graph

regularization.

Intuitively, if node 𝑥 in the first graph is linked to node 𝑦 in the second

graph, then the clusters they belong to should related. This

observation leads to our cross-graph regularization.

Experiments

Codes are available at: https://github.com/flyingdoog/DMGC