Background p Network A ubiquitous data structure to model the - PowerPoint PPT Presentation
BiNE : Bi partite N etwork E mbedding ACM SIGIR 2018, July 8, Ann Arbor Michigan, U.S.A. MingGao * , Leihui Chen * , Xiangnan He + , Aoying Zhou * * East China Normal University + National University of Singapore Background p Network A
BiNE : Bi partite N etwork E mbedding ACM SIGIR 2018, July 8, Ann Arbor Michigan, U.S.A. MingGao * , Leihui Chen * , Xiangnan He + , Aoying Zhou * * East China Normal University + National University of Singapore
Background p Network Ø A ubiquitous data structure to model the relationships between entities p Network embedding Ø Crucial to obtain the representations for vertices Ø Helpful to many applications, such as vertex labeling , link prediction , recommendation , and clustering , etc. Homogeneous Network Heterogeneous Network ü Social network ü Item adoption ü Collaboration ü Web visiting network ü Question ü Transportation answering network ü … ü … 2
Drawbacks of Existing Works for Bipartite Networks p Homogeneous network embedding: Ø Ignore type information of vertices (e.g., Node2vec, DeepWalk, etc.) Ø Ignore key characteristic of bipartite network -- power-law distribution of vertex degrees Heterogeneous network embedding: Ø MetaPath2vec [Dong et al, KDD’17] treats explicit and implicit relations as contributing equally 3
Outline p Background & Motivations p Proposed Method p Experiments and Results p Conclusions
BiNE: Bipartite Network Embedding p Two Characteristics of BiNE Ø Modeling the explicit and implicit relations simultaneously Ø A biased and self-adaptive random walk generator & " " % ! 2 ∶ . ∪ * → ℝ 7 " " & " # |(| ! # & # # % # & " $ & $ # ! $ % $ .2 .3 .5 1 .7 & $ $ … … .4 .3 .5 .1 .2 |U| .2 .6 .5 .9 .1 Capture explicit … … … … … relations + = (. , * , W) BiNE Jointly model explicit Input and implicit relations |(| & " " % ! " " & " # Obtain implicit .2 .1 .2 1 .7 & # # % # ! # relations .3 .4 .5 .5 .7 |*| & " $ & $ # ! $ % $ .1 .6 .5 .9 .1 % % ! ! & $ $ " " " " … … … … … … … ! # % # % # ! # ! $ ! $ % $ % $ … … … … 5
Modeling Explicit Relations (Observed links) p Original network space The joint probability between vertices ! " and # $ is defined as: p Preserving the local proximity Minimizing the difference (KL- divergence) between the two distributions: p Embedding space The joint probability between vertices ! " and # $ is estimated as: 6
Modeling Implicit Relations (High-order relations) p Constructing Corpus of Vertex & " " % ! " " Sequences & " # & # # % # ! # Ø Construct U-U and V-V networks & " $ & $ # ! $ % $ & $ $ … … Ø Run Self-adaptive random walker 1) # of walks starting from a vertex depends on its centrality score. % % ! ! " " " " 2) Length of a vertex sequence is % # % # ! # ! # controlled by a stop probability. % $ % $ ! $ ! $ p Optimizing a point-wise … … … … classification loss to capture the high-order correlations 7
Capturing the High-order Relations p Assumption: vertices frequently co-occurred in the same context of a sequence should be assigned to similar embeddings . A. Taking corpus of users ! " as example , given a sequence # , $% (=2) and a vertex & ' : & ( & ) & * & + & , & - & . & / #: & ' B. Sample High-quality and Diverse Negatives with Locality Sensitive Hashing C. (LSH) 8
Joint Optimization p A joint optimization framework Explicit relations Implicit relations 9
Outline p Background & Motivations p Proposed Method p Experiments and Results p Conclusions 10
Experimental Setting-up p Tasks Ø Two tasks: link prediction (classification) & recommendation (ranking) p Datasets and Metrics p Research Questions RQ1 Performance of BiNE compared to representative baselines Ø RQ2 Is the implicit relations helpful? Ø RQ3 Effect of random walk generator Ø 11
Baselines p Network embedding methods p Recommendation methods Ø DeepWalk [ Perozzi et al KDD 2014 ] Ø BPR [Rendle et al UAI 2009] Ø LINE [ Tang et al WWW 2015 ] Ø RankALS [Takács et al Recsys 2012] Ø Node2vec [ Grover et al KDD 2016 ] Ø FISMauc [Kabbur et al KDD 2013] Ø Metapath2vec++ [ Dong et al KDD 2017 ] p Link Prediction methods [ Xia et al ASONAM 2012 ] Ø JC (Jaccard coefficient) Ø AA (Adamic/Adar) Ø Katz (Katz index) Ø PA (Preferential attachment) 12
RQ1: Performance of Link Prediction Observations: 1. Data-dependent supervised manner is more advantageous. 2. Positive e ff ect of modeling both explicit and implicit relations into the embedding process. 3. E ff ectiveness of modeling the explicit and implicit relations in di fff erent ways. 13
RQ2: Performance of Recommendation Observations: 1. Positive effect of considering information of weight 2. Importance of focusing on the higher-order proximities among vertices 3. Jointly training is superior to separately training + post-processing 14
Utility of Implicit Relations (RQ2) Observation: Modeling high-order implicit relations is e ff ective to complement with explicit relation modeling. 15
Random Walk Generator (RQ3) Observation: The biased and self-adaptive random walk generator contributes to learning better vertex embeddings. 16
Random Walk Generator (RQ3) Distribution of vertex degree � Our Generator: DeepWalk Generator: (c) Self-Adaptive generator Observation: The biased and self-adaptive random walk generator contributes to learning better vertex embeddings. 17
Case Study 19
Conclusions p Conclusions Ø Propose a dedicated approach for embedding bipartite networks Ø Jointly model both the explicit relations and higher-order implicit relations Ø Extensive experiments on several tasks of link prediction, recommendation, and visualization p Future work Ø Extend our BiNE method to model auxiliary side info Ø Investigate how to efficiently refresh embeddings for dynamic bipartite networks Ø Network embedding + adversarial training 20
Acknowledgments p Ming Gao ( �� ) p National Natural Science Foundation of China (East China Normal University) p The Press of East China Normal University p Leihui Chen ( ��� ) (East China Normal University) p National Research Foundation, Prime Minister’s Office, p Aoying Zhou ( ��� ) Singapore (East China Normal University) 25
Code available: Thank You for Your Attention
Negative Sampling p Optimizing a point-wise classification loss Ø p ( ! " | ! # ) can be approximate as: LSH-based Ø Following the similar formulations, we can get the counterparts for the conditional probability p ( $ | % # ) 23
LSH-based Negative Sampling p LSH-based negative sampling method Ø For a center vertex ! " , high-quality negatives should be the vertices that are dissimilar from ! " Frequency-based or LSH-based negative popularity-based sampling sampling Strategy High frequency objects Word Useless words Embedding Dissimilar objects Network Popular items or active users Embedding 24
Experimental Results p Performance of BiNE with different negative sampling strategies. Observations: 1. Two methods show roughly equivalent performance in most case. 2. However, there are situations (see VisualizeUS) in which LSH- based sampling method uses dissimilar information obtained from user behavior data can generate more reasonable negative samples 29
Recommend
More recommend
Explore More Topics
Stay informed with curated content and fresh updates.