AspEm: Embedding Learning by Aspects in Heterogeneous Information Networks

Similar documents
Mining Trusted Information in Medical Science: An Information Network Approach

User Guided Entity Similarity Search Using Meta-Path Selection in Heterogeneous Information Networks

CS249: ADVANCED DATA MINING

GraphGAN: Graph Representation Learning with Generative Adversarial Nets

This Talk. Map nodes to low-dimensional embeddings. 2) Graph neural networks. Deep learning architectures for graphstructured

PTE : Predictive Text Embedding through Large-scale Heterogeneous Text Networks

Network embedding. Cheng Zheng

Personalized Entity Recommendation: A Heterogeneous Information Network Approach

Query-Based Outlier Detection in Heterogeneous Information Networks

Mining Query-Based Subnetwork Outliers in Heterogeneous Information Networks

arxiv: v1 [cs.si] 12 Jan 2019

Community Preserving Network Embedding

Graph Classification in Heterogeneous

MetaGraph2Vec: Complex Semantic Path Augmented Heterogeneous Network Embedding

CS224W: Analysis of Networks Jure Leskovec, Stanford University

Link Prediction for Social Network

Effective Latent Space Graph-based Re-ranking Model with Global Consistency

node2vec: Scalable Feature Learning for Networks

Heterogeneous Graph-Based Intent Learning with Queries, Web Pages and Wikipedia Concepts

Link prediction in multiplex bibliographical networks

Personalized Recommendations using Knowledge Graphs. Rose Catherine Kanjirathinkal & Prof. William Cohen Carnegie Mellon University

Know your neighbours: Machine Learning on Graphs

TransPath: Representation Learning for Heterogeneous Information Networks via Translation Mechanism

Meta-path based Multi-Network Collective Link Prediction

Challenges in Multiresolution Methods for Graph-based Learning

Scalable Network Analysis

Integrating Meta-Path Selection with User-Preference for Top-k Relevant Search in Heterogeneous Information Networks

arxiv: v1 [cs.si] 17 Oct 2017

Meta-Graph Based HIN Spectral Embedding: Methods, Analyses, and Insights

Citation Prediction in Heterogeneous Bibliographic Networks

Chapter 1 Introduction

arxiv: v2 [cs.lg] 12 Feb 2019

Movie Recommender System - Hybrid Filtering Approach

Big Data Analytics Influx of data pertaining to the 4Vs, i.e. Volume, Veracity, Velocity and Variety

Reduce and Aggregate: Similarity Ranking in Multi-Categorical Bipartite Graphs

Holistic Analysis of Multi-Source, Multi- Feature Data: Modeling and Computation Challenges

Hierarchical Mixed Neural Network for Joint Representation Learning of Social-Attribute Network

Developing Focused Crawlers for Genre Specific Search Engines

Network Lasso: Clustering and Optimization in Large Graphs

c 2014 by Xiao Yu. All rights reserved.

Network Representation Learning Enabling Network Analytics and Inference in Vector Space. Peng Cui, Wenwu Zhu, Daixin Wang Tsinghua University

Measuring and Evaluating Dissimilarity in Data and Pattern Spaces

Online Social Networks and Media

Deep Learning on Graphs

Are Meta-Paths Necessary? Revisiting Heterogeneous Graph Embeddings

Graph Cube: On Warehousing and OLAP Multidimensional Networks

5/13/2009. Introduction. Introduction. Introduction. Introduction. Introduction

Link Prediction in Graph Streams

Deep Learning on Graphs

GraphNet: Recommendation system based on language and network structure

Data Mining: Dynamic Past and Promising Future

Combining Text Embedding and Knowledge Graph Embedding Techniques for Academic Search Engines

Relation Structure-Aware Heterogeneous Information Network Embedding

Meta-path based Multi-Network Collective Link Prediction

CODENSE v

Graph Mining and Social Network Analysis

Use of KNN for the Netflix Prize Ted Hong, Dimitris Tsamis Stanford University

Sparse Estimation of Movie Preferences via Constrained Optimization

Heterogeneous (Information) Networks. CS 6604: Data Mining Large Networks and Time-Series Paper Presentation Prashant Chandrasekar 11/01/17

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

Thanks to Jure Leskovec, Anand Rajaraman, Jeff Ullman

More Efficient Classification of Web Content Using Graph Sampling

CS 124/LINGUIST 180 From Languages to Information

Modeling Movie Character Networks with Random Graphs

QUINT: On Query-Specific Optimal Networks

CS 124/LINGUIST 180 From Languages to Information

Entity Retrieval over Structured Data

Limitations of XPath & XQuery in an Environment with Diverse Schemes

Mining of Massive Datasets Jure Leskovec, Anand Rajaraman, Jeff Ullman Stanford University Infinite data. Filtering data streams

Fast Contextual Preference Scoring of Database Tuples

Zurich Open Repository and Archive. Private Cross-page Movie Recommendations with the Firefox add-on OMORE

Estimating Sizes of Social Networks via Biased Sampling

Semi-Supervised Clustering with Partial Background Information

CS 124/LINGUIST 180 From Languages to Information

Using Machine Learning to Optimize Storage Systems

over Multi Label Images

On Smart Query Routing: For Distributed Graph Querying with Decoupled Storage

arxiv: v1 [cs.si] 3 Oct 2017

Bring Semantic Web to Social Communities

MODES v1.1 User Manual

Chapter 8 Prototype System Based on Heterogeneous Network

Pseudo-Implicit Feedback for Alleviating Data Sparsity in Top-K Recommendation

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

Deep Attributed Network Embedding

RiMOM Results for OAEI 2009

An Empirical Analysis of Communities in Real-World Networks

An Exploratory Journey Into Network Analysis A Gentle Introduction to Network Science and Graph Visualization

Unsupervised Multi-view Nonlinear Graph Embedding

SCALABLE KNOWLEDGE BASED AGGREGATION OF COLLECTIVE BEHAVIOR

Discovering Complex Matchings across Web Query Interfaces: A Correlation Mining Approach

arxiv: v1 [cs.si] 14 Mar 2017

Rich feature hierarchies for accurate object detection and semantic segmentation

Predicting Social Links for New Users across Aligned Heterogeneous Social Networks

CS224W Project Write-up Static Crawling on Social Graph Chantat Eksombatchai Norases Vesdapunt Phumchanit Watanaprakornkul

Semi-supervised Methods for Graph Representation

Fast Inbound Top- K Query for Random Walk with Restart

Method to Study and Analyze Fraud Ranking In Mobile Apps

A Composite Trust based Public Key Management in MANETs

Mining Significant Graph Patterns by Leap Search

arxiv: v1 [cs.lg] 19 May 2018

Transcription:

AspEm: Embedding Learning by Aspects in Heterogeneous Information Networks Yu Shi, Huan Gui, Qi Zhu, Lance Kaplan, Jiawei Han University of Illinois at Urbana-Champaign (UIUC) Facebook Inc. U.S. Army Research Laboratory

Wong UC Berkeley Christine ScienceNicodemus In real Wong world applications, objects of different types can have different Kaecilius UC Berkeley ScienceNicodemus relations, Wong Stanford Stephen which form Universityheterogeneous information networks (HINs). Kaecilius Berkeley, Urbana, CA IL Berkeley, Urbana, CA IL Christine Mordo UC Berkeley Stanford Typed nodes: objects University Kaecilius Berkeley, CA Mordo Mordo Stephen Stanford Typed edges: relations University Stephen Computer Computer Computer Science UIUC Nicodemus UIUC 2 [movie] [genre] [user] [director] [user] reviews [movie] [movie] is of [genre] [director] directs [movie] musical (genre) A toy movie reviewing network musical (genre)

Christine 3 Heterogeneous information networks (HINs) are ubiquitous. Nicodemus Wong Mordo Kaecilius Stephen UIUC Christine Nicodemus Wong Mordo UC Berkeley IMDb Network Bibliographical Network Biomedical Network Kaecilius Stephen Stanford University UIUC Urbana, IL Christine Berkeley, CA Nicodemus Wong Mordo Computer Science UC Berkeley Kaecilius Stanford University Economic Graph Stephen Social Network Facebook Open Graph Figures partly from http://xren7.web.engr.illinois.edu/yahoo-dais_award.html, http://www.garrygolden.com/2015/07/06/linkedin-economic-graph-technosolutionism, and http://www.businessinsider.com/explainer-what-exactly-is-the-social-graph-2012-3

4 Input network Network embedding has been heavily studied recently as a representation learning method. Embed represent nodes by vectors in the embedding space The learned vectors can be used as features in downstream applications Node classification Link Prediction Community detection Recommendation Embedding space Figures from Perozzi, et al. Deepwalk: Online learning of social representations. In KDD, 2014.

5 We are motivated to study the problem of Embedding Learning in Heterogeneous Information Networks (HINs). What would happen when we embed nodes of various types into the embedding space? Input network Embedding space

6 While the heterogeneity in HINs carries rich information, it also poses special challenges. If we embed all nodes into the same metric space musical (genre) musical (genre) likes musical, so he should be close to musical. likes Spielberg, so he should be close to Spielberg. One embedding space Spielberg is semantically dissimilar to musical, so their S. Spielberg (director) S. Spielberg (director) embeddings are far apart. As a result, turns out to be close to neither. we would suffer from information loss due to the incompatibility among node and edge types. It is of interest to develop embedding method that can alleviate this problem i.e., preserve s preference for both musicals and Spielberg s movies.

7 We alleviate the problem of information loss due to the incompatibility by embedding representative aspects in an HIN, where a representative aspect is a unit representing one compatible semantic facet of the HIN, and propose the AspEm framework (short for Aspect Embedding)

The AspEm Framework Overview 8 Formally, we define an aspect of an HIN by a connected subgraph of its schema. user director movie genre actor An HIN Its schema (an abstraction of the type information)

The AspEm Framework Overview 9 Formally, we define an aspect of an HIN by a connected subgraph of its schema user user director Schema level movie movie genre An HIN user director movie Network level genre actor Its schema Two example aspects of the HIN

S. Spielberg (director) Multiple embedding spaces One embedding space S. Spielberg (director) The AspEm Framework Overview One embedding space 10 S. Spielberg (director) musical S. Spielberg (genre) AspEm first selects a set of aspects, then embeds each(director) aspect into its own metric space S. Spielberg (director) user director Space reflecting the director aspect of the IMDb network Space reflecting the genre aspect of the IMDb network movie musical (genre) musical (genre) Select aspects One embedding space S. Spielberg (director) user movie Embed into respective spaces Multiple embedding spaces S. Spielberg (director) Space reflecting the director aspect of the IMDb network musical (genre) S. Spielberg (director) genre Space reflecting the genre aspect of the IMDb network S. Spielberg (director) and finally, for any node involved in multiple aspects, concatenates to build its final embedding. Space reflecting the director aspect of the IMDb network Multiple embedding spaces

The AspEm Framework Aspect Selection 11 How to select aspects? When supervision in downstream application is available, one may choose the set of aspects that perform the best in the downstream application. When supervision is not available, can we still select a set of representative aspects? Yes. Since a representative aspect corresponds to one compatible semantic facet, we should be able to find them by dataset-wide statistics that measures incompatibility. In other words, the selected aspects should not have incompatible node types and edge types within themselves.

The AspEm Framework Aspect Selection 12 To quantify such incompatibility, we propose the following Jaccard coefficient based measure for aspect incompatibility. An aspect A base aspect The incompatibility of an aspect is aggregated from all its base sub-aspects. A central node The incompatibility of a base aspect depends on the inconsistency (γ) observed by its central nodes. The inconsistency (γ) captures the difference in reachability via different edge types. This measure satisfies properties: non-negativity, monotonicity, and convexity w.r.t. aspects. We will further validate its effectiveness by experiments.

The AspEm Framework Aspect Selection 13 AspEm selects a set of aspects with incompatibility below a specified threshold. More general (bigger), more incompatibility More specific (smaller), less incompatibility

The AspEm Framework Embedding One Aspect 14 To embed each selected representative aspect: AspEm is a flexible framework, and one can choose their favorite network embedding algorithm. In our instantiation, we adapt the LINE (WWW 15) [1] algorithm and further distinguish edge types. The probability inferred from embedding: The empirical probability observed in data: Minimizing the difference between the two probabilities is equivalent to minimizing the following objective function: [1] Tang, et al. Line: Large-scale information network embedding. In WWW, 2015.

Experiments 15 Datasets and evaluation tasks DBLP: a bibliographical network in the computer science domain: Link prediction task: author identification to infer the authors of a paper. Classification task: inferring the research group and the research area of authors. IMDb: a movie reviewing network. Link prediction task: predicting if a user will review a movie.

Experiments 16 Aspects selected by AspEm DBLP: APRTV and APY IMDb: UMA, UMD and UMG Aspect 1: APRTV Aspect 2: APY Aspect 1: UMA Aspect 2: UMD Aspect 3: UMG

[2] Perozzi, et al. Deepwalk: Online learning of social representations. In KDD, 2014. [3] Grover, et al. node2vec: Scalable feature learning for networks. In KDD, 2016. [4] Tang, et al. Line: Large-scale information network embedding. In WWW, 2015. Experiments 17 Baselines SVD: a matrix factorization based method. DeepWalk [2]: a homogeneous network embedding method, which samples multiple walks starting from each node. Equivalent to node2vec [3] under default parameters. LINE [4]: a homogeneous network embedding method, which considers first-order and second-order neighbors. OneSpace: a heterogeneous network embedding method and an ablated version of AspEm. It uses heterogeneous negative sampling to distinguish node types, but do not model aspects or embed into multiple metric spaces.

Experiments 18 AspEm uniformly outperformed all four baselines in both link prediction and classification tasks. In particular, AspEm yielded better results than OneSpace, which confirms our intuition that incompatibility can exist among aspects, and explicitly modeling aspects can help better preserve the semantics of an HIN.

Experiments 19 Are the representative aspects selected by the incompatibility measure of AspEm really good? We exhaust and experiment with all comparable combinations of aspects in the DBLP link prediction task. Selected by AspEm The set selected by AspEm indeed perform the best.

Experiments 20 Parameter study: In the DBLP link prediction task, the performance grows as embedding dimension or number of edge sampled increases at first. The change becomes less significant when dimension reaches 100, and number of edges sampled reaches 1,000 million. Dimension of embedding space Dimension of embedding space Number of edges sampled Number of edges sampled

Summary 21 We provide an insight that an HIN can have multiple representative aspects that do not align with each other. We thereby identify that embedding algorithms employing only one metric space may suffer from information loss due to such incompatibility. We propose a flexible HIN embedding framework, named AspEm, that can mitigate the information loss by modeling aspects. We propose a representative aspect selection method for AspEm using statistics of HINs without additional supervision. Code available at https://github.com/ysyushi/aspem.