Hierarchical Mixed Neural Network for Joint Representation Learning of Social-Attribute Network
|
|
- Kelley Short
- 5 years ago
- Views:
Transcription
1 Hierarchical Mixed Neural Network for Joint Representation Learning of Social-Attribute Network Weizheng Chen 1(B), Jinpeng Wang 2, Zhuoxuan Jiang 1, Yan Zhang 1, and Xiaoming Li 1 1 School of Electronics Engineering and Computer Science, Peking University, Beijing, China cwz.pku@gmail.com, jzhx1211@gmail.com, zhyzhy001@gmail.com, lxm.at.pku@gmail.com 2 Microsoft Research, Beijing, China jinpwa@microsoft.com Abstract. Most existing network representation learning (NRL) methods are designed for homogeneous network, which only consider topological properties of networks. However, in real-world networks, text or categorical attributes are usually associated with nodes, providing another description for networks in a different perspective. In this paper, we present a joint learning approach which learns the representations of nodes and attributes in the same low-dimensional vector space simultaneously. Particularly, we show that more discriminative node representations can be acquired by leveraging attribute features. The experiments conducted on three social-attribute network datasets demonstrate that our model outperforms several state-of-the-art baselines significantly for node classification task and network visualization task. Keywords: Social-attribute network Representation learning Joint learning 1 Introduction The growth of online social media produces massive amounts of user-generated content, such as tweets posted in Twitter, personal profile in LinkedIn. These data and social relations among users make up complex heterogeneous information network [16], which is an effective organization form of multi-source data. Mining such heterogeneous information network is crucial for various research tasks and commercial applications, for example, node classification [9] and product recommendation [15]. During the last few years, representation learning, also known as embedding, has become a promising and powerful tool in network analysis area. Since the density of typical social network is usually quite small in the real world, c Springer International Publishing AG 2017 J. Kim et al. (Eds.): PAKDD 2017, Part I, LNAI 10234, pp , DOI: /
2 Hierarchical Mixed Neural Network for Joint Representation Learning 239 traditional network representation such as adjacency matrix suffers from the data sparsity problem. Thus we can t apply most statistical machine learning algorithms to solve network analysis task directly. To overcome this problem, many NRL methods which aim to project the nodes into a low-dimensional continuous vector space have been proposed. Among those representative methods, DeepWalk [14], LINE [17], GraRep [3] and node2vec [8] learn general node representations which are not tuned for specific task for homogeneous network by only considering the structural features. However, both text and categorical attributes play an crucial role in real-life networks, e.g., papers or authors in citation networks are associated with corresponding text content, users in Twitter or Facebook have profiles with categorical attributes such as gender and job. It is necessary to uncover the potential effect of nodes attributes in NRL process. To utilize the rich text content information of nodes, Yang et al. [20] presents Text-associated DeepWalk (TADW) to learn network representations from both network structure and text attributes in an inductive matrix completion framework. Because of high dimensionality of text attributes (namely words), singularvalue decomposition is performed on the node-attribute matrix to get robust attribute features in TADW. However, TADW has two serious weaknesses: 1. Unlike text attributes, categorical attributes space is low-dimensional, e.g., only up to dozens of demographic attributes appears in the mobile social network [6] and Twitter network [4], which makes TADW unsuitable in such a scenario. 2. The performance of TADW falls fast if texts of some nodes are missing. Note that, text information is often incomplete. For example, in online social media, some users text features are difficult to obtain due to their privacy settings or they actually never publish any texts. To overcome the above problems, in this paper, we propose Social-Attribute Network Representation Learning (SANRL), a scalable joint NRL framework which preserve both structural and attribute information in the unified representations. Compared with TADW, which only learn representations of nodes, our model also learn representations of attributes in the same low dimensional vector space. A hierarchical mixed neural network model is adopted to model the interactive relationship between the nodes and the attributes. We conduct experiments with three real-world network datasets, including a social network with categorical attributes, a co-author network with text attributes and a citation network with text attributes. In summary, this paper has the following three major contributions: 1. We propose a network representation learning model for the heterogeneous social-attribute networks, which can handle either categorical attributes or text attributes. Our model can make use of limited attribute information by using a coupled architecture, which makes it more flexible in real scenarios. 2. Our model can map the nodes and the attributes to the same space, which provide meaningful features for various applications. 3. The experimental results show that our model outperforms other competitive baselines significantly.
3 240 W. Chen et al. 2 Related Work Most existing unsupervised NRL models only consider the network structure information. These models are mainly based on two classical assumptions. The first is only suitable for the undirected networks, called smoothness assumption, in which two linked nodes should have close representations. Laplacian Eigenmaps [1], LINE with first-order proximity, SDNE [19] all adopt the smoothness assumption. The second is also applicable to the directed networks, in which two nodes should have close representations if they have similar neighbors. Here, the neighbors can be high-order. DeepWalk, GraRep and node2vec all adopt this assumption. To incorporate the text attribute information, TADW adopts the inductive matrix completion technique to incorporates the tfidf matrix into Deep- Walk. As far as we know, how to use the categorical attributes of nodes has not been studied yet. Recently, learning network representations in a semi-supervised way has drawn many attentions by incorporating the label information [11, 18, 21]. All these models are trained in a transductive way, which means that they use a combination of an unsupervised NRL model and a classifier trained on the labeled nodes. However, all these models only learn the distributed representations of the nodes. Unlike semi-supervised NRL models which leverage the label information to get more discriminative node representation, our motivation is to enhance node representation with rich categorical or text attributes via jointly projecting the nodes and the attributes into the same vector space. In this paper, we follow the line of the unsupervised NRL models. In the future, we will explore the semi-supervised extension of our model by adding a classifier to it. 3 Problem Definition We adopt a social-attribute network (SAN) [7] G =(V,E,A,C) to represent an attributed network in this work. V is the set of social nodes, each representing a data object. E is the set of social links, each representing an edge between two social nodes. A is the set of attribute nodes, each representing an attribute. C is the set of attribute links, each representing an affiliation relationship between an attribute node and a social node. For simplification, we use A(v) ={a (a, v) C} to represent the attribute sets of node v V. All attribute links are undirected, while social links can be directed or undirected and weighted or unweighted depending upon the data type. Moreover, the weight of attribute links are defined as relative importance of certain attribute for corresponding social node. As shown on the left side of Fig. 1, we present a sample SAN which has five social nodes and four attribute nodes. Our research target is to a learn informative continuous vector representation u R d for each social node of a SAN, where d V. In vector space R d, both structural information and attribute information should be preserved. Then the learned low-dimensional representations can used as input features to a variety of machine learning models such as logistic regression for classification task, k-means for clustering task. For
4 Hierarchical Mixed Neural Network for Joint Representation Learning 241 Attribute layer A vector space a1 a2 a3 a4 a1 v1 a4 v 1 v 2 v 3 v 4 a 1 a 2 a 3 a 4 Representation learning a2 v2 a3 v3 v2 v4 v5 v1 v3 v4 v5 attribute node social node Social layer attribute link social link u v or u a Fig. 1. Illustration of learning representations jointly for a toy social-attribute network. Fig. 2. Mixed binary Huffman tree used in SANRL. simplification and clarification, in the following paper, when we say node, we refer to a social node. And when we say attribute, we refer to an attribute node. 4 SANRL Model 4.1 Loss Function for Structural Information We adapt DeepWalk to learn representations of social nodes based on the assumption that nodes which have similar neighborhoods will have similar representations. Here, neighborhoods refer to both direct neighbors and higher order neighbors. Next, we give a brief outline of DeepWalk. Given a social network G = (V,E) (e.g., social layer of Fig. 1), Deep- Walk generates many node sequences as training data. More specifically, staring from a node v 1, DeepWalk generates γ fixed-length sequences of nodes S v1 k = {v 1,v 2,...,v t } through random walk for 1 k γ. Repeat the above process for every node v V.ThenwegetasetS seq which contains γ V node sequences. By feeding shuffled S seq to Skip-Gram [12], an efficient and scalable method for learning word representations, the following objective loss function will be minimized: O 1 = log P (S seq ) = log P (Sk) v 1 k γ = log P (v i w,...,v i 1,v i+1,...,v i+w v i ) 1 k γ v i Sk v = log P (v j v i ), 1 k γ v i Sk v i w j i+w,j i where w is the size of the sliding window and P (v j v i ) is formulated using softmax function: P (v j v i )= exp(u v T j u vi ) exp(u v T u vi ), (2) where u v and u v are the input and output representations of social node v. (1)
5 242 W. Chen et al. In the context of SAN, we replace Eq. (2) with the following Eq. (3) ino 1 to take attribute nodes into consideration: P (v j v i )= exp(u T v exp(u v T j u vi ) u vi )+ a A exp(u T a u vi ). (3) Here u a and u a are the input and output representations of attribute node a. In addition, the original DeepWalk is only applicable to unweighted networks. Therefore, weighted random walk is adopted in the sampling process to handle weighted network in our model. 4.2 Loss Function for Attribute Information Considering a social node v with categorical attribute sets A(v), we use the co-occurrence patterns to model P (A(v)) log P (A(v)) = {log P (a i v) + log P (v a i )+ log P (a j a i )}, (4) a i A(v) a j A(v),j i where log P (a i v j ), log P (v j a i ) and log P (a j a i ) are defined as: P (a i v j )= P (v j a i )= P (a j a i )= exp(u T a i u vj ) exp(u T v u vj )+ a A exp(u T v j u ai ) exp(u T v u ai )+ a A exp(u T a j u ai ) exp(u T v u ai )+ a A exp(u T a u vj ), (5) exp(u T a u ai ), (6) exp(u T a u ai ). (7) The above three kinds of conditional probability can capture different similarities in the SAN. Firstly, by maximizing P (a i v j ), nodes with many similar attributes will tend to have close representations, which is the key idea of the PV-DBOW model [10]. Secondly, by maximizing P (a j a i ), paradigmatic relations will be modeled. As a consequence, attributes with many similar contexts will tend to have close representations, which is the key idea of the Word2Vec model [12]. Finally, by maximizing P (v j a i ), syntagmatic relations will be modeled. Then attributes often co-occur will tend to have close representations. Unlike categorical attributes, word order and word frequency are essential properties of text attributes, which are not captured in Eq. (4). To make Eq. (4) suitable for both categorical and text attributes, we replace A(v) by a attribute set sampled from A(v). More specifically, given a node v V, we generate γ fixed-length sequences of attributes Tk v = {a 1,...,a w } through random sampling for 1 k γ. For text attributes, Tk v is a text window whose length is w sampled from the corresponding document of v. But for categorical attributes,
6 Hierarchical Mixed Neural Network for Joint Representation Learning 243 Tk v is generated by sampling w attributes independently from A(v). Using T to represent the γ V attribute sequences acquired in the preceding procedure, we define the following objective loss function for attribute information: O 2 = log P (T ) = = 1 k γ log P (T v k ) 1 k γ a i Tk v {log P (a i v) + log P (v a i )+ a j A(v),j i log P (a j a i )}. (8) 4.3 Loss Function of SANRL To integrate both structure and attribute information into a joint representation learning framework, we use a weighted linear combination of O 1 and O 2 to formulate our objective loss function of the SANRL model: O joint = O 1 + λo 2, (9) where λ is a trade-off parameter. If the network has rich discriminative attributes, λ should be a big number. Otherwise, a small λ is suitable. By minimizing O joint, we can get two resulting vectors u v and u v for each node v, and two resulting vectors u a and u a for each attribute a. Finally, u v will be used as the feature vector of v. 4.4 Learning and Complexity Analysis The optimization scheme of SANRL model is similar to DeepWalk, in which Skip- Gram is applied to maximize the co-occurrence probability among node-node pairs, node-attribute pairs, attribute-attribute pairs and attribute-node pairs. In practice, all the representation vectors are initialized randomly at first. Then node sequences and attribute sequences are sampled alternately and iteratively as the input data streams. In our implementation, the effect of λ is reflected by controlling the sampling probability of attribute sequences. By using the back-propagation algorithm to estimate the derivatives, Eq. (9) can be optimized by adopting the asynchronous stochastic gradient descent (ASGD) algorithm. Directly computing softmax functions defined in Eqs. (3), (5), (6) and (7) is very expensive, so we use hierarchical softmax technique to speed up training. In previous works [5,13], hierarchical softmax is widely used to train a similar neural network, in which only one type of objects such as words or nodes are mapped to the leaves of one binary Huffman tree. However, the application of hierarchical softmax in our model is very different from them. As shown in Fig. 2, every social node and attribute node are associated with one leaf in a same single mixed Huffman tree. And the input of this neural network can be either the representation vector of a node or an attribute. Thus given an arbitrary training
7 244 W. Chen et al. instance whose form is an input-output pair, the computational complexity of computing P (output input) can be reduced from O( V + A ) to O(log( V + A )). Overall, the time complexity of SANRLis γ V d(t + λw) log( V + A ) Because A is much more smaller than V in most cases, the computational complexity of SANRL is still acceptable. 5 Experiments 5.1 Dataset Weibo Dataset. We have crawled 294,634 users detailed information from Sina Weibo, the most popular microblogging service in China, including profiles and 3,183,187 following relationships among them. In our Weibo dataset, the user profile contains four demographic variables: Gender: male; female. Age: 1 11; 12 17; 18 29; 30 44; 45 59; 60+. Education: literature, history; natural science; engineering; economics; medicine; art. Job: Internet industry; creative design industry; cultural media industry; public service industry; manufacturing industry; scientific research industry; pharmaceutical industry; business management industry; Here, we regard every possible value of each demographic variable as a categorical attribute. Moreover, an user may have multiple education or job attributes but values in gender and age are exclusive. We build two SANs based on Weibo dataset for node classification task, i.e., the Weibo-education network and the Weibo-job network. In the Weibo-education network, education attributes are treated as labels, which are not used in the representation learning process. The Weibo-job network is organised in the same manner. DBLP Dataset. We use DBLP-four-area dataset provided in [9] to build a weighted co-author network. This data contains 20 major computer conferences from four related areas, i.e., data mining, database, information retrieval and machine learning, and 27,199 authors and all their publications in these conferences. If two authors have co-authored a paper, we add an undirected edge between them. The weight of the edge is the number of their collaborative papers. Finally, the number of edges in 66,832. The titles of all the paper published by one author is recognized as his or her text attributes. The size of the word vocabulary is 12,091. If an author publishes a paper in a certain conference, the research areas of this conference will be added to the author s label set. 5.2 Compared Algorithms DeepWalk [14]. DeepWalk is the first work which adopts the neural network language model to solve NRL problem.
8 Hierarchical Mixed Neural Network for Joint Representation Learning 245 LINE[17]. LINE can learn two representation vectors for each node by optimizing two carefully designed objective function that preserves the first-order proximity and second-order proximity. Then the two representations are concatenated as the final representation. LDA[2]. Latent Dirichlet Allocation (LDA) is a classical probabilistic topic model. Each node can be represented as a topic distribution vector. TADW [20]. TADW is a state-of-the-art NRL algorithm based on matrix decomposition. First, Singular-value decomposition is performed on the tfidf matrix to get robust text features of nodes. Then, the text features and a node relation matrix are fed to an inductive matrix completion framework to get node representations. SANRL. Our proposed method. For SANRL, the sliding window size w = 10, the length of each node sequence t = 40, number of node sequences for per node γ = 80. We set the trade-off parameter λ = 2 for two Weibo networks and λ = 8 for DBLP-author network. SANRL doc. A simplified version of SANRL, in which the final objective loss function is Eq. (8). We treat the attribute sets of a node as its pseudo document. Parameter settings of SANRL doc is same to SANRL. 5.3 Node Classification Table 1. Macro-F1 (%) of node classification on the Weibo-education network. % labeled nodes 10% 20% 30% 40% 50% 60% 70% 80% 90% DeepWalk LINE SANRL doc SANRL Table 2. Macro-F1 (%) of node classification on the Weibo-job network. % labeled nodes 10% 20% 30% 40% 50% 60% 70% 80% 90% DeepWalk LINE SANRL doc SANRL Following the settings in previous works [14, 17], we also use the multi-label node classification task to evaluate the quality of the representation vectors learned by different models. The one-vs-the-rest logistic regression classifier implemented in LibLinear is used in our experiments. The Macro-F1 is chosen as the evaluation metric. We follow the suggested parameter settings in the original
9 246 W. Chen et al. Table 3. Macro-F1 (%) of node classification the DBLP-author network. % labeled nodes 10% 20% 30% 40% 50% 60% 70% 80% 90% DeepWalk LINE SANRL doc SANRL LDA TADW papers for the baseline models. Note that, since there is no text attributes in the Weibo dataset, LDA and TADW are inapplicable to Weibo. Tables 1, 2, and 3 show the results of classification with different training ratios on three networks when d = 128 for all the models. All reported results are averaged over 10 runs. Firstly, we observe that SANRL always significantly outperform other baselines. Compared with DeepWalk, SANRL achieves nearly 6% and 10% improvement on two Weibo networks and 13% improvement on DBLP-author network, which proves that the network represent can be enhanced with either categorical attributes or text attributes. Secondly, on the DBLP-author network, the two content-based methods, SANRL doc and LDA, perform better than the structure-based methods, Deep- Walk and LINE. By benefiting from incorporating attributes information and structure information, SANRL and TADW both outperform other four methods. But the relative improvement of SANRL over TADW is around 3.5% since attributes information and structure information is better balanced with a tunable trade-off parameter λ in SANRL. 5.4 Parameter Sensitivity We also explore the sensitivity of the performance w.r.t. the dimension d and the trade-off parameter λ. Here, we take the DBLP-author network as an example. By setting the training ratio to 20%, we report the grid search results over d and λ on the DBLP-author network in Fig. 3. We observe that SANRL achieves best performance when λ =8andd = 256. To further investigate the effect of increasing dimension d, we use Fig. 4 to show the Micro-F1 and Macro-F1 curves of different models. By varying d from 64 to 1024, we can see that SANRL consistently performs best. In SANRL, we add a constraint to DeepWalk to let nodes with similar attributes have similar representations. Our motivation is based on the statistics shown in Fig. 5. The Jaccard coefficient is calculated between the attribute sets of two social nodes sampled from the network at random. Then we compute the conditional probability that two nodes share a same label given the corresponding discretized Jaccard coefficient. Our assumption is validated on all
10 Hierarchical Mixed Neural Network for Joint Representation Learning Micro-F Macro-F Micro-F Macro-F d=64 d = 512 d=128 d = 1024 d= λ 0.68 d=64 d=512 d=128 d = 1024 d= λ 0.60 DeepWalk LDA LINE SANRLdoc TADW SANRL Dimension 0.60 DeepWalk LDA LINE SANRLdoc TADW SANRL Dimension Fig. 3. Grid search over dimension d and λ. Fig. 4. Performance over dimension d Weibo education DeepWalk TADW LINE SANRL Probability of sharing labels [0, 0.2) [0.2, 0.4) [0.4, 0.6) [0.6, 0.8) [0.8, 1.0] Weibo job [0, 0.2) [0.2, 0.4) [0.4, 0.6) [0.6, 0.8) [0.8, 1.0] DBLP author [0, 0.2) [0.2, 0.4) [0.4, 0.6) [0.6, 0.8) [0.8, 1.0] Jaccard coefficient Macro-F α Fig. 5. Illustration of interdependencies between labels of users and attributes of users. Fig. 6. Parameter sensitivity w.r.t α three networks. Note that the DBLP-author network has much more discriminative attributes than those two Weibo networks. That s why a smaller λ is more suitable for the Weibo dataset. Next, we will test the robustness of SANRL in an incomplete data scenario. We random choose some users and remove their text contents. We use α to represent the percentage of users whose texts are removed. LDA and SANRL doc could not work in this situation. By ranging α from 0.5 to 0.9, we report the Macro-F1 of different models when the training ratio is 0.1 in Fig. 6. Obviously, SANRL has greater robustness than TADW whose performance is very poor. The reason lies with the fact that the text content matrix calculated from the tf-idf matrix which has many rows with all zeros is uninformative. In contrast, SANRL benefits from limited text information by using a coupled design. 5.5 Network Visualization Network visualization is an essential task in network analysis area. In this part, we use the t-sne package which takes the node representation vectors as inputs to generate network layouts in a 2D space. Our target is to compare the properties of the layouts of DeepWalk and SANRL qualitatively. Because our previous three networks are not mono-labeled, we build a paper-citation network by using
11 248 W. Chen et al. a large-scale DBLP database 1. We select three computer conferences to represent different research fields: SIGMOD from database, ICML from machine learning, and ACL from natural language processing. Papers published on these three conferences and citation relationships among them are extracted to build this DBLP-citation network. We also use words in the title of each paper as their text attributes. Finally, this network has 1280 papers from SIGMOD, 568 papers from ICML, 1,206 papers from ACL, 6,647 edges and 3,814 unique attributes. Fig. 7. t-sne 2D representations on the DBLP-citation network. We use blue, green and red to indicate papers from SIGMOD, ICML and ACL respectively. Words are colored black (Color figure online) Under the same parameter configuration, the 2-D layouts of the DBLPcitation network are shown in Fig. 7. Three different colors, blue, green and red are used to indicate papers from SIGMOD, ICML and ACL respectively. Attributes are colored black. In Fig. 7, (a) and (b) are generated by only feeding the node representations to the t-sne toolkit. And (c) and (d) are generated by feeding the node representations and attributes representations together to the t-sne toolkit. We have three observations: First, We observe that node distributions in (b) and (a) are very similar, but groups in (b) seem to be more tighter. Second, after plotting the papers and the words in a same space simultaneously in (c), the words are spread throughout the space, but most of them are densely populated in the center while different paper groups are surrounding the word groups. Finally, it is obvious that each paper groups are becoming more denser and separable in (d) which is obtained by removing words from (c). These observations show that the distinguishability of node representations is improved significantly by jointing learning the representations of their associated attributes in SANRL model. 1 Citation.
12 Paper title Hierarchical Mixed Neural Network for Joint Representation Learning 249 Table 4. Top-10 related words for selected papers. Recommended words Dynamic multidimensional histograms Histogram, filesystem, stholes, braid, histograms, lag, partiqle, frequencies, filtered, multidimensional Computing weakest readings Weakest, readings, ambiguities, scope, semantically, formalisms, polysemic, hole, dominance, distributions 5.6 Case Study To explore the correlations among node vectors and attribute vectors learned in SANRL, we provide a case study on the DBLP-citation dataset. We recommend top-10 most related words to papers by calculating the cosine similarly between the paper vector and the word vector. As shown in Table 4, we can see that though the titles of two selected papers both have only three words, SANRL can find much richer related words by considering the citation relationship between the papers and the interactive relationship between the papers and the words. 6 Conclusion and Future Work In this paper, we propose SANRL, an efficient model which integrates both structure and attribute information into the NRL task. By embedding nodes and attributes into the same vector space, the quality of the node representations are improved significantly. The experimental results show that SANRL outperforms competitive baselines for different data mining tasks. We strive to adapt SANRL to learn more discriminative representations by using the semisupervised learning technique in the future. Acknowledgments. This work is supported by 973 Program with Grant No. 2014CB Yan Zhang is supported by NSFC with Grant Nos and , and MOE-RCOE with Grant No. 2016ZD201. We thank the anonymous reviewers for their comments. References 1. Belkin, M., Niyogi, P.: Laplacian eigenmaps and spectral techniques for embedding and clustering. In: NIPS, vol. 14, pp (2001) 2. Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent Dirichlet allocation. J. Mach. Learn. Res. 3, (2003) 3. Cao, S., Lu, W., Xu, Q.: GraRep: learning graph representations with global structural information. In: CIKM, pp ACM (2015) 4. Culotta, A., Ravi, N.K., Cutler, J.: Predicting the demographics of Twitter users from website traffic data. In: ICWSM. AAAI Press, Menlo Park (2015, in press)
13 250 W. Chen et al. 5. Djuric, N., Wu, H., Radosavljevic, V., Grbovic, M., Bhamidipati, N.: Hierarchical neural language models for joint representation of streaming documents and their content. In: WWW, pp (2015) 6. Dong, Y., Yang, Y., Tang, J., Yang, Y., Chawla, N.V.: Inferring user demographics and social strategies in mobile social networks. In: SIGKDD, pp ACM (2014) 7. Gong, N.Z., Talwalkar, A., Mackey, L., Huang, L., Shin, E.C.R., Stefanov, E., Shi, E.R., Song, D.: Joint link prediction and attribute inference using a social-attribute network. ACM TIST 5(2), 27 (2014) 8. Grover, A., Leskovec, J.: node2vec: scalable feature learning for networks. In: SIGKDD, pp ACM, New York (2016) 9. Ji, M., Sun, Y., Danilevsky, M., Han, J., Gao, J.: Graph regularized transductive classification on heterogeneous information networks. In: Balcázar, J.L., Bonchi, F., Gionis, A., Sebag, M. (eds.) ECML PKDD LNCS (LNAI), vol. 6321, pp Springer, Heidelberg (2010). doi: / Le, Q.V., Mikolov, T.: Distributed representations of sentences and documents (2014). arxiv preprint arxiv: Li, J., Zhu, J., Zhang, B.: Discriminative deep random walk for network classification. In: ACL, pp (2016) 12. Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., Dean, J.: Distributed representations of words and phrases and their compositionality. In: NIPS, pp (2013) 13. Mnih, A., Hinton, G.E.: A scalable hierarchical distributed language model. In: Advances in Neural Information Processing Systems, pp (2009) 14. Perozzi, B., Al-Rfou, R., Skiena, S.: DeepWalk: online learning of social representations. In: SIGKDD, pp ACM (2014) 15. Shi, C., Zhang, Z., Luo, P., Yu, P.S., Yue, Y., Wu, B.: Semantic path based personalized recommendation on weighted heterogeneous information networks. In: CIKM, pp ACM (2015) 16. Sun, Y., Han, J.: Mining heterogeneous information networks: a structural analysis approach. ACM SIGKDD Explor. Newslett. 14(2), (2013) 17. Tang, J., Qu, M., Wang, M., Zhang, M., Yan, J., Mei, Q.: Line: large-scale information network embedding. In: WWW. ACM (2015) 18. Tu, C., Zhang, W., Liu, Z., Sun, M.: Max-margin DeepWalk: discriminative learning of network representation. In: IJCAI, pp (2016) 19. Wang, D., Cui, P., Zhu, W.: Structural deep network embedding. In: SIGKDD, pp , ACM, New York (2016) 20. Yang, C., Liu, Z., Zhao, D., Sun, M., Chang, E.Y.: Network representation learning with rich text information. In: IJCAI, pp AAAI Press (2015) 21. Yang, Z., Cohen, W., Salakhutdinov, R.: Revisiting semi-supervised learning with graph embeddings (2016). arxiv preprint arxiv:
Network embedding. Cheng Zheng
Network embedding Cheng Zheng Outline Problem definition Factorization based algorithms --- Laplacian Eigenmaps(NIPS, 2001) Random walk based algorithms ---DeepWalk(KDD, 2014), node2vec(kdd, 2016) Deep
More informationDeepWalk: Online Learning of Social Representations
DeepWalk: Online Learning of Social Representations ACM SIG-KDD August 26, 2014, Rami Al-Rfou, Steven Skiena Stony Brook University Outline Introduction: Graphs as Features Language Modeling DeepWalk Evaluation:
More informationPTE : Predictive Text Embedding through Large-scale Heterogeneous Text Networks
PTE : Predictive Text Embedding through Large-scale Heterogeneous Text Networks Pramod Srinivasan CS591txt - Text Mining Seminar University of Illinois, Urbana-Champaign April 8, 2016 Pramod Srinivasan
More informationA United Approach to Learning Sparse Attributed Network Embedding
2018 IEEE International Conference on Data Mining A United Approach to Learning Sparse Attributed Network Embedding Hao Wang 1, Enhong Chen 1,*,QiLiu 1, Tong Xu 1, Dongfang Du 2, Wen Su 2, Xiaopeng Zhang
More informationGraphGAN: Graph Representation Learning with Generative Adversarial Nets
The 32 nd AAAI Conference on Artificial Intelligence (AAAI 2018) New Orleans, Louisiana, USA GraphGAN: Graph Representation Learning with Generative Adversarial Nets Hongwei Wang 1,2, Jia Wang 3, Jialin
More informationnode2vec: Scalable Feature Learning for Networks
node2vec: Scalable Feature Learning for Networks A paper by Aditya Grover and Jure Leskovec, presented at Knowledge Discovery and Data Mining 16. 11/27/2018 Presented by: Dharvi Verma CS 848: Graph Database
More informationarxiv: v1 [cs.si] 12 Jan 2019
Predicting Diffusion Reach Probabilities via Representation Learning on Social Networks Furkan Gursoy furkan.gursoy@boun.edu.tr Ahmet Onur Durahim onur.durahim@boun.edu.tr arxiv:1901.03829v1 [cs.si] 12
More informationMetaGraph2Vec: Complex Semantic Path Augmented Heterogeneous Network Embedding
MetaGraph2Vec: Complex Semantic Path Augmented Heterogeneous Network Embedding Daokun Zhang 1, Jie Yin 2, Xingquan Zhu 3, and Chengqi Zhang 1 1 Centre for Artificial Intelligence, FEIT, University of Technology
More informationUnsupervised Multi-view Nonlinear Graph Embedding
Unsupervised Multi-view Nonlinear Graph Embedding Jiaming Huang,2, Zhao Li, Vincent W. Zheng 3, Wen Wen 2, Yifan Yang, Yuanmi Chen Alibaba Group, Hangzhou, China 2 School of Computers, Guangdong University
More informationRelation Structure-Aware Heterogeneous Information Network Embedding
Relation Structure-Aware Heterogeneous Information Network Embedding Yuanfu Lu 1, Chuan Shi 1, Linmei Hu 1, Zhiyuan Liu 2 1 Beijing University of osts and Telecommunications, Beijing, China 2 Tsinghua
More informationPaper2vec: Combining Graph and Text Information for Scientific Paper Representation
Paper2vec: Combining Graph and Text Information for Scientific Paper Representation Soumyajit Ganguly and Vikram Pudi International Institute of Information Technology Hyderabad, India. soumyajit.ganguly@research.iiit.ac.in,
More informationarxiv: v1 [cs.si] 14 Mar 2017
SNE: Signed Network Embedding Shuhan Yuan 1, Xintao Wu 2, and Yang Xiang 1 1 Tongji University, Shanghai, China, Email:{4e66,shxiangyang}@tongji.edu.cn 2 University of Arkansas, Fayetteville, AR, USA,
More informationHARP: Hierarchical Representation Learning for Networks
HARP: Hierarchical Representation Learning for Networks Haochen Chen Stony Brook University haocchen@cs.stonybrook.edu Yifan Hu Yahoo! Research yifanhu@oath.com Bryan Perozzi Google Research bperozzi@acm.org
More informationDomain-specific user preference prediction based on multiple user activities
7 December 2016 Domain-specific user preference prediction based on multiple user activities Author: YUNFEI LONG, Qin Lu,Yue Xiao, MingLei Li, Chu-Ren Huang. www.comp.polyu.edu.hk/ Dept. of Computing,
More informationNon-Linear Smoothed Transductive Network Embedding with Text Information
JMLR: Workshop and Conference Proceedings 63:1 16, 2016 ACML 2016 Non-Linear Smoothed Transductive Network Embedding with Text Information Weizheng Chen 1 cwz.pku@gmail.com Xia Zhang 1 zhangxia9403@gmail.com
More informationA Novel deep learning models for Cold Start Product Recommendation using Micro blogging Information
A Novel deep learning models for Cold Start Product Recommendation using Micro blogging Information Chunchu.Harika, PG Scholar, Department of CSE, QIS College of Engineering and Technology, Ongole, Andhra
More informationCombining Review Text Content and Reviewer-Item Rating Matrix to Predict Review Rating
Combining Review Text Content and Reviewer-Item Rating Matrix to Predict Review Rating Dipak J Kakade, Nilesh P Sable Department of Computer Engineering, JSPM S Imperial College of Engg. And Research,
More informationLearning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li
Learning to Match Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li 1. Introduction The main tasks in many applications can be formalized as matching between heterogeneous objects, including search, recommendation,
More informationSupervised classification of law area in the legal domain
AFSTUDEERPROJECT BSC KI Supervised classification of law area in the legal domain Author: Mees FRÖBERG (10559949) Supervisors: Evangelos KANOULAS Tjerk DE GREEF June 24, 2016 Abstract Search algorithms
More informationLink Prediction for Social Network
Link Prediction for Social Network Ning Lin Computer Science and Engineering University of California, San Diego Email: nil016@eng.ucsd.edu Abstract Friendship recommendation has become an important issue
More informationDeep Attributed Network Embedding
Deep Attributed Network Embedding Hongchang Gao, Heng Huang Department of Electrical and Computer Engineering University of Pittsburgh, USA hongchanggao@gmail.com, heng.huang@pitt.edu Abstract Network
More informationLinked Document Classification by Network Representation Learning
Linked Document Classification by Network Representation Learning Yue Zhang 1,2, Liying Zhang 1,2 and Yao Liu 1* 1 Institute of Scientific and Technical Information of China, Beijing, China 2 School of
More informationDeep Learning on Graphs
Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 22 March 2017 joint work with Max Welling (University of Amsterdam) BDL Workshop @ NIPS 2016
More informationThis Talk. Map nodes to low-dimensional embeddings. 2) Graph neural networks. Deep learning architectures for graphstructured
Representation Learning on Networks, snap.stanford.edu/proj/embeddings-www, WWW 2018 1 This Talk 1) Node embeddings Map nodes to low-dimensional embeddings. 2) Graph neural networks Deep learning architectures
More informationA Brief Review of Representation Learning in Recommender 赵鑫 RUC
A Brief Review of Representation Learning in Recommender Systems @ 赵鑫 RUC batmanfly@qq.com Representation learning Overview of recommender systems Tasks Rating prediction Item recommendation Basic models
More informationThe Algorithm of Automatic Text Summarization Based on Network Representation Learning
The Algorithm of Automatic Text Summarization Based on Network Representation Learning Xinghao Song 1, Chunming Yang 1,3(&), Hui Zhang 2, and Xujian Zhao 1 1 School of Computer Science and Technology,
More informationA Unified Framework for Community Detection and Network Representation Learning
JOURNAL OF L A T E X CLASS FILES, VOL. 14, NO. 8, AUGUST 2015 1 A Unified Framework for Community Detection and Network Representation Learning Cunchao Tu, Xiangkai Zeng, Hao Wang, Zhengyan Zhang, Zhiyuan
More informationNetwork Representation Learning with Rich Text Information
Network Representation Learning with Rich Text Information Cheng Yang 1,, Zhiyuan Liu 1,, Deli Zhao, Maosong Sun 1, Edward Y. Chang 1 Department of Computer Science and Technology, State Key Lab on Intelligent
More informationGraphNet: Recommendation system based on language and network structure
GraphNet: Recommendation system based on language and network structure Rex Ying Stanford University rexying@stanford.edu Yuanfang Li Stanford University yli03@stanford.edu Xin Li Stanford University xinli16@stanford.edu
More informationEffective Latent Space Graph-based Re-ranking Model with Global Consistency
Effective Latent Space Graph-based Re-ranking Model with Global Consistency Feb. 12, 2009 1 Outline Introduction Related work Methodology Graph-based re-ranking model Learning a latent space graph A case
More informationFast or furious? - User analysis of SF Express Inc
CS 229 PROJECT, DEC. 2017 1 Fast or furious? - User analysis of SF Express Inc Gege Wen@gegewen, Yiyuan Zhang@yiyuan12, Kezhen Zhao@zkz I. MOTIVATION The motivation of this project is to predict the likelihood
More informationHetPathMine: A Novel Transductive Classification Algorithm on Heterogeneous Information Networks
HetPathMine: A Novel Transductive Classification Algorithm on Heterogeneous Information Networks Chen Luo 1,RenchuGuan 1, Zhe Wang 1,, and Chenghua Lin 2 1 College of Computer Science and Technology, Jilin
More informationAspEm: Embedding Learning by Aspects in Heterogeneous Information Networks
AspEm: Embedding Learning by Aspects in Heterogeneous Information Networks Yu Shi, Huan Gui, Qi Zhu, Lance Kaplan, Jiawei Han University of Illinois at Urbana-Champaign (UIUC) Facebook Inc. U.S. Army Research
More informationVideo annotation based on adaptive annular spatial partition scheme
Video annotation based on adaptive annular spatial partition scheme Guiguang Ding a), Lu Zhang, and Xiaoxu Li Key Laboratory for Information System Security, Ministry of Education, Tsinghua National Laboratory
More information3D model classification using convolutional neural network
3D model classification using convolutional neural network JunYoung Gwak Stanford jgwak@cs.stanford.edu Abstract Our goal is to classify 3D models directly using convolutional neural network. Most of existing
More informationText Document Clustering Using DPM with Concept and Feature Analysis
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 2, Issue. 10, October 2013,
More informationJOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS. Puyang Xu, Ruhi Sarikaya. Microsoft Corporation
JOINT INTENT DETECTION AND SLOT FILLING USING CONVOLUTIONAL NEURAL NETWORKS Puyang Xu, Ruhi Sarikaya Microsoft Corporation ABSTRACT We describe a joint model for intent detection and slot filling based
More informationDeep Learning on Graphs
Deep Learning on Graphs with Graph Convolutional Networks Hidden layer Hidden layer Input Output ReLU ReLU, 6 April 2017 joint work with Max Welling (University of Amsterdam) The success story of deep
More informationTransPath: Representation Learning for Heterogeneous Information Networks via Translation Mechanism
Article TransPath: Representation Learning for Heterogeneous Information Networks via Translation Mechanism Yang Fang *, Xiang Zhao and Zhen Tan College of System Engineering, National University of Defense
More informationSemi-supervised Data Representation via Affinity Graph Learning
1 Semi-supervised Data Representation via Affinity Graph Learning Weiya Ren 1 1 College of Information System and Management, National University of Defense Technology, Changsha, Hunan, P.R China, 410073
More informationEmpirical Evaluation of RNN Architectures on Sentence Classification Task
Empirical Evaluation of RNN Architectures on Sentence Classification Task Lei Shen, Junlin Zhang Chanjet Information Technology lorashen@126.com, zhangjlh@chanjet.com Abstract. Recurrent Neural Networks
More informationStructural Deep Embedding for Hyper-Networks
The Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18) Structural Deep Embedding for Hyper-Networks Ke Tu, 1 Peng Cui, 1 Xiao Wang, 1 Fei Wang, 2 Wenwu Zhu 1 1 Department of Computer Science
More informationPseudo-Implicit Feedback for Alleviating Data Sparsity in Top-K Recommendation
Pseudo-Implicit Feedback for Alleviating Data Sparsity in Top-K Recommendation Yun He, Haochen Chen, Ziwei Zhu, James Caverlee Department of Computer Science and Engineering, Texas A&M University Department
More information[Gidhane* et al., 5(7): July, 2016] ISSN: IC Value: 3.00 Impact Factor: 4.116
IJESRT INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY AN EFFICIENT APPROACH FOR TEXT MINING USING SIDE INFORMATION Kiran V. Gaidhane*, Prof. L. H. Patil, Prof. C. U. Chouhan DOI: 10.5281/zenodo.58632
More informationarxiv: v1 [cs.mm] 12 Jan 2016
Learning Subclass Representations for Visually-varied Image Classification Xinchao Li, Peng Xu, Yue Shi, Martha Larson, Alan Hanjalic Multimedia Information Retrieval Lab, Delft University of Technology
More informationAn efficient face recognition algorithm based on multi-kernel regularization learning
Acta Technica 61, No. 4A/2016, 75 84 c 2017 Institute of Thermomechanics CAS, v.v.i. An efficient face recognition algorithm based on multi-kernel regularization learning Bi Rongrong 1 Abstract. A novel
More informationPOI2Vec: Geographical Latent Representation for Predicting Future Visitors
: Geographical Latent Representation for Predicting Future Visitors Shanshan Feng 1 Gao Cong 2 Bo An 2 Yeow Meng Chee 3 1 Interdisciplinary Graduate School 2 School of Computer Science and Engineering
More informationLINE: Large-scale Information Network Embedding
LINE: Large-scale Information Network Embedding Jian Tang 1, Meng Qu 2, Mingzhe Wang 2, Ming Zhang 2, Jun Yan 1, Qiaozhu Mei 3 1 Microsoft Research Asia, {jiatang, junyan}@microsoft.com 2 School of EECS,
More informationCOLD-START PRODUCT RECOMMENDATION THROUGH SOCIAL NETWORKING SITES USING WEB SERVICE INFORMATION
COLD-START PRODUCT RECOMMENDATION THROUGH SOCIAL NETWORKING SITES USING WEB SERVICE INFORMATION Z. HIMAJA 1, C.SREEDHAR 2 1 PG Scholar, Dept of CSE, G Pulla Reddy Engineering College, Kurnool (District),
More informationGraRep: Learning Graph Representations with Global Structural Information
GraRep: Learning Graph Representations with Global Structural Information Shaosheng Cao Xidian University shelsoncao@gmail.com Wei Lu Singapore University of Technology and Design luwei@sutd.edu.sg Qiongkai
More informationANRL: Attributed Network Representation Learning via Deep Neural Networks
ANRL: Attributed Network Representation Learning via Deep Neural Networks Zhen Zhang 1,2, Hongxia Yang 3, Jiajun Bu 1, Sheng Zhou 1,2, Pinggang Yu 1,2, Jianwei Zhang 3, Martin Ester 4, Can Wang 1 1 College
More informationNetwork Representation Learning Enabling Network Analytics and Inference in Vector Space. Peng Cui, Wenwu Zhu, Daixin Wang Tsinghua University
Network Representation Learning Enabling Network Analytics and Inference in Vector Space Peng Cui, Wenwu Zhu, Daixin Wang Tsinghua University 2 A Survey on Network Embedding Peng Cui, Xiao Wang, Jian Pei,
More informationECS289: Scalable Machine Learning
ECS289: Scalable Machine Learning Cho-Jui Hsieh UC Davis Sept 22, 2016 Course Information Website: http://www.stat.ucdavis.edu/~chohsieh/teaching/ ECS289G_Fall2016/main.html My office: Mathematical Sciences
More informationSemi-Supervised Clustering with Partial Background Information
Semi-Supervised Clustering with Partial Background Information Jing Gao Pang-Ning Tan Haibin Cheng Abstract Incorporating background knowledge into unsupervised clustering algorithms has been the subject
More informationChannel Locality Block: A Variant of Squeeze-and-Excitation
Channel Locality Block: A Variant of Squeeze-and-Excitation 1 st Huayu Li Northern Arizona University Flagstaff, United State Northern Arizona University hl459@nau.edu arxiv:1901.01493v1 [cs.lg] 6 Jan
More informationAdversarial Network Embedding
Adversarial Network Embedding Quanyu Dai 1, Qiang Li 1,2, Jian Tang 3,4, Dan Wang 1 1 Department of Computing, The Hong Kong Polytechnic University, Hong Kong 2 School of Software, FEIT, The University
More informationarxiv: v1 [cs.lg] 12 Mar 2015
LINE: Large-scale Information Network Embedding Jian Tang 1, Meng Qu 2, Mingzhe Wang 2, Ming Zhang 2, Jun Yan 1, Qiaozhu Mei 3 1 Microsoft Research Asia, {jiatang, junyan}@microsoft.com 2 School of EECS,
More informationA Deep Relevance Matching Model for Ad-hoc Retrieval
A Deep Relevance Matching Model for Ad-hoc Retrieval Jiafeng Guo 1, Yixing Fan 1, Qingyao Ai 2, W. Bruce Croft 2 1 CAS Key Lab of Web Data Science and Technology, Institute of Computing Technology, Chinese
More informationImproving Image Segmentation Quality Via Graph Theory
International Symposium on Computers & Informatics (ISCI 05) Improving Image Segmentation Quality Via Graph Theory Xiangxiang Li, Songhao Zhu School of Automatic, Nanjing University of Post and Telecommunications,
More informationCS229 Final Project: Predicting Expected Response Times
CS229 Final Project: Predicting Expected Email Response Times Laura Cruz-Albrecht (lcruzalb), Kevin Khieu (kkhieu) December 15, 2017 1 Introduction Each day, countless emails are sent out, yet the time
More informationStructural Deep Network Embedding
Structural Deep Network Embedding Daixin Wang, Peng Cui, Wenwu Zhu Tsinghua National Laboratory for Information Science and Technology Department of Computer Science and Technology, Tsinghua University.
More informationFirst Place Solution for NLPCC 2018 Shared Task User Profiling and Recommendation
First Place Solution for NLPCC 2018 Shared Task User Profiling and Recommendation Qiaojing Xie, Yuqian Wang, Zhenjing Xu, Kaidong Yu, Chen Wei (B), and ZhiChen Yu Turing Robot, Beijing, China {xieqiaojing,wangyuqian,weichen}@uzoo.cn
More informationUnsupervised User Identity Linkage via Factoid Embedding
Unsupervised User Identity Linkage via Factoid Embedding Wei Xie, Xin Mu, Roy Ka-Wei Lee, Feida Zhu and Ee-Peng Lim Living Analytics Research Centre Singapore Management University, Singapore Email: {weixie,roylee.2013,fdzhu,eplim}@smu.edu.sg
More informationMicroblog Sentiment Analysis with Emoticon Space Model
Fei Jiang, Yiqun Liu, etc.. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY : 1 05. 2015 Microblog Sentiment Analysis with Emoticon Space Model Fei Jiang 1, Yiqun Liu 1,, (Senior Member, CCF), Huanbo Luan 1,
More informationRepresentation Learning for Scale-free Networks
Representation Learning for Scale-free Networks Rui Feng, Yang Yang, Wenjie Hu, Fei Wu, and Yueting Zhuang College of Computer Science and Technology, Zhejiang University, China Corresponding author: yangya@zju.edu.cn
More informationOverview Citation. ML Introduction. Overview Schedule. ML Intro Dataset. Introduction to Semi-Supervised Learning Review 10/4/2010
INFORMATICS SEMINAR SEPT. 27 & OCT. 4, 2010 Introduction to Semi-Supervised Learning Review 2 Overview Citation X. Zhu and A.B. Goldberg, Introduction to Semi- Supervised Learning, Morgan & Claypool Publishers,
More informationData Mining. Introduction. Hamid Beigy. Sharif University of Technology. Fall 1395
Data Mining Introduction Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1395 1 / 21 Table of contents 1 Introduction 2 Data mining
More informationComment Extraction from Blog Posts and Its Applications to Opinion Mining
Comment Extraction from Blog Posts and Its Applications to Opinion Mining Huan-An Kao, Hsin-Hsi Chen Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan
More informationQuery Independent Scholarly Article Ranking
Query Independent Scholarly Article Ranking Shuai Ma, Chen Gong, Renjun Hu, Dongsheng Luo, Chunming Hu, Jinpeng Huai SKLSDE Lab, Beihang University, China Beijing Advanced Innovation Center for Big Data
More informationA FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen
A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS Kuan-Chuan Peng and Tsuhan Chen School of Electrical and Computer Engineering, Cornell University, Ithaca, NY
More informationVisoLink: A User-Centric Social Relationship Mining
VisoLink: A User-Centric Social Relationship Mining Lisa Fan and Botang Li Department of Computer Science, University of Regina Regina, Saskatchewan S4S 0A2 Canada {fan, li269}@cs.uregina.ca Abstract.
More informationMax-Margin DeepWalk: Discriminative Learning of Network Representation
Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI-16) Max-Margin DeepWalk: Discriminative Learning of Network Representation Cunchao Tu 1,2, Weicheng Zhang
More informationCPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016
CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2016 A2/Midterm: Admin Grades/solutions will be posted after class. Assignment 4: Posted, due November 14. Extra office hours:
More informationLimitations of Matrix Completion via Trace Norm Minimization
Limitations of Matrix Completion via Trace Norm Minimization ABSTRACT Xiaoxiao Shi Computer Science Department University of Illinois at Chicago xiaoxiao@cs.uic.edu In recent years, compressive sensing
More informationDeep Learning for Face Recognition. Xiaogang Wang Department of Electronic Engineering, The Chinese University of Hong Kong
Deep Learning for Face Recognition Xiaogang Wang Department of Electronic Engineering, The Chinese University of Hong Kong Deep Learning Results on LFW Method Accuracy (%) # points # training images Huang
More informationHARP: Hierarchical Representation Learning for Networks
HARP: Hierarchical Representation Learning for Networks ABSTRACT Haochen Chen Stony Brook University haocchen@cs.stonybrook.edu Yifan Hu Yahoo! Research yifanhu@yahoo.com We present HARP, a novel method
More informationBuilding Corpus with Emoticons for Sentiment Analysis
Building Corpus with Emoticons for Sentiment Analysis Changliang Li 1(&), Yongguan Wang 2, Changsong Li 3,JiQi 1, and Pengyuan Liu 2 1 Kingsoft AI Laboratory, 33, Xiaoying West Road, Beijing 100085, China
More informationCS224W: Social and Information Network Analysis Project Report: Edge Detection in Review Networks
CS224W: Social and Information Network Analysis Project Report: Edge Detection in Review Networks Archana Sulebele, Usha Prabhu, William Yang (Group 29) Keywords: Link Prediction, Review Networks, Adamic/Adar,
More informationIntroduction to Text Mining. Hongning Wang
Introduction to Text Mining Hongning Wang CS@UVa Who Am I? Hongning Wang Assistant professor in CS@UVa since August 2014 Research areas Information retrieval Data mining Machine learning CS@UVa CS6501:
More informationData Mining. Introduction. Hamid Beigy. Sharif University of Technology. Fall 1394
Data Mining Introduction Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1394 1 / 20 Table of contents 1 Introduction 2 Data mining
More informationFast Network Embedding Enhancement via High Order Proximity Approximation
Fast Network Embedding Enhancement via High Order Proximity Approximation Cheng Yang 1, Maosong Sun 1,2, Zhiyuan Liu 1,2, Cunchao Tu 1,2 1 Department of Computer Science and Technology, Tsinghua University,
More informationLabel Informed Attributed Network Embedding
Label Informed Attributed Network Embedding Xiao Huang, Jundong Li, Xia Hu, Department of Computer Science and Engineering, Texas A&M University, College Station, TX, USA Computer Science and Engineering,
More informationSEMI-SUPERVISED LEARNING (SSL) for classification
IEEE SIGNAL PROCESSING LETTERS, VOL. 22, NO. 12, DECEMBER 2015 2411 Bilinear Embedding Label Propagation: Towards Scalable Prediction of Image Labels Yuchen Liang, Zhao Zhang, Member, IEEE, Weiming Jiang,
More informationTriRank: Review-aware Explainable Recommendation by Modeling Aspects
TriRank: Review-aware Explainable Recommendation by Modeling Aspects Xiangnan He, Tao Chen, Min-Yen Kan, Xiao Chen National University of Singapore Presented by Xiangnan He CIKM 15, Melbourne, Australia
More informationUsing Machine Learning to Optimize Storage Systems
Using Machine Learning to Optimize Storage Systems Dr. Kiran Gunnam 1 Outline 1. Overview 2. Building Flash Models using Logistic Regression. 3. Storage Object classification 4. Storage Allocation recommendation
More informationA Survey on Postive and Unlabelled Learning
A Survey on Postive and Unlabelled Learning Gang Li Computer & Information Sciences University of Delaware ligang@udel.edu Abstract In this paper we survey the main algorithms used in positive and unlabeled
More informationCANE: Context-Aware Network Embedding for Relation Modeling
CANE: Context-Aware Network Embedding for Relation Modeling Cunchao Tu 1,2, Han Liu 3, Zhiyuan Liu 1,2, Maosong Sun 1,2 1 Department of Computer Science and Technology, State Key Lab on Intelligent Technology
More informationjldadmm: A Java package for the LDA and DMM topic models
jldadmm: A Java package for the LDA and DMM topic models Dat Quoc Nguyen School of Computing and Information Systems The University of Melbourne, Australia dqnguyen@unimelb.edu.au Abstract: In this technical
More informationAero-engine PID parameters Optimization based on Adaptive Genetic Algorithm. Yinling Wang, Huacong Li
International Conference on Applied Science and Engineering Innovation (ASEI 215) Aero-engine PID parameters Optimization based on Adaptive Genetic Algorithm Yinling Wang, Huacong Li School of Power and
More informationA Taxonomy of Semi-Supervised Learning Algorithms
A Taxonomy of Semi-Supervised Learning Algorithms Olivier Chapelle Max Planck Institute for Biological Cybernetics December 2005 Outline 1 Introduction 2 Generative models 3 Low density separation 4 Graph
More informationMining Web Data. Lijun Zhang
Mining Web Data Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Introduction Web Crawling and Resource Discovery Search Engine Indexing and Query Processing Ranking Algorithms Recommender Systems
More informationNovel Lossy Compression Algorithms with Stacked Autoencoders
Novel Lossy Compression Algorithms with Stacked Autoencoders Anand Atreya and Daniel O Shea {aatreya, djoshea}@stanford.edu 11 December 2009 1. Introduction 1.1. Lossy compression Lossy compression is
More informationSpatial Latent Dirichlet Allocation
Spatial Latent Dirichlet Allocation Xiaogang Wang and Eric Grimson Computer Science and Computer Science and Artificial Intelligence Lab Massachusetts Tnstitute of Technology, Cambridge, MA, 02139, USA
More informationGraph Matching Iris Image Blocks with Local Binary Pattern
Graph Matching Iris Image Blocs with Local Binary Pattern Zhenan Sun, Tieniu Tan, and Xianchao Qiu Center for Biometrics and Security Research, National Laboratory of Pattern Recognition, Institute of
More informationFacial Expression Classification with Random Filters Feature Extraction
Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey mren@cs.toronto.edu Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle
More informationA REVIEW ON SEARCH BASED FACE ANNOTATION USING WEAKLY LABELED FACIAL IMAGES
A REVIEW ON SEARCH BASED FACE ANNOTATION USING WEAKLY LABELED FACIAL IMAGES Dhokchaule Sagar Patare Swati Makahre Priyanka Prof.Borude Krishna ABSTRACT This paper investigates framework of face annotation
More informationFastText. Jon Koss, Abhishek Jindal
FastText Jon Koss, Abhishek Jindal FastText FastText is on par with state-of-the-art deep learning classifiers in terms of accuracy But it is way faster: FastText can train on more than one billion words
More informationGraph Classification in Heterogeneous
Title: Graph Classification in Heterogeneous Networks Name: Xiangnan Kong 1, Philip S. Yu 1 Affil./Addr.: Department of Computer Science University of Illinois at Chicago Chicago, IL, USA E-mail: {xkong4,
More informationA new Graph constructor for Semi-supervised Discriminant Analysis via Group Sparsity
2011 Sixth International Conference on Image and Graphics A new Graph constructor for Semi-supervised Discriminant Analysis via Group Sparsity Haoyuan Gao, Liansheng Zhuang, Nenghai Yu MOE-MS Key Laboratory
More informationRandom projection for non-gaussian mixture models
Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,
More information