Representation Learning of Knowledge Graphs with Hierarchical Types

Size: px
Start display at page:

Download "Representation Learning of Knowledge Graphs with Hierarchical Types"

Transcription

1 Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI-16) Representation Learning of Knowledge Graphs with Hierarchical Types Ruobing Xie, 1 Zhiyuan Liu, 1,2 Maosong Sun 1,2 1 Department of Computer Science and Technology, State Key Lab on Intelligent Technology and Systems, National Lab for Information Science and Technology, Tsinghua University, Beijing, China 2 Jiangsu Collaborative Innovation Center for Language Ability, Jiangsu Normal University, Xuzhou China Abstract Representation learning of knowledge graphs aims to encode both entities and relations into a continuous low-dimensional vector space. Most existing methods only concentrate on learning representations with structured information located in triples, regardless of the rich information located in hierarchical types of entities, which could be collected in most knowledge graphs. In this paper, we propose a novel method named Type-embodied Knowledge Representation Learning (TKRL) to take advantages of hierarchical entity types. We suggest that entities should have multiple representations in different types. More specifically, we consider hierarchical types as projection matrices for entities, with two type encoders designed to model hierarchical structures. Meanwhile, type information is also utilized as relation-specific type constraints. We evaluate our models on two tasks including knowledge graph completion and triple classification, and further explore the performances on long-tail dataset. Experimental results show that our models significantly outperform all baselines on both tasks, especially with long-tail distribution. It indicates that our models are capable of capturing hierarchical type information which is significant when constructing representations of knowledge graphs. The source code of this paper can be obtained from 1 Introduction Knowledge graphs (KGs) such as Freebase, DBpedia and YAGO provide effective structured information and have been crucial in information retrieval and question answering. A typical KG is usually represented as multi-relational data with enormous triple facts in the form of (head entity, relation, tail entity), abridged as (h, r, t). As KG size increases, KG applications become more challenging due to data sparsity and computational inefficiency. Corresponding author: Z. Liu (liuzy@tsinghua.edu.cn) To address these problems, representation learning (RL), which aims to project both entities and relations into a continuous low-dimensional semantic space, is blooming and widely utilized in knowledge completion, fusion and inference. It significantly improves the ability of KGs in cognition and reasoning [Bordes et al., 2013; Yang et al., 2014; Dong et al., 2014; Neelakantan et al., 2015]. Many methods have been proposed on RL for KGs, among which the translation-based models are simple and effective with the state-of-the-art performances. Unfortunately, most conventional methods merely focus on the structured information in triples, paying less attention to the rich information located in hierarchical types of entities. Fig. 1 shows a triple instance combined with part of its hierarchical types sampled from Freebase, in which the solid lines indicate the most significant roles that head and tail play in this triple. music award book TV artist award_nominee author written_work TV_subject ( William Shakespeare, book/author/works_written, Romeo and Juliet ) Figure 1: Example of entity hierarchical types in Freebase. It is intuitive that entities possessing multiple types should have various representations in different scenarios. To take advantages of entity types, we propose a novel and effective RL method for KGs named Type-embodied Knowledge Representation Learning (TKRL). In TKRL, we follow the assumption in TransE [Bordes et al., 2013], considering relations as translating operations between head and tail. For each triple (h, r, t), h and t are first projected to their corresponding type spaces in this relation as h rh and t rt, according to the type-specific projection matrices constructed with two hierarchical type encoders. TKRL is then optimized by minimizing the energy function of E(h, r, t) = h rh + r t rt. Moreover, type information is also considered as type constraints in both training and evaluation. We evaluate the TKRL model on benchmark datasets of 2965

2 Freebase with knowledge graph completion and triple classification. Experimental results show that TKRL has significant and consistent improvements compared to all baselines on both tasks, especially with long-tail distribution, which confirms the capability of TKRL modeling KGs. 2 Related Work 2.1 Translation-based Models Recent years there are great advances in representation learning for knowledge graphs. TransE [Bordes et al., 2013] projects both entities and relations into the same continuous low-dimensional vector space, interpreting relations as translating operations between head and tail entities. TransE assumes that the embedding of tail t should be in the neighbourhood of h+r. The energy function is defined as follows: E(h, r, t) = h + r t. (1) TransE is effective and efficient, while the simple translating operation may lead to problems when modeling more complicated relations like 1-to-N, N-to-1 and N-to-N. To address this issue, TransH [Wang et al., 2014b] interprets relations as translating operations on relation-specific hyperplanes, allowing entities to have different distances in different relations. TransR [Lin et al., 2015b] directly models entities and relations in separate entity and relation spaces, projecting entities from entity space to relation-specific space when judging the distance between entities. TransD [Ji et al., 2015] proposes dynamic mapping matrix constructed via both entity and relation, considering the diversity of entities as well as relations simultaneously. Besides, [Lin et al., 2015a; Gu et al., 2015] also encode multiple-step relation path information into KG representation learning. However, These models only concentrate on information in triples, ignoring rich information in entity types, which will be considered in TKRL model. 2.2 Multi-source Information Learning Multi-source information like textual information and type information, considered as supplements for the structured information embedded in triples, is significant for RL in KGs. NTN [Socher et al., 2013] encodes 3-way tensors into neural network and represents an entity as the average of word embeddings in its entity name. [Wang et al., 2014a; Zhong et al., 2015] encode both entities and words into a joint continuous vector space by alignment models using Wikipedia anchors, entity names or entity descriptions. DKRL [Xie et al., 2016] proposes description-based representations for entities constructed from entity descriptions with CBOW or CNN, which is capable of modeling entities in zero-shot scenario. Rich information located in hierarchical entity types is also significant for KGs, while it has just attracted attention. [Krompaß et al., 2015] considers entity types as hard constraints in latent variable models for KGs. However, the type information is not explicitly encoded into KG representations, and their method doesn t consider the hierarchical structure of entity types. Moreover, hard constraints may have issues with noises and incompleteness in type information, which is pretty common in real-world KGs. Therefore, we propose the TKRL model to overcome these shortages. To the best of our knowledge, TKRL is the first method which explicitly encodes type information into multiple representations in KGs with the help of hierarchical structures. 3 Methodology To utilize rich information located in entity types, we take type information into consideration when constructing projection matrices of entities, with two hierarchical type encoders modeling hierarchical structures. Moreover, type information is also utilized as relation-specific type constraints in both training and evaluation. 3.1 Hierarchical Type Structure Hierarchical type information, which implies different roles an entity may play in different scenarios, is of great significance for representation learning in knowledge graphs. Most typical knowledge graphs (e.g. Freebase and DBpedia) possess their own entity type information or could collect it from large encyclopedias like Wikipedia through entity alignment. These types are usually constructed with hierarchical structures, in which different granularities of semantic concepts are considered as sub-types in different layers. Most entities have more than one hierarchical type. Fig. 1 shows a brief example of the hierarchical type structure. Taking a hierarchical type c with k layers for instance, c (i) is the i-th sub-type of c. We consider the most precise subtype to be the first layer and the most general sub-type to be the last layer, while each sub-type c (i) has only one parent sub-type c (i+1). Walking through the bottom-up path in hierarchical structure, we can get the representation of hierarchical type as c = {c (1),c (2),...,c (k) }. 3.2 Overall Architecture Existing translation-based models perform well in knowledge graphs, but few of them make full use of the rich information located in entity types. TransE represents entities and relations in a low-dimensional vector space and interprets relations as translating operations between entities. However, TransE has issues when modeling N-to-1, 1-to-N and N-to-N relations, since each entity has only one representation in every scenario. For instance, William Shakespeare has a variety of types (e.g. book/author, award/award nominee and music/artist) and shows different attributes under different types. We believe that every entity in different scenarios, as the reflections of itself from various angles, should have different representations. To implement the multiple representations of entities in different types, we propose the TKRL model. More specifically, we set type-specific projection matrices M c constructed from the hierarchical structures for each type c, and then represent both h and t under the projections of the specific types c rh and c rt which head and tail should belong to in this relation. The energy function is defined as follows: E(h, r, t) = M rh h + r M rt t, (2) 2966

3 in which M rh and M rt are different projection matrices for h and t. We propose two hierarchical type encoders to construct those projection matrices. 3.3 Hierarchical Type Encoders To encode the hierarchical type information into representation learning, we first propose a general form of type encoder to construct projection matrices for each entity. Secondly, two advanced encoders are proposed to take advantages of the internal connections in hierarchical structures and the prior knowledge in relation-specific type information. General Form of Type Encoder Most entities in KGs have more than one type, which could be utilized as supplementary information when representing entities. We propose a general form of type encoder, in which the projection matrix M e for entity e will be the weighted summation of all type matrices: M e = 1 M c1 + 2 M c2 + + n M cn, (3) where n is the number of types entity e has, c i is the i-th type e belongs to, M ci is the projection matrix of c i, and i represents the corresponding weight for c i. Those weights could be set according to the influence c i has on e measured by some statistical features like type frequency. Through the general type encoder, all projection matrices for entity e will be the same in different scenarios. However, entities should have different representations to emphasize attributes of more importance in different scenarios. Fortunately, the relation-specific type information in KGs, which provides the possible type(s) an entity may belong to in a specific relation, could help for multiple entity representations. To take advantage of this information, the projection matrix M rh in a specific triple (h, r, t) will be: P n M rh = P im ci 1, ci 2 C n, i = rh (4) i 0, c i /2 C rh where C rh represents the type set of head in relation r given by the relation-specific type information. Projection matrices M rt for entities in position of tail will be of the same form as those for entities in head. M c is the projection matrix for type c, which could be constructed by the following two encoders. Recursive Hierarchy Encoder To further improve the representation of projection matrix M c by mining the latent information located in hierarchical type structures, we propose the Recursive Hierarchy Encoder (RHE). Inspired by [Hu et al., 2015], each sub-type (i.e. layer) in hierarchy is represented as a projection matrix with different granularities. During the projection process, entities (e.g. William Shakespeare) will be first mapped to the more general sub-type space (e.g. book) and then be sequentially mapped to the more precise sub-type space (e.g. book/author). The matrix M c is designed as follows: my M c = M c (i) = M c (1)M c (2)...M c (m), (5) in which m is the number of layers for type c in the hierarchical structure, while M c (i) represents the projection matrix of the i-th sub-type c (i). h M ch (m) M ch (m-1) r t M ct (m) M ch (1) M ct (1) h ch (a) RHE t ct M ct (m-1) h ch h iβ i M (i) ch r (b) WHE Figure 2: Hierarchical Type Encoders t iβ i M (i) ct Weighted Hierarchy Encoder RHE proposes a recursive method of building hierarchical type projection matrices. However, different granularities of sub-types in hierarchical structures may vary in significance when mapping entities. In this case, we propose the Weighted Hierarchy Encoder (WHE) to consider weights in hierarchy. In WHE, sub-types are represented as projection matrices too. However, instead of using recursive operation to encode different granularities of sub-types, we sum up those projection matrices with different weights to represent the hierarchical type matrix M c as follows: mx M c = im c (i) = 1 M c (1) + + m M c (m), (6) in which m is the number of layers in the hierarchical structure, M c (i) is the projection matrix of c (i), while i is the corresponding weight of c (i). We design a proportional-declined weighting strategy between c (i) and c (i+1) as follows: mx i : i+1 =(1 ) :, i =1, (7) in which we set 2 (0, 0.5). The strategy indicates that the more precise sub-type c (i) is, the higher weight i will be, thus the greater influence c (i) will have on M c. 3.4 Objective Formalization We formalize a margin-based score function with negative sampling as objective for training: L = X X max( + E(h, r, t)) (h,r,t)2t (h 0,r 0,t 0 )2T 0 (8) t ct E(h 0,r 0,t 0 ), 0), where E(h, r, t) is the energy function score of positive triple and E(h 0,r 0,t 0 ) is that of negative triple. >0 is a hyperparameter of margin. T 0 stands for the negative sampling set of T. Since there are no explicit negative triples in knowledge graphs, T 0 is constructed as follows: T 0 ={(h 0,r,t) h 0 2 E}[{(h, r, t 0 ) t 0 2 E} [{(h, r 0,t) r 0 2 R}, (h, r, t) 2 T, (9) 2967

4 in which the head or tail in a positive triple is randomly replaced by any other entity in E. Differed from [Bordes et al., 2013; Lin et al., 2015b], we also add relation replacements to negative sampling for better performances in relation prediction. Moreover, the new triples after replacements will not be considered as negative samples if they are already in T. 3.5 Type Information as Constraints We use hierarchical type encoders to build type-specific projection matrices mapping entities to different semantic vector spaces in different scenarios. Besides we can use type information not only as projection matrices but also as constraints with the help of relation-specific type information. We propose two methods for using type information as constraints in both training and evaluation. Type Constraints in Training In training, we select negative entity samples from all entities in E with equal probability. In this case, entities sharing the same types tend to cluster and have similar representations, which actually becomes the main cause of errors in entity prediction. To solve this problem, we propose a method for negative sampling named Soft Type Constraint (STC). STC improves the probability of selecting entities which have the same types constrained by relation-specific type information, and thus broadens the distances between similar entities. Instead of setting hard constraints as in [Krompaß et al., 2015] which may disturb the clustering of similar entities, we utilize soft constraints in which the probability of entities with the same types being selected is as follows: P (e 0 2 E c )= (k + 1) E c, k 2 N, (10) E + k E c where c is the golden type for e in this triple, E c 2 E is the entity set in which all entities have type c, with E c to be the number of entities in E c. k is a hyper-parameter that a smaller k means a softer type constraint. It indicates that the probability of selecting entities in E c is k times bigger than the probability of selecting entities which are not. With STC in negative sampling, we can balance the diversity as well as similarity between entities with the same types to get better performances. Type Constraints in Evaluation Besides type constraints in training, we can also utilize the information as type constraints in evaluation (TCE). It is intuitive that the head and tail entities should follow the type constraints provided by relation-specific type information. We can simply remove the candidates which don t follow the type constraints in evaluation. However, the performances of TCE are mostly based on the correctness and completeness of relation-specific type information, since type constraints could be incomplete or even wrong, which may lead to misses in prediction. In evaluation, we will show the power of TCE in auxiliary experiments. 3.6 Optimization and Implementation Details The TKRL model can be stated as a parameter set = (E, R, M), in which E and R stand for the embeddings of entities and relations, and M stands for the projection matrices of all sub-types. Optimization and Model Initialization The TKRL model is optimized with mini-batch stochastic gradient descent (SGD), while chain rule is applied to update all parameters. The sub-type projection matrix set M could be initialized randomly or by identity matrix. Both E and R could be either initialized randomly or be pre-trained by existing translation-based models like TransE. Implementation Details We look through KGs to collect type instances for entities and relation-specific type information. To decrease the influence of incompleteness located in KGs, we also follow the local closed-world assumptions (LCWA) proposed in [Krompaß et al., 2015] as a supplementary method. In LCWA, all entities appear in head (or tail) with the same relation should be allocated with the same type. Following [Wang et al., 2014b], we implement two strategies (i.e. unif and bern ) for replacing head or tail with equal or different probabilities when selecting negative samples. For the consideration of efficiency, we employ a multi-thread version for learning. 4 Experiments 4.1 Datasets and Experiment Settings Datasets In this paper, we first use a typical knowledge graph FB15K [Bordes et al., 2013] to evaluate our models on knowledge graph completion and triple classification. FB15K is a dataset extracted from Freebase [Bollacker et al., 2008] which contains 14,951 entities, 1,345 relations and 592,213 triples in total. Following [Bordes et al., 2013], we split those triples into train, validation and test set. As for the type information, we collect all type instances for entities in FB15K located in type/instance field, as well as the relation-specific type information located in rdfschema#domain and rdf-schema#range fields. We also filter the type instances which never appear in extracted relationspecific type information, since those types have little influence on relations in FB15K. After filtering, we have 571 types. All entities in FB15K have at least one hierarchical type and the average number of types is approximately 8. Entities and relations in FB15K are limited that they should have at least 100 mentions in Freebase [Bordes et al., 2013]. However, relations and entities in real-world KGs are much more sparse than FB15K due to the long tail. To further present the advantages of our models in real-world distribution, we construct FB15K+, which contains the same entities in FB15K and almost all relations between those entities. We only discard relations which just appear once, for they can t exist both in train and test set. The new-added 6,607 triples are split into train and test set where every relation is mentioned at least once in train set. FB15K+ have 806 types after filtering. The statistics of datasets are listed in Table 1. Experiment Settings In evaluation, we implement TransE and TransR for comparison. For TransE, we improve their dissimilarity measure with L 1 -norm and replace relations as well as entities during negative sampling. We also use bern to replace head or tail with different probability following [Wang 2968

5 Table 1: Statistics of datasets Dataset #Rel #Ent #Train #Valid #Test FB15K 1,345 14, ,142 50,000 59,071 FB15K+ 1,855 14, ,446 50,000 62,374 et al., 2014b]. For TransR, we directly use the released code given in [Lin et al., 2015b] and utilize replacements of relations in negative sampling for better performances in relation prediction. Both TransE and TransR are trained with the best parameters reported in their papers. For other baselines including RESCAL [Nickel et al., 2011; 2012], SE [Bordes et al., 2011], SME [Bordes et al., 2012; 2014] and LFM [Jenatton et al., 2012], we directly use the results reported in [Lin et al., 2015b]. We train TKRL model with mini-batch SGD. As for parameters, we select the batch size B among {20, 240, 1200, 4800}, and margin among {0.5, 1.0, 1.5, 2.0}. We also set the dimensions of entity and relation to be the same n, and all projection matrices are set to be n n. For learning rate, we could select a fixed rate following [Bordes et al., 2013; Lin et al., 2015b] or design a flexible learning rate which will descend through iteration. For WHE, we select the descending weight between sub-types among {0.1, 0.15, 0.2}. For k in type constraints, we select among {5, 10, 15}. The optimal configurations of our models are: B = 4800, = 1.0, = 0.1, k = 10, with designed by a linear-declined strategy which ranges from to TKRL and TransR are trained with entities and relations initialized by pre-trained TransE (unif) model. For a fair comparison, all models are trained under the same dimension n = Knowledge Graph Completion Evaluation Protocal Knowledge graph completion aims to complete a triple (h, r, t) when one of h, r, t is missing, which is used in [Bordes et al., 2011; 2012; 2013]. Two measures are considered as our evaluation metrics: (1) mean rank of correct entities or relations; (2) proportion of correct answers ranked in top 10 (for entities) or top 1 (for relations). We also follow the two evaluation settings named raw and filter. We conduct our evaluation on FB15K and divide the task into two sub-tasks: entity prediction and relation prediction. For a fair comparison, evaluation conditions are the same for all models. We also evaluate on the method of type constraint in evaluation (TCE) and a new dataset with long-tail distribution as auxiliary experiments. Entity Prediction The results of entity prediction are shown in Table 2. From the results we observe that: (1) Both RHE and WHE significantly outperform all baselines in mean rank and Hits@10. It indicates that the hierarchical type information, which is successfully encoded into entity and relation embeddings, could improve the representation learning of knowledge graphs. (2) WHE+STC achieves the best performance with approximately 6.2% improvement compared to TransR in Hits@10, and such improvement provided by Soft Type Constraint (STC) can also be found in RHE. It is because that STC increases the probability of entities with the same types as the golden one being selected during negative sampling, which widens the distances between entities sharing the same types, and thus lowers the errors caused by those similar entities. However, STC has side effects that it may result in higher mean rank, since some wrong-predicted instances with extremely high rank will significantly increase mean rank. (3) Type information, either in form of projection matrices or type constraints, could provide significant supplements for RL in KGs. Table 2: Evaluation results on entity prediction Metric Mean Rank Hits@10(%) Raw Filter Raw Filter RESCAL SE SME (linear) SME (bilinear) LFM TransE TransR TKRL (RHE) TKRL (WHE) TKRL (RHE+STC) TKRL (WHE+STC) Relation Prediction The results of relation prediction are shown in Table 3. We implement two typical models including TransE and TransR as baselines. From Table 3 we observe that: (1) Both RHE and WHE significantly outperform TransE and TransR in mean rank and Hits@10, and RHE achieves the best performance. It proves that RHE is better in relation prediction while WHE is better in entity prediction. (2) STC lowers the performances on relation prediction since wider distances between entities with the same types may confuse the entity clustering. In spite of this, all models with STC still outperform TransE, which indicates the positive effects of type information as constraints. Table 3: Evaluation results on relation prediction Metric Mean Rank Hits@1(%) Raw Filter Raw Filter TransE TransR TKRL (RHE) TKRL (WHE) TKRL (RHE+STC) TKRL (WHE+STC) Type Constraints in Evaluation Type constraints in training have been proved to be effective, while type constraints in evaluation (TCE) could be utilized to achieve even better performances in entity prediction, on 2969

6 Relation Frequency Test Number Table 4: Evaluation results on long-tail distribution for Entity (%) for Relation (%) TransE TransR TKRL (WHE) TransE TransR TKRL (WHE) f r <= 10 1, (+4.4) 38.1 (+10.1) (+3.8) 21.5 (+8.3) f r <= 100 4, (+4.6) 57.9 (+8.0) (+4.8) 54.3 (+8.6) f r <= , (+3.0) 71.6 (+5.5) (+4.5) 77.8 (+6.9) total 62, (+5.3) 69.2 (+7.3) (+8.4) 89.7 (+9.3) condition that the relation-specific type information is relatively complete. For a fair comparison, we implement baselines with the helps of both STC and TCE. Results in Table 5 show that: (1) All models have better performances with TCE compared to those corresponding results without TCE shown in Table 2, and the improvements will be more significant when combined with STC. It is because that TCE removes the candidates which don t follow the type constraints, while STC sharpens the differences between similar entities. (2) TKRL models outperform all baselines even when compared with the enhanced versions with STC, which implies the significance of hierarchical type encoders. Table 5: Evaluation results on entity prediction with TCE Metric Mean Rank Hits@10(%) Raw Filter Raw Filter TransE+TCE TransR+TCE TransE+STC+TCE TransR+STC+TCE TKRL (RHE+STC+TCE) TKRL (WHE+STC+TCE) Knowledge Graph Completion with Long Tail Representation learning of real-world KGs suffers from the long-tail distribution. We construct FB15K+, which contains almost all relations between entities in FB15K as well as the corresponding triples, to simulate the distribution in realworld KGs. From Table 4 we can observe that: (1) WHE significantly and consistently outperforms TransE and TransR in all conditions even without STC. (2) WHE achieves 5.8% and 4.5% improvements on entity and relation prediction compared to TransR with f r <= 10, while it achieves 2.0% and 0.9% improvements with all triples. It demonstrates that TKRL takes advantages over TransR especially with lowfrequency relations, and thus is more robust when modeling KGs with real-world distribution. 4.3 Triple Classification Triple classification aims to confirm whether a triple (h, r, t) is correct or not. This binary classification task has been explored in [Socher et al., 2013; Wang et al., 2014b; Lin et al., 2015b] for evaluation. Evaluation Protocal We evaluate this task on FB15K. Since FB15K has no explicit negative triples, we construct the negative triples following the same protocol used in [Socher et al., 2013]. The classification strategy is conducted as follows: We set different relation-specific thresholds r for each relation. For a triple (h, r, t), if the dissimilarity score of E(h, r, t) is below r, the triple is then predicted to be positive and otherwise negative. The relation-specific thresholds r are optimized by maximizing the classification accuracies in all triples with the corresponding r on the validation set. Results Evaluation results on triple classification are shown in Table 6. From Table 6 we observe that: (1) TKRL models outperform all baselines, and WHE+STC achieves the best performance, which confirms the advantages TKRL has over baselines in triple classification. (2) STC improves the performances of both RHE and WHE, which indicates that sharpening the dissimilarity between entities with the same types is significantly helpful for triple classification. Table 6: Evaluation results on triple classification Methods Accuracy(%) TransE 85.7 TransR 86.4 TKRL (RHE) 86.9 TKRL (WHE) 87.1 TKRL (RHE+STC) 88.4 TKRL (WHE+STC) Conclusion and Future Work In this paper, we propose TKRL model for representation learning of knowledge graphs with hierarchical types. We consider type information as projection matrices for entities, which are constructed with two hierarchical type encoders. Moreover, type information is also regarded as constraints in training and evaluation. In experiments, we evaluate our models on two tasks including knowledge graph completion and triple classification. Experimental results show that type information is significant in both tasks especially with longtail distribution, and TKRL model is capable of encoding hierarchical type information into KG representations. We will explore the following research directions in future: (1) TKRL model only considers type information into representation learning of KGs, while there is rich information in the form of images and texts which could be integrated to our model. We will explore the advantages of those rich information in future. (2) More hierarchical type structures such as Wikipedia categories could be introduced to bring in deeper 2970

7 hierarchical information, while the hierarchical type encoders could be further improved with more sophisticated algorithms designed for hierarchical structures. Acknowledgments This work is supported by the 973 Program (No. 2014CB340501), the National Natural Science Foundation of China (NSFC No and ) and the Tsinghua University Initiative Scientific Research Program ( ). References [Bollacker et al., 2008] Kurt Bollacker, Colin Evans, Praveen Paritosh, Tim Sturge, and Jamie Taylor. Freebase: a collaboratively created graph database for structuring human knowledge. In Proceedings of KDD, pages , [Bordes et al., 2011] Antoine Bordes, Jason Weston, Ronan Collobert, and Yoshua Bengio. Learning structured embeddings of knowledge bases. In Proceedings of AAAI, pages , [Bordes et al., 2012] Antoine Bordes, Xavier Glorot, Jason Weston, and Yoshua Bengio. Joint learning of words and meaning representations for open-text semantic parsing. In Proceedings of AISTATS, pages , [Bordes et al., 2013] Antoine Bordes, Nicolas Usunier, Alberto Garcia-Duran, Jason Weston, and Oksana Yakhnenko. Translating embeddings for modeling multirelational data. In Proceedings of NIPS, pages , [Bordes et al., 2014] Antoine Bordes, Xavier Glorot, Jason Weston, and Yoshua Bengio. A semantic matching energy function for learning with multi-relational data. Machine Learning, 94(2): , [Dong et al., 2014] Xin Dong, Evgeniy Gabrilovich, Geremy Heitz, Wilko Horn, Ni Lao, Kevin Murphy, Thomas Strohmann, Shaohua Sun, and Wei Zhang. Knowledge vault: A web-scale approach to probabilistic knowledge fusion. In Proceedings of KDD, pages , [Gu et al., 2015] Kelvin Gu, John Miller, and Percy Liang. Traversing knowledge graphs in vector space. In Proceedings of EMNLP, pages , [Hu et al., 2015] Zhiting Hu, Poyao Huang, Yuntian Deng, Yingkai Gao, and Eric P Xing. Entity hierarchy embedding. In Proceedings of ACL, volume 1, pages , [Jenatton et al., 2012] Rodolphe Jenatton, Nicolas L Roux, Antoine Bordes, and Guillaume R Obozinski. A latent factor model for highly multi-relational data. In Proceedings of NIPS, pages , [Ji et al., 2015] Guoliang Ji, Shizhu He, Liheng Xu, Kang Liu, and Jun Zhao. Knowledge graph embedding via dynamic mapping matrix. In Proceedings of ACL, pages , [Krompaß et al., 2015] Denis Krompaß, Stephan Baier, and Volker Tresp. Type-constrained representation learning in knowledge graphs. In Proceedings of ISWC, pages [Lin et al., 2015a] Yankai Lin, Zhiyuan Liu, Huanbo Luan, Maosong Sun, Siwei Rao, and Song Liu. Modeling relation paths for representation learning of knowledge bases. In Proceedings of EMNLP, pages , [Lin et al., 2015b] Yankai Lin, Zhiyuan Liu, Maosong Sun, Yang Liu, and Xuan Zhu. Learning entity and relation embeddings for knowledge graph completion. In Proceedings of AAAI, [Neelakantan et al., 2015] Arvind Neelakantan, Benjamin Roth, and Andrew McCallum. Compositional vector space models for knowledge base completion. Proceedings of EMNLP, [Nickel et al., 2011] Maximilian Nickel, Volker Tresp, and Hans-Peter Kriegel. A three-way model for collective learning on multi-relational data. In Proceedings of ICML, pages , [Nickel et al., 2012] Maximilian Nickel, Volker Tresp, and Hans-Peter Kriegel. Factorizing yago: scalable machine learning for linked data. In Proceedings of WWW, pages , [Socher et al., 2013] Richard Socher, Danqi Chen, Christopher D Manning, and Andrew Ng. Reasoning with neural tensor networks for knowledge base completion. In Proceedings of NIPS, pages , [Wang et al., 2014a] Zhen Wang, Jianwen Zhang, Jianlin Feng, and Zheng Chen. Knowledge graph and text jointly embedding. In Proceedings of EMNLP, pages , [Wang et al., 2014b] Zhen Wang, Jianwen Zhang, Jianlin Feng, and Zheng Chen. Knowledge graph embedding by translating on hyperplanes. In Proceedings of AAAI, pages , [Xie et al., 2016] Ruobing Xie, Zhiyuan Liu, Jia Jia, Huanbo Luan, and Maosong Sun. Representation learning of knowledge graphs with entity descriptions. In Proceedings of AAAI, [Yang et al., 2014] Bishan Yang, Wen-tau Yih, Xiaodong He, Jianfeng Gao, and Li Deng. Embedding entities and relations for learning and inference in knowledge bases. Proceedings of ICLR, [Zhong et al., 2015] Huaping Zhong, Jianwen Zhang, Zhen Wang, Hai Wan, and Zheng Chen. Aligning knowledge and text embeddings by entity descriptions. In Proceedings of EMNLP, pages ,

Knowledge Graph Embedding with Numeric Attributes of Entities

Knowledge Graph Embedding with Numeric Attributes of Entities Knowledge Graph Embedding with Numeric Attributes of Entities Yanrong Wu, Zhichun Wang College of Information Science and Technology Beijing Normal University, Beijing 100875, PR. China yrwu@mail.bnu.edu.cn,

More information

From One Point to a Manifold: Knowledge Graph Embedding for Precise Link Prediction

From One Point to a Manifold: Knowledge Graph Embedding for Precise Link Prediction Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI-16) From One Point to a Manifold: Knowledge Graph Embedding for Precise Link Prediction Han Xiao, Minlie

More information

Knowledge graph embeddings

Knowledge graph embeddings Knowledge graph embeddings Paolo Rosso, Dingqi Yang, Philippe Cudré-Mauroux Synonyms Knowledge base embeddings RDF graph embeddings Definitions Knowledge graph embeddings: a vector representation of entities

More information

Image-embodied Knowledge Representation Learning

Image-embodied Knowledge Representation Learning -embodied Knowledge Representation Learning Ruobing Xie 1, Zhiyuan Liu 1,2, Huanbo Luan 1, Maosong un 1,2 1 Department of Computer cience and Technology, tate Key Lab on Intelligent Technology and ystems,

More information

Transition-based Knowledge Graph Embedding with Relational Mapping Properties

Transition-based Knowledge Graph Embedding with Relational Mapping Properties Transition-based Knowledge Graph Embedding with Relational Mapping Properties Miao Fan,, Qiang Zhou, Emily Chang, Thomas Fang Zheng, CSLT, Tsinghua National Laboratory for Information Science and Technology,

More information

Knowledge Graph Embedding Translation Based on Constraints

Knowledge Graph Embedding Translation Based on Constraints Journal of Information Hiding and Multimedia Signal Processing c 2017 ISSN 2073-4212 Ubiquitous International Volume 8, Number 5, September 2017 Knowledge Graph Embedding Translation Based on Constraints

More information

Knowledge Transfer for Out-of-Knowledge-Base Entities: A Graph Neural Network Approach

Knowledge Transfer for Out-of-Knowledge-Base Entities: A Graph Neural Network Approach Knowledge Transfer for Out-of-Knowledge-Base Entities: A Graph Neural Network Approach Takuo Hamaguchi, 1 Hidekazu Oiwa, 2 Masashi Shimbo, 1 and Yuji Matsumoto 1 1 Nara Institute of Science and Technology,

More information

Row-less Universal Schema

Row-less Universal Schema Row-less Universal Schema Patrick Verga & Andrew McCallum College of Information and Computer Sciences University of Massachusetts Amherst {pat, mccallum}@cs.umass.edu Abstract Universal schema jointly

More information

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li Learning to Match Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li 1. Introduction The main tasks in many applications can be formalized as matching between heterogeneous objects, including search, recommendation,

More information

arxiv: v1 [cs.cl] 3 Jun 2015

arxiv: v1 [cs.cl] 3 Jun 2015 Traversing Knowledge Graphs in Vector Space Kelvin Gu Stanford University kgu@stanford.edu John Miller Stanford University millerjp@stanford.edu Percy Liang Stanford University pliang@cs.stanford.edu arxiv:1506.01094v1

More information

Partial Knowledge in Embeddings

Partial Knowledge in Embeddings Partial Knowledge in Embeddings R.V.Guha... Abstract. Representing domain knowledge is crucial for any task. There has been a wide range of techniques developed to represent this knowledge, from older

More information

arxiv: v1 [cs.ai] 11 Aug 2015

arxiv: v1 [cs.ai] 11 Aug 2015 Type-Constrained Representation Learning in Knowledge Graphs Denis Krompaß 1,2, Stephan Baier 1, and Volker Tresp 1,2 arxiv:1508.02593v1 [cs.ai] 11 Aug 2015 Siemens AG, Corporate Technology, Munich, Germany

More information

Cross-lingual Knowledge Graph Alignment via Graph Convolutional Networks

Cross-lingual Knowledge Graph Alignment via Graph Convolutional Networks Cross-lingual Knowledge Graph Alignment via Graph Convolutional Networks Zhichun Wang, Qingsong Lv, Xiaohan Lan, Yu Zhang College of Information Science and Technique Beijing Normal University, Beijing,

More information

ProjE: Embedding Projection for Knowledge Graph Completion

ProjE: Embedding Projection for Knowledge Graph Completion Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17) ProjE: Embedding Projection for Knowledge Graph Completion Baoxu Shi and Tim Weninger University of Notre Dame Email:

More information

Type-Constrained Representation Learning in Knowledge Graphs

Type-Constrained Representation Learning in Knowledge Graphs Type-Constrained Representation Learning in Knowledge Graphs Denis Krompaß 1,2, Stephan Baier 1, and Volker Tresp 1,2 Siemens AG, Corporate Technology, Munich, Germany Denis.Krompass@siemens.com Volker.Tresp@siemens.com

More information

Entity Alignment between Knowledge Graphs Using Attribute Embeddings

Entity Alignment between Knowledge Graphs Using Attribute Embeddings Entity Alignment between Knowledge Graphs Using Attribute Embeddings Bayu Distiawan Trsedya and Jianzhong Qi and Rui Zhang School of Computing and Information Systems, The University of Melbourne {btrisedya@student.,

More information

Holographic Embeddings of Knowledge Graphs

Holographic Embeddings of Knowledge Graphs CBMM Memo No. 39 November 16, 2015 Holographic Embeddings of Knowledge Graphs by Maximilian Nickel, Lorenzo Rosasco and Tomaso Poggio Abstract Learning embeddings of entities and relations is an efficient

More information

A Multimodal Translation-Based Approach for Knowledge Graph Representation Learning

A Multimodal Translation-Based Approach for Knowledge Graph Representation Learning A Multimodal Translation-Based Approach for Knowledge Graph Representation Learning Hatem Mousselly-Sergieh, Teresa Botschen, Iryna Gurevych, Stefan Roth Ubiquitous Knowledge Processing (UKP) Lab Visual

More information

arxiv: v2 [cs.cl] 19 Aug 2015

arxiv: v2 [cs.cl] 19 Aug 2015 Traversing Knowledge Graphs in Vector Space Kelvin Guu Stanford University kguu@stanford.edu John Miller Stanford University millerjp@stanford.edu Percy Liang Stanford University pliang@cs.stanford.edu

More information

Compositional Learning of Embeddings for Relation Paths in Knowledge Bases and Text

Compositional Learning of Embeddings for Relation Paths in Knowledge Bases and Text Compositional Learning of Embeddings for Relation Paths in Knowledge Bases and Text Kristina Toutanova Microsoft Research Redmond, WA, USA Xi Victoria Lin Computer Science & Engineering University of Washington

More information

A Brief Review of Representation Learning in Recommender 赵鑫 RUC

A Brief Review of Representation Learning in Recommender 赵鑫 RUC A Brief Review of Representation Learning in Recommender Systems @ 赵鑫 RUC batmanfly@qq.com Representation learning Overview of recommender systems Tasks Rating prediction Item recommendation Basic models

More information

Outline. Morning program Preliminaries Semantic matching Learning to rank Entities

Outline. Morning program Preliminaries Semantic matching Learning to rank Entities 112 Outline Morning program Preliminaries Semantic matching Learning to rank Afternoon program Modeling user behavior Generating responses Recommender systems Industry insights Q&A 113 are polysemic Finding

More information

Neural Symbolic Machines: Learning Semantic Parsers on Freebase with Weak Supervision

Neural Symbolic Machines: Learning Semantic Parsers on Freebase with Weak Supervision Neural Symbolic Machines: Learning Semantic Parsers on Freebase with Weak Supervision Anonymized for review Abstract Extending the success of deep neural networks to high level tasks like natural language

More information

DeepWalk: Online Learning of Social Representations

DeepWalk: Online Learning of Social Representations DeepWalk: Online Learning of Social Representations ACM SIG-KDD August 26, 2014, Rami Al-Rfou, Steven Skiena Stony Brook University Outline Introduction: Graphs as Features Language Modeling DeepWalk Evaluation:

More information

Automatic Domain Partitioning for Multi-Domain Learning

Automatic Domain Partitioning for Multi-Domain Learning Automatic Domain Partitioning for Multi-Domain Learning Di Wang diwang@cs.cmu.edu Chenyan Xiong cx@cs.cmu.edu William Yang Wang ww@cmu.edu Abstract Multi-Domain learning (MDL) assumes that the domain labels

More information

Embedding Multimodal Relational Data

Embedding Multimodal Relational Data Embedding Multimodal Relational Data Pouya Pezeshkpour University of California Irvine, CA pezeshkp@uci.edu Liyan Chen University of California Irvine, CA liyanc@uci.edu Sameer Singh University of California

More information

Efficient and Expressive Knowledge Base Completion Using Subgraph Feature Extraction

Efficient and Expressive Knowledge Base Completion Using Subgraph Feature Extraction Efficient and Expressive Knowledge Base Completion Using Subgraph Feature Extraction Matt Gardner and Tom Mitchell Carnegie Mellon University mg1@cs.cmu.edu, tom.mitchell@cmu.edu Abstract We explore some

More information

Combining Text Embedding and Knowledge Graph Embedding Techniques for Academic Search Engines

Combining Text Embedding and Knowledge Graph Embedding Techniques for Academic Search Engines Combining Text Embedding and Knowledge Graph Embedding Techniques for Academic Search Engines SemDeep-4, Oct. 2018 Gengchen Mai Krzysztof Janowicz Bo Yan STKO Lab, University of California, Santa Barbara

More information

Empirical Evaluation of RNN Architectures on Sentence Classification Task

Empirical Evaluation of RNN Architectures on Sentence Classification Task Empirical Evaluation of RNN Architectures on Sentence Classification Task Lei Shen, Junlin Zhang Chanjet Information Technology lorashen@126.com, zhangjlh@chanjet.com Abstract. Recurrent Neural Networks

More information

On the Representation and Embedding of Knowledge Bases Beyond Binary Relations

On the Representation and Embedding of Knowledge Bases Beyond Binary Relations Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI-16) On the Representation and Embedding of Knowledge Bases Beyond Binary Relations Jianfeng Wen 1,, Jianxin

More information

A Korean Knowledge Extraction System for Enriching a KBox

A Korean Knowledge Extraction System for Enriching a KBox A Korean Knowledge Extraction System for Enriching a KBox Sangha Nam, Eun-kyung Kim, Jiho Kim, Yoosung Jung, Kijong Han, Key-Sun Choi KAIST / The Republic of Korea {nam.sangha, kekeeo, hogajiho, wjd1004109,

More information

On2Vec: Embedding-based Relation Prediction for Ontology Population

On2Vec: Embedding-based Relation Prediction for Ontology Population On2Vec: Embedding-based Relation Prediction for Ontology Population Muhao Chen Yingtao Tian Xuelu Chen Zijun Xue Carlo Zaniolo Abstract Populating ontology graphs represents a long-standing problem for

More information

Channel Locality Block: A Variant of Squeeze-and-Excitation

Channel Locality Block: A Variant of Squeeze-and-Excitation Channel Locality Block: A Variant of Squeeze-and-Excitation 1 st Huayu Li Northern Arizona University Flagstaff, United State Northern Arizona University hl459@nau.edu arxiv:1901.01493v1 [cs.lg] 6 Jan

More information

Effective Latent Space Graph-based Re-ranking Model with Global Consistency

Effective Latent Space Graph-based Re-ranking Model with Global Consistency Effective Latent Space Graph-based Re-ranking Model with Global Consistency Feb. 12, 2009 1 Outline Introduction Related work Methodology Graph-based re-ranking model Learning a latent space graph A case

More information

GraphGAN: Graph Representation Learning with Generative Adversarial Nets

GraphGAN: Graph Representation Learning with Generative Adversarial Nets The 32 nd AAAI Conference on Artificial Intelligence (AAAI 2018) New Orleans, Louisiana, USA GraphGAN: Graph Representation Learning with Generative Adversarial Nets Hongwei Wang 1,2, Jia Wang 3, Jialin

More information

Show, Discriminate, and Tell: A Discriminatory Image Captioning Model with Deep Neural Networks

Show, Discriminate, and Tell: A Discriminatory Image Captioning Model with Deep Neural Networks Show, Discriminate, and Tell: A Discriminatory Image Captioning Model with Deep Neural Networks Zelun Luo Department of Computer Science Stanford University zelunluo@stanford.edu Te-Lin Wu Department of

More information

Semi-supervised Data Representation via Affinity Graph Learning

Semi-supervised Data Representation via Affinity Graph Learning 1 Semi-supervised Data Representation via Affinity Graph Learning Weiya Ren 1 1 College of Information System and Management, National University of Defense Technology, Changsha, Hunan, P.R China, 410073

More information

arxiv: v1 [cs.lg] 23 Jul 2018

arxiv: v1 [cs.lg] 23 Jul 2018 LinkNBed: Multi-Graph Representation Learning with Entity Linkage Rakshit Trivedi College of Computing Georgia Tech Christos Faloutsos SCS, CMU and Amazon.com Bunyamin Sisman Amazon.com Hongyuan Zha College

More information

KBLRN: End-to-End Learning of Knowledge Base Representations with Latent, Relational, and Numerical Features

KBLRN: End-to-End Learning of Knowledge Base Representations with Latent, Relational, and Numerical Features KBLRN: End-to-End Learning of Knowledge Base Representations with Latent, Relational, and Numerical Features More recently, the research community has focused on efficient machine learning models that

More information

TriRank: Review-aware Explainable Recommendation by Modeling Aspects

TriRank: Review-aware Explainable Recommendation by Modeling Aspects TriRank: Review-aware Explainable Recommendation by Modeling Aspects Xiangnan He, Tao Chen, Min-Yen Kan, Xiao Chen National University of Singapore Presented by Xiangnan He CIKM 15, Melbourne, Australia

More information

KADE: Aligning Knowledge Base and Document Embedding Models using Regularized Multi-Task Learning

KADE: Aligning Knowledge Base and Document Embedding Models using Regularized Multi-Task Learning KADE: Aligning Knowledge Base and Document Embedding Models using Regularized Multi-Task Learning Matthias Baumgartner 1, Wen Zhang 2,3, Bibek Paudel 1, Daniele Dell Aglio 1, Huajun Chen 2,3, and Abraham

More information

RiMOM Results for OAEI 2009

RiMOM Results for OAEI 2009 RiMOM Results for OAEI 2009 Xiao Zhang, Qian Zhong, Feng Shi, Juanzi Li and Jie Tang Department of Computer Science and Technology, Tsinghua University, Beijing, China zhangxiao,zhongqian,shifeng,ljz,tangjie@keg.cs.tsinghua.edu.cn

More information

Location-Aware Web Service Recommendation Using Personalized Collaborative Filtering

Location-Aware Web Service Recommendation Using Personalized Collaborative Filtering ISSN 2395-1621 Location-Aware Web Service Recommendation Using Personalized Collaborative Filtering #1 Shweta A. Bhalerao, #2 Prof. R. N. Phursule 1 Shweta.bhalerao75@gmail.com 2 rphursule@gmail.com #12

More information

Combining Review Text Content and Reviewer-Item Rating Matrix to Predict Review Rating

Combining Review Text Content and Reviewer-Item Rating Matrix to Predict Review Rating Combining Review Text Content and Reviewer-Item Rating Matrix to Predict Review Rating Dipak J Kakade, Nilesh P Sable Department of Computer Engineering, JSPM S Imperial College of Engg. And Research,

More information

Learning Semantic Entity Representations with Knowledge Graph and Deep Neural Networks and its Application to Named Entity Disambiguation

Learning Semantic Entity Representations with Knowledge Graph and Deep Neural Networks and its Application to Named Entity Disambiguation Learning Semantic Entity Representations with Knowledge Graph and Deep Neural Networks and its Application to Named Entity Disambiguation Hongzhao Huang 1 and Larry Heck 2 Computer Science Department,

More information

Scheme of Big-Data Supported Interactive Evolutionary Computation

Scheme of Big-Data Supported Interactive Evolutionary Computation 2017 2nd International Conference on Information Technology and Management Engineering (ITME 2017) ISBN: 978-1-60595-415-8 Scheme of Big-Data Supported Interactive Evolutionary Computation Guo-sheng HAO

More information

A Deep Relevance Matching Model for Ad-hoc Retrieval

A Deep Relevance Matching Model for Ad-hoc Retrieval A Deep Relevance Matching Model for Ad-hoc Retrieval Jiafeng Guo 1, Yixing Fan 1, Qingyao Ai 2, W. Bruce Croft 2 1 CAS Key Lab of Web Data Science and Technology, Institute of Computing Technology, Chinese

More information

The Open University s repository of research publications and other research outputs. Search Personalization with Embeddings

The Open University s repository of research publications and other research outputs. Search Personalization with Embeddings Open Research Online The Open University s repository of research publications and other research outputs Search Personalization with Embeddings Conference Item How to cite: Vu, Thanh; Nguyen, Dat Quoc;

More information

Latent Distance Models and. Graph Feature Models Dr. Asja Fischer, Prof. Jens Lehmann

Latent Distance Models and. Graph Feature Models Dr. Asja Fischer, Prof. Jens Lehmann Latent Distance Models and Graph Feature Models 2016-02-09 Dr. Asja Fischer, Prof. Jens Lehmann Overview Last lecture we saw how neural networks/multi-layer perceptrons (MLPs) can be applied to SRL. After

More information

Representation Learning using Multi-Task Deep Neural Networks for Semantic Classification and Information Retrieval

Representation Learning using Multi-Task Deep Neural Networks for Semantic Classification and Information Retrieval Representation Learning using Multi-Task Deep Neural Networks for Semantic Classification and Information Retrieval Xiaodong Liu 12, Jianfeng Gao 1, Xiaodong He 1 Li Deng 1, Kevin Duh 2, Ye-Yi Wang 1 1

More information

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS Kuan-Chuan Peng and Tsuhan Chen School of Electrical and Computer Engineering, Cornell University, Ithaca, NY

More information

Predicting answer types for question-answering

Predicting answer types for question-answering Predicting answer types for question-answering Ivan Bogatyy Google Inc. 1600 Amphitheatre pkwy Mountain View, CA 94043 bogatyy@google.com Abstract We tackle the problem of answer type prediction, applying

More information

Conflict Graphs for Parallel Stochastic Gradient Descent

Conflict Graphs for Parallel Stochastic Gradient Descent Conflict Graphs for Parallel Stochastic Gradient Descent Darshan Thaker*, Guneet Singh Dhillon* Abstract We present various methods for inducing a conflict graph in order to effectively parallelize Pegasos.

More information

Linking Entities in Chinese Queries to Knowledge Graph

Linking Entities in Chinese Queries to Knowledge Graph Linking Entities in Chinese Queries to Knowledge Graph Jun Li 1, Jinxian Pan 2, Chen Ye 1, Yong Huang 1, Danlu Wen 1, and Zhichun Wang 1(B) 1 Beijing Normal University, Beijing, China zcwang@bnu.edu.cn

More information

Proceedings of the International MultiConference of Engineers and Computer Scientists 2018 Vol I IMECS 2018, March 14-16, 2018, Hong Kong

Proceedings of the International MultiConference of Engineers and Computer Scientists 2018 Vol I IMECS 2018, March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong TABLE I CLASSIFICATION ACCURACY OF DIFFERENT PRE-TRAINED MODELS ON THE TEST DATA

More information

Multi-Stage Rocchio Classification for Large-scale Multilabeled

Multi-Stage Rocchio Classification for Large-scale Multilabeled Multi-Stage Rocchio Classification for Large-scale Multilabeled Text data Dong-Hyun Lee Nangman Computing, 117D Garden five Tools, Munjeong-dong Songpa-gu, Seoul, Korea dhlee347@gmail.com Abstract. Large-scale

More information

Recurrent Convolutional Neural Networks for Scene Labeling

Recurrent Convolutional Neural Networks for Scene Labeling Recurrent Convolutional Neural Networks for Scene Labeling Pedro O. Pinheiro, Ronan Collobert Reviewed by Yizhe Zhang August 14, 2015 Scene labeling task Scene labeling: assign a class label to each pixel

More information

Semantic Website Clustering

Semantic Website Clustering Semantic Website Clustering I-Hsuan Yang, Yu-tsun Huang, Yen-Ling Huang 1. Abstract We propose a new approach to cluster the web pages. Utilizing an iterative reinforced algorithm, the model extracts semantic

More information

S-MART: Novel Tree-based Structured Learning Algorithms Applied to Tweet Entity Linking

S-MART: Novel Tree-based Structured Learning Algorithms Applied to Tweet Entity Linking S-MART: Novel Tree-based Structured Learning Algorithms Applied to Tweet Entity Linking Yi Yang * and Ming-Wei Chang # * Georgia Institute of Technology, Atlanta # Microsoft Research, Redmond Traditional

More information

Translation-based Recommendation: A Scalable Method for Modeling Sequential Behavior

Translation-based Recommendation: A Scalable Method for Modeling Sequential Behavior Translation-based Recommendation: A Scalable Method for Modeling Sequential Behavior Ruining He, Wang-Cheng Kang, Julian McAuley UC San Diego {r4he, wckang, jmcauley}@ucsd.edu Abstract Modeling the complex

More information

Top-N Recommendations from Implicit Feedback Leveraging Linked Open Data

Top-N Recommendations from Implicit Feedback Leveraging Linked Open Data Top-N Recommendations from Implicit Feedback Leveraging Linked Open Data Vito Claudio Ostuni, Tommaso Di Noia, Roberto Mirizzi, Eugenio Di Sciascio Polytechnic University of Bari, Italy {ostuni,mirizzi}@deemail.poliba.it,

More information

Grounded Compositional Semantics for Finding and Describing Images with Sentences

Grounded Compositional Semantics for Finding and Describing Images with Sentences Grounded Compositional Semantics for Finding and Describing Images with Sentences R. Socher, A. Karpathy, V. Le,D. Manning, A Y. Ng - 2013 Ali Gharaee 1 Alireza Keshavarzi 2 1 Department of Computational

More information

A Recommender System Based on Improvised K- Means Clustering Algorithm

A Recommender System Based on Improvised K- Means Clustering Algorithm A Recommender System Based on Improvised K- Means Clustering Algorithm Shivani Sharma Department of Computer Science and Applications, Kurukshetra University, Kurukshetra Shivanigaur83@yahoo.com Abstract:

More information

arxiv: v6 [cs.lg] 4 Jul 2018

arxiv: v6 [cs.lg] 4 Jul 2018 Convolutional 2D Knowledge Graph Embeddings Tim Dettmers Università della Svizzera italiana tim.dettmers@gmail.com Pasquale Minervini Pontus Stenetorp Sebastian Riedel University College London {p.minervini,p.stenetorp,s.riedel}@cs.ucl.ac.uk

More information

Heterogeneous Graph-Based Intent Learning with Queries, Web Pages and Wikipedia Concepts

Heterogeneous Graph-Based Intent Learning with Queries, Web Pages and Wikipedia Concepts Heterogeneous Graph-Based Intent Learning with Queries, Web Pages and Wikipedia Concepts Xiang Ren, Yujing Wang, Xiao Yu, Jun Yan, Zheng Chen, Jiawei Han University of Illinois, at Urbana Champaign MicrosoD

More information

Collaborative Filtering using a Spreading Activation Approach

Collaborative Filtering using a Spreading Activation Approach Collaborative Filtering using a Spreading Activation Approach Josephine Griffith *, Colm O Riordan *, Humphrey Sorensen ** * Department of Information Technology, NUI, Galway ** Computer Science Department,

More information

arxiv: v1 [cs.ai] 1 Sep 2018

arxiv: v1 [cs.ai] 1 Sep 2018 Improving Visual Relationship Detection using Semantic Modeling of Scene Descriptions Stephan Baier 1, Yunpu Ma 1,2, and Volker Tresp 1,2 1 Ludwig Maximilian University, 80538 Munich, Germany stephan.baier@campus.lmu.de

More information

Presented by: Dimitri Galmanovich. Petros Venetis, Alon Halevy, Jayant Madhavan, Marius Paşca, Warren Shen, Gengxin Miao, Chung Wu

Presented by: Dimitri Galmanovich. Petros Venetis, Alon Halevy, Jayant Madhavan, Marius Paşca, Warren Shen, Gengxin Miao, Chung Wu Presented by: Dimitri Galmanovich Petros Venetis, Alon Halevy, Jayant Madhavan, Marius Paşca, Warren Shen, Gengxin Miao, Chung Wu 1 When looking for Unstructured data 2 Millions of such queries every day

More information

Sentiment Classification of Food Reviews

Sentiment Classification of Food Reviews Sentiment Classification of Food Reviews Hua Feng Department of Electrical Engineering Stanford University Stanford, CA 94305 fengh15@stanford.edu Ruixi Lin Department of Electrical Engineering Stanford

More information

Improving the Efficiency of Fast Using Semantic Similarity Algorithm

Improving the Efficiency of Fast Using Semantic Similarity Algorithm International Journal of Scientific and Research Publications, Volume 4, Issue 1, January 2014 1 Improving the Efficiency of Fast Using Semantic Similarity Algorithm D.KARTHIKA 1, S. DIVAKAR 2 Final year

More information

Active Learning for Multi-relational Data Construction

Active Learning for Multi-relational Data Construction Active Learning for Multi-relational Data Construction Akihiro Kishimoto IBM Research - Ireland Damastown Industrial Estate Mulhuddart, Dublin 15, Ireland AKIHIROK@ie.ibm.com Adi Botea IBM Research - Ireland

More information

FastText. Jon Koss, Abhishek Jindal

FastText. Jon Koss, Abhishek Jindal FastText Jon Koss, Abhishek Jindal FastText FastText is on par with state-of-the-art deep learning classifiers in terms of accuracy But it is way faster: FastText can train on more than one billion words

More information

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

The Comparative Study of Machine Learning Algorithms in Text Data Classification* The Comparative Study of Machine Learning Algorithms in Text Data Classification* Wang Xin School of Science, Beijing Information Science and Technology University Beijing, China Abstract Classification

More information

The Results of Falcon-AO in the OAEI 2006 Campaign

The Results of Falcon-AO in the OAEI 2006 Campaign The Results of Falcon-AO in the OAEI 2006 Campaign Wei Hu, Gong Cheng, Dongdong Zheng, Xinyu Zhong, and Yuzhong Qu School of Computer Science and Engineering, Southeast University, Nanjing 210096, P. R.

More information

Metric Learning for Large-Scale Image Classification:

Metric Learning for Large-Scale Image Classification: Metric Learning for Large-Scale Image Classification: Generalizing to New Classes at Near-Zero Cost Florent Perronnin 1 work published at ECCV 2012 with: Thomas Mensink 1,2 Jakob Verbeek 2 Gabriela Csurka

More information

POI2Vec: Geographical Latent Representation for Predicting Future Visitors

POI2Vec: Geographical Latent Representation for Predicting Future Visitors : Geographical Latent Representation for Predicting Future Visitors Shanshan Feng 1 Gao Cong 2 Bo An 2 Yeow Meng Chee 3 1 Interdisciplinary Graduate School 2 School of Computer Science and Engineering

More information

Music Recommendation with Implicit Feedback and Side Information

Music Recommendation with Implicit Feedback and Side Information Music Recommendation with Implicit Feedback and Side Information Shengbo Guo Yahoo! Labs shengbo@yahoo-inc.com Behrouz Behmardi Criteo b.behmardi@criteo.com Gary Chen Vobile gary.chen@vobileinc.com Abstract

More information

LRLW-LSI: An Improved Latent Semantic Indexing (LSI) Text Classifier

LRLW-LSI: An Improved Latent Semantic Indexing (LSI) Text Classifier LRLW-LSI: An Improved Latent Semantic Indexing (LSI) Text Classifier Wang Ding, Songnian Yu, Shanqing Yu, Wei Wei, and Qianfeng Wang School of Computer Engineering and Science, Shanghai University, 200072

More information

Explore Co-clustering on Job Applications. Qingyun Wan SUNet ID:qywan

Explore Co-clustering on Job Applications. Qingyun Wan SUNet ID:qywan Explore Co-clustering on Job Applications Qingyun Wan SUNet ID:qywan 1 Introduction In the job marketplace, the supply side represents the job postings posted by job posters and the demand side presents

More information

Detecting Bone Lesions in Multiple Myeloma Patients using Transfer Learning

Detecting Bone Lesions in Multiple Myeloma Patients using Transfer Learning Detecting Bone Lesions in Multiple Myeloma Patients using Transfer Learning Matthias Perkonigg 1, Johannes Hofmanninger 1, Björn Menze 2, Marc-André Weber 3, and Georg Langs 1 1 Computational Imaging Research

More information

Deep MIML Network. Abstract

Deep MIML Network. Abstract Deep MIML Network Ji Feng and Zhi-Hua Zhou National Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210023, China Collaborative Innovation Center of Novel Software Technology

More information

Finding Tiny Faces Supplementary Materials

Finding Tiny Faces Supplementary Materials Finding Tiny Faces Supplementary Materials Peiyun Hu, Deva Ramanan Robotics Institute Carnegie Mellon University {peiyunh,deva}@cs.cmu.edu 1. Error analysis Quantitative analysis We plot the distribution

More information

Facial Expression Classification with Random Filters Feature Extraction

Facial Expression Classification with Random Filters Feature Extraction Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey mren@cs.toronto.edu Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Data Matrices and Vector Space Model Denis Helic KTI, TU Graz Nov 6, 2014 Denis Helic (KTI, TU Graz) KDDM1 Nov 6, 2014 1 / 55 Big picture: KDDM Probability

More information

Outlier Detection Using Unsupervised and Semi-Supervised Technique on High Dimensional Data

Outlier Detection Using Unsupervised and Semi-Supervised Technique on High Dimensional Data Outlier Detection Using Unsupervised and Semi-Supervised Technique on High Dimensional Data Ms. Gayatri Attarde 1, Prof. Aarti Deshpande 2 M. E Student, Department of Computer Engineering, GHRCCEM, University

More information

Network embedding. Cheng Zheng

Network embedding. Cheng Zheng Network embedding Cheng Zheng Outline Problem definition Factorization based algorithms --- Laplacian Eigenmaps(NIPS, 2001) Random walk based algorithms ---DeepWalk(KDD, 2014), node2vec(kdd, 2016) Deep

More information

Opinion Mining by Transformation-Based Domain Adaptation

Opinion Mining by Transformation-Based Domain Adaptation Opinion Mining by Transformation-Based Domain Adaptation Róbert Ormándi, István Hegedűs, and Richárd Farkas University of Szeged, Hungary {ormandi,ihegedus,rfarkas}@inf.u-szeged.hu Abstract. Here we propose

More information

End-To-End Spam Classification With Neural Networks

End-To-End Spam Classification With Neural Networks End-To-End Spam Classification With Neural Networks Christopher Lennan, Bastian Naber, Jan Reher, Leon Weber 1 Introduction A few years ago, the majority of the internet s network traffic was due to spam

More information

Web Information Retrieval using WordNet

Web Information Retrieval using WordNet Web Information Retrieval using WordNet Jyotsna Gharat Asst. Professor, Xavier Institute of Engineering, Mumbai, India Jayant Gadge Asst. Professor, Thadomal Shahani Engineering College Mumbai, India ABSTRACT

More information

arxiv: v1 [cs.si] 12 Jan 2019

arxiv: v1 [cs.si] 12 Jan 2019 Predicting Diffusion Reach Probabilities via Representation Learning on Social Networks Furkan Gursoy furkan.gursoy@boun.edu.tr Ahmet Onur Durahim onur.durahim@boun.edu.tr arxiv:1901.03829v1 [cs.si] 12

More information

Structured Learning. Jun Zhu

Structured Learning. Jun Zhu Structured Learning Jun Zhu Supervised learning Given a set of I.I.D. training samples Learn a prediction function b r a c e Supervised learning (cont d) Many different choices Logistic Regression Maximum

More information

CIS UDEL Working Notes on ImageCLEF 2015: Compound figure detection task

CIS UDEL Working Notes on ImageCLEF 2015: Compound figure detection task CIS UDEL Working Notes on ImageCLEF 2015: Compound figure detection task Xiaolong Wang, Xiangying Jiang, Abhishek Kolagunda, Hagit Shatkay and Chandra Kambhamettu Department of Computer and Information

More information

Link Prediction for Social Network

Link Prediction for Social Network Link Prediction for Social Network Ning Lin Computer Science and Engineering University of California, San Diego Email: nil016@eng.ucsd.edu Abstract Friendship recommendation has become an important issue

More information

EFFICIENT ATTRIBUTE REDUCTION ALGORITHM

EFFICIENT ATTRIBUTE REDUCTION ALGORITHM EFFICIENT ATTRIBUTE REDUCTION ALGORITHM Zhongzhi Shi, Shaohui Liu, Zheng Zheng Institute Of Computing Technology,Chinese Academy of Sciences, Beijing, China Abstract: Key words: Efficiency of algorithms

More information

Learning Spatiotemporal-Aware Representation for POI Recommendation

Learning Spatiotemporal-Aware Representation for POI Recommendation Learning Spatiotemporal-Aware Representation for POI Recommendation Bei Liu 1, Tieyun Qian 1, Bing Liu 2, Liang Hong 3, Zhenni You 1, Yuxiang Li 1 1 State Key Laboratory of Software Engineering, Wuhan

More information

arxiv: v1 [cs.cv] 6 Jul 2016

arxiv: v1 [cs.cv] 6 Jul 2016 arxiv:607.079v [cs.cv] 6 Jul 206 Deep CORAL: Correlation Alignment for Deep Domain Adaptation Baochen Sun and Kate Saenko University of Massachusetts Lowell, Boston University Abstract. Deep neural networks

More information

Supervised classification of law area in the legal domain

Supervised classification of law area in the legal domain AFSTUDEERPROJECT BSC KI Supervised classification of law area in the legal domain Author: Mees FRÖBERG (10559949) Supervisors: Evangelos KANOULAS Tjerk DE GREEF June 24, 2016 Abstract Search algorithms

More information

Tag Based Image Search by Social Re-ranking

Tag Based Image Search by Social Re-ranking Tag Based Image Search by Social Re-ranking Vilas Dilip Mane, Prof.Nilesh P. Sable Student, Department of Computer Engineering, Imperial College of Engineering & Research, Wagholi, Pune, Savitribai Phule

More information

Multi-label Classification. Jingzhou Liu Dec

Multi-label Classification. Jingzhou Liu Dec Multi-label Classification Jingzhou Liu Dec. 6 2016 Introduction Multi-class problem, Training data (x $, y $ ) ( ), x $ X R., y $ Y = 1,2,, L Learn a mapping f: X Y Each instance x $ is associated with

More information

MULTI-SCALE OBJECT DETECTION WITH FEATURE FUSION AND REGION OBJECTNESS NETWORK. Wenjie Guan, YueXian Zou*, Xiaoqun Zhou

MULTI-SCALE OBJECT DETECTION WITH FEATURE FUSION AND REGION OBJECTNESS NETWORK. Wenjie Guan, YueXian Zou*, Xiaoqun Zhou MULTI-SCALE OBJECT DETECTION WITH FEATURE FUSION AND REGION OBJECTNESS NETWORK Wenjie Guan, YueXian Zou*, Xiaoqun Zhou ADSPLAB/Intelligent Lab, School of ECE, Peking University, Shenzhen,518055, China

More information