WIKIMATCHER: LEVERAGING WIKIPEDIA FOR ONTOLOGY ALIGNMENT

Size: px
Start display at page:

Download "WIKIMATCHER: LEVERAGING WIKIPEDIA FOR ONTOLOGY ALIGNMENT"

Transcription

1 WIKIMATCHER: LEVERAGING WIKIPEDIA FOR ONTOLOGY ALIGNMENT A thesis submitted in partial fulfillment of the requirements for the degree of Master of Science in Computer Engineering By HELENA B. MCCURDY B.S., Wright State University, Wright State University

2 WRIGHT STATE UNIVERSITY GRADUATE SCHOOL April 27, 2016 I HEREBY RECOMMEND THAT THE THESIS PREPARED UNDER MY SUPER- VISION BY Helena B McCurdy ENTITLED WikiMatcher: Leveraging Wikipedia for Ontology Alignment BE ACCEPTED IN PARTIAL FULFILLMENT OF THE REQUIRE- MENTS FOR THE DEGREE OF Master of Science in Computer Engineering. Michelle Cheatham, Ph.D. Thesis Director Committee on Final Examination Mateen M. Rizki, Ph.D. Chair, Department of Computer Science and Engineering Michelle Cheatham, Ph.D. Mateen M. Rizki, Ph.D. Krishnaprasad Thirunarayan, Ph.D. Robert E.W. Fyffe, Ph.D. Vice President for Research and Dean of the Graduate School

3 ABSTRACT McCurdy, Helena. M.S. Department of Computer Science and Engineering, Wright State University, WikiMatcher: Leveraging Wikipedia for Ontology Alignment. As the Semantic Web grows, so does the number of ontologies used to structure the data within it. Aligning these ontologies is critical to fully realizing the potential of the web. Previous work in ontology alignment has shown that even alignment systems utilizing basic string similarity metrics can produce useful matches. Researchers speculate that including semantic as well as syntactic information inherent in entity labels can further improve alignment results. This paper examines that hypothesis by exploring the utility of using Wikipedia as a source of semantic information. Various elements of Wikipedia are considered, including article content, page terms, and search snippets. The utility of each information source is analyzed and a composite system, WikiMatcher, is created based on this analysis. The performance of WikiMatcher is compared to that of a basic stringbased alignment system on two established alignment benchmarks and two other real-world datasets. The extensive evaluation shows that although WikiMatcher performs similarly to that of the string metric overall, it is able to find many matches with no syntactic similarity between labels. This performance seems to be driven by Wikipedia s query resolution and page redirection system, rather than by the particular information from Wikipedia that is used to compare entities. iii

4 Contents 1 Introduction and Motivation Ontology Parts Ontology Alignment Motivation Background Wikipedia Based Matchers Evaluation of Ontology Alignment Systems Top Performers Approach Preprocessing Knowledge Gathering Alignment Generation Evaluation and Analysis Articles Snippets Page Terms When A Wikipedia-based Approach Is Viable Conclusion and Future Work 45 Bibliography 47 iv

5 List of Figures 1.1 An example ontology snippet describing athletic teams An example ontology snippet describing athletic teams An example of the response from the Wikipedia API after a snippet query (2) is made An example of the response from the Wikipedia API after the pages terms query (3) is made An example of the response from the Wikipedia API after the search query (4) is made An example of the information given to determine if a match was correct or incorrect Wikipedia page hit counts for the anatomy, conference, hydro, and geo datasets Precision scores for each dataset across each of the three tests Recall scores for each dataset across each of the three tests v

6 List of Tables 2.1 Results of top alignment systems and basic string equivalency at OAEI 2015 on conference track Results of top alignment systems and basic string equivalency at OAEI 2015 on anatomy track Results of Levenstein alignment Results of WikiMatcher article alignment Sample matches of WikiMatcher article alignment Results of WikiMatcher snippet alignment Sample matches of WikiMatcher snippet alignment Results of WikiMatcher terms alignment Sample matches of WikiMatcher terms alignment vi

7 ACKNOWLEDGEMENTS I would like to express my gratitude to my adviser and mentor, Dr. Michelle Cheatham, for giving me the inspiration and guidance in working on this project and thesis. I would also like to thank my family and friends for their enduring support during my time as an athlete and a student. Without their support I would not be where I am today. Lastly, thank you to Hermanus Botha for always being there for me throughout my two years of graduate school. vii

8 1 Introduction and Motivation It has become increasingly difficult to effectively manage data, knowledge, and information in the World Wide Web. This is due to the vast amount of data added to the web on a daily basis. No one person, or machine for that matter, has the ability to read or process it all let alone understand it all and be able to utilize it effectively and efficiently. However, with the introduction and implementation of The Semantic Web, we will be able to do just that. The Semantic Web was first proposed by Tim Berners Lee in The primary motivation is to provide the current web with the ability to contain more structured knowledge [Berners-Lee et al. 2001]. The Semantic Web enables computational machines to handle more intelligent tasks and reason about data without explicit input or assistance from a user. Semantics in the web helps with expressing meaning and context of information which allows for novel ways of making existing data work for us. To realize the vision of the Semantic Web, ontologies are used to store the meaning, context, and the structure of data within the web. Specifically, an ontology is a representation of the concepts in a 1

9 domain and the relationships between them. Ontologies allow for the sharing and reuse of a common understanding of information structure and domain knowledge between both software and people. Ontologies also make it easy to separate domain knowledge from operational knowledge and facilitate the analysis of established domain knowledge [Noy et al. 2001] so there can be universal acceptance. It is important to have a common understanding of information structure and be able to reuse domain knowledge since many specific concepts share the same base concepts. For example, every family is different, however the underlying concept is always the same. There are parents and children and all of them are people. Using the same base ontology to describe all families ensures the ability to access, extract, and combine the same data in the same manner for every family. 1.1 Ontology Parts Ontologies consist of classes, individuals, and properties. Figure 1.1 shows a basic example of an ontology describing athletic teams. An ontology describes a specific domain by making use of the components mentioned above. Classes within an ontology form a hierarchy which describe entities that have similar characteristics. Classes in the provided athletic team example include Thing, Person, Athlete, Coach, Team, and Sport. This ontology shows that Athlete and Coach are both subclasses of Person. A subclass is a more specific representation of its superclass. An individual within an ontology is an instance of a class. An example is Jane Smith who is an instance of type Athlete. Jane Smith has a specific name, position, and age of type string, string, and integer, respectively. Classes and individ- 2

10 uals in an ontology are very much the same as classes and instances within Object Oriented Programming. All athletes, regardless of team and sport will have the same descriptive features which are contained within the Athlete class. There are two types of properties Thing Person subclassof String hasposition subclassof subclassof Int hasage Athlete hasname coaches ismemberof Coach hasname String String Team String hasrules participatesin Sport istype String Figure 1.1: An example ontology snippet describing athletic teams that can be found in an ontology: object properties and data properties. One individual can be related to another with object properties. For example, an individual of type Coach coaches an individual of type Athlete where coaches is the object property in this example. Other object properties from the athletic team example include ismemberof and participatesin. Data properties relate an instance to a literal value. An example is the relationship hasage between Athlete and integer. Other data properties from the athletic team ontology example include hasname and hasposition. 3

11 1.2 Ontology Alignment There is no single correct way to design and model an ontology for a given domain. Furthermore, two ontologies describing different domains could possibly include the same objects within them. Also, many ontology design decisions must be made by designers with varied backgrounds and experiences. Their differing experiences and opinions could result in them creating different ontologies for the same domain. The end result is that even with two ontologies that represent the same domain, the overall layout and object names may not be the same. This is because engineering new ontologies is not a deterministic process. For example, Figure 1.2 shows another ontology snippet of athletic teams from the point of view of a league organizer. Both example ontology snippets indicate that athletic teams have athletes and coaches; however, each one does so differently. The goal of ontology alignment is to determine when an entity in one ontology is semantically related to an entity in another ontology [Euzenat et al. 2007]. A typical ontology alignment system will accept two different ontologies. It will then compare all or a selection of entities within each ontology and compile a set of matches, usually with a confidence score. An example of a match from the two athletic team example ontology snippets include Athlete to Player. In both ontology snippets that class represents a person who plays for a sports team. The difference between the two ontologies is that in one ontology snippet the position of the person is important whereas in the other the boolean value of fees paid is important. The ability to accurately align ontologies without substantial manual involvement is 4

12 Thing subclassof Employee hassalary float Coach subclassof Athletic Trainer assignedto String istype Sports Team totalfeesowed float hasparticipants feespaid hasname Boolean Players String hasage Int Figure 1.2: An example ontology snippet describing athletic teams essential to the development of applications that leverage the potential of the Semantic Web. There are various common methods used for ontology alignment, each of which utilizes a different aspect of the information that can be gathered from the ontologies being aligned [Euzenat et al. 2004]. One method, terminological, uses the syntactic relatedness between entity names or other stings gathered from each entity. Previous work showed that much of the semantic meaning inherent in current ontologies is encapsulated in the names (i.e. labels) that ontology designers assign to entities. For this reason, ontology matching algorithms based solely on lexical comparison of entity labels perform surprisingly well 5

13 [Cheatham and Hitzler 2013] [Cheatham and Hitzler 2014]. An example of a terminological match between the two ontology snippets is the match of Coach from example 1 to Coach from example 2. Another method for ontology alignment is to compare the structural or design similarity between two ontologies. This can be the internal (the value range of attributes) or external (the relationship of one entity to other entities) structure of the ontologies. An example of a structural match from the two example ontologies is Athlete from example 1 to Player from example 2. This is because both Athlete and Player are very similar in what defines them and both hasage and hasname can be easily matched first. A third method for ontology alignment is to consider the semantic meaning or interpretation of entity labels. Ontology alignment researchers frequently speculate that leveraging external knowledge sources such as thesauri, dictionaries, encyclopedias, or search engine results could improve the performance of current alignment algorithms [Cheatham and Hitzler 2013]. An example of an alignment that could be generated from gathering external knowledge is Team from ontology 1 to Sports Team from ontology 2. For example, if Wikipedia is used the article of team discusses the article of sports team which could lead to a match. In this work we explore the utility of using Wikipedia as an external knowledge source to enable the semantic comparison of entities by an ontology alignment algorithm. In particular, we seek to answer the following questions: What, if any, information from Wikipedia is useful for ontology alignment? How can this information be used in an automated alignment system? 6

14 Is Wikipedia useful for aligning some types of ontologies but not others? How can we distinguish between the two cases? 1.3 Motivation Wikipedia was chosen because it provides a vast amount of data written by both specialists in every field and by average people. Wikipedia is the largest known encyclopedia in the world. At the time this document was written, there are 5,109,197 unique articles in the English version of Wikipedia and over 35 million articles in 291 different languages across all versions of Wikipedia. The site has 27,800,340 unique users and there are currently 133,529 active users who create, edit, and validate pages in the encyclopedia. On average, 20,000 new articles are added to Wikipedia every month, covering any and every topic. Up to date statistics can be found on Wikipedia:Statistics 1. In addition to the web interface, Wikipedia provides access to this vast amount of information through an easily accessible API. While this is not the first time Wikipedia has been incorporated into an alignment algorithm, this work improves upon previous work by systematically comparing different approaches to leveraging the data available on Wikipedia (e.g. simple text parsing, article snippet analysis, page term overlap). This work analyzes performance on a wider variety of datasets, including both general and domain-specific data

15 2 Background The importance of entity labels for ontology alignment is widely known. Many current ontology alignment tools use string similarity metrics based on labels to determine relatedness, e.g. [Cheatham and Hitzler 2013], [Stoilos et al. 2005], and all of the systems mentioned in [Shvaiko and Euzenat 2005]. Many researchers have also suggested that adding more knowledge by using the labels to consult an external knowledge source can improve alignment quality. Approaches along this line include domain-aware ontology matching [Slabbekoorn et al. 2012a] and matching using the WordNet thesaurus [Miller 1995]. An ontology alignment system using a domain-aware method first attempts to identify the domain of a source ontology so that it can restrict matches to only entities that exist within that domain. This type of matcher typically consults an external data source relevant to the domain of interest to gather more information about each entity, which then informs the matching process. This method shows improvement over domain-unaware methods in many situations, [Slabbekoorn et al. 2012b], [Ritze and Paulheim 2011], [Shamdasani 8

16 et al. 2011]. It obviously relies on the existence of a relevant machine-accessible domainspecific knowledge source. Many such sources are available for biomedical topics because that community has found semantic web technologies to be a valuable tool [Pesquita et al. 2009]. Examples of biomedical ontologies include the 514 ontologies in BioPortal 1. Using synonyms from WordNet to expand the possible equivalent terms for an entity label is one of the most common thoughts by alignment researchers regarding leveraging external knowledge sources [Lin and Sandkuhl 2008]. However, we have shown in previous work [Cheatham and Hitzler 2013] that the results of this approach are generally poor. The problem is two-fold: for common words the large number and topic variety of synonyms add significant noise to the alignment process, which increases false positives. On the other hand, the coverage of WordNet is very limited, to the extent that it doesn t contain enough technical/scientific words or jargon to be of any use when aligning domain-specific ontologies. The difficulties inherent in automatically determining the subject domain of an ontology and finding an appropriate domain-specific knowledge source, along with the limitations of WordNet for use in ontology matching led us to consider Wikipedia as a possible intermediate knowledge source that contains both domain-specific and general-purpose information

17 2.1 Wikipedia Based Matchers Several other researchers have considered Wikipedia for evaluating the semantic similarity between two concepts, including the developers of WikiRelate [Strube and Ponzetto 2006], BLOOMS [Jain et al. 2010], Wikipedia Linked-Based Measure (WLM) [Witten and Milne 2008], and WikiMatch [Hertling and Paulheim 2012]. These alignment systems differ in either how Wikipedia was utilized or how the data gathered from Wikipedia was analyzed. The WikiRelate approach uses Wikipedia s page category hierarchy to compute relatedness between pages corresponding to entities. The approach computes similarity values using a path-length algorithm that determines the shortest and most informative path between pages. Though the category approach outperformed approaches using WordNet, it was found that Wikipedia combined with both Google and WordNet performed best on benchmark datasets. Similarly, BLOOMS also utilizes Wikipedia s category hierarchy. A forest (group of trees) is created from the categories returned by Wikipedia for a particular search term and is compared to the forest for a second term using a standard set similarity metric. BLOOMS performs well in respect to other alignments systems on both linked open data as well as on benchmark datasets because it is able to utilize the noisy data from the Wikipedia categories effectively. WLM differs from the previous two approaches by using internal Wikipedia links as opposed to using categories. It uses both the links pointing to a Wikipedia page and links pointing outward from the page. These sets of links are then compared to determine relatedness. WLM differs from many other Wikipedia-based approaches in its response to failed queries. Generally if a query does not return a specific 10

18 article, a system will give up and use an alternative similarity metric that does not rely on information from Wikipedia. Instead, WLM uses the Wikipedia general search function to attempt to return related data in this situation. Finally WikiMatch utilizes Wikipedia s articles and language links. It uses no structural information from within the ontology itself. WikiMatch computes a similarity score based on the titles of the articles pulled. It does not beat state of the art systems but does perform better than the benchmark of the string metric approach on OAEI datasets. To further explain these existing Wikipedia based matchers, a simple example using the label player (game) and the label athlete will be walked through. For both WikiRelate and BLOOMS the categories of both labels will be retrieved from Wikipedia. The categories for Athlete include Sports Terminology and Sports Competitors. The category for Player (game) is Game Terminology. The BLOOMS method then pulls the parent categories to a depth of four. A parent category for Sports Terminology is Game Terminology, meaning many of the parent categories will be overlapping for these two labels. This will result in a higher confidence score for the match. In the case of WikiRelate, the shortest category path between the two pages is found, namely from Player (game) to Game Terminology to Sports Terminology to Athlete. From this path a similary score is calculated. WLM gathers the internal links for both pages rather than the categories. Some of the links gathered for Athlete include: sport, professional, some particularly well known athletes, and different types of sports that athletes participate in. Some of the links gathered for Player (game) include: game, gamer, player of the match, and some well known 11

19 games. A match confidence score is then computed by measuring the overlap of these internal links. Lastly, WikiMatch pulls the page title and article for each of the labels, fragments, and comments found within the ontologies being aligned. In the case of this simple example the articles and page titles Athlete and Player (Game) would be found. Then the language links to translated articles are pulled for each page title, examples for Athlete include Atleta, Atleto, Sportler, among others. No language links exist for Player (game). This information is then used to produce a confidence value for the match between the entities. The work presented here differs from these previous efforts in the type of data available on Wikipedia that is used for similarity computation. Rather than focusing on categories or links, this work evaluates the utility of article text, snippet text, and page terms for ontology alignment. This work also explores the utility of secondary information sources available on Wikipedia when an exact page can not be found for a particular entity label. 2.2 Evaluation of Ontology Alignment Systems The Ontology Alignment Evaluation Initiative 2 (OAEI) is a worldwide initiative to evalute the inpact and usefullness of ontology alignment systems. The OAEI hosts a yearly workshop at which ontology alignment systems are tested on various dataset tracks and their strengths and weaknesses are evaluated. Each alignment system and its results on said tracks are then published for further evaluation. The OAEI provides a platform for

20 researchers to share what they have learned and to improve upon the overall processes and techniques of ontology alignment. As mentioned, the OAEI consists of eight different tracks, which include: Benchmark - this synthetic dataset varies from year to year. It is designed to have wide coverage and to expose the strength of weaknesses of alignment systems. It is designed by systematically modifying established ontologies. Anatomy - a real world dataset consisting of two biomedical ontologies. One describes the anatomy of a human, and the other describes the anatomy of a mouse. Both ontologies are contain around 3000 entities and have a high number of overlapping terms. Conference - consists of 16 small real world datasets focused on the domain of conference organization. These datasets were created by gathering information about conference organization from a wide variety of sources. Multifarm - consists of a subset of the datasets from the conference track that have been translated into nine different languages. This is useful for evaluating an alignment system s multilingual capabilities. Large BioMedical Ontologies (largebio) - consists of three large biomedial datasets: the Foundational Model of Anatomy (FMA), SNOMED CT, and the National Cancer Institute Thesaurus (NCI). This track tests the scalability of alignment systems, as each dataset contains tens of thousands of semantically rich classes. 13

21 Interactive matching evaluation (interactive) - user interaction is incorporated into ontology alignment systems in an attempt to further improve overall alignment results. The datasets used are from the anatomy, conference, and largebio tracks. This track was introduced when researchers found alignment tools are slowly hitting the upper bound of performance for fully automated systems. 3 Instance Matching - consists of synthetically generate datasets that focus on the matching of instances. Ontology Alignment for Query Answering (oa4qa) - this track simulates an ontologybased data access (ODBA) use case for ontology alignment, in which a database is organized according to one ontology and a user is querying based on a different ontologies. The ontologies used are from the conference track. For the purpose of this work, both the conference and anatomy datasets will be used to evaluate the performance of Wikimatcher. These two data were chosen because they provide a dataset with frequent general terms and a dataset with very specific terms. This work is applicable to any of the other tracks that involve entities that might have Wikipedia pages. If the foreign language versions of Wikipedia were also used, then the multifarm track would also be relevant. Reference alignments for each track are provided to verify each pairwise match generated by alignment systems. When evaluating the effectiveness of an ontology alignment system on each of the OAEI tracks, the metrics used are precision, recall and f-measure

22 Precision is computed using Equation 2.1, where TP is true positives and FP is false positives. Precision measures the fraction of matches made that are correct. It is important when an alignment has many potentially viable matches that need to be narrowed down. precision = T P T P + F P (2.1) Recall is computed using Equation 2.2, where TP is true positives and FN is false negatives. Recall measures the fraction of correct matches made in comparison to all possible correct matches established by the reference alignment. Recall is important when an alignment system is trying to generate match possibilities to be filtered down in later stages of the alignment system. recall = T P T P + F N (2.2) F-measure is the overall accuracy of an alignment when both precision and recall is equally taken into account. That is, it is the harmonic mean of precision and recall. Equation 2.3 is used to compute f-measure for a generated alignment. f measure = 2 precision recall precision + recall = 2 T P 2 T P + F P + F Nl (2.3) Many of the tracks in OAEI contain matches that can be found by an exact label match. To remove the exact matching as a factor in the results, OAEI uses a metric called recall+. Recall+ is computed the same way as recall however, all exact matches are removed before the number of true positives and false negatives are found. This metric determines an 15

23 alignment system s effectiveness in finding non-trivial matches. Each of these scores results in a value between 0 and 1, where a higher value means a better performing system. The combination of these metrics provide a thorough insight into how each alignment system performs on each of the tracks. 2.3 Top Performers There were 22 total participants in the OAEI 2015 workshop, 14 participated in the conference track and 15 in the anatomy track [Cheatham et al. 2015]. Of the 14 participants 11 performed better overall than the basic string metric alignment. The top three performers and the basic string metric on the conference track are as seen in Table 2.1. Of the 15 participants on the anatomy track, nine preformed better than the basic string metric alignment. The top three performers and the basic string metric on the anatomy track are as seen in Table 2.2. The top performing system for both the conference and anatomy track was AgreementMakerLite (AML). AgreementMakerLite by [Faria et al. 2015] is based on the AgreementMarker ontology system [Cruz et al. 2009]. It utilizes both external knowledge and internal knowledge gathered form the ontologies being aligned. AML has nine different modules that can be used. The collection of modules that are employed depends on the input ontologies. The modules include: translation of foreign languages, string matching metrics, word matching algorithms, structural matching, property matching, and matching utilizing external knowledge sources. The external knowledge sources include three different bio-medical 16

24 Table 2.1: Results of top alignment systems and basic string equivalency at OAEI 2015 on conference track F-Measure Precision Recall AML Mamba XMAP StringEquiv Table 2.2: Results of top alignment systems and basic string equivalency at OAEI 2015 on anatomy track F-Measure Precision Recall AML XMAP LogMapBio StringEquiv ontologies and WordNet. Whereas other systems employ one or two of the matching techniques, AML brings them all together, producing top alignment results across all OAEI tracks. Another top performer in both the conference and anatomy tracks was XMap (extensible matching) [Djeddi et al. ]. XMap utilizes three different stages of alignment: terminological, structural, and alignment. The terminological layer combines finding string similarity of entity names with semantic data gathered from within the ontologies being aligned. The structural layer combines computing similarity between the external structural hierarchy of datasets, along with computing the similarity of internal structure of concepts. The alignment stage combines the information and matches gathered in the previous two stages to produce a final alignment. XMap performs well in both the anatomy 17

25 and conference tracks. XMap produces results with a higher precision in the conference track than any other system. The Mamba alignment system also performed very well on the conference dataset [Meilicke ]. Mamba utilizes the labels and logical entities to make hypotheses and assumptions about matches. It then uses Markov Logic to identify constraints and formulate a final alignment. An in-depth description of this approach can be found in [Meilicke and Stuckenschmidt 2015]. On the anatomy dataset LogMapBio also performed well, with a f-measure. LogMapBio is a variation of the LogMap system [Cuenca Grau and Jimenez-Ruiz 2011]. The LogMap alignment system uses a combination of lexical indexation, structural indexation, computation of initial anchor mappings, and mapping repair and discovery. LogMapBio extends LogMap by including the use of the external knowledge source Bio- Portal. BioPortal provides LogMapBio with the top bio-medical ontologies to pull knowledge from. This configuration allows LogMap to perform well on biomedical datasets. The work in this thesis is largely orthogonal to these full-featured alignment systems. The work presented here could be combined with existing complete alignment systems by providing additional semantic knowledge about entities being aligned. For instance, a Wikipedia-based matcher could be easily added as a tenth module in AML, an additional stage in XMap (which would allow false positives generated from Wikipedia to be culled in later stages), or an additional external knowledge source in LogMapBio. Achieving strong results via a Wikipedia-based approach to ontology alignment could therefore potentially 18

26 be used to improve the performance of the top-performing alignment systems when aligning ontologies pairs with entities likely to be covered on Wikipedia. 19

27 3 Approach WikiMatcher is a proof-of-concept alignment system that leverages knowledge from Wikipedia in order to generate an alignment between two ontologies. In keeping with the exploratory nature of this matcher, scalability is not currently a main design driver the algorithm compares every element in one ontology to every element in the other ontology. For each pair of entities, the labels are used to query Wikipedia in order to gather additional information about the concepts. This information is then used as input to a similarity metric. If the similarity of the two entities is above a threshold, they are considered equivalent and added to the overall alignment. Wikimatcher s alignment process can be thought of in terms of three distinct stages: preprocessing, knowledge gathering, and similarity computation. In the preprocessing phase, the entity labels are extracted and normalized. These normalized labels are then used to query Wikipedia during the knowledge gathering phase. Finally, the information gathered is used to compare entities and generate the overall alignment. The remainder of 20

28 this section describes each of these operations in more detail. 3.1 Preprocessing The primary goals of the preprocessing phase are to mitigate differences between ontologies that arise due to different naming convensions (e.g. camelcase versus underscores as dividers) and to produce a version of the entity label that is most likely to generate a relevant search result on Wikipedia. The OWLAPI 1 is first used to extract all of the classes, data properties, object properties, and individuals in an ontology. Custom code then determines the appropriate label for each entity. Labels are extracted based either on the entity s URI (i.e. using the substring after the last #) or, if the URI does not contain a meaningful label, the rdf-schema label attribute. Labels are then tokenized and put into lowercase (e.g. work address becomes work address ). 3.2 Knowledge Gathering Once an appropriate label for each entity is available, it is used to query Wikipedia in order to collect more information about the meaning, context, and use of that label. Wikipedia provides a useful API called the MediaWiki Action API, which we use extensively in Wiki- Matcher. The MediaWiki Action API is a web service which provides access to the contents of any Wikipedia page via an HTTP request

29 The uniqueness of WikiMatcher lies in the variety of information it uses from Wikipedia in order to improve alignment accuracy. In particular, the contents of the relevant Wikipedia article, a snippet of the article s content, and a list of other page terms that are relevant to the article are collected for each label. Specifically, an article is the complete plain text of a desired Wikipedia page. An article snippet is the first three sentences of an article. Page terms include the page title, all redirects (alternative titles that point to the same Wikipedia article), and aliases (terms associated with a page). If any of that content is not available then the results of Wikipedia s search query provides alternative data. The query made to the MediaWiki API is an HTTP request which includes the entity label to query for, as well as parameters that specify the desired data to be gathered from the Wikipedia page. In order to facilitate the replication of the results presented here, or the utilization of this approach by other ontology alignment researchers, the exact queries made to the MediaWiki API in this work are given below: ARTICLE URL = h t t p s : / / en. w i k i p e d i a. org /w/ a p i. php? a c t i o n = 1 query&prop = p a g e p r o p s e x t r a c t s&f o r m a t =xml&e x p l a i n t e x t =& e x s e c t i o n f o r m a t = p l a i n&ppprop = d i s a m b i g u a t i o n&r a w c o n t i n u e=& t i t l e s = + LABEL + & r e d i r e c t s =&maxlag =5 ARTICLE SNIPPET URL = h t t p s : / / en. w i k i p e d i a. org /w/ a p i. php? 2 a c t i o n = query&prop = p a g e p r o p s e x t r a c t s&f o r m a t =xml& e x s e n t e n c e s =2& e x i n t r o=& e x p l a i n t e x t =&e x s e c t i o n f o r m a t = p l a i n &ppprop = d i s a m b i g u a t i o n&r a w c o n t i n u e=& t i t l e s = + LABEL + & 22

30 r e d i r e c t s =&maxlag =5 PAGE TERMS URL = h t t p s : / / en. w i k i p e d i a. org /w/ a p i. php? a c t i o n = 3 query&prop = p a g e p r o p s p a g e t e r m s r e d i r e c t s&f o r m a t =xml& ppprop = d i s a m b i g u a t i o n&wbptterms = a l i a s&r d p r o p = t i t l e & rdnamespace=0& r d l i m i t =max&r a w c o n t i n u e=& t i t l e s = + LABEL + & r e d i r e c t s =&maxlag =5 WIKI SEARCH URL = h t t p s : / / en. w i k i p e d i a. org /w/ a p i. php? a c t i o n 4 = query& l i s t = s e a r c h&f o r m a t =xml&s r s e a r c h = + LABEL + & srnamespace=0& s r i n f o = s u g g e s t i o n&s r p r o p = s n i p p e t%7 C i s f i l e m a t c h&s r l i m i t =10& r a w c o n t i n u e=& r e d i r e c t s =&maxlag =5 The first query (1), collects the complete article data. This is encyclopedic information about the entity label, which provides a broad overview of the concept and its relation to other concepts. Specifically, each Wikipedia article is a comprehensive summary which includes references and titles of related topics. 3 The response from the MediaWiki API is very similar to the response for the snippet query demonstrated in Figure 3.1, however the extract field would contain the entire article. The second query (2) collects the snippet text for the article (the first three sentences of the introductory paragraph), which is meant to provide enough context for users of Wikipedia to get a sense of the article s overall content. An alignment system based on snippet data would reduce the amount of data that it would be necessary to gather for

31 each entity and normalize the information available for each entity, which might result in more directly comparable data related to each entity. An example of the reply sent from Wikipedia in response to a request for the snippet data about the topic femur can be found in Figure 3.1. Figure 3.1: An example of the response from the Wikipedia API after a snippet query (2) is made. The next query (3) retrieves the page title, redirects and aliases of the entity label. This can be thought of as a set of synonyms for the label. Using this set of terms rather than article or snippet content further reduces the amount of text involved and potentially increases the information content of each word, which may improve matching precision. An example of the Wikipedia response to this type of query can be found in Figure 3.2. Each type of page term is separated into its own section of the response and needs to be parsed and combined. These terms contain alternate spellings, common incorrect spellings, 24

32 common alternative names, and synonyms for a page title. Figure 3.2: An example of the response from the Wikipedia API after the pages terms query (3) is made. 25

33 For each query (of the first three types) made to the MediaWiki API, there are four possible responses: found, redirected, ambiguous and missing. We now describe each of these: Found: A page is found with an exact or near-exact match (Wikipedia will perform simple normalization steps 4 ), thus the desired data is returned. Redirected: The given label redirects to an alternate, yet related, page title. In this case the desired data is still returned. Ambiguous: The given label is ambiguous and therefore has multiple possibly-related pages. In this case, no data is returned for the article and snippet queries (1, 2), and WikiMatcher launches a general search query (4) in an attempt to gather some information about the entity label. On the other hand, the page terms query (3) still returns the desired data in this situation since the terms associated with the page are still relevant synonyms. Missing: No page is found for the given label, and therefore no data is returned. In this case a general search query (4) is always attempted. There are only two possible results for the general search query (4) from the MediaWiki API: the response will either contain data or it won t. If the response does contain data, it consists of a list of relevant page names along with a short extract of the part of each page where the query term appears. An example general search query response that contains

34 data is illustrated in Figure 3.3. In this example, the response contains three page results for the label Gut epithelium. Each of these consists of a page name and a short extract of the part of the page where the term Gut epithelium appears. Figure 3.3: An example of the response from the Wikipedia API after the search query (4) is made. If the general search response contains data (as it does in the example for gut epithelium ), it is parsed differently depending on whether the original goal was query (1), (2), or (3). When the original goal was to collect the article (1) or snippet (2) data, the snippet data for all of the pages within the general search results are combined and returned as the desired data. If the original goal was to retrieve the page terms (3), the titles of all of the pages within the general search results (up to a limit of ten) are combined and returned. This is because the the combined snippets are the best approximation available to the arti- 27

35 cle extract and snippet extract, as they will allow a similar type of comparison to be made during the alignment process. Similarly, the page titles correlate to the list of terms gathered from the page term query in the sense that they can be used to compute values for the same type of similarity metrics. 3.3 Alignment Generation Similarity between two entities is calculated differently depending on what knowledge from Wikipedia is being used in the alignment process. The various possibilities are discussed in detail in the appropriate subsection within Chapter 4. Regardless of which similarity metric is used, the resulting value for each entity pair is in the range [0,1], with 0 representing no similarity and 1 representing perfect similarity according to that metric. As mentioned previously, all entities from the first ontology are compared to all entities from the second ontology. As this process unfolds, the best possible match (i.e. the one with the highest similarity) for every entity from both ontologies is tracked. After all pair-wise comparisons have been made, this list of best matches for each entity is filtered to produce the alignment. For every entity, if the confidence value associated with its best match is greater than or equal to a predetermined threshold then the match is added to the alignment. In the event of a tie (i.e. there are multiple matches with the same confidence value) for an entity, none of those matches are included in the alignment. The rationale for this is that any choice made between two matches with equal similarity would be arbitrary, and precision would likely suffer. 28

36 4 Evaluation and Analysis The evaluation procedure for this work was designed to explore the performance of using Wikipedia for ontology alignment, as implemented in WikiMatcher, compared to that of a non-semantic label comparison method. Crucially, this approach is not compared against a full-featured alignment system, because it is meant to result in a component that could be used by any existing system, in particular to develop a set of potential matches that can then be pruned down by more computationally intensive procedures. Rather than simply judging the performance of the best available configuration of the WikiMatcher system, the goal of this section is to evaluate the contribution of each component of the system so that the end result will be an effective ontology matcher (or matcher component) that contains no extraneous elements. In support of this, three different tests were performed in order to evaluate the utility of article data, snippet data, and page terms (i.e. title, redirects, and aliases) to the alignment process. Each of these approaches reduces the volume of information available about each entity while also (theoretically) reducing 29

37 the amount of noise introduced to the matching process Datasets Four different test cases were used to evaluate the effectiveness of WikiMatcher, henceforth known as anatomy, conference, hydro, and geo. Two of the test sets used here are from the Ontology Alignment Evaluation Initiative (OAEI). The first test case, anatomy, is from the OAEI anatomy track, which consists of two bio-medical ontologies: one describing mouse anatomy and the other describing human anatomy. The second test set, conference, is from the OAEI conference track, which consists of seven ontologies describing the organization of a conference. The other two test sets, geology realms and hydro, are real-world ontology matching problems. The geology case contains a pair of ontologies. The first is the Environment Ontology 1 (EnvO) which contains entities describing biomes, environmental features, and environmental material. The second is a compilation of ontology snippets from NASA s Semantic Web for Earth and Environmental Ontology 2 (SWEET) realm ontology collection. SWEET was originally a taxonomy of terms from the Global Change Master Directory that has evolved into a shallow ontology. These ontology snippets describe ocean, land surface, terrestrial hydrosphere, atmosphere, and geosphere, among other geology related topics. For this work they have been merged into a single ontology. The hydro dataset consists of six different ontologies that contain terms from the surface hydrographic domain. The Surface Water Ontology is based upon an analy

38 sis of the National Hydrography Dataset (NHD) of the US Geological Survey [Varanka and Usery 2015]. A subset of the EnvO ontology which relates to water-based environments was considered, i.e. those entities related to the EnvO class hydrographic feature (ENVO ) [Buttigieg et al. 2013]. HY Feature attempts to model hydrographic systems governed by disparate global geographic entities and authorities in a uniform way. For this work, the portion of the model related to the HY SurfaceHydroFeature was used [Dornblut and Atkinson 2014]. The Surface Water Network Ontology was developed by a group of geographers and ontologists during a modeling session in The model contains surface water features and corresponding containing features within the terrain [Sinha et al. 2014]. The HydroGazetteer was developed to support semantic gazetteer functions involving topology. It includes hydrographic surface water entities and spatial relationships between them, with a focus on topological links [Vijayasankaran 2015]. The final ontology in this dataset is the realmhydro module of NASA s SWEET ontology Evaluation Techniques The performance of WikiMatcher on each of the three tests for each of the four test sets has been compared to that of an alignment system based solely on the Levenstein string similarity metric. Levenstein computes the edit distance between two strings, which is the number of insertions, deletions, and substitutions required to transform one string into another. The baseline alignment system compares all entities from one ontology to all entities from the other and keeps the best match for each entity. Any entity for which there 31

39 are multiple best matches is not included in the alignment. The results of the Levenstein alignments are shown in Table 4.1. The metrics used to evaluate WikiMatcher include precision, recall and f-measure. The results tables included in this chapter will not include f-measure and recall scores for both the geo and hydro datasets. This is because there is no verified gold-standard reference alignment for these test cases that can be used to verify the results generated by WikiMatcher. This means there is no false negative value, so recall and f-measure can not be computed. To verify the results of hydro and geo in this work, an individual with no connection to this work was consulted and used to confirm the matches generated. This person was given the entity labels matched along with the relations that could be extracted from within the ontologies to determine if the match was correct or incorrect. An example of the information provided to this individual is presented in Figure 4.1. Figure 4.1: An example of the information given to determine if a match was correct or incorrect. 32

40 Table 4.1: Results of Levenstein alignment Anatomy Conference Hydro Geo True Positives False Positives False Negatives NA NA F-Measure NA NA Precision Recall NA NA 4.1 Articles There are three levels of possible similarity when entire Wikipedia articles are used for comparing two entities. Obviously, if the labels of both entities are identical, they will return the same article if they are used to query Wikipedia. Because this would be a waste of time and bandwidth, entities with syntactically equal labels are declared equivalent with a confidence value of 1.0 without doing any knowledge gathering. If there is only one such equivalent match 3, then this relationship will be included in the final alignment during the alignment generation phase. Even if two labels are not syntactically equal, they may still return the same article when used to query Wikipedia due to the normalization and page redirects the query system employs automatically. In this case there is again no need to compare the article text, since it is obviously identical. For these situations, WikiMatcher assigns a confidence value of 0.9. The exact confidence value chosen is not critical the important point is to represent 3 Intuitively, it seems that exact lexical matches would always be unique since a single ontology will not use the same label for two entities. This is indeed the case when WikiMatcher is employed; however, some alignment systems remove stopwords when preprocessing entity labels, in which case multiple distinct labels may be conflated (e.g. has and might both be converted to ). 33

41 that this type of match involves slightly more uncertainty than one in which the labels of both entities are identical. Because WikiMatcher only keeps the match with the highest confidence value for each entity, this has the effect of preferring string similarity above background knowledge. In practice we have found that this avoids cases in which the system removes a correct match in favor of an incorrect one or does not generate any match for an entity (because multiple potential matches have the same confidence value) even when there is an obvious best answer. In the final case, queries to Wikipedia for the entity labels return two different articles. In this situation we employ a basic word presence based method to compare the articles [Pang et al. 2002]. If the article returned when entity A s label contains the label for entity B within its text and vice versa, a confidence value of 0.8 is assigned to that entity pair. If the text inclusion is only present in one direction (i.e. the article returned for entity A contains the label for entity B within its text, but the article returned for entity B does not mention entity A), then the confidence value is set to 0.7. While our results show that this basic technique performs quite well, we do plan to evaluate the effectiveness of more advanced text comparison methods in the future. The results of the article alignment are shown in Table 4.2. Of particular interest in this table are the results on the anatomy test set. Levenstein has a higher precision, while Wiki- Matcher based on articles has a slightly higher recall. This is a recurring theme: leveraging information available on Wikipedia in the ontology alignment process can uncover matches that have no string similarity. In the case of anatomy, recall+ was computed and found to 34

42 Table 4.2: Results of WikiMatcher article alignment Anatomy Conference Hydro Geo True Positives False Positives False Negatives NA NA F-Measure NA NA Precision Recall NA NA be 0.15 in comparison to Levenstein s recall+ score of Some notable matches made by WikiMatcher when using articles include that of adenohypophysis to anterior lobe of the pituitary gland, midbrain to mesencephalon, and brachiocephalic trunk to innominate artery. A non-inclusive selection of other notable matches made by WikiMatcher using articles that do not have strong syntactic similarity is included in Table 4.3. In general, WikiMatcher is capable of using articles to uncover more non-obvious true positives than a string-based approach, but with a corresponding increase in false positives that keeps F-measure essentially the same between the two methods. This is a very useful result, as later phases in the alignment process can draw on established techniques, such as inconsistency checking and repair [Meilicke et al. 2007] to reduce the number of false positives, but it is difficult to find viable methods to identify more true positives. Of the true positive matches made by the article alignment for the anatomy dataset, 933 came from the entity labels being syntactically equal. Of the non-trivial matches 83 came from entity labels that had an exact page or was redirected to an exact page. No other information returned from Wikipedia resulted in a match being made. 35

43 Another important point that can be gleaned from Table 4.2 is the rather dismal performance of WikiMatcher on the conference test set. WikiMatcher found roughly the same number of correct matches as Levenstein while identifying more than twice as many false positives. Furthermore, WikiMatcher made only one unique match ( to ), whereas Levenstein made four (organization to organisation (three times, for different ontology pairs) and sponsorship to sponzorship). This is empirical support for something other researchers in this area have mentioned previously: there is likely no such thing as a single approach to ontology alignment that performs well for all types of alignment tasks [Cheatham and Hitzler 2013; Eckert et al. 2009]. The one unique match made by Wiki- Matcher came from a finding the exact page in Wikipedia and having the other label redirected to the same page. The general problem on the conference test set for WikiMatcher is the generality of the labels involved words like paper have a much wider variety of meanings than terms like gut epithelium. In the cases of hydro and geo, WikiMatcher had more true positives than Levenstein. However, Levenstein had fewer false positives. All of the matches made by Levenstein for both hydro and geo were exact syntactic matches. Though WikiMatcher allowed for more false positives it was able to find interesting non-trivial matches including floodbank to levee for hydro and red clay to uitsol for geo. 36

44 Table 4.3: Sample matches of WikiMatcher article alignment. profunda femoris artery = deep femoral artery lienal vein = splenic vein kidney cortex = renal cortex obliquus externus abdominis = external oblique muscle white fat = white adipose tissue forebrain = prosencephalon midbrain = mesencephalon synovial joint = diarthrosis triquetral = triangular bone xiphisternum = xiphoid process podzol = spodosol floodbank = levee 4.2 Snippets As with the article case, when snippets are used to augment the information available about an entity, multiple levels of similarity are possible. As was done with articles, if the labels for two entities are identical, they are matched with a confidence of 1.0 without bothering to query Wikipedia. If the labels are different but they return the same snippet text, they are matched with a confidence of 0.99 (in order to avoid confounding exact syntactic matches). The remaining possibility is that two entity labels return different snippet text. In the case of complete articles, the text was compared based on the presence or absence of the potentially-matching entity s label in the article. This is not suitable for snippet comparison the shorter length of the text resulted in an inordinate number of false negatives. Instead, snippets are compared based on a standard bag-of-words string similarity metric. A term vector is generated for each snippet (after removing stopwords 4 ) and the Jaccard

45 Table 4.4: Results of WikiMatcher snippet alignment Anatomy Conference Hydro Geo True Positives False Positives False Negatives NA NA F-Measure NA NA Precision Recall NA NA similarity coefficient is employed to compute the similarity between the two vectors. This value is used as the confidence value of a match between the two entities. Similar to the article configuration, the snippet-based alignment has a better recall than Levenstein for the anatomy dataset. Table 4.5 shows notable non-obvious matches made based on snippets. These matches were all overlooked by the Levenstein based matcher. The recall+ score for anatomy improved to 0.18 from the article alignment which is twice the recall+ score that Levenstein achieved. The snippet-based approach performs worse than Levenstein for the conference dataset in terms of both precision and recall. Wiki- Matcher also performs worse than Levenstein in precision for both hydro and geo. Though, as with the article alignment, for hydro and geo the snippet method has more true positives than Levenstein. Most of these extra matches are non-trivial. Using snippets produces more true positives than using the full text of the Wikipedia articles, but snippets also result in more false positives on all datasets besides hydro. For the anatomy dataset the snippet method allowed for 19 true positive matches, with only 9 false positive matches, to be found using the results from the search-search results. This is an improvement over the article method for the search results. For the geo dataset, more 38

46 Table 4.5: Sample matches of WikiMatcher snippet alignment. cranium = skull auditory tube = eustachian tube glomerular capillary endothelium = endothelium of the glomerular capillary liver sinusoid = hepatic sinusoid profunda brachii artery = superior profunda artery sublingual gland = sublingual salivary gland intercostales = intercostal muscle occipital cortex = occipital lobe ventricular septum = interventricular septum atrial septum = interatrial septum downstream = is downstream to organic material = organic matter false positives than true positives resulted from using the search query. This points to the search query having some potential however, more work is needed to better analyze the data gathered. 4.3 Page Terms The page terms test is similar to the snippet test, however the term vectors are generated from the page terms data returned from Wikipedia rather than from the snippet data. Jaccard similarity coefficient is again used to compute the similarity between the two term vectors, which is then used as the confidence value of the potential match. As with the previous two alignments, using page terms from Wikipedia to augment entity labels results in an alignment with better recall than a Levenstein based approach on the anatomy dataset. Some of the notable non-syntactically similar matches identified through the use of terms are shown in Table 4.7. Once again, the results of WikiMatcher 39

47 Table 4.6: Results of WikiMatcher terms alignment Anatomy Conference Hydro Geo True Positives False Positives False Negatives NA NA F-Measure NA NA Precision Recall NA NA in this configuration perform worse than the Levenstein baseline on the conference dataset with respect to both precision and recall. The results of this test on the anatomy and hydro datasets indicate that basing entity similarity on page terms may be a more accurate approach than using either snippets or full articles in many cases. The primary reason for this seems to be that page terms remain useful even when a particular entity label is not associated with a specific Wikipedia page and a general search must be conducted (Case (4) in the description from Chapter 3). More than 400 pairwise entity comparisons from the anatomy test set devolved to this general search approach. The results from this type of query to Wikipedia contain limited textual data but all possible page titles, which is a major boon for the term-based approach. Matches produced by the term-based approach in these situations were correct ten times as often as they were incorrect for the anatomy test set. This approach also reduced the number of false positives for both the anatomy dataset and the geo dataset. 40

48 Table 4.7: Sample matches of WikiMatcher terms alignment. cerebellar vermis = vermis forebrain = prosencephalon spinal ganglion = dorsal root ganglion spiral organ = organ of corti brachiocephalic trunk = innominate artery bony labyrinth = osseous labyrinth lateral geniculate nucleus = external geniculate body cervical vertebra 2 = c2 vertebra stomach mucosa = gastric mucosa profunda femoris artery = deep femoral artery ocean trench = deep sea trench waterbody = body of water 4.4 When A Wikipedia-based Approach Is Viable Looking at all three tests, the results suggest that WikiMatcher performs significantly better when more pages related to the entities in the ontologies to be aligned are found in Wikipedia (as opposed to a query resulting in a list of potentially matching pages or devolving to a general text search). The utility of any external knowledge source for ontology alignment is constrained by its coverage of the ontologies subject areas. Figure 4.2 shows the degree of coverage within Wikipedia for the anatomy, conference, hydro, and geo datasets. As mentioned previously, when a query is made to Wikipedia for a particular term, there are several possible results. In the best case, an article exists that exactly matches the query term, and that article is returned. The next-best outcome occurs when the query term does not have an article associated with it directly, but the term does exist in Wikipedia as a redirect to another page. In this case that redirect article is returned. Sometimes 41

49 Figure 4.2: Wikipedia page hit counts for the anatomy, conference, hydro, and geo datasets Wikipedia does not have a specific article indexed for a query term. In these ambiguous situations, rather than returning a single article, a list of possibly relevant articles is returned. In the worst case, Wikipedia does not have the query term indexed to any articles. In this case, a text search is performed on all articles, similar to a Google search, and the top articles that mention the query term are returned in a list. It is important to realize that each of these levels generally adds more noise to the query results, which is likely to decrease the accuracy of an alignment system that relies on the queries. It is evident from Figure 4.2 that the query results are much more specific for the anatomy test set than for the other test sets. Nearly 40 percent of queries for anatomy labels return a single article (either exactly or after a redirection), while this is the case for only about 25 percent of conference terms, 27 percent of hydro terms, and 29 percent of geo 42

50 terms. This is likely to be a part of the reason for the large difference in performance of WikiMatcher on the different test sets. When looking at the type of Wikipedia response to queries for entities in potential matches, it is noticeable that results vary considerably based on the type of query response from Wikipedia. In the anatomy test set, many interesting matching came from cases in which either one of the entities resulted in an exact page hit and the other a redirected page hit or both resulted in a redirected page hit. This is because pages in Wikipedia use redirects to apply many names (synonyms) to an article. This allows syntactically unrelated terms to be matched in a highly accurate way. An example is brown fat, which redirects to the page found for brown adipose tissue. This pattern holds for the hydro and geo datasets as well: Interesting matches tended to come from a found-redirected or redirected-redirected case. Examples from geo include podzol to spodosol and red clay to ultisol. An example from hydro includes floodbank to levee. While the type of response to Wikipedia queries is a good predictor of accuracy, the particular information used from Wikipedia is not. Neither precision nor recall varies much regardless of whether the articles, snippets, or page terms from Wikipedia are used for the similarity calculation. This is evident in Figure 4.3 and Figure 4.4. This again suggests that the mere existence of Wikipedia pages for the entity labels within the ontologies being aligned (and the redirection system that sends requests for similar concepts to the same page) is the biggest impact on overall performance, rather than the particular information used from those pages. 43

51 Figure 4.3: Precision scores for each dataset across each of the three tests. Figure 4.4: Recall scores for each dataset across each of the three tests. 44

POMap results for OAEI 2017

POMap results for OAEI 2017 POMap results for OAEI 2017 Amir Laadhar 1, Faiza Ghozzi 2, Imen Megdiche 1, Franck Ravat 1, Olivier Teste 1, and Faiez Gargouri 2 1 Paul Sabatier University, IRIT (CNRS/UMR 5505) 118 Route de Narbonne

More information

WeSeE-Match Results for OEAI 2012

WeSeE-Match Results for OEAI 2012 WeSeE-Match Results for OEAI 2012 Heiko Paulheim Technische Universität Darmstadt paulheim@ke.tu-darmstadt.de Abstract. WeSeE-Match is a simple, element-based ontology matching tool. Its basic technique

More information

HotMatch Results for OEAI 2012

HotMatch Results for OEAI 2012 HotMatch Results for OEAI 2012 Thanh Tung Dang, Alexander Gabriel, Sven Hertling, Philipp Roskosch, Marcel Wlotzka, Jan Ruben Zilke, Frederik Janssen, and Heiko Paulheim Technische Universität Darmstadt

More information

CroLOM: Cross-Lingual Ontology Matching System

CroLOM: Cross-Lingual Ontology Matching System CroLOM: Cross-Lingual Ontology Matching System Results for OAEI 2016 Abderrahmane Khiat LITIO Laboratory, University of Oran1 Ahmed Ben Bella, Oran, Algeria abderrahmane khiat@yahoo.com Abstract. The current

More information

ADOM: Arabic Dataset for Evaluating Arabic and Cross-lingual Ontology Alignment Systems

ADOM: Arabic Dataset for Evaluating Arabic and Cross-lingual Ontology Alignment Systems ADOM: Arabic Dataset for Evaluating Arabic and Cross-lingual Ontology Alignment Systems Abderrahmane Khiat 1, Moussa Benaissa 1, and Ernesto Jiménez-Ruiz 2 1 LITIO Laboratory, University of Oran1 Ahmed

More information

Ontology Matching with CIDER: Evaluation Report for the OAEI 2008

Ontology Matching with CIDER: Evaluation Report for the OAEI 2008 Ontology Matching with CIDER: Evaluation Report for the OAEI 2008 Jorge Gracia, Eduardo Mena IIS Department, University of Zaragoza, Spain {jogracia,emena}@unizar.es Abstract. Ontology matching, the task

More information

YAM++ Results for OAEI 2013

YAM++ Results for OAEI 2013 YAM++ Results for OAEI 2013 DuyHoa Ngo, Zohra Bellahsene University Montpellier 2, LIRMM {duyhoa.ngo, bella}@lirmm.fr Abstract. In this paper, we briefly present the new YAM++ 2013 version and its results

More information

FCA-Map Results for OAEI 2016

FCA-Map Results for OAEI 2016 FCA-Map Results for OAEI 2016 Mengyi Zhao 1 and Songmao Zhang 2 1,2 Institute of Mathematics, Academy of Mathematics and Systems Science, Chinese Academy of Sciences, Beijing, P. R. China 1 myzhao@amss.ac.cn,

More information

Semantic Web. Ontology Alignment. Morteza Amini. Sharif University of Technology Fall 95-96

Semantic Web. Ontology Alignment. Morteza Amini. Sharif University of Technology Fall 95-96 ه عا ی Semantic Web Ontology Alignment Morteza Amini Sharif University of Technology Fall 95-96 Outline The Problem of Ontologies Ontology Heterogeneity Ontology Alignment Overall Process Similarity (Matching)

More information

Semantic Interactive Ontology Matching: Synergistic Combination of Techniques to Improve the Set of Candidate Correspondences

Semantic Interactive Ontology Matching: Synergistic Combination of Techniques to Improve the Set of Candidate Correspondences Semantic Interactive Ontology Matching: Synergistic Combination of Techniques to Improve the Set of Correspondences Jomar da Silva 1, Fernanda Araujo Baião 1, Kate Revoredo 1, and Jérôme Euzenat 2 1 Graduated

More information

INFORMATION RETRIEVAL SYSTEM: CONCEPT AND SCOPE

INFORMATION RETRIEVAL SYSTEM: CONCEPT AND SCOPE 15 : CONCEPT AND SCOPE 15.1 INTRODUCTION Information is communicated or received knowledge concerning a particular fact or circumstance. Retrieval refers to searching through stored information to find

More information

RiMOM Results for OAEI 2008

RiMOM Results for OAEI 2008 RiMOM Results for OAEI 2008 Xiao Zhang 1, Qian Zhong 1, Juanzi Li 1, Jie Tang 1, Guotong Xie 2 and Hanyu Li 2 1 Department of Computer Science and Technology, Tsinghua University, China {zhangxiao,zhongqian,ljz,tangjie}@keg.cs.tsinghua.edu.cn

More information

RiMOM Results for OAEI 2009

RiMOM Results for OAEI 2009 RiMOM Results for OAEI 2009 Xiao Zhang, Qian Zhong, Feng Shi, Juanzi Li and Jie Tang Department of Computer Science and Technology, Tsinghua University, Beijing, China zhangxiao,zhongqian,shifeng,ljz,tangjie@keg.cs.tsinghua.edu.cn

More information

ALIN Results for OAEI 2017

ALIN Results for OAEI 2017 ALIN Results for OAEI 2017 Jomar da Silva 1, Fernanda Araujo Baião 1, and Kate Revoredo 1 Graduated Program in Informatics, Department of Applied Informatics Federal University of the State of Rio de Janeiro

More information

warwick.ac.uk/lib-publications

warwick.ac.uk/lib-publications Original citation: Zhao, Lei, Lim Choi Keung, Sarah Niukyun and Arvanitis, Theodoros N. (2016) A BioPortalbased terminology service for health data interoperability. In: Unifying the Applications and Foundations

More information

Ontology Based Prediction of Difficult Keyword Queries

Ontology Based Prediction of Difficult Keyword Queries Ontology Based Prediction of Difficult Keyword Queries Lubna.C*, Kasim K Pursuing M.Tech (CSE)*, Associate Professor (CSE) MEA Engineering College, Perinthalmanna Kerala, India lubna9990@gmail.com, kasim_mlp@gmail.com

More information

Using AgreementMaker to Align Ontologies for OAEI 2010

Using AgreementMaker to Align Ontologies for OAEI 2010 Using AgreementMaker to Align Ontologies for OAEI 2010 Isabel F. Cruz, Cosmin Stroe, Michele Caci, Federico Caimi, Matteo Palmonari, Flavio Palandri Antonelli, Ulas C. Keles ADVIS Lab, Department of Computer

More information

Proposal for Implementing Linked Open Data on Libraries Catalogue

Proposal for Implementing Linked Open Data on Libraries Catalogue Submitted on: 16.07.2018 Proposal for Implementing Linked Open Data on Libraries Catalogue Esraa Elsayed Abdelaziz Computer Science, Arab Academy for Science and Technology, Alexandria, Egypt. E-mail address:

More information

ServOMap and ServOMap-lt Results for OAEI 2012

ServOMap and ServOMap-lt Results for OAEI 2012 ServOMap and ServOMap-lt Results for OAEI 2012 Mouhamadou Ba 1, Gayo Diallo 1 1 LESIM/ISPED, Univ. Bordeaux Segalen, F-33000, France first.last@isped.u-bordeaux2.fr Abstract. We present the results obtained

More information

A Session-based Ontology Alignment Approach for Aligning Large Ontologies

A Session-based Ontology Alignment Approach for Aligning Large Ontologies Undefined 1 (2009) 1 5 1 IOS Press A Session-based Ontology Alignment Approach for Aligning Large Ontologies Editor(s): Name Surname, University, Country Solicited review(s): Name Surname, University,

More information

The Results of Falcon-AO in the OAEI 2006 Campaign

The Results of Falcon-AO in the OAEI 2006 Campaign The Results of Falcon-AO in the OAEI 2006 Campaign Wei Hu, Gong Cheng, Dongdong Zheng, Xinyu Zhong, and Yuzhong Qu School of Computer Science and Engineering, Southeast University, Nanjing 210096, P. R.

More information

Open Research Online The Open University s repository of research publications and other research outputs

Open Research Online The Open University s repository of research publications and other research outputs Open Research Online The Open University s repository of research publications and other research outputs Social Web Communities Conference or Workshop Item How to cite: Alani, Harith; Staab, Steffen and

More information

BLOOMS on AgreementMaker: results for OAEI 2010

BLOOMS on AgreementMaker: results for OAEI 2010 BLOOMS on AgreementMaker: results for OAEI 2010 Catia Pesquita 1, Cosmin Stroe 2, Isabel F. Cruz 2, Francisco M. Couto 1 1 Faculdade de Ciencias da Universidade de Lisboa, Portugal cpesquita at xldb.di.fc.ul.pt,

More information

The Semantic Planetary Data System

The Semantic Planetary Data System The Semantic Planetary Data System J. Steven Hughes 1, Daniel J. Crichton 1, Sean Kelly 1, and Chris Mattmann 1 1 Jet Propulsion Laboratory 4800 Oak Grove Drive Pasadena, CA 91109 USA {steve.hughes, dan.crichton,

More information

Falcon-AO: Aligning Ontologies with Falcon

Falcon-AO: Aligning Ontologies with Falcon Falcon-AO: Aligning Ontologies with Falcon Ningsheng Jian, Wei Hu, Gong Cheng, Yuzhong Qu Department of Computer Science and Engineering Southeast University Nanjing 210096, P. R. China {nsjian, whu, gcheng,

More information

Building an effective and efficient background knowledge resource to enhance ontology matching

Building an effective and efficient background knowledge resource to enhance ontology matching Building an effective and efficient background knowledge resource to enhance ontology matching Amina Annane 1,2, Zohra Bellahsene 2, Faiçal Azouaou 1, Clement Jonquet 2,3 1 Ecole nationale Supérieure d

More information

A GML SCHEMA MAPPING APPROACH TO OVERCOME SEMANTIC HETEROGENEITY IN GIS

A GML SCHEMA MAPPING APPROACH TO OVERCOME SEMANTIC HETEROGENEITY IN GIS A GML SCHEMA MAPPING APPROACH TO OVERCOME SEMANTIC HETEROGENEITY IN GIS Manoj Paul, S. K. Ghosh School of Information Technology, Indian Institute of Technology, Kharagpur 721302, India - (mpaul, skg)@sit.iitkgp.ernet.in

More information

LogMap: Logic-based and Scalable Ontology Matching

LogMap: Logic-based and Scalable Ontology Matching LogMap: Logic-based and Scalable Ontology Matching Ernesto Jiménez-Ruiz and Bernardo Cuenca Grau Department of Computer Science, University of Oxford {ernesto,berg}@cs.ox.ac.uk Abstract. In this paper,

More information

Lily: Ontology Alignment Results for OAEI 2009

Lily: Ontology Alignment Results for OAEI 2009 Lily: Ontology Alignment Results for OAEI 2009 Peng Wang 1, Baowen Xu 2,3 1 College of Software Engineering, Southeast University, China 2 State Key Laboratory for Novel Software Technology, Nanjing University,

More information

A method for recommending ontology alignment strategies

A method for recommending ontology alignment strategies A method for recommending ontology alignment strategies He Tan and Patrick Lambrix Department of Computer and Information Science Linköpings universitet, Sweden This is a pre-print version of the article

More information

What you have learned so far. Interoperability. Ontology heterogeneity. Being serious about the semantic web

What you have learned so far. Interoperability. Ontology heterogeneity. Being serious about the semantic web What you have learned so far Interoperability Introduction to the Semantic Web Tutorial at ISWC 2010 Jérôme Euzenat Data can be expressed in RDF Linked through URIs Modelled with OWL ontologies & Retrieved

More information

WEIGHTING QUERY TERMS USING WORDNET ONTOLOGY

WEIGHTING QUERY TERMS USING WORDNET ONTOLOGY IJCSNS International Journal of Computer Science and Network Security, VOL.9 No.4, April 2009 349 WEIGHTING QUERY TERMS USING WORDNET ONTOLOGY Mohammed M. Sakre Mohammed M. Kouta Ali M. N. Allam Al Shorouk

More information

A Tagging Approach to Ontology Mapping

A Tagging Approach to Ontology Mapping A Tagging Approach to Ontology Mapping Colm Conroy 1, Declan O'Sullivan 1, Dave Lewis 1 1 Knowledge and Data Engineering Group, Trinity College Dublin {coconroy,declan.osullivan,dave.lewis}@cs.tcd.ie Abstract.

More information

Balanced Large Scale Knowledge Matching Using LSH Forest

Balanced Large Scale Knowledge Matching Using LSH Forest Balanced Large Scale Knowledge Matching Using LSH Forest 1st International KEYSTONE Conference IKC 2015 Coimbra Portugal, 8-9 September 2015 Michael Cochez * Vagan Terziyan * Vadim Ermolayev ** * Industrial

More information

3.4 Data-Centric workflow

3.4 Data-Centric workflow 3.4 Data-Centric workflow One of the most important activities in a S-DWH environment is represented by data integration of different and heterogeneous sources. The process of extract, transform, and load

More information

3 Classifications of ontology matching techniques

3 Classifications of ontology matching techniques 3 Classifications of ontology matching techniques Having defined what the matching problem is, we attempt at classifying the techniques that can be used for solving this problem. The major contributions

More information

MEASURING SEMANTIC SIMILARITY BETWEEN WORDS AND IMPROVING WORD SIMILARITY BY AUGUMENTING PMI

MEASURING SEMANTIC SIMILARITY BETWEEN WORDS AND IMPROVING WORD SIMILARITY BY AUGUMENTING PMI MEASURING SEMANTIC SIMILARITY BETWEEN WORDS AND IMPROVING WORD SIMILARITY BY AUGUMENTING PMI 1 KAMATCHI.M, 2 SUNDARAM.N 1 M.E, CSE, MahaBarathi Engineering College Chinnasalem-606201, 2 Assistant Professor,

More information

A Developer s Guide to the Semantic Web

A Developer s Guide to the Semantic Web A Developer s Guide to the Semantic Web von Liyang Yu 1. Auflage Springer 2011 Verlag C.H. Beck im Internet: www.beck.de ISBN 978 3 642 15969 5 schnell und portofrei erhältlich bei beck-shop.de DIE FACHBUCHHANDLUNG

More information

Papers for comprehensive viva-voce

Papers for comprehensive viva-voce Papers for comprehensive viva-voce Priya Radhakrishnan Advisor : Dr. Vasudeva Varma Search and Information Extraction Lab, International Institute of Information Technology, Gachibowli, Hyderabad, India

More information

ODGOMS - Results for OAEI 2013 *

ODGOMS - Results for OAEI 2013 * ODGOMS - Results for OAEI 2013 * I-Hong Kuo 1,2, Tai-Ting Wu 1,3 1 Industrial Technology Research Institute, Taiwan 2 yihonguo@itri.org.tw 3 taitingwu@itri.org.tw Abstract. ODGOMS is a multi-strategy ontology

More information

NUS-I2R: Learning a Combined System for Entity Linking

NUS-I2R: Learning a Combined System for Entity Linking NUS-I2R: Learning a Combined System for Entity Linking Wei Zhang Yan Chuan Sim Jian Su Chew Lim Tan School of Computing National University of Singapore {z-wei, tancl} @comp.nus.edu.sg Institute for Infocomm

More information

MapPSO Results for OAEI 2010

MapPSO Results for OAEI 2010 MapPSO Results for OAEI 2010 Jürgen Bock 1 FZI Forschungszentrum Informatik, Karlsruhe, Germany bock@fzi.de Abstract. This paper presents and discusses the results produced by the MapPSO system for the

More information

DNA Certification Programs Overview (as of 10 June 2006)

DNA Certification Programs Overview (as of 10 June 2006) DNA Certification Programs Overview (as of 10 June 2006) General Statement: It should be noted that OHi considers each of the DNA Activations as a Sacred Ceremony in addition to being excellent healing

More information

BabelNet: The automatic construction, evaluation and application of a wide-coverage multilingual semantic network

BabelNet: The automatic construction, evaluation and application of a wide-coverage multilingual semantic network BabelNet: The automatic construction, evaluation and application of a wide-coverage multilingual semantic network Roberto Navigli, Simone Paolo Ponzetto What is BabelNet a very large, wide-coverage multilingual

More information

Semantic Interoperability. Being serious about the Semantic Web

Semantic Interoperability. Being serious about the Semantic Web Semantic Interoperability Jérôme Euzenat INRIA & LIG France Natasha Noy Stanford University USA 1 Being serious about the Semantic Web It is not one person s ontology It is not several people s common

More information

2 The IBM Data Governance Unified Process

2 The IBM Data Governance Unified Process 2 The IBM Data Governance Unified Process The benefits of a commitment to a comprehensive enterprise Data Governance initiative are many and varied, and so are the challenges to achieving strong Data Governance.

More information

Theme Identification in RDF Graphs

Theme Identification in RDF Graphs Theme Identification in RDF Graphs Hanane Ouksili PRiSM, Univ. Versailles St Quentin, UMR CNRS 8144, Versailles France hanane.ouksili@prism.uvsq.fr Abstract. An increasing number of RDF datasets is published

More information

YAM++ : A multi-strategy based approach for Ontology matching task

YAM++ : A multi-strategy based approach for Ontology matching task YAM++ : A multi-strategy based approach for Ontology matching task Duy Hoa Ngo, Zohra Bellahsene To cite this version: Duy Hoa Ngo, Zohra Bellahsene. YAM++ : A multi-strategy based approach for Ontology

More information

0.1 Knowledge Organization Systems for Semantic Web

0.1 Knowledge Organization Systems for Semantic Web 0.1 Knowledge Organization Systems for Semantic Web 0.1 Knowledge Organization Systems for Semantic Web 0.1.1 Knowledge Organization Systems Why do we need to organize knowledge? Indexing Retrieval Organization

More information

AML Results for OAEI 2015

AML Results for OAEI 2015 AML Results for OAEI 2015 Daniel Faria 1, Catarina Martins 2, Amruta Nanavaty 4, Daniela Oliveira 2, Booma Sowkarthiga 4, Aynaz Taheri 4, Catia Pesquita 2,3, Francisco M. Couto 2,3, and Isabel F. Cruz

More information

Cluster-based Similarity Aggregation for Ontology Matching

Cluster-based Similarity Aggregation for Ontology Matching Cluster-based Similarity Aggregation for Ontology Matching Quang-Vinh Tran 1, Ryutaro Ichise 2, and Bao-Quoc Ho 1 1 Faculty of Information Technology, Ho Chi Minh University of Science, Vietnam {tqvinh,hbquoc}@fit.hcmus.edu.vn

More information

International Journal of Computer Science Trends and Technology (IJCST) Volume 3 Issue 4, Jul-Aug 2015

International Journal of Computer Science Trends and Technology (IJCST) Volume 3 Issue 4, Jul-Aug 2015 RESEARCH ARTICLE OPEN ACCESS Multi-Lingual Ontology Server (MOS) For Discovering Web Services Abdelrahman Abbas Ibrahim [1], Dr. Nael Salman [2] Department of Software Engineering [1] Sudan University

More information

Anchor-Profiles for Ontology Mapping with Partial Alignments

Anchor-Profiles for Ontology Mapping with Partial Alignments Anchor-Profiles for Ontology Mapping with Partial Alignments Frederik C. Schadd Nico Roos Department of Knowledge Engineering, Maastricht University, Maastricht, The Netherlands Abstract. Ontology mapping

More information

TECHNOLOGY BRIEF: CA ERWIN DATA PROFILER. Combining Data Profiling and Data Modeling for Better Data Quality

TECHNOLOGY BRIEF: CA ERWIN DATA PROFILER. Combining Data Profiling and Data Modeling for Better Data Quality TECHNOLOGY BRIEF: CA ERWIN DATA PROFILER Combining Data Profiling and Data Modeling for Better Data Quality Table of Contents Executive Summary SECTION 1: CHALLENGE 2 Reducing the Cost and Risk of Data

More information

Annotating Spatio-Temporal Information in Documents

Annotating Spatio-Temporal Information in Documents Annotating Spatio-Temporal Information in Documents Jannik Strötgen University of Heidelberg Institute of Computer Science Database Systems Research Group http://dbs.ifi.uni-heidelberg.de stroetgen@uni-hd.de

More information

Chapter 27 Introduction to Information Retrieval and Web Search

Chapter 27 Introduction to Information Retrieval and Web Search Chapter 27 Introduction to Information Retrieval and Web Search Copyright 2011 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 27 Outline Information Retrieval (IR) Concepts Retrieval

More information

Testing the Impact of Pattern-Based Ontology Refactoring on Ontology Matching Results

Testing the Impact of Pattern-Based Ontology Refactoring on Ontology Matching Results Testing the Impact of Pattern-Based Ontology Refactoring on Ontology Matching Results Ondřej Šváb-Zamazal 1, Vojtěch Svátek 1, Christian Meilicke 2, and Heiner Stuckenschmidt 2 1 University of Economics,

More information

Guide for the international tekom certification examinations

Guide for the international tekom certification examinations Guide for the international tekom certification examinations 1 Obtaining the tekom-certificate First Level 1.1 The tekom-certificate First Level at a glance The tekom-certificate First Level is the first

More information

AOT / AOTL Results for OAEI 2014

AOT / AOTL Results for OAEI 2014 AOT / AOTL Results for OAEI 2014 Abderrahmane Khiat 1, Moussa Benaissa 1 1 LITIO Lab, University of Oran, BP 1524 El-Mnaouar Oran, Algeria abderrahmane_khiat@yahoo.com moussabenaissa@yahoo.fr Abstract.

More information

PRIOR System: Results for OAEI 2006

PRIOR System: Results for OAEI 2006 PRIOR System: Results for OAEI 2006 Ming Mao, Yefei Peng University of Pittsburgh, Pittsburgh, PA, USA {mingmao,ypeng}@mail.sis.pitt.edu Abstract. This paper summarizes the results of PRIOR system, which

More information

ResPubliQA 2010

ResPubliQA 2010 SZTAKI @ ResPubliQA 2010 David Mark Nemeskey Computer and Automation Research Institute, Hungarian Academy of Sciences, Budapest, Hungary (SZTAKI) Abstract. This paper summarizes the results of our first

More information

Domain-specific Concept-based Information Retrieval System

Domain-specific Concept-based Information Retrieval System Domain-specific Concept-based Information Retrieval System L. Shen 1, Y. K. Lim 1, H. T. Loh 2 1 Design Technology Institute Ltd, National University of Singapore, Singapore 2 Department of Mechanical

More information

Category Theory in Ontology Research: Concrete Gain from an Abstract Approach

Category Theory in Ontology Research: Concrete Gain from an Abstract Approach Category Theory in Ontology Research: Concrete Gain from an Abstract Approach Markus Krötzsch Pascal Hitzler Marc Ehrig York Sure Institute AIFB, University of Karlsruhe, Germany; {mak,hitzler,ehrig,sure}@aifb.uni-karlsruhe.de

More information

Knowledge Representations. How else can we represent knowledge in addition to formal logic?

Knowledge Representations. How else can we represent knowledge in addition to formal logic? Knowledge Representations How else can we represent knowledge in addition to formal logic? 1 Common Knowledge Representations Formal Logic Production Rules Semantic Nets Schemata and Frames 2 Production

More information

Patent Terminlogy Analysis: Passage Retrieval Experiments for the Intellecutal Property Track at CLEF

Patent Terminlogy Analysis: Passage Retrieval Experiments for the Intellecutal Property Track at CLEF Patent Terminlogy Analysis: Passage Retrieval Experiments for the Intellecutal Property Track at CLEF Julia Jürgens, Sebastian Kastner, Christa Womser-Hacker, and Thomas Mandl University of Hildesheim,

More information

AROMA results for OAEI 2009

AROMA results for OAEI 2009 AROMA results for OAEI 2009 Jérôme David 1 Université Pierre-Mendès-France, Grenoble Laboratoire d Informatique de Grenoble INRIA Rhône-Alpes, Montbonnot Saint-Martin, France Jerome.David-at-inrialpes.fr

More information

Semantic Annotation for Semantic Social Networks. Using Community Resources

Semantic Annotation for Semantic Social Networks. Using Community Resources Semantic Annotation for Semantic Social Networks Using Community Resources Lawrence Reeve and Hyoil Han College of Information Science and Technology Drexel University, Philadelphia, PA 19108 lhr24@drexel.edu

More information

Alignment Results of SOBOM for OAEI 2009

Alignment Results of SOBOM for OAEI 2009 Alignment Results of SBM for AEI 2009 Peigang Xu, Haijun Tao, Tianyi Zang, Yadong, Wang School of Computer Science and Technology Harbin Institute of Technology, Harbin, China xpg0312@hotmail.com, hjtao.hit@gmail.com,

More information

Acquiring Experience with Ontology and Vocabularies

Acquiring Experience with Ontology and Vocabularies Acquiring Experience with Ontology and Vocabularies Walt Melo Risa Mayan Jean Stanford The author's affiliation with The MITRE Corporation is provided for identification purposes only, and is not intended

More information

everyone s minds are the same questions that Credo the Mentor asked in, Then What?:

everyone s minds are the same questions that Credo the Mentor asked in, Then What?: Web 3.0 - The Semantic Web Cometh What Happens When the Read-Write Web Begins to Think? By Jason Ohler 2008 From Info Tease to Info Overload When the web first appeared at my university I used a very early

More information

Semantic Web. Ontology Alignment. Morteza Amini. Sharif University of Technology Fall 94-95

Semantic Web. Ontology Alignment. Morteza Amini. Sharif University of Technology Fall 94-95 ه عا ی Semantic Web Ontology Alignment Morteza Amini Sharif University of Technology Fall 94-95 Outline The Problem of Ontologies Ontology Heterogeneity Ontology Alignment Overall Process Similarity Methods

More information

Lightweight Transformation of Tabular Open Data to RDF

Lightweight Transformation of Tabular Open Data to RDF Proceedings of the I-SEMANTICS 2012 Posters & Demonstrations Track, pp. 38-42, 2012. Copyright 2012 for the individual papers by the papers' authors. Copying permitted only for private and academic purposes.

More information

PORTAL RESOURCES INFORMATION SYSTEM: THE DESIGN AND DEVELOPMENT OF AN ONLINE DATABASE FOR TRACKING WEB RESOURCES.

PORTAL RESOURCES INFORMATION SYSTEM: THE DESIGN AND DEVELOPMENT OF AN ONLINE DATABASE FOR TRACKING WEB RESOURCES. PORTAL RESOURCES INFORMATION SYSTEM: THE DESIGN AND DEVELOPMENT OF AN ONLINE DATABASE FOR TRACKING WEB RESOURCES by Richard Spinks A Master s paper submitted to the faculty of the School of Information

More information

Ontology Matching as Regression Problem

Ontology Matching as Regression Problem Ontology Matching as Regression Problem Nadia Alboukaey, Ammar Joukhadar Faculty of information technology engineering-damascus university Syrian Arab Republic iteng.nadia@gmail.com ajoukhadar@el-ixir.com

More information

ANNUAL REPORT Visit us at project.eu Supported by. Mission

ANNUAL REPORT Visit us at   project.eu Supported by. Mission Mission ANNUAL REPORT 2011 The Web has proved to be an unprecedented success for facilitating the publication, use and exchange of information, at planetary scale, on virtually every topic, and representing

More information

Probabilistic Information Integration and Retrieval in the Semantic Web

Probabilistic Information Integration and Retrieval in the Semantic Web Probabilistic Information Integration and Retrieval in the Semantic Web Livia Predoiu Institute of Computer Science, University of Mannheim, A5,6, 68159 Mannheim, Germany livia@informatik.uni-mannheim.de

More information

Heading-Based Sectional Hierarchy Identification for HTML Documents

Heading-Based Sectional Hierarchy Identification for HTML Documents Heading-Based Sectional Hierarchy Identification for HTML Documents 1 Dept. of Computer Engineering, Boğaziçi University, Bebek, İstanbul, 34342, Turkey F. Canan Pembe 1,2 and Tunga Güngör 1 2 Dept. of

More information

TEXT PREPROCESSING FOR TEXT MINING USING SIDE INFORMATION

TEXT PREPROCESSING FOR TEXT MINING USING SIDE INFORMATION TEXT PREPROCESSING FOR TEXT MINING USING SIDE INFORMATION Ms. Nikita P.Katariya 1, Prof. M. S. Chaudhari 2 1 Dept. of Computer Science & Engg, P.B.C.E., Nagpur, India, nikitakatariya@yahoo.com 2 Dept.

More information

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK REVIEW PAPER ON IMPLEMENTATION OF DOCUMENT ANNOTATION USING CONTENT AND QUERYING

More information

Semantic text features from small world graphs

Semantic text features from small world graphs Semantic text features from small world graphs Jurij Leskovec 1 and John Shawe-Taylor 2 1 Carnegie Mellon University, USA. Jozef Stefan Institute, Slovenia. jure@cs.cmu.edu 2 University of Southampton,UK

More information

Identifying Important Communications

Identifying Important Communications Identifying Important Communications Aaron Jaffey ajaffey@stanford.edu Akifumi Kobashi akobashi@stanford.edu Abstract As we move towards a society increasingly dependent on electronic communication, our

More information

0.1 Upper ontologies and ontology matching

0.1 Upper ontologies and ontology matching 0.1 Upper ontologies and ontology matching 0.1.1 Upper ontologies Basics What are upper ontologies? 0.1 Upper ontologies and ontology matching Upper ontologies (sometimes also called top-level or foundational

More information

Document Retrieval using Predication Similarity

Document Retrieval using Predication Similarity Document Retrieval using Predication Similarity Kalpa Gunaratna 1 Kno.e.sis Center, Wright State University, Dayton, OH 45435 USA kalpa@knoesis.org Abstract. Document retrieval has been an important research

More information

New Approach to Graph Databases

New Approach to Graph Databases Paper PP05 New Approach to Graph Databases Anna Berg, Capish, Malmö, Sweden Henrik Drews, Capish, Malmö, Sweden Catharina Dahlbo, Capish, Malmö, Sweden ABSTRACT Graph databases have, during the past few

More information

A Survey Of Different Text Mining Techniques Varsha C. Pande 1 and Dr. A.S. Khandelwal 2

A Survey Of Different Text Mining Techniques Varsha C. Pande 1 and Dr. A.S. Khandelwal 2 A Survey Of Different Text Mining Techniques Varsha C. Pande 1 and Dr. A.S. Khandelwal 2 1 Department of Electronics & Comp. Sc, RTMNU, Nagpur, India 2 Department of Computer Science, Hislop College, Nagpur,

More information

Ontology Alignment Evaluation Initiative: Six Years of Experience

Ontology Alignment Evaluation Initiative: Six Years of Experience Ontology Alignment Evaluation Initiative: Six Years of Experience Jérôme Euzenat 1, Christian Meilicke 2, Heiner Stuckenschmidt 2, Pavel Shvaiko 3, and Cássia Trojahn 1 1 INRIA & LIG, Grenoble, France

More information

An Evaluation of Geo-Ontology Representation Languages for Supporting Web Retrieval of Geographical Information

An Evaluation of Geo-Ontology Representation Languages for Supporting Web Retrieval of Geographical Information An Evaluation of Geo-Ontology Representation Languages for Supporting Web Retrieval of Geographical Information P. Smart, A.I. Abdelmoty and C.B. Jones School of Computer Science, Cardiff University, Cardiff,

More information

Part I: Data Mining Foundations

Part I: Data Mining Foundations Table of Contents 1. Introduction 1 1.1. What is the World Wide Web? 1 1.2. A Brief History of the Web and the Internet 2 1.3. Web Data Mining 4 1.3.1. What is Data Mining? 6 1.3.2. What is Web Mining?

More information

Rank Preserving Clustering Algorithms for Paths in Social Graphs

Rank Preserving Clustering Algorithms for Paths in Social Graphs University of Waterloo Faculty of Engineering Rank Preserving Clustering Algorithms for Paths in Social Graphs LinkedIn Corporation Mountain View, CA 94043 Prepared by Ziyad Mir ID 20333385 2B Department

More information

Ontology matching using vector space

Ontology matching using vector space University of Wollongong Research Online University of Wollongong in Dubai - Papers University of Wollongong in Dubai 2008 Ontology matching using vector space Zahra Eidoon University of Tehran, Iran Nasser

More information

Context Sensitive Search Engine

Context Sensitive Search Engine Context Sensitive Search Engine Remzi Düzağaç and Olcay Taner Yıldız Abstract In this paper, we use context information extracted from the documents in the collection to improve the performance of the

More information

Technical Brief: Domain Risk Score Proactively uncover threats using DNS and data science

Technical Brief: Domain Risk Score Proactively uncover threats using DNS and data science Technical Brief: Domain Risk Score Proactively uncover threats using DNS and data science 310 Million + Current Domain Names 11 Billion+ Historical Domain Profiles 5 Million+ New Domain Profiles Daily

More information

Information Retrieval

Information Retrieval Information Retrieval CSC 375, Fall 2016 An information retrieval system will tend not to be used whenever it is more painful and troublesome for a customer to have information than for him not to have

More information

Business Rules Extracted from Code

Business Rules Extracted from Code 1530 E. Dundee Rd., Suite 100 Palatine, Illinois 60074 United States of America Technical White Paper Version 2.2 1 Overview The purpose of this white paper is to describe the essential process for extracting

More information

Efficient, Scalable, and Provenance-Aware Management of Linked Data

Efficient, Scalable, and Provenance-Aware Management of Linked Data Efficient, Scalable, and Provenance-Aware Management of Linked Data Marcin Wylot 1 Motivation and objectives of the research The proliferation of heterogeneous Linked Data on the Web requires data management

More information

On the Feasibility of Using OWL 2 DL Reasoners for Ontology Matching Problems

On the Feasibility of Using OWL 2 DL Reasoners for Ontology Matching Problems On the Feasibility of Using OWL 2 DL Reasoners for Ontology Matching Problems Ernesto Jiménez-Ruiz, Bernardo Cuenca Grau, and Ian Horrocks Department of Computer Science, University of Oxford {ernesto,berg,ian.horrocks}@cs.ox.ac.uk

More information

DSSim-ontology mapping with uncertainty

DSSim-ontology mapping with uncertainty DSSim-ontology mapping with uncertainty Miklos Nagy, Maria Vargas-Vera, Enrico Motta Knowledge Media Institute (Kmi) The Open University Walton Hall, Milton Keynes, MK7 6AA, United Kingdom mn2336@student.open.ac.uk;{m.vargas-vera,e.motta}@open.ac.uk

More information

Usability Testing Report of College of Liberal Arts & Sciences (CLAS) Website

Usability Testing Report of College of Liberal Arts & Sciences (CLAS) Website Usability Testing Report of College of Liberal Arts & Sciences (CLAS) Website Submitted to: Ceily Hamilton, Director of Information Technology Alex Chapin, Executive Director of Academic Technology Compiled

More information

Creating Large-scale Training and Test Corpora for Extracting Structured Data from the Web

Creating Large-scale Training and Test Corpora for Extracting Structured Data from the Web Creating Large-scale Training and Test Corpora for Extracting Structured Data from the Web Robert Meusel and Heiko Paulheim University of Mannheim, Germany Data and Web Science Group {robert,heiko}@informatik.uni-mannheim.de

More information

Engagement Portal. Physician Engagement User Guide Press Ganey Associates, Inc.

Engagement Portal. Physician Engagement User Guide Press Ganey Associates, Inc. Engagement Portal Physician Engagement User Guide 2015 Press Ganey Associates, Inc. Contents Logging In... 3 Summary Dashboard... 4 Results For... 5 Filters... 6 Summary Page Engagement Tile... 7 Summary

More information