Time to evaluate: Targeting Annotation Tools

Size: px
Start display at page:

Download "Time to evaluate: Targeting Annotation Tools"

Transcription

1 Time to evaluate: Targeting Annotation Tools Peyman Sazedj 1 and H. Sofia Pinto 1 Instituto Superior Técnico, Departamento de Eng. Informática, Av. Rovisco Pais, Lisboa, Portugal psaz@ist.utl.pt, sofia.pinto@dei.ist.utl.pt Abstract. One of the problems newcomers to the annotation area face is which tool to choose. In order to compare available tools one must evaluate them according to an evaluation framework. In this paper we propose an evaluation framework, that introduces general tool-focused criteria interface and general and annotation-focused criteria - metadata and procedure. We describe how the evaluation of a set of tools was performed and the results of such a comparison. Our results may not only be useful to newcomers, but also to experts who may use them to seize new opportunities for development. 1 Introduction and Motivation The term annotation either designates metadata, or a procedure (that produces metadata). Based on this definition, an annotation tool is a tool that either allows to manipulate annotation metadata, that operates an annotation procedure, or both. Annotation metadata can have a textual, ontological or linguistic nature. Annotations are associated to resources, which can take several forms: web pages, images, video. If we take these dimensions there is already a considerable wealth of tools that fall into the annotation tool category. In our case we are interested in ontological annotation tools that deal with web pages. Moreover, there are several other dimensions that characterize and distinguish annotation tools. Therefore, one of the problems newcomers to the annotation area face is which tool to choose. Although some initial attempts to describe some of the available tools have been made [1], these attempts aimed only to list and describe some functionality issues (languages offered, kind of annotation procedure supported - manual or automatic -, etc.) However, due to the vast array of different features provided by these tools it would be interesting to rank them according to a given set of features. In order to compare available tools one must evaluate them according to an evaluation framework. In this paper we propose an evaluation framework, section 3 that introduces domain specific and domain independent criteria. These criteria range from general tool-focused criteria interface and general to annotationfocused criteria - metadata and procedure. We describe how the evaluation of a set of tools was performed, section 4 and the results of such a comparison, section 5 and take some initial conclusions from our findings, section 6. 37

2 2 Choosing the Tool Set In order to select among existing annotation tools, we chose two simple selection criteria. The first criterion captures the requirement of providing explicit formal meaning to annotations. From among the technologies we have seen around so far, only ontology-based tools assign semantics to their data. Thus, we were only concerned with the evaluation of ontology-based annotation tools. The diversity of resources available on the web, lead us to impose a second criterion. Since the largest part of the world wide web is composed of web pages, and since web pages are the main medium for hosting other types of resources, we were mostly interested in the annotation of web pages. Based on these two criteria, a set of 5 tools was selected: Kim Plugin 1.05 [2] (Sirma Inc.), Melita 2.0 [3] (University of Sheffield), MnM 2.1 [4] (KMI), Ontomat 0.8 [5] (AIFB) and C-Pankow [6] (AIFB). 1 Even though Ontomat is supposed to support automatic annotation, due to technical reasons, we could only operate it in manual mode. Melita and MnM are semi-automatic annotation tools based on the Amilcare Information Extraction Engine [7]. Kim and C-Pankow are fully automatic and unsupervised tools. Kim is a semantic information platform, where only the Kim annotation plugin and its annotations were evaluated. With regard to C-Pankow, we evaluated its web-based implementation 2. 3 Evaluation Framework We propose an evaluation framework for the evaluation of annotation tools, based on a set of well-defined criteria. Our hope is to define each criterion as clearly as possible, encouraging a rigorous and fair evaluation. For that purpose, a metric is associated with each criterion, quantifying its measurement as accurately as possible. Since perfect quantification rather exceeds our scope, our metrics are mostly significant to a relative extent. In other words, the final quantification of a tool can only be meaningful when compared to the quantification of another tool, evaluated by the same framework according to the same set of criteria. Moreover, we try to identify as many criteria as possible, representing the diverse issues dealt with by the area of semantic annotation to a fair extent. We propose a set of 20 criteria which are classified into four dimensions: general, interface, metadata and procedure. We argue that any criterion is either domain-independent ordomain-specific. In turn, domain-independent criteria are divided into interface and general criteria, whereas domain-specific criteria are divided into metadata and procedure criteria. Even though some of our interface criteria may apparently seem to have domain-specific metrics, the criteria themselves are essentially domain independent and their metrics can be adapted to any domain of choice. Based on their metrics, our criteria may be classified into three types: 1 Smore 5.0 was not evaluated due to lack of support of annotation metadata

3 1. Feature-based criteria 2. Set-dependent criteria 3. Set-independent criteria Feature-based criteria are those which verify the existence of specific features or functionalities. Most of our domain-specific criteria are of this type. Setdependent criteria are those which depend on some property inherent to the whole set of tools under evaluation. In other words, the same tool may score differently if evaluated within distinct sets of tools. Set-independent criteria are based on mathematical metrics which are independent of the tool set. In table 1 we present a list of all criteria. Feature-based criteria are indicated with (f), while set-dependent criteria are marked with (d). Criteria with no marks are set-independent. The asterisk indicates that a criterion is only applicable to tools with support for automatic annotation. Interface Metadata Procedure General Self-documentation Association (f) Expressiveness (f) Documentation Simplicity Flexibility (f) Input (f) Scalability (f) Timeliness Heterogeneity (f) Output (f) Stability Usability (d) Integrity (f) Precision* Interoperability (d) Recall* Scope (f) Reliability* Speed* Table 1. All criteria, classified into four dimensions. Since we classified our criteria into one of three types, we present a detailed example of each type within the remainder of this section. Table 2 summarizes our domain-specific criteria. Due to lack of space we omit general and interface criteria tables. Heterogeneity is a feature-based criterion, defined as the quality of being able to combine diverse types of knowledge. The criterion defines the following features: (1) The tool permits the simultaneous loading of multiple Ontologies; (2) The tool supports the annotation of the same unit with multiple concepts. The metric is simply based on assigning a score of one point per feature. Since the criterion defines two features, a tool may score zero, one or two points depending on whether it fulfills none, one or both features. Interoperability is a set-dependent criterion, defined as the ability to exchange and use information among different tools. Its metric is defined as: F r ( + F w ) 1 F r.max F w.max 2 100% Consider a set of tools T.Atoolt T is considered inter-operable with a tool t T where t t, if it can write in a format which t can read and if it can read the formats which t can write. In order to be 100% interoperable with all tools, a tool t needs to be able to read all formats which the tools t T,t t can write and needs to be able to write a subset of formats F, 39

4 Name Short Definition Range Metric Association The way an annotation [0,2] 1 point per feature: is associated with the annotated (1) Tool supports external annotations. (2) Tool supports internal annotations. resource. Flexibility Integrity The quality of being adaptable or variable. The capacity of ensuring the soundness and integrity of an annotation. Scope The scope of an annotation corresponds to the parts of a resource to which the annotation applies. Expressiveness The capacity of fully conveying the meaning of the text through annotations. Input The different types of resources which are supported as input to the annotation procedure. Output The output of the annotation procedure. It is verified whether the tool Precision Recall Reliability Speed writes annotations in a format usable on the Semantic Web. The precision of an algorithm is measured by the certainty of the algorithm in producing correct results. The recall of an algorithm is measured by the percentage of correct solutions found while discarding incorrect solutions. The reliability of the algorithm in producing the expected results. [0,3] 1 point per feature: (1) It is possible to delete previously created annotations. (2) There is a component for editing ontologies. (3) It is possible to define a custom namespace to be used with the annotations. [0,3] 1 point per feature: (1) Tool verifies the domain and target constrains of a relation. (2) Tool certifies the correctness of an association over time. (3) Tool verifies the existence of an association. [0,3] 1 point per feature: (1) Annotate the minimum unit of information. (2) Annotate any multiple of that minimum unit. (3) Annotate the resource as a whole, instead of the multiple of all units. [0,5] 1 point per feature: (1) Tool supports annotation with concepts. (2) Tool supports annotation with relations. (3) Tool supports the instantiation of concepts. (4) Tool supports the instantiation of relations. (5) Tool associates mark-up with instances. [0,3] 1 point per feature: (1) Web page (2)Textdocument (3) Image [0,2] 1 point per feature: (1) Annotations are written in a formalism strictly conforming to a W3C standard (2) Annotations are unambiguously associated with ontological concepts [0,1] [0,1] [0,1] The speed of an annotation algorithm [0,1] is the ratio of the number of annotations produced, by the time it took to produce the results. P = C T C - Number of correct annotations T - Total number of annotations R = C M C - Number of correct annotations M - Ideal number of annotations R = 2 P R P + R P - Precision of the algorithm R - Recall of the algorithm S =1 T T + A A - Number of named entities analyzed for annotation T - Time it took to produce annotations Table 2. Summarized definition of metadata and procedure criteria (annotation-specific criteria), excluding the Heterogeneity and Interoperability criteria which are explained within the main text of this article. 40

5 such that any tool t is able to read at least one of the formats f F. Based on this discussion, F r.max is the cardinality of the set of all formats which can be written by annotation tools different from the one under evaluation. F w.max is the cardinality of a minimum set F of formats such that any tool (other than the one under evaluation) should be able to read one of the formats f F. F r is the number of different annotation formats the tool can read and F w is the number of different formats the tool can write. Simplicity is a set-independent criterion, defined as the quality of being simple, intuitive and easy to understand. Its metric is based on a set of subcriteria where each sub-criterion is assigned a rating of 1 or 0, depending on whether the criterion applies or not: S = L c + E f + C l + C o The learning curve (L c ) designates the ease of learning the tool. If it is assigned a value of 1, then the tool can be learned within a few hours (typically less than 5),otherwise,thetoolismorecomplex and eventually the help of an expert is needed in order to understand some of the provided features. By learning a tool, we mean not only how to produce annotations, but also how to work with most of the features provided. Thus, tools with many features typically have a higher learning curve. By efficiency (E f ) we designate the simplicity of producing annotations, once the tool has been learned. A heuristic for measuring efficiency is whether the tool allows to create a valid annotation, in less operations than most other tools. Even so, additional details have to be taken into account: Some tools provide error-checking while others don t. On the long term, tools that do provide error-checking may be more time saving than tools that don t, even though the latter may create annotations in less steps. We understand that this criterion is slightly subjective as it is. Clearness (C l ) determines how well a novice user gets along with simple tasks. This criterion is different from learning curve, because it is not concerned with understanding all the features and functionalities of the tool. Instead, the question is whether it is easy to locate a specific feature. For example, giving a novice user the task of opening an ontology from disk, the question is whether the interface is clear enough for him to easily perform this task. Finally, consistency (C o ) measures whether the tool is in conformance with actual standards and whether it is consistent with itself. For example, if there are different icons to perform the same operation, the tool is inconsistent. It also has to be taken into account whether other tools that perform similar operations, use the same icons or not. The same is true for the hierarchy and names of menus. 4 Evaluation Procedure Having defined all criteria, the next step was to devise how the evaluation of the selected tools should be carried out. It was clear that not all tools could be evaluated according to all criteria. For example, our precision, recall and reliability criteria are only applicable to semi-automatic and automatic annotation tools 41

6 and cannot be applied to manual annotation tools (Ontomat). MnM, Melita, Kim and C-Pankow were evaluated according to all criteria. A word must be said concerning the applicability of our stability criterion with regard to remote services, since both Kim and C-Pankow were tested as remote services. Let s suppose the invocation of a service sometimes fails. Due to lack of evidence, we cannot draw any conclusion regarding the cause of failure. The problem can either lie at the remote service, or at our internet connection, or elsewhere. Just in case a service never fails, we can assume that it is stable. Although most criteria were of simple application, some criteria required detailed preparation. The usability criterion was measured by means of Jacob Nielsen s heuristic evaluation method [8], with the help of a team of three expert evaluators. All other criteria were measured by a single domain expert. Tools with support for automatic annotation were tested against a set of corpora which were manually tagged. Within the following sections we describe the corpora and the ontologies which we used. We also describe an inter-annotation experiment which was carried out. 4.1 Corpora The first corpus was the Planet Visits corpus, courtesy of the Knowledge Media Institute (KMI), a well known corpus, since it comes included with some annotation tools. We selected a set of 30 web pages, each containing one news article. The particularity of this corpus is that all news are about visits, either by someone from the KMI or by some external entity to the KMI. The documents of this corpus may be considered unstructured text. Our second unstructured corpus was a smaller collection of news articles extracted from the Baha i International Community world news service. 3 The corpus consists of three larger articles, with a total count of 1582 words. Finally a third tabular corpus was created specifically for the evaluation of semi-automatic annotation tools, based on the observation that the performance of semi-automatic tools depends on the structure of documents. Since annotation with semi-automatic tools was not successful on the previous corpora, we created this tabular corpus hoping to obtain useful results. The corpus consists of a single table of 100 rows 4, and two columns, where the annotation task is as simple as classifying the words of the second column, with the corresponding ontological concepts of the first column. Concerning the two sample entries in table 3, Siscog would have to be annotated as an Organization and Germany as a Location. Organization: Siscog Location: Germany Table 3. Small extract of the tabular corpus. The original corpus contains 100 rows The corpus was later divided into two tables of 50 rows each, one for training and the other for testing. 42

7 4.2 Ontologies After choosing the corpora, specific ontologies had to be designed for the annotation of each corpus. This task was a tedious one, due to the apparent lack of interoperability between tools. Among the tools with support for automatized annotation, KIM works with its own meta-ontology, C-Pankow and MnM support ontologies formalized in RDF[9], and Melita works with its own proprietary ontology language. Additionally, Melita only supports annotation with ontological concepts, whereas MnM only supports annotation with ontological relations, therefore we were forced to redesign the same ontology for use with each tool. AsimpleVisits ontology was designed for the annotation of the Planet Visits corpus. The ontology contains 11 concepts and 14 relations such as location, date, duration, organization, visitor, visited, host, etc. For use with the tabular corpus, a very simple Conference ontology was designed. The corpus only contains the names of persons, organizations, locations and dates, therefore we created a simple ontology containing these four concepts. Finally, for use with the Baha i News corpus, the WordNet 5 ontology and Kim s own Kimo 6 ontology were used. 4.3 Inter-annotation experiment It was clear that semi-automatic and automatic annotation tools had to be tested against manually tagged corpora and that the manual tagging of a corpus would not be that obvious at all. To gain more insight, we conducted a simple inter-annotation experiment. Two knowledge engineers were asked to annotate all named entities of a subset of 1433 words of the Planet Visits corpus with the Visits ontology. The annotations of both experts were compared with the help of the kappa-statistic [10, 11] which corrects for chance agreement. Two important results can be extracted from this experiment. An agreement-rate of 0.79 was obtained, when only considering entities annotated by both experts, showing that the classification with ontological concepts was quite well defined. But when also considering terms annotated by only one of the experts (and left blank by the other), we obtained a much lower rate of This leads us to conclude that the task of named entity recognition was quite loosely defined. We also believe that sometimes entities may have been ignored due to lack of relevance. We borrow the notion of relevance [12] to describe this phenomenon. For example, it may be asked how relevant it is to annotate distinct references to the same named entity. For example, if the name Paul Cohen appears several times within the same paragraph or sentence, does it have to be annotated each time? To give yet another example, consider the annotation of a date such as 11/08/2003. Is it relevant to say that 11 is a number, or does it suffice to annotate it as a day? Our experience shows that the answer to this question is

8 apparently a complex one, mostly depending on the purpose of the annotation task and the structure of the ontology. Our main conclusion here is that unless the notion of relevance is clearly defined for the annotation task, the agreement rate can be expected to degrade. 5 Results We start this section by presenting a quantitative evaluation of semi-automatic and automatic annotation procedures, according to the traditional Precision, Recall and F1 measures [12]. Even though these measures are incorporated into the precision, recall and reliability criteria of our evaluation framework, we present them separately due to their important role in the past and for the sake of facilitating comparison with previous evaluations in this field. Please refer to table 2 in section 3 for the definition of these criteria. We then summarize the full scope of our results, according to the four dimensions of the proposed evaluation framework and highlight some of the most important technical drawbacks and features of the evaluated tools. At the end we present a summary of features which may guide the newcomer in his choice. 5.1 The famous Precision, Recall and F1 Measures Much care must be given to the definition of evaluation set-ups, in order to produce reproducible and meaningful results. Table 4 defines the different setups which were used with each tool. Please refer to sections 4.1 and 4.2 for the description of the choice of corpora and ontologies. Corpora / Tools Melita 2.0 MnM 2.1 Kim 1.05 C-Pankow Planet Visits Visits Visits Kimo Wordnet Baha i News - - Kimo Kimo Tabular Conference Conference Conference Conference Table 4. Different evaluation set-ups. For each tool and corpus we indicate the ontology which was used to annotate the corpus. The first corpus to be tested was the Planet Visits corpus. To our disappointment, semi-automatic tools, Melita and MnM, failed entirely in learning to replicate annotations on this corpus. Even though much larger corpora could have produced some results, we concluded that the tools were generally not adequate for unstructured corpora (typical websites). Based on the results of the previous experiments, we attempted a much simpler corpus: the tabular corpus. Annotation of this corpus produced successful results. In order to have equal bases of comparison among all tools, we also tested Kim and C-Pankow on this corpus with the same ontology. Results are presented in table 5. C-Pankow is excluded from this table since it only produced one annotation on this corpus. Table 4 further indicates that automatic tools Kim and C-Pankow were successfully evaluated on both unstructured news corpora. For Kim we obtained an 44

9 Corpora / Tools Melita 2.0 MnM 2.1 Kim 1.05 Precision Recall F Table 5. The traditional Precision, Recall and F1 measures given in percentages for the tabular corpus annotated with the conference ontology F1 score of 59% (P=63.8%, R=54.8%) on the Planet Visits corpus and a score of 64.9% (P=70.6%, R=60%) on the Baha i corpus, whereas C-Pankow scored 16.4% (P=21.2%, R=13.4%) and 32.9% (P=36.4%, R=30%) on those corpora. 5.2 The new evaluation framework According to the proposed evaluation framework, we distinguish among two annotation-specific dimensions: procedure and metadata. Even though the traditional measures discussed in the previous section are an integral part of procedure criteria, we suggest that a holistic view includes other important criteria as well. In conformance with the procedure criteria presented earlier (Table 2 of Sect. 3), summarized results of the procedure dimension as well as all other dimensions will be presented at the end of this section (Fig. 1). We can see that both semi-automatic tools produce similar results, since they are based on the same information extraction engine. With regard to automatic tools, scores in Figure 1 are based on the annotation of the Baha i corpus. When comparing the results of Kim with those of semi-automatic tools, the future of the latter tools is clearly under question taking into account the simplicity of the former. Although C-Pankow s results fall slightly behind, it promises to work well on domain-specific corpora [6] where Kim s upper-ontology is of little use. In the arena of metadata, Ontomat clearly scores best. Interestingly, semiautomatic tools Melita and MnM obtain the same medium scores, whereas automatic tools Kim and C-Pankow fall slightly behind. The equal scores of the semi-automatic tools reflect the fact that they have similar metadata components and that they both author annotations in a similar mark-up based on XML [13]. Even so, both have an interoperability of 0: the XML language lacks formal semantics, therefore the semantics of the annotations are those intended by the authors of the tools. Since both tools follow different naming conventions for the creation of the XML tags (annotations), there is no formal way of establishing that the annotations of one tool are the same as the annotations of the other. For example, considering an ontology which has the concept House as a subclass of Thing, whenever an entity is annotated as a house, Melita annotates the entity with the tag <house>, whereas MnM creates the tag <thouse>. The tools follow different naming conventions and due to the lack of semantics of the XML language, there is no way of establishing that both tags refer to the same concept. We point out that Ontomat s annotations are of high quality, formalized as external annotations in OWL [14]. Additionally, they may be embedded into web pages, simulating internal annotations. Although not an essential requirement 45

10 of an annotation tool, Ontomat s ontology creation component is also extremely handy. With regard to automatic annotation tools, their low scores can be justified by the fact that they are simple interfaces to remote annotation services, and as such, they don t have a component for editing annotations or ontologies. Low scores in this dimension are mostly due to features that could be present in the tools but are not present in their current state. While such scores might be discouraging to newcomers, experts and researchers may interpret them as new opportunities for development. Kim distinguishes itself as the tool which scores best in the scope criterion. This is due to the fact that it can mark-up all occurrences of the same entity, as belonging to the same instance of the ontology. Likewise, tools were evaluated according to the domain-independent dimensions (Fig. 1). The final evaluation scores of each dimension are given by the arithmetic mean of the normalized scores of its criteria. We recall that the scores have no absolute meaning, but rather a relative one. Therefore it is only meaningful to compare tools with similar functionality. Concerning full-featured manual tools, the ideal tool should aim to combine the simplicity of Melita s interface, with the annotation-related features of Ontomat. Interestingly all tools scored medium with respect to our general criteria. Melita has a very good documentation, but scores low in scalability. On the contrary, Ontomat is scalable but has very little documentation (only a simple tutorial is available). Also, it has several stability issues, which is not surprising since the tool is officially marked as being in alpha development stage. MnM, on the other hand, scores medium in all general criteria Interface Metadata Procedure (manual) Procedure (automatic) General Melita MnM Ontomat Kim C-Pankow Fig. 1. Comparison of annotation tools. Concerning the speed of annotation procedures, it was not possible to apply our speed criterion due to the fact that we had no way of finding out which named entities each tool analyzed for annotation. Therefore we have no real indicator for the speed of the tools. Even so, we can affirm that semi-automatic procedures only took several seconds of processing. Kim usually took 4 seconds per page (of approximately 150 words), whereas C-Pankow took several minutes for the same page. 46

11 Finally, a brief reference for newcomers is presented in Table 6. Due to lack of space, only some of the most important aspects of our framework are included. With regard to a feature, a classification of ++ indicates excellence, a single + indicates the feature exists and is fine, +- stands for medium to poor and - stands for absent. In case a feature is not applicable to a given tool, this is indicated with na. 6 Conclusion and Future Work In this paper we present an evaluation framework for ontological annotation tools. A set of 5 tools were evaluated using this framework. One of the main problems faced by all annotation tools is lack of interoperability. Given the fact that probably one tool is not enough to perform perfect annotation of the web and that probably one should aim at a combined multi-tool approach this is a severe problem. Regarding manual annotation tools our main conclusion is that they do not provide a good enough compromise: some are too focused on the interface, while others focus too much on the metadata. They lack a good compromise on both issues. Regarding semi-automatic tools the main conclusion is they only work with structured corpora, which is not the most common case in the current web. Finally, regarding automatic annotation tools current results are rather impressive, but it is difficult to assess how they could scale. For instance it is difficult to foresee how Kim would perform in domain specific webpages. Therefore, to our understanding, they still have a long way to go before they can be fully used to annotate the current web. The long term goal of the work reported in this paper is to contribute to the automatic annotation field. Our plans include trying to improve one of current tools. Criteria / Tools Melita 2.0 MnM 2.1 Ontomat 0.8 Kim 1.05 C-Pankow Documentation Simplicity of the tool Support for Internal Annotations Support for External Annotations Ontology Editor Load multiple ontologies Annotate with multiple concepts Constraint checking na na Interoperability Annotation with concepts Annotation with relations Ontology Population Manual Annotation Semi-Automatic Annotation Automatic Annotation Table 6. Brief summary of features which may be interesting to newcomers. 47

12 Acknowledgments Our special thanks go to Bruno Grilo and Rudi Araujo (inter-annotation experiment) and to Claudio Gil and Vitor Oliveira (expert usability analysis). We extend our gratitude to Philipp Cimiano for his constant availability and help with C-Pankow. We would also like to thank Aditya Kalyanpur, Borislav Popov, Enrico Motta, Fabio Ciravegna, Jeff Heflin and Siegfried Handschuh for their replies. References 1. A. Gómez-Pérez, M. Fernandéz-López, and O. Corcho, Ontological Engineering., Springer-Verlag London Limited, B. Popov, et al, Kim - semantic annotation platform., in ISWC, pp , F. Ciravegna, et al, User-system cooperation in document annotation based on information extraction., in EKAW, pp , M. Vargas-Vera, et al, Mnm: Ontology driven semi-automatic and automatic support for semantic markup., in EKAW, pp , S. Handschuh, S. Staab, and F. Ciravegna, S-cream - semi-automatic creation of metadata., in EKAW, pp , P. Cimiano, G. Ladwig, and S. Staab, Gimme the context: context-driven automatic semantic annotation with c-pankow., in WWW, pp , F. Ciravegna, et al, Amilcare: adaptive information extraction for document annotation., in SIGIR, pp , J. Nielsen, Finding usability problems through heuristic evaluation., in CHI, pp , Rdf specification, J. Cohen, A coecient of agreement for nominal scales. Educational and Psychological measurements, volume 20, no. 1, pp , J. Carletta, Assessing agreement on classification tasks: The kappa statistic. Computational Linguistics, volume 22, no. 2, pp , C. J. van Rijsbergen, Information Retrieval., Butterworth, Xml specification, Owl specification, 48

Requirements for Information Extraction for Knowledge Management

Requirements for Information Extraction for Knowledge Management Requirements for Information Extraction for Knowledge Management Philipp Cimiano*, Fabio Ciravegna, John Domingue, Siegfried Handschuh*, Alberto Lavelli +, Steffen Staab*, Mark Stevenson AIFB, University

More information

Annotation for the Semantic Web During Website Development

Annotation for the Semantic Web During Website Development Annotation for the Semantic Web During Website Development Peter Plessers and Olga De Troyer Vrije Universiteit Brussel, Department of Computer Science, WISE, Pleinlaan 2, 1050 Brussel, Belgium {Peter.Plessers,

More information

GoNTogle: A Tool for Semantic Annotation and Search

GoNTogle: A Tool for Semantic Annotation and Search GoNTogle: A Tool for Semantic Annotation and Search Giorgos Giannopoulos 1,2, Nikos Bikakis 1,2, Theodore Dalamagas 2, and Timos Sellis 1,2 1 KDBSL Lab, School of ECE, Nat. Tech. Univ. of Athens, Greece

More information

A tutorial report for SENG Agent Based Software Engineering. Course Instructor: Dr. Behrouz H. Far. XML Tutorial.

A tutorial report for SENG Agent Based Software Engineering. Course Instructor: Dr. Behrouz H. Far. XML Tutorial. A tutorial report for SENG 609.22 Agent Based Software Engineering Course Instructor: Dr. Behrouz H. Far XML Tutorial Yanan Zhang Department of Electrical and Computer Engineering University of Calgary

More information

An Architecture for Semantic Enterprise Application Integration Standards

An Architecture for Semantic Enterprise Application Integration Standards An Architecture for Semantic Enterprise Application Integration Standards Nenad Anicic 1, 2, Nenad Ivezic 1, Albert Jones 1 1 National Institute of Standards and Technology, 100 Bureau Drive Gaithersburg,

More information

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY A PATH FOR HORIZING YOUR INNOVATIVE WORK REVIEW PAPER ON IMPLEMENTATION OF DOCUMENT ANNOTATION USING CONTENT AND QUERYING

More information

In this project, I examined methods to classify a corpus of s by their content in order to suggest text blocks for semi-automatic replies.

In this project, I examined methods to classify a corpus of  s by their content in order to suggest text blocks for semi-automatic replies. December 13, 2006 IS256: Applied Natural Language Processing Final Project Email classification for semi-automated reply generation HANNES HESSE mail 2056 Emerson Street Berkeley, CA 94703 phone 1 (510)

More information

ResPubliQA 2010

ResPubliQA 2010 SZTAKI @ ResPubliQA 2010 David Mark Nemeskey Computer and Automation Research Institute, Hungarian Academy of Sciences, Budapest, Hungary (SZTAKI) Abstract. This paper summarizes the results of our first

More information

Benchmarking Textual Annotation Tools for the Semantic Web

Benchmarking Textual Annotation Tools for the Semantic Web Benchmarking Textual Annotation Tools for the Semantic Web Diana Maynard Dept of Computer Science, University of Sheffield, Sheffield, UK {diana}@dcs.shef.ac.uk Abstract This paper investigates the state

More information

Formal Verification for safety critical requirements From Unit-Test to HIL

Formal Verification for safety critical requirements From Unit-Test to HIL Formal Verification for safety critical requirements From Unit-Test to HIL Markus Gros Director Product Sales Europe & North America BTC Embedded Systems AG Berlin, Germany markus.gros@btc-es.de Hans Jürgen

More information

Annotation Science From Theory to Practice and Use Introduction A bit of history

Annotation Science From Theory to Practice and Use Introduction A bit of history Annotation Science From Theory to Practice and Use Nancy Ide Department of Computer Science Vassar College Poughkeepsie, New York 12604 USA ide@cs.vassar.edu Introduction Linguistically-annotated corpora

More information

Extracting knowledge from Ontology using Jena for Semantic Web

Extracting knowledge from Ontology using Jena for Semantic Web Extracting knowledge from Ontology using Jena for Semantic Web Ayesha Ameen I.T Department Deccan College of Engineering and Technology Hyderabad A.P, India ameenayesha@gmail.com Khaleel Ur Rahman Khan

More information

Using idocument for Document Categorization in Nepomuk Social Semantic Desktop

Using idocument for Document Categorization in Nepomuk Social Semantic Desktop Using idocument for Document Categorization in Nepomuk Social Semantic Desktop Benjamin Adrian, Martin Klinkigt,2 Heiko Maus, Andreas Dengel,2 ( Knowledge-Based Systems Group, Department of Computer Science

More information

RiMOM Results for OAEI 2009

RiMOM Results for OAEI 2009 RiMOM Results for OAEI 2009 Xiao Zhang, Qian Zhong, Feng Shi, Juanzi Li and Jie Tang Department of Computer Science and Technology, Tsinghua University, Beijing, China zhangxiao,zhongqian,shifeng,ljz,tangjie@keg.cs.tsinghua.edu.cn

More information

7. Relational Calculus (Part I) 7.1 Introduction

7. Relational Calculus (Part I) 7.1 Introduction 7. Relational Calculus (Part I) 7.1 Introduction We established earlier the fundamental role of relational algebra and calculus in relational databases (see 5.1). More specifically, relational calculus

More information

Ontology Extraction from Tables on the Web

Ontology Extraction from Tables on the Web Ontology Extraction from Tables on the Web Masahiro Tanaka and Toru Ishida Department of Social Informatics, Kyoto University. Kyoto 606-8501, JAPAN mtanaka@kuis.kyoto-u.ac.jp, ishida@i.kyoto-u.ac.jp Abstract

More information

- What we actually mean by documents (the FRBR hierarchy) - What are the components of documents

- What we actually mean by documents (the FRBR hierarchy) - What are the components of documents Purpose of these slides Introduction to XML for parliamentary documents (and all other kinds of documents, actually) Prof. Fabio Vitali University of Bologna Part 1 Introduce the principal aspects of electronic

More information

Running Head: APPLIED KNOWLEDGE MANAGEMENT. MetaTech Consulting, Inc. White Paper

Running Head: APPLIED KNOWLEDGE MANAGEMENT. MetaTech Consulting, Inc. White Paper Applied Knowledge Management 1 Running Head: APPLIED KNOWLEDGE MANAGEMENT MetaTech Consulting, Inc. White Paper Application of Knowledge Management Constructs to the Massive Data Problem Jim Thomas July

More information

Usability of Annotation Tools: comparison and outlook

Usability of Annotation Tools: comparison and outlook Usability of Annotation Tools: comparison and outlook Diana Maynard 1, Martin Dzbor 2, and Siegfried Handschuh 3 1 Department of Computer Science, University of Sheffield Regent Court, 211 Portobello Street,

More information

Executing Evaluations over Semantic Technologies using the SEALS Platform

Executing Evaluations over Semantic Technologies using the SEALS Platform Executing Evaluations over Semantic Technologies using the SEALS Platform Miguel Esteban-Gutiérrez, Raúl García-Castro, Asunción Gómez-Pérez Ontology Engineering Group, Departamento de Inteligencia Artificial.

More information

Creating a Classifier for a Focused Web Crawler

Creating a Classifier for a Focused Web Crawler Creating a Classifier for a Focused Web Crawler Nathan Moeller December 16, 2015 1 Abstract With the increasing size of the web, it can be hard to find high quality content with traditional search engines.

More information

A Study on Website Quality Models

A Study on Website Quality Models International Journal of Scientific and Research Publications, Volume 4, Issue 12, December 2014 1 A Study on Website Quality Models R.Anusha Department of Information Systems Management, M.O.P Vaishnav

More information

Handout 9: Imperative Programs and State

Handout 9: Imperative Programs and State 06-02552 Princ. of Progr. Languages (and Extended ) The University of Birmingham Spring Semester 2016-17 School of Computer Science c Uday Reddy2016-17 Handout 9: Imperative Programs and State Imperative

More information

6.2 DATA DISTRIBUTION AND EXPERIMENT DETAILS

6.2 DATA DISTRIBUTION AND EXPERIMENT DETAILS Chapter 6 Indexing Results 6. INTRODUCTION The generation of inverted indexes for text databases is a computationally intensive process that requires the exclusive use of processing resources for long

More information

Semantic Web in a Constrained Environment

Semantic Web in a Constrained Environment Semantic Web in a Constrained Environment Laurens Rietveld and Stefan Schlobach Department of Computer Science, VU University Amsterdam, The Netherlands {laurens.rietveld,k.s.schlobach}@vu.nl Abstract.

More information

NeOn Methodology for Building Ontology Networks: a Scenario-based Methodology

NeOn Methodology for Building Ontology Networks: a Scenario-based Methodology NeOn Methodology for Building Ontology Networks: a Scenario-based Methodology Asunción Gómez-Pérez and Mari Carmen Suárez-Figueroa Ontology Engineering Group. Departamento de Inteligencia Artificial. Facultad

More information

CPSC 444 Project Milestone III: Prototyping & Experiment Design Feb 6, 2018

CPSC 444 Project Milestone III: Prototyping & Experiment Design Feb 6, 2018 CPSC 444 Project Milestone III: Prototyping & Experiment Design Feb 6, 2018 OVERVIEW... 2 SUMMARY OF MILESTONE III DELIVERABLES... 2 1. Blog Update #3 - Low-fidelity Prototyping & Cognitive Walkthrough,

More information

Designing a System Engineering Environment in a structured way

Designing a System Engineering Environment in a structured way Designing a System Engineering Environment in a structured way Anna Todino Ivo Viglietti Bruno Tranchero Leonardo-Finmeccanica Aircraft Division Torino, Italy Copyright held by the authors. Rubén de Juan

More information

OWL-CM : OWL Combining Matcher based on Belief Functions Theory

OWL-CM : OWL Combining Matcher based on Belief Functions Theory OWL-CM : OWL Combining Matcher based on Belief Functions Theory Boutheina Ben Yaghlane 1 and Najoua Laamari 2 1 LARODEC, Université de Tunis, IHEC Carthage Présidence 2016 Tunisia boutheina.yaghlane@ihec.rnu.tn

More information

Motivating Ontology-Driven Information Extraction

Motivating Ontology-Driven Information Extraction Motivating Ontology-Driven Information Extraction Burcu Yildiz 1 and Silvia Miksch 1, 2 1 Institute for Software Engineering and Interactive Systems, Vienna University of Technology, Vienna, Austria {yildiz,silvia}@

More information

Ontology Extraction from Heterogeneous Documents

Ontology Extraction from Heterogeneous Documents Vol.3, Issue.2, March-April. 2013 pp-985-989 ISSN: 2249-6645 Ontology Extraction from Heterogeneous Documents Kirankumar Kataraki, 1 Sumana M 2 1 IV sem M.Tech/ Department of Information Science & Engg

More information

Shrey Patel B.E. Computer Engineering, Gujarat Technological University, Ahmedabad, Gujarat, India

Shrey Patel B.E. Computer Engineering, Gujarat Technological University, Ahmedabad, Gujarat, India International Journal of Scientific Research in Computer Science, Engineering and Information Technology 2018 IJSRCSEIT Volume 3 Issue 3 ISSN : 2456-3307 Some Issues in Application of NLP to Intelligent

More information

Survey of Semantic Annotation Platforms

Survey of Semantic Annotation Platforms Survey of Semantic Annotation Platforms Lawrence Reeve College of Information Science and Technology Drexel University Philadelphia, PA 19104 USA larry.reeve@drexel.edu Hyoil Han College of Information

More information

SEMANTIC SOLUTIONS FOR OIL & GAS: ROLES AND RESPONSIBILITIES

SEMANTIC SOLUTIONS FOR OIL & GAS: ROLES AND RESPONSIBILITIES SEMANTIC SOLUTIONS FOR OIL & GAS: ROLES AND RESPONSIBILITIES Jeremy Carroll, Ralph Hodgson, {jeremy,ralph}@topquadrant.com This paper is submitted to The W3C Workshop on Semantic Web in Energy Industries

More information

Towards Semantic Data Mining

Towards Semantic Data Mining Towards Semantic Data Mining Haishan Liu Department of Computer and Information Science, University of Oregon, Eugene, OR, 97401, USA ahoyleo@cs.uoregon.edu Abstract. Incorporating domain knowledge is

More information

Web Semantic Annotation Using Data-Extraction Ontologies

Web Semantic Annotation Using Data-Extraction Ontologies Web Semantic Annotation Using Data-Extraction Ontologies A Dissertation Proposal Presented to the Department of Computer Science Brigham Young University In Partial Fulfillment of the Requirements for

More information

Making Ontology Documentation with LODE

Making Ontology Documentation with LODE Proceedings of the I-SEMANTICS 2012 Posters & Demonstrations Track, pp. 63-67, 2012. Copyright 2012 for the individual papers by the papers' authors. Copying permitted only for private and academic purposes.

More information

XETA: extensible metadata System

XETA: extensible metadata System XETA: extensible metadata System Abstract: This paper presents an extensible metadata system (XETA System) which makes it possible for the user to organize and extend the structure of metadata. We discuss

More information

Study of Design Issues on an Automated Semantic Annotation System

Study of Design Issues on an Automated Semantic Annotation System Abstract Study of Design Issues on an Automated Semantic Annotation System Yihong Ding Departments of Computer Science Brigham Young University ding@cs.byu.edu Reference to this article should be made

More information

Conceptual document indexing using a large scale semantic dictionary providing a concept hierarchy

Conceptual document indexing using a large scale semantic dictionary providing a concept hierarchy Conceptual document indexing using a large scale semantic dictionary providing a concept hierarchy Martin Rajman, Pierre Andrews, María del Mar Pérez Almenta, and Florian Seydoux Artificial Intelligence

More information

A Top-Down Visual Approach to GUI development

A Top-Down Visual Approach to GUI development A Top-Down Visual Approach to GUI development ROSANNA CASSINO, GENNY TORTORA, MAURIZIO TUCCI, GIULIANA VITIELLO Dipartimento di Matematica e Informatica Università di Salerno Via Ponte don Melillo 84084

More information

A Framework for Source Code metrics

A Framework for Source Code metrics A Framework for Source Code metrics Neli Maneva, Nikolay Grozev, Delyan Lilov Abstract: The paper presents our approach to the systematic and tool-supported source code measurement for quality analysis.

More information

Ontology Merging: on the confluence between theoretical and pragmatic approaches

Ontology Merging: on the confluence between theoretical and pragmatic approaches Ontology Merging: on the confluence between theoretical and pragmatic approaches Raphael Cóbe, Renata Wassermann, Fabio Kon 1 Department of Computer Science University of São Paulo (IME-USP) {rmcobe,renata,fabio.kon}@ime.usp.br

More information

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION 6.1 INTRODUCTION Fuzzy logic based computational techniques are becoming increasingly important in the medical image analysis arena. The significant

More information

A QUERY BY EXAMPLE APPROACH FOR XML QUERYING

A QUERY BY EXAMPLE APPROACH FOR XML QUERYING A QUERY BY EXAMPLE APPROACH FOR XML QUERYING Flávio Xavier Ferreira, Daniela da Cruz, Pedro Rangel Henriques Department of Informatics, University of Minho Campus de Gualtar, 4710-057 Braga, Portugal flavioxavier@di.uminho.pt,

More information

Applying the Semantic Web Layers to Access Control

Applying the Semantic Web Layers to Access Control J. Lopez, A. Mana, J. maria troya, and M. Yague, Applying the Semantic Web Layers to Access Control, IEEE International Workshop on Web Semantics (WebS03), pp. 622-626, 2003. NICS Lab. Publications: https://www.nics.uma.es/publications

More information

Retrieval Evaluation. Hongning Wang

Retrieval Evaluation. Hongning Wang Retrieval Evaluation Hongning Wang CS@UVa What we have learned so far Indexed corpus Crawler Ranking procedure Research attention Doc Analyzer Doc Rep (Index) Query Rep Feedback (Query) Evaluation User

More information

Dynamic Ontology Evolution

Dynamic Ontology Evolution Dynamic Evolution Fouad Zablith Knowledge Media Institute (KMi), The Open University. Walton Hall, Milton Keynes, MK7 6AA, United Kingdom. f.zablith@open.ac.uk Abstract. Ontologies form the core of Semantic

More information

Mymory: Enhancing a Semantic Wiki with Context Annotations

Mymory: Enhancing a Semantic Wiki with Context Annotations Mymory: Enhancing a Semantic Wiki with Context Annotations Malte Kiesel, Sven Schwarz, Ludger van Elst, and Georg Buscher Knowledge Management Department German Research Center for Artificial Intelligence

More information

DITA for Enterprise Business Documents Sub-committee Proposal Background Why an Enterprise Business Documents Sub committee

DITA for Enterprise Business Documents Sub-committee Proposal Background Why an Enterprise Business Documents Sub committee DITA for Enterprise Business Documents Sub-committee Proposal Background Why an Enterprise Business Documents Sub committee Documents initiate and record business change. It is easy to map some business

More information

Web Design for the Semantic Web

Web Design for the Semantic Web Web Design for the Semantic Web Peter Plessers *, Olga De Troyer Vrije Universiteit Brussel, Department of Computer Science, WISE, Pleinlaan 2, 1050 Brussel, Belgium {Peter.Plessers, Olga.DeTroyer}@vub.ac.be

More information

Harmonization of usability measurements in ISO9126 software engineering standards

Harmonization of usability measurements in ISO9126 software engineering standards Harmonization of usability measurements in ISO9126 software engineering standards Laila Cheikhi, Alain Abran and Witold Suryn École de Technologie Supérieure, 1100 Notre-Dame Ouest, Montréal, Canada laila.cheikhi.1@ens.etsmtl.ca,

More information

Categorizing Migrations

Categorizing Migrations What to Migrate? Categorizing Migrations A version control repository contains two distinct types of data. The first type of data is the actual content of the directories and files themselves which are

More information

Semantic Exploitation of Engineering Models: An Application to Oilfield Models

Semantic Exploitation of Engineering Models: An Application to Oilfield Models Semantic Exploitation of Engineering Models: An Application to Oilfield Models Laura Silveira Mastella 1,YamineAït-Ameur 2,Stéphane Jean 2, Michel Perrin 1, and Jean-François Rainaud 3 1 Ecole des Mines

More information

Conformance Requirements Guideline Version 0.1

Conformance Requirements Guideline Version 0.1 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 Editors: Conformance Requirements Guideline Version 0.1 Aug 22, 2001 Lynne Rosenthal (lynne.rosenthal@nist.gov)

More information

3.4 Data-Centric workflow

3.4 Data-Centric workflow 3.4 Data-Centric workflow One of the most important activities in a S-DWH environment is represented by data integration of different and heterogeneous sources. The process of extract, transform, and load

More information

2 Data Reduction Techniques The granularity of reducible information is one of the main criteria for classifying the reduction techniques. While the t

2 Data Reduction Techniques The granularity of reducible information is one of the main criteria for classifying the reduction techniques. While the t Data Reduction - an Adaptation Technique for Mobile Environments A. Heuer, A. Lubinski Computer Science Dept., University of Rostock, Germany Keywords. Reduction. Mobile Database Systems, Data Abstract.

More information

METEOR-S Web service Annotation Framework with Machine Learning Classification

METEOR-S Web service Annotation Framework with Machine Learning Classification METEOR-S Web service Annotation Framework with Machine Learning Classification Nicole Oldham, Christopher Thomas, Amit Sheth, Kunal Verma LSDIS Lab, Department of CS, University of Georgia, 415 GSRC, Athens,

More information

Ontology Matching with CIDER: Evaluation Report for the OAEI 2008

Ontology Matching with CIDER: Evaluation Report for the OAEI 2008 Ontology Matching with CIDER: Evaluation Report for the OAEI 2008 Jorge Gracia, Eduardo Mena IIS Department, University of Zaragoza, Spain {jogracia,emena}@unizar.es Abstract. Ontology matching, the task

More information

Development of an Ontology-Based Portal for Digital Archive Services

Development of an Ontology-Based Portal for Digital Archive Services Development of an Ontology-Based Portal for Digital Archive Services Ching-Long Yeh Department of Computer Science and Engineering Tatung University 40 Chungshan N. Rd. 3rd Sec. Taipei, 104, Taiwan chingyeh@cse.ttu.edu.tw

More information

Graphical Notation for Topic Maps (GTM)

Graphical Notation for Topic Maps (GTM) Graphical Notation for Topic Maps (GTM) 2005.11.12 Jaeho Lee University of Seoul jaeho@uos.ac.kr 1 Outline 2 Motivation Requirements for GTM Goals, Scope, Constraints, and Issues Survey on existing approaches

More information

Software Language Engineering of Architectural Viewpoints

Software Language Engineering of Architectural Viewpoints Software Language Engineering of Architectural Viewpoints Elif Demirli and Bedir Tekinerdogan Department of Computer Engineering, Bilkent University, Ankara 06800, Turkey {demirli,bedir}@cs.bilkent.edu.tr

More information

Chapter 2 Overview of the Design Methodology

Chapter 2 Overview of the Design Methodology Chapter 2 Overview of the Design Methodology This chapter presents an overview of the design methodology which is developed in this thesis, by identifying global abstraction levels at which a distributed

More information

Data Curation Profile Human Genomics

Data Curation Profile Human Genomics Data Curation Profile Human Genomics Profile Author Profile Author Institution Name Contact J. Carlson N. Brown Purdue University J. Carlson, jrcarlso@purdue.edu Date of Creation October 27, 2009 Date

More information

66 defined. The main constraint in the MUC conferences is that applications are to be developed in a short time (e.g. one month). The MUCs represent a

66 defined. The main constraint in the MUC conferences is that applications are to be developed in a short time (e.g. one month). The MUCs represent a User-System Cooperation in Document Annotation based on Information Extraction 65 Fabio Ciravegna 1, Alexiei Dingli 1, Daniela Petrelli 2 and Yorick Wilks 1 1 Department of Computer Science, University

More information

EFFICIENT INTEGRATION OF SEMANTIC TECHNOLOGIES FOR PROFESSIONAL IMAGE ANNOTATION AND SEARCH

EFFICIENT INTEGRATION OF SEMANTIC TECHNOLOGIES FOR PROFESSIONAL IMAGE ANNOTATION AND SEARCH EFFICIENT INTEGRATION OF SEMANTIC TECHNOLOGIES FOR PROFESSIONAL IMAGE ANNOTATION AND SEARCH Andreas Walter FZI Forschungszentrum Informatik, Haid-und-Neu-Straße 10-14, 76131 Karlsruhe, Germany, awalter@fzi.de

More information

Human Intelligence in the Process of Semantic Content Creation

Human Intelligence in the Process of Semantic Content Creation World Wide Web (2010) 13:33 59 DOI 10.1007/s11280-009-0078-0 Human Intelligence in the Process of Semantic Content Creation Katharina Siorpaes Elena Simperl Received: 27 March 2009 / Revised: 16 August

More information

On the automatic classification of app reviews

On the automatic classification of app reviews Requirements Eng (2016) 21:311 331 DOI 10.1007/s00766-016-0251-9 RE 2015 On the automatic classification of app reviews Walid Maalej 1 Zijad Kurtanović 1 Hadeer Nabil 2 Christoph Stanik 1 Walid: please

More information

Semantic Annotation using Horizontal and Vertical Contexts

Semantic Annotation using Horizontal and Vertical Contexts Semantic Annotation using Horizontal and Vertical Contexts Mingcai Hong, Jie Tang, and Juanzi Li Department of Computer Science & Technology, Tsinghua University, 100084. China. {hmc, tj, ljz}@keg.cs.tsinghua.edu.cn

More information

A Semantic Role Repository Linking FrameNet and WordNet

A Semantic Role Repository Linking FrameNet and WordNet A Semantic Role Repository Linking FrameNet and WordNet Volha Bryl, Irina Sergienya, Sara Tonelli, Claudio Giuliano {bryl,sergienya,satonelli,giuliano}@fbk.eu Fondazione Bruno Kessler, Trento, Italy Abstract

More information

Information mining and information retrieval : methods and applications

Information mining and information retrieval : methods and applications Information mining and information retrieval : methods and applications J. Mothe, C. Chrisment Institut de Recherche en Informatique de Toulouse Université Paul Sabatier, 118 Route de Narbonne, 31062 Toulouse

More information

Towards the Semantic Web

Towards the Semantic Web Towards the Semantic Web Ora Lassila Research Fellow, Nokia Research Center (Boston) Chief Scientist, Nokia Venture Partners LLP Advisory Board Member, W3C XML Finland, October 2002 1 NOKIA 10/27/02 -

More information

Optimized Implementation of Logic Functions

Optimized Implementation of Logic Functions June 25, 22 9:7 vra235_ch4 Sheet number Page number 49 black chapter 4 Optimized Implementation of Logic Functions 4. Nc3xe4, Nb8 d7 49 June 25, 22 9:7 vra235_ch4 Sheet number 2 Page number 5 black 5 CHAPTER

More information

Verification of Multiple Agent Knowledge-based Systems

Verification of Multiple Agent Knowledge-based Systems Verification of Multiple Agent Knowledge-based Systems From: AAAI Technical Report WS-97-01. Compilation copyright 1997, AAAI (www.aaai.org). All rights reserved. Daniel E. O Leary University of Southern

More information

Models, Tools and Transformations for Design and Evaluation of Interactive Applications

Models, Tools and Transformations for Design and Evaluation of Interactive Applications Models, Tools and Transformations for Design and Evaluation of Interactive Applications Fabio Paternò, Laila Paganelli, Carmen Santoro CNUCE-C.N.R. Via G.Moruzzi, 1 Pisa, Italy fabio.paterno@cnuce.cnr.it

More information

An Annotation Tool for Semantic Documents

An Annotation Tool for Semantic Documents An Annotation Tool for Semantic Documents (System Description) Henrik Eriksson Dept. of Computer and Information Science Linköping University SE-581 83 Linköping, Sweden her@ida.liu.se Abstract. Document

More information

Semantic Web-Based Document: Editing and Browsing in AktiveDoc

Semantic Web-Based Document: Editing and Browsing in AktiveDoc Semantic Web-Based Document: Editing and Browsing in AktiveDoc Vitaveska Lanfranchi 1, Fabio Ciravegna 1, and Daniela Petrelli 2 1 Department of Computer Science, University of Sheffield, Regent Court,

More information

Ontology Refinement and Evaluation based on is-a Hierarchy Similarity

Ontology Refinement and Evaluation based on is-a Hierarchy Similarity Ontology Refinement and Evaluation based on is-a Hierarchy Similarity Takeshi Masuda The Institute of Scientific and Industrial Research, Osaka University Abstract. Ontologies are constructed in fields

More information

GLEIF Global LEI Data Quality Report Dictionary

GLEIF Global LEI Data Quality Report Dictionary GLEIF Global LEI Data Quality Report Dictionary Global LEI Data Quality Report Dictionary 2 19 Contents Data Quality Report Glossary... 3 1. Chapter 1: Preface... 4 1.1. Purpose of the Data Quality Report...

More information

MULTIMEDIA DATABASES OVERVIEW

MULTIMEDIA DATABASES OVERVIEW MULTIMEDIA DATABASES OVERVIEW Recent developments in information systems technologies have resulted in computerizing many applications in various business areas. Data has become a critical resource in

More information

User Experience Research Report: Heuristic Evaluation

User Experience Research Report: Heuristic Evaluation User Experience Research Report: Heuristic Evaluation SI 622 003 Group 3: Yunqi Hu, Diane Pham, Chieh-Lin Wu, Ruofan Zhang Date: March 31, 2016 Word Count: 2,610 Table of Contents Executive Summary...

More information

IBM Watson Application Developer Workshop. Watson Knowledge Studio: Building a Machine-learning Annotator with Watson Knowledge Studio.

IBM Watson Application Developer Workshop. Watson Knowledge Studio: Building a Machine-learning Annotator with Watson Knowledge Studio. IBM Watson Application Developer Workshop Lab02 Watson Knowledge Studio: Building a Machine-learning Annotator with Watson Knowledge Studio January 2017 Duration: 60 minutes Prepared by Víctor L. Fandiño

More information

Chapter 6 Evaluation Metrics and Evaluation

Chapter 6 Evaluation Metrics and Evaluation Chapter 6 Evaluation Metrics and Evaluation The area of evaluation of information retrieval and natural language processing systems is complex. It will only be touched on in this chapter. First the scientific

More information

Using Attention and Context Information for Annotations in a Semantic Wiki

Using Attention and Context Information for Annotations in a Semantic Wiki Using Attention and Context Information for Annotations in a Semantic Wiki Malte Kiesel, Sven Schwarz, Ludger van Elst, and Georg Buscher Knowledge Management Department German Research Center for Artificial

More information

The Trustworthiness of Digital Records

The Trustworthiness of Digital Records The Trustworthiness of Digital Records International Congress on Digital Records Preservation Beijing, China 16 April 2010 1 The Concept of Record Record: any document made or received by a physical or

More information

IJCSC Volume 5 Number 1 March-Sep 2014 pp ISSN

IJCSC Volume 5 Number 1 March-Sep 2014 pp ISSN Movie Related Information Retrieval Using Ontology Based Semantic Search Tarjni Vyas, Hetali Tank, Kinjal Shah Nirma University, Ahmedabad tarjni.vyas@nirmauni.ac.in, tank92@gmail.com, shahkinjal92@gmail.com

More information

Framework for Sense Disambiguation of Mathematical Expressions

Framework for Sense Disambiguation of Mathematical Expressions Proc. 14th Int. Conf. on Global Research and Education, Inter-Academia 2015 JJAP Conf. Proc. 4 (2016) 011609 2016 The Japan Society of Applied Physics Framework for Sense Disambiguation of Mathematical

More information

Enterprise Multimedia Integration and Search

Enterprise Multimedia Integration and Search Enterprise Multimedia Integration and Search José-Manuel López-Cobo 1 and Katharina Siorpaes 1,2 1 playence, Austria, 2 STI Innsbruck, University of Innsbruck, Austria {ozelin.lopez, katharina.siorpaes}@playence.com

More information

IST MPEG-4 Video Compliant Framework

IST MPEG-4 Video Compliant Framework IST MPEG-4 Video Compliant Framework João Valentim, Paulo Nunes, Fernando Pereira Instituto de Telecomunicações, Instituto Superior Técnico, Av. Rovisco Pais, 1049-001 Lisboa, Portugal Abstract This paper

More information

Usability Evaluation of Tools for Nomadic Application Development

Usability Evaluation of Tools for Nomadic Application Development Usability Evaluation of Tools for Nomadic Application Development Cristina Chesta (1), Carmen Santoro (2), Fabio Paternò (2) (1) Motorola Electronics S.p.a. GSG Italy Via Cardinal Massaia 83, 10147 Torino

More information

Ontosophie: A Semi-Automatic System for Ontology Population from Text

Ontosophie: A Semi-Automatic System for Ontology Population from Text Ontosophie: A Semi-Automatic System for Ontology Population from Text Tech Report kmi-04-19 David Celjuska and Dr. Maria Vargas-Vera Ontosophie: A Semi-Automatic System for Ontology Population from Text

More information

Data analysis using Microsoft Excel

Data analysis using Microsoft Excel Introduction to Statistics Statistics may be defined as the science of collection, organization presentation analysis and interpretation of numerical data from the logical analysis. 1.Collection of Data

More information

Starting Ontology Development by Visually Modeling an Example Situation - a User Study

Starting Ontology Development by Visually Modeling an Example Situation - a User Study Starting Ontology Development by Visually Modeling an Example Situation - a User Marek Dudáš 1, Vojtěch Svátek 1, Miroslav Vacura 1,2, and Ondřej Zamazal 1 1 Department of Information and Knowledge Engineering,

More information

Satisfactory Peening Intensity Curves

Satisfactory Peening Intensity Curves academic study Prof. Dr. David Kirk Coventry University, U.K. Satisfactory Peening Intensity Curves INTRODUCTION Obtaining satisfactory peening intensity curves is a basic priority. Such curves will: 1

More information

Get the most value from your surveys with text analysis

Get the most value from your surveys with text analysis SPSS Text Analysis for Surveys 3.0 Specifications Get the most value from your surveys with text analysis The words people use to answer a question tell you a lot about what they think and feel. That s

More information

The Application Research of Semantic Web Technology and Clickstream Data Mart in Tourism Electronic Commerce Website Bo Liu

The Application Research of Semantic Web Technology and Clickstream Data Mart in Tourism Electronic Commerce Website Bo Liu International Conference on Education Technology, Management and Humanities Science (ETMHS 2015) The Application Research of Semantic Web Technology and Clickstream Data Mart in Tourism Electronic Commerce

More information

The Semantic Web: A New Opportunity and Challenge for Human Language Technology

The Semantic Web: A New Opportunity and Challenge for Human Language Technology The Semantic Web: A New Opportunity and Challenge for Human Language Technology Kalina Bontcheva and Hamish Cunningham Department of Computer Science, University of Sheffield 211 Portobello St, Sheffield,

More information

FOAM Framework for Ontology Alignment and Mapping Results of the Ontology Alignment Evaluation Initiative

FOAM Framework for Ontology Alignment and Mapping Results of the Ontology Alignment Evaluation Initiative FOAM Framework for Ontology Alignment and Mapping Results of the Ontology Alignment Evaluation Initiative Marc Ehrig Institute AIFB University of Karlsruhe 76128 Karlsruhe, Germany ehrig@aifb.uni-karlsruhe.de

More information

A B2B Search Engine. Abstract. Motivation. Challenges. Technical Report

A B2B Search Engine. Abstract. Motivation. Challenges. Technical Report Technical Report A B2B Search Engine Abstract In this report, we describe a business-to-business search engine that allows searching for potential customers with highly-specific queries. Currently over

More information

Supporting Documentation and Evolution of Crosscutting Concerns in Business Processes

Supporting Documentation and Evolution of Crosscutting Concerns in Business Processes Supporting Documentation and Evolution of Crosscutting Concerns in Business Processes Chiara Di Francescomarino supervised by Paolo Tonella dfmchiara@fbk.eu - Fondazione Bruno Kessler, Trento, Italy Abstract.

More information