Information Intelligence: Metadata for Information Discovery, Access, and Integration
|
|
- Juliana Cook
- 6 years ago
- Views:
Transcription
1 Information Intelligence: Metadata for Information Discovery, Access, and Integration Randall Hauch, Alex Miller, Rob Cardwell MetaMatrix St. Louis, Missouri {rhauch, amiller, Abstract Integrating enterprise information requires an accurate, precise and complete understanding of the disparate data sources, the needs of the information consumers, and how these map to the semantic business concepts of the enterprise. We describe how MetaMatrix captures and manages this metadata through the use of the OMG s MOF architecture and multiple domain-specific modeling languages, and how this semantic and syntactic metadata is then used for a variety of purposes, including accessing data in real-time from the underlying enterprise systems, integrating it, and returning it as information expected by consumers. 1. Turning Data Into Information Information access and integration is the process of locating and extracting data in real-time from multiple disparate sources, relating it, and producing a unified representation matching the syntactic and semantic expectation of the recipient. This process is achieved by leveraging metadata that describes the syntax and semantics of each data source and how the data is related. This definition contains four critical facets: Locating and extracting data in real-time from multiple disparate sources For decades we have been deconstructing the information that we use to communicate and interact with each other and persisting it in many different forms. Most of those systems were built to support specific applications, and thus intentionally deal with a subset of the enterprise s data. Business needs require that the data in these disparate silos be accessible beyond the systems that were originally designed to use it, and that this data be available for reuse by the newer enterprise applications and processes. An important aspect of this is that many of these disparate data silos change continuously, making real-time access to that data a requirement. Using and leveraging metadata describing the syntax and semantics of each data source Metadata attempts to define and capture the syntax and semantics of how information is represented as data. In other words, metadata helps us understand the technical specifics of how data is represented or Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. SIGMOD 2005 June 14-16, 2005, Baltimore, Maryland, USA Copyright 2005 ACM /05/06 $5.00. stored, how it can be accessed, and perhaps most importantly, what the data means and how it relates to concepts that people (within a certain context) understand. A primary challenge of information integration is capturing metadata that is accurate, complete and precise, while also making the maintenance of that metadata straightforward and simple for users. Relate, join and synthesize the data Extracting data from multiple sources can be challenging, given the variety of information storage systems, interfaces, and languages used to interact with those systems. Further combining and synthesizing data from these systems requires the use of algorithms that are resilient in the face of varying data formats, poor data quality, and semantic value mismatches. Finally, users expect to work with large volumes of data with performance comparable to a single source, regardless of the difficulties of the integration process. Produce a unified representation matching the syntactic and semantic expectation of the recipient This is perhaps the most challenging part of information integration, and it is the part that distinguishes information integration from federated query or mediation. Each information consumer has their own unique context, requirements, and information needs. The more closely the data can be provided in the form expected by the recipient, the less work they must perform to extract value. The key is abstracting the recipient from how and where the data is obtained and transformed into information deemed useful by that recipient. Abstraction reduces the time and effort required to connect multiple systems to the information integration service, thereby increasing its cost effectiveness. Why is information integration even necessary? An enterprise that is able to break down the barriers surrounding the disparate silos can extract maximum benefit from its enterprise data, providing unified views of business-critical data (such as customers, products, orders, etc.) and leveraging existing information in new ways. In addition, breaking down the barriers allows the business to detect redundancy and appropriately manage their operational costs. Removing redundancy also allows a business to more quickly respond to change as one system must be changed rather than many. Information integration can be achieved a number of ways. Data warehouses and data marts perform the steps a priori to construct a centralized representation that is closer to the recipient, allowing the recipient to access and process the copied and cleansed data to find the information they are seeking. One substantial benefit of this approach is typically fast access, since the information is all collocated and preprocessed. Another advantage is that transactional systems are isolated from decision support systems, 793
2 allowing them to be managed independently. However, businesses today need ever-fresher data, increasing the cost and complexity of maintaining a warehouse. Additionally, warehouses are often structured and optimized for specific types of processing, not necessarily for providing the information in the form expected by one or more recipients with different needs and requirements [1]. Enterprise Information Integration (EII) performs the integration in real time on an as-needed basis. Performing integration in real time means that the EII system must often provide a complete bidirectional abstraction, not just a stylized layer for retrieval in a certain context. This capability includes transactional inserts, updates, and deletes. 2. Historical Perspective Large and medium size enterprises have data stored in different systems using a variety of technologies. One major reason for this is that new systems are often developed in conjunction with their own new data stores, yet systems can remain in production for decades. Additionally, changes in corporate structure lead to new data stores for an enterprise, often with similar purposes as existing systems. Finally, new technologies such as the Internet, , and wireless provide benefits to users that drive new kinds of stores. The result is that the typical enterprise has many sources of data that are disjoint, dissimilar, overlapping, and even duplicate. Integrating data that exists in these different types of systems requires understanding what data is managed by each system, and how the data relates to the semantic terms and concepts expressed by the information consumer. 2.1 Modeling Technology In the 60s, 70s, and 80s, the technology to efficiently persist, manage, access and update large amounts of data matured. This included hierarchical, relational, and object-oriented database management systems. As these systems could be configured to support different data schemas, tools were needed to help data architects design data structures. But describing the different data schemas required different concepts, different terminology, and different notations to best match the underlying systems that would host the data. Mylopoulos [2] provides a brief history of several of the major modeling approaches, including: Physical models used the same data representations and structures used by the software applications (b-trees, arrays, lists, variables, etc.), and consequently required the programmer/modeler to deal expressly with the conflicting concerns of computational efficiency and application and information simplicity. Logical models offered abstract mathematical constructs (sets, arrays, relations) to hide implementation details, but the constructs were limited and flat, and still required mapping the semantic concepts that people deal with into the mathematical constructs. Conceptual models used more expressive techniques to capture the same semantic terminology (customer, account, employee, etc.) used by people in the enterprise to model the application, and were able to organize the information using abstraction mechanisms (generalization, aggregation, classification, normalization, parameterization, etc.). 2.2 Data Management and Integration There are many good surveys and histories of the different technologies involved in data management, and consequently we only wish to provide a brief synopsis. Data management systems (mainframes and hierarchical systems) initially enabled large enterprises to persist storage for large enterprises. With the advent of relational database management systems that could be run on smaller and cheaper hardware, many small- and medium-sized companies were able to take advantage of this technology. Until the late 90s, attempts to integrate decoupled and disjoint systems ( silos of data ) often followed one of two approaches. Data warehousing extracts the data from its master locations, transforms it into a common schema, and loads the newly transformed data into a single, centralized store used for highvolume read-only decision support activities. As such, data warehouse schemas are optimized for such queries, often at significant expense to updates. Data marts are similar to data warehouses, but are often smaller scale and involve data for only parts of an enterprise. The other common approach to integrate disparate silos was Enterprise Application Integration (EAI), which attempted to integrate the software applications that sit on top of the data stores. Early integration architectures used point-to-point communication that was extremely complicated with even moderate numbers of systems. Subsequently, hub and spoke integration architectures required the creation of a common representation that was mutually agreed upon, but arriving at a common representation that accurately and completely satisfied the structure and semantic requirements of all systems involved proved costly and difficult. Perhaps more fundamentally, integrating the software applications lost sight of the fact that the enterprise data is what needs to be exchanged between the systems, and involving the software applications only complicated the integration efforts. 3. Disparate Data and Disparate Metadata Perhaps the most fundamental reason why data sources have proliferated is because people have different information needs brought about by the different roles they play. One solution does not fit all, and probably never will. Thus, people have continually looked to new approaches and new technologies that can help them solve their specific information needs in ways that are more efficient, effective, and effortless. So why, then, do people expect their metadata to be expressed and managed in a single, common and ubiquitous language? In fact, metadata is merely another form of data, and thus people bring to metadata management the same types of bias, context, and expectation as they do to data management. Consider the many modeling systems that use a single modeling language (ER, UML, XSD, Object-Role modeling, etc.). While such tools are often very good at modeling a specific type of system, users become familiar with the tool and techniques and eventually attempt to use it to model other types of systems. Unfortunately, if the system has different constructs and concepts, modeling that system will require the person to adapt the modeling language in ways that often not easily repeatable or sharable. The resulting models form isolated silos of metadata that are not only less accurate, precise and complete, but also less usable and subject to disagreement. 794
3 Figure 1. Systems have their own semantic concepts that are best modeled and represented with domain-specific languages. Therefore, if the lessons of data management and integration can be applied to metadata representation and management, then metadata, like data, will continue to be disparate and disjoint. Consequently, effective definition and management of metadata requires the ability to describe the metadata using the semantic constructs and terminology most appropriate for the system being modeled. In other words, if different types of systems are to be modeled effectively, accurately and precisely, then practical metadata management must have a foundation with support for different and domain-specific modeling languages, or metamodels (See Figure 1). Some systems import metadata expressed in any metamodel and transform it into a single, generic representation or metamodel [3,4]. This is often the case for systems that, following import, automatically use and process this unified and homogenized metadata the homogeneous metadata representation makes it Figure 2. The MetaBase Modeler, an integrated model development environment. significantly easier to process. However, one major disadvantage of this approach when used for metadata management is that once the metadata is imported, its representation is no longer in the domain-specific language that captures the specific structure and semantics of the system, and so it becomes difficult for users to view, interpret and maintain the metadata. 4. Model Development and Integration The MetaMatrix MetaBase product provides an integrated environment for modeling different types of data and information systems. MetaBase includes a modeling workbench, called the MetaBase Modeler (see Figure 2), where models of different systems can be viewed, edited, validated and related. MetaBase also includes a repository that provides standard configuration management capabilities, and a virtual catalog that can integrate disparate metadata repositories to enable searching, reporting, and discovery of metadata across the enterprise (see Figure 3). The Modeler is an integrated model development environment where models of different types of systems can be created, viewed, manipulated and managed in common ways. For example, the Modeler provides integrated model validation, searching across all models, support for a number of diagram types, undo/redo, events, and various wizards to import, update or export metadata. The Modeler also provides integrated support for refactoring to change models (delete, rename, move, etc.) and automatically view and accept changes to dependent models. Also, the Modeler provides a common mechanism for managing descriptions. However, the Modeler does not support these features with one or even a couple of modeling languages. The Modeler supports multiple domain-specific languages (or metamodels), including relational, XML documents, XML Schema Documents (XSD) 1,
4 Figure 3. MetaBase: modeling all aspects of information, and using metadata for real-time integration. web services, UML2 (object-oriented) 2, and relationships. The Modeler uses the OMG s Meta Object Facility (MOF) 3 architecture, which defines how multiple modeling languages can be defined so that they are interoperable (see Figure 4). The Modeler then adds on top of the MOF architecture the ability to drive the runtime behavior of the modeling environment with these different modeling languages. As such, new domain-specific modeling languages can be added at any time, and the runtime behavior adapts to include the new languages throughout the environment. Additionally, the metamodels can be extended to add custom properties on each of the language s constructs. Because the Modeler natively supports many domain-specific languages in a single user interface workbench environment, users can find, view, relate and transform metadata that exists in separate models and that is expressed in different modeling languages. For example, a relational model of a database can be related to the UML conceptual/logical model, and to the XML schemas used to represent the data in an XML format. In other cases, a relational model may be generated, for example, from a UML or ER conceptual representation that was imported from IBM Rational Rose, Popkin System Architect, or ERwin. These types of general relationships are useful to track dependencies, usage, realization, refinements, etc., and in fact the types of relationships can be extended. This ability to capture mappings between models that form different contexts of the enterprise information provides substantial flexibility and power to define how information can be transformed from one form to another. These transformation definitions are detailed mapping relationships that address, among other things, data type mismatches, schema and structural mismatches (e.g., combining data from two columns into one), joins between sources, unions of multiple sources, procedural logic, and bi-directional transformations (querying and updates). The most common approach is to create different layers of metadata, each with different purposes. Physical layer The lowest layer directly and accurately models the data sources themselves. Because these models directly correlate with the data source, the metadata can be used to identify and understand the data source itself. Business concept layer The next layer sits above the physical layer, and represents the business constructs and terminology. To each of these business concepts are added the transformation detail that defines how the data in the physical layer is related to the business concept. Because these business constructs represent the semantic terminology of the enterprise and because they are mapped with executable transformations to the underlying enterprise data, these are essentially reusable data components. If desirable, the transformations can be defined as supporting updates, meaning that inserts, updates and deletes may be pushed to the data sources through the business layers. Consumer layer This layer builds upon the reusable semantic constructs of the business layer and defines views that are specific to different categories of information consumer. Thus, this layer maps the semantic enterprise concepts back into a particular literal representation needed by the consumer and may take the form of relational tables, relational procedures, XML documents, or web services. In fact, the relational-to- XML mapping technology is able to transform any relational structure into any XML document structure (see Figure 1). Again, the transformations can be defined as supporting updates. 2 and and 796
5 Figure 4. OMG s MOF Architecture and support for multiple metamodels The Modeler is built on top of the Eclipse platform 4, which is an open-source tool-integration framework used by numerous commercial and open-source projects. This plug-in based architecture encourages decoupling various software components, provides standard and reusable components for common functionality (file management, repository integration, editors, view management, software updates, etc.), and provides a mechanism by which new functionality and new components (new importers, exporters, wizards, metamodels) can be added to the platform. 5. Using the Metadata The MetaBase metadata management system can be used to model many different types of systems and multiple views of those systems within a single environment with multiple modeling languages. These different and disparate models can be validated, related, shared, controlled, reused, automatically transformed into alternative metadata representations, and assembled into consistent and complete packages of metadata. These metadata packages are easily managed and deployable, and can be leveraged by a variety of tools for a number of purposes and activities. 5.1 Information Discovery This metadata, once captured, can be used to search, discover and understand what data is available and how the enterprise data is related. Being able to compare dissimilar data structures provides the ability to identify similarities and to provide harmonized views or representations. Additionally, the metadata can be used to help rationalize the existing data assets to find redundancies and eliminate duplicate silos of data. 5.2 Metadata Access and Integration The packages of complete and consistent metadata can be leveraged by applications that drive their behavior with metadata. 4 These applications come in many forms. For example, data access portals may use the metadata to completely define the behavior and presentation of the portal. Metadata-aware applications consume and integrate metadata and data from a variety of sources, incorporating and relating the metadata from enterprise sources with the information available on the enterprise intranet or the Internet. A single and extensible framework for modeling and managing the metadata for multiple types of systems makes it possible for each of these applications to use custom views and representations of the integrated metadata. 5.3 Information Access and Integration The MetaMatrix Server is an advanced integration engine that provides unified access to integrated information by consuming metadata about data sources, consumer-specific views of information, and the relationships between them. A metadata package is simply deployed to a MetaMatrix Server to create a virtual database (VDB), which appears to client applications as a typical database or web service. Multiple virtual databases can be deployed simultaneously, making it possible to support information consumers with different expectations and views of the same underlying enterprise data. Clients connect to and communicate with the virtual databases through standard mechanisms like Java Database Connectivity (JDBC) 5, Open Database Connectivity (ODBC), Simple Object Access Protocol (SOAP) 6 over HTTP 7, or SOAP over Java Message Service (JMS) 8, and interact with the VDB through SQL, XQuery 9 or web service operations. As such, client applications are abstracted away from the underlying data sources and how to communicate and interact with them
6 Driving the information integration with metadata has a number of advantages: Optimization is left to the engine The models declare what data exists and how the business concept and consumer layers relate to the data. The MetaMatrix integration engine is responsible for determining the actual process to execute the integration activity. Optimizations may include the use of a variety of algorithms, the removal of parts of the abstraction inconsequential to the result, and the enlistment of underlying data sources to perform the integration work in the optimal location for minimal data flow. Caching policies are declarative Different consumers have different expectations about performance and the liveliness of data. MetaMatrix can provide caching policies that vary between users to satisfy their requirements. Security Access to information can be controlled through flexible security policies that, again, are independent of the metadata. MetaMatrix can leverage and integrate with existing security infrastructure (single sign-on, LDAP directories, and data source specific mechanisms). Safety and control MetaMatrix can be configured to constrain and limit the loads on the underlying sources. This may include caching, but also may involve defining limits on the connectors to the underlying data sources, such as allowed access paths. Traceability MetaMatrix provides several ways to track usage of information by user and by data source. This information can be used to understand and view access histories and to enable financial policies such as billing. 6. Case Study A large financial services company, like many organizations, has numerous and often-duplicate data sources tightly bound to specific applications. Without metadata management, the organization was challenged to understand the data it had or where it resided. In particular, the organization wanted to improve the availability of vital reference data, including somewhat static information describing assets and account entities used in trade processing transactions, business intelligence, risk management, and reporting. Inconsistent and incomplete reference data can cause trade delays and trade failures, and problems caused by faulty reference data result in increased operational risk, lost revenue opportunities, and expensive manual trade duplication and reconciliation processes. Having many legacy silos of reference data complicates the process of data retrieval, normalization and aggregation. The organization used MetaMatrix technology to move from data silos and batch processing to centralized information access in real-time. The organization modeled the existing silos of reference data and created virtual models with specific views of the reference data needed by transaction processing, reporting, and business intelligence applications. These models were then packaged as virtual databases and deployed to the MetaMatrix Server integration engine, resulting in the ability to eliminate inefficient batch processing jobs and giving desktop trading applications ondemand access to integrated information. Processing that previously occurred in a nightly batch operation now occurs in MetaMatrix on both an ad-hoc and a batch request basis. Frequently changing information is available in real-time while static data is updated only as needed, leveraging existing data stores and data warehouses. 7. Conclusion Integrating disparate enterprise data sources is one of the most pressing problems facing enterprises today. These different sources of information contain different structures and contexts of data that have their own semantics, but the different types of sources and different views of information needed by different consumers require an adaptive and flexible approach to modeling information. MetaMatrix employs a modeling environment that supports multiple domain-specific languages that are appropriate for different types of systems, and that effectively model the available information, the enterprise business concepts, and the different views needed by consumers. This metadata is then deployed to the MetaMatrix Server and used to model-drive the on-demand access and integration of information. 8. Acknowledgements The authors would like to thank Bob Scanlon, Shawn Curtiss, and Michael Lang for their helpful comments. 9. References [1] Hellerstein, J., Stonebraker, M., and Caccia, R., Independent, Open Enterprise Data Integration. IEEE Data Engineering Bulleting, 22(1): 43-49(1999) [2] Mylopoulos, John, Information Modeling in the Time of the Revolution. ACM Information Systems, 23, 3-4 (May 1998), [3] Bernstein, P., Melnik, S., Petropoulos, M., Quix, C., Industrial-Strength Schema Matching, ACM SIGMOD Record, 33, 4, (Dec 2004), [4] McComb, D., Semantics in Business Systems, Morgan Kaufman Publishers,
Powering EII with MOA
Powering EII with MOA Brad Wright Randall M. Hauch January 8, 2004 Topics What is an MOA and why is it important? EII as a case study for MOA MOA in action, a demonstration Attributes of an MOA Miscellaneous
More informationTeiid Designer User Guide 7.5.0
Teiid Designer User Guide 1 7.5.0 1. Introduction... 1 1.1. What is Teiid Designer?... 1 1.2. Why Use Teiid Designer?... 2 1.3. Metadata Overview... 2 1.3.1. What is Metadata... 2 1.3.2. Editing Metadata
More informationFull file at
Chapter 2 Data Warehousing True-False Questions 1. A real-time, enterprise-level data warehouse combined with a strategy for its use in decision support can leverage data to provide massive financial benefits
More informationTeiid Designer User Guide 7.7.0
Teiid Designer User Guide 1 7.7.0 1. Introduction... 1 1.1. What is Teiid Designer?... 1 1.2. Why Use Teiid Designer?... 2 1.3. Metadata Overview... 2 1.3.1. What is Metadata... 2 1.3.2. Editing Metadata
More informationUsing MDA to Integrate Corporate Data into an SOA
Using MDA to Integrate Corporate Data into an SOA Randall M. Hauch Chief Architect presented at MDA, SOA and Web Services: Delivering the Integrated Enterprise Practice, not Promise MetaMatrix Products
More informationTools to Develop New Linux Applications
Tools to Develop New Linux Applications IBM Software Development Platform Tools for every member of the Development Team Supports best practices in Software Development Analyst Architect Developer Tester
More informationmetamatrix enterprise data services platform
metamatrix enterprise data services platform Bridge the Gap with Data Services Leaders of high-priority application projects in businesses and government agencies are looking to complete projects efficiently,
More informationMetaMatrix Enterprise Data Services Platform
MetaMatrix Enterprise Data Services Platform MetaMatrix Overview Agenda Background What it does Where it fits How it works Demo Q/A 2 Product Review: Problem Data Challenges Difficult to implement new
More informationMETADATA INTERCHANGE IN SERVICE BASED ARCHITECTURE
UDC:681.324 Review paper METADATA INTERCHANGE IN SERVICE BASED ARCHITECTURE Alma Butkovi Tomac Nagravision Kudelski group, Cheseaux / Lausanne alma.butkovictomac@nagra.com Dražen Tomac Cambridge Technology
More informationTECHNOLOGY BRIEF: CA ERWIN DATA PROFILER. Combining Data Profiling and Data Modeling for Better Data Quality
TECHNOLOGY BRIEF: CA ERWIN DATA PROFILER Combining Data Profiling and Data Modeling for Better Data Quality Table of Contents Executive Summary SECTION 1: CHALLENGE 2 Reducing the Cost and Risk of Data
More informationThe Value of Data Modeling for the Data-Driven Enterprise
Solution Brief: erwin Data Modeler (DM) The Value of Data Modeling for the Data-Driven Enterprise Designing, documenting, standardizing and aligning any data from anywhere produces an enterprise data model
More informationHow a Federated Identity Service Turns Identity into a Business Enabler, Not an IT Bottleneck
How a Federated Identity Service Turns Identity into a Business Enabler, Not an IT Bottleneck Add Agility, Flexibility, and Responsiveness into Your Enterprise Delivering Identity the Way Your Business
More informationThe Open Group SOA Ontology Technical Standard. Clive Hatton
The Open Group SOA Ontology Technical Standard Clive Hatton The Open Group Releases SOA Ontology Standard To Increase SOA Adoption and Success Rates Ontology Fosters Common Understanding of SOA Concepts
More informationIntroduction to Federation Server
Introduction to Federation Server Alex Lee IBM Information Integration Solutions Manager of Technical Presales Asia Pacific 2006 IBM Corporation WebSphere Federation Server Federation overview Tooling
More informationCA ERwin Data Profiler
PRODUCT BRIEF: CA ERWIN DATA PROFILER CA ERwin Data Profiler CA ERWIN DATA PROFILER HELPS ORGANIZATIONS LOWER THE COSTS AND RISK ASSOCIATED WITH DATA INTEGRATION BY PROVIDING REUSABLE, AUTOMATED, CROSS-DATA-SOURCE
More informationThe Model-Driven Semantic Web Emerging Standards & Technologies
The Model-Driven Semantic Web Emerging Standards & Technologies Elisa Kendall Sandpiper Software March 24, 2005 1 Model Driven Architecture (MDA ) Insulates business applications from technology evolution,
More informationMeaning & Concepts of Databases
27 th August 2015 Unit 1 Objective Meaning & Concepts of Databases Learning outcome Students will appreciate conceptual development of Databases Section 1: What is a Database & Applications Section 2:
More informationMicrosoft SharePoint Server 2013 Plan, Configure & Manage
Microsoft SharePoint Server 2013 Plan, Configure & Manage Course 20331-20332B 5 Days Instructor-led, Hands on Course Information This five day instructor-led course omits the overlap and redundancy that
More informationSemantics, Metadata and Identifying Master Data
Semantics, Metadata and Identifying Master Data A DataFlux White Paper Prepared by: David Loshin, President, Knowledge Integrity, Inc. Once you have determined that your organization can achieve the benefits
More informationTechno Expert Solutions An institute for specialized studies!
Getting Started Course Content of IBM Cognos Data Manger Identify the purpose of IBM Cognos Data Manager Define data warehousing and its key underlying concepts Identify how Data Manager creates data warehouses
More informationA WHITE PAPER By Silwood Technology Limited
A WHITE PAPER By Silwood Technology Limited Delivering metadata transparency for Enterprise Application packages Executive Summary Enterprise Resource Planning (ERP) packages such as SAP, J.D. Edwards
More informationUNIT I. Introduction
UNIT I Introduction Objective To know the need for database system. To study about various data models. To understand the architecture of database system. To introduce Relational database system. Introduction
More informationComposite Data Virtualization Maximizing Value from Enterprise Data Warehouse Investments
Composite Data Virtualization Maximizing Value from Enterprise Data Warehouse Investments Composite Software August 2012 TABLE OF CONTENTS MAXIMIZING VALUE FROM ENTERPRISE DATA WAREHOUSE INVESTMENTS...
More informationCA ERwin Data Modeler r7.3
PRODUCT BRIEF: CA ERWIN DATA MODELER R7.3 CA ERwin Data Modeler r7.3 CA ERWIN DATA MODELER (CA ERWIN DM) IS AN INDUSTRY-LEADING DATA MODELING SOLUTION THAT ENABLES YOU TO CREATE AND MAINTAIN DATABASES,
More informationCONSOLIDATING RISK MANAGEMENT AND REGULATORY COMPLIANCE APPLICATIONS USING A UNIFIED DATA PLATFORM
CONSOLIDATING RISK MANAGEMENT AND REGULATORY COMPLIANCE APPLICATIONS USING A UNIFIED PLATFORM Executive Summary Financial institutions have implemented and continue to implement many disparate applications
More informationManaging Data Resources
Chapter 7 Managing Data Resources 7.1 2006 by Prentice Hall OBJECTIVES Describe basic file organization concepts and the problems of managing data resources in a traditional file environment Describe how
More informationFig 1.2: Relationship between DW, ODS and OLTP Systems
1.4 DATA WAREHOUSES Data warehousing is a process for assembling and managing data from various sources for the purpose of gaining a single detailed view of an enterprise. Although there are several definitions
More informationFrom business need to implementation Design the right information solution
From business need to implementation Design the right information solution Davor Gornik (dgornik@us.ibm.com) Product Manager Agenda Relational design Integration design Summary Relational design Data modeling
More informationVocabulary Harvesting Using MatchIT. By Andrew W Krause, Chief Technology Officer
July 31, 2006 Vocabulary Harvesting Using MatchIT By Andrew W Krause, Chief Technology Officer Abstract Enterprises and communities require common vocabularies that comprehensively and concisely label/encode,
More information1 Dulcian, Inc., 2001 All rights reserved. Oracle9i Data Warehouse Review. Agenda
Agenda Oracle9i Warehouse Review Dulcian, Inc. Oracle9i Server OLAP Server Analytical SQL Mining ETL Infrastructure 9i Warehouse Builder Oracle 9i Server Overview E-Business Intelligence Platform 9i Server:
More informationchallenges in domain-specific modeling raphaël mannadiar august 27, 2009
challenges in domain-specific modeling raphaël mannadiar august 27, 2009 raphaël mannadiar challenges in domain-specific modeling 1/59 outline 1 introduction 2 approaches 3 debugging and simulation 4 differencing
More informationTeiid Designer User Guide 7.8.0
Teiid Designer User Guide 1 7.8.0 1. Introduction... 1 1.1. What is Teiid Designer?... 1 1.2. Metadata Overview... 2 1.2.1. What is Metadata... 2 1.2.2. Business and Technical Metadata... 4 1.2.3. Design-Time
More informationThe Top Five Reasons to Deploy Software-Defined Networks and Network Functions Virtualization
The Top Five Reasons to Deploy Software-Defined Networks and Network Functions Virtualization May 2014 Prepared by: Zeus Kerravala The Top Five Reasons to Deploy Software-Defined Networks and Network Functions
More informationGrid Computing Systems: A Survey and Taxonomy
Grid Computing Systems: A Survey and Taxonomy Material for this lecture from: A Survey and Taxonomy of Resource Management Systems for Grid Computing Systems, K. Krauter, R. Buyya, M. Maheswaran, CS Technical
More informationCA ERwin Data Modeler
PRODUCT BRIEF: CA ERWIN DATA MODELER CA ERwin Data Modeler CA ERWIN DATA MODELER (CA ERWIN DM) IS AN INDUSTRY-LEADING DATA MODELING SOLUTION THAT ENABLES YOU TO CREATE AND MAINTAIN DATABASES, DATA WAREHOUSES
More informationFast Innovation requires Fast IT
Fast Innovation requires Fast IT Cisco Data Virtualization Puneet Kumar Bhugra Business Solutions Manager 1 Challenge In Data, Big Data & Analytics Siloed, Multiple Sources Business Outcomes Business Opportunity:
More information1 DATAWAREHOUSING QUESTIONS by Mausami Sawarkar
1 DATAWAREHOUSING QUESTIONS by Mausami Sawarkar 1) What does the term 'Ad-hoc Analysis' mean? Choice 1 Business analysts use a subset of the data for analysis. Choice 2: Business analysts access the Data
More informationA Tutorial on The Jini Technology
A tutorial report for SENG 609.22 Agent Based Software Engineering Course Instructor: Dr. Behrouz H. Far A Tutorial on The Jini Technology Lian Chen Introduction Jini network technology provides a simple
More informationCA ERwin Data Modeler r9 Rick Alaras N.A. Channel Account Manager
ERwin r9 CA ERwin Data Modeler r9 Rick Alaras N.A. Channel Account Manager In today s data-driven economy, there is an increasing disconnect between consumers and providers of data DATA VOLUMES INCREASING
More informationComputation Independent Model (CIM): Platform Independent Model (PIM): Platform Specific Model (PSM): Implementation Specific Model (ISM):
viii Preface The software industry has evolved to tackle new approaches aligned with the Internet, object-orientation, distributed components and new platforms. However, the majority of the large information
More informationFINANCIAL REGULATORY REPORTING ACROSS AN EVOLVING SCHEMA
FINANCIAL REGULATORY REPORTING ACROSS AN EVOLVING SCHEMA MODELDR & MARKLOGIC - DATA POINT MODELING MARKLOGIC WHITE PAPER JUNE 2015 CHRIS ATKINSON Contents Regulatory Satisfaction is Increasingly Difficult
More informationOpen Source egovernment Reference Architecture. Cory Casanave, President. Data Access Technologies, Inc.
Open Source egovernment Reference Architecture Cory Casanave, President www.enterprisecomponent.com Slide 1 What we will cover OsEra OsEra Overview Model to Integrate From business model to execution Synthesis
More informationOracle Warehouse Builder 10g Release 2 Integrating Packaged Applications Data
Oracle Warehouse Builder 10g Release 2 Integrating Packaged Applications Data June 2006 Note: This document is for informational purposes. It is not a commitment to deliver any material, code, or functionality,
More informationSemantic Information Modeling for Federation (SIMF)
Purpose Semantic Information Modeling for Federation (SIMF) Overview V0.2-04/21/2011 The Architecture Ecosystem SIG of the Object Management Group (OMG) is in the process of drafting an RFP focused on
More informationIBM Rational Application Developer for WebSphere Software, Version 7.0
Visual application development for J2EE, Web, Web services and portal applications IBM Rational Application Developer for WebSphere Software, Version 7.0 Enables installation of only the features you need
More informationData Warehousing. New Features in SAS/Warehouse Administrator Ken Wright, SAS Institute Inc., Cary, NC. Paper
Paper 114-25 New Features in SAS/Warehouse Administrator Ken Wright, SAS Institute Inc., Cary, NC ABSTRACT SAS/Warehouse Administrator 2.0 introduces several powerful new features to assist in your data
More informationCHAPTER 2: DATA MODELS
Database Systems Design Implementation and Management 12th Edition Coronel TEST BANK Full download at: https://testbankreal.com/download/database-systems-design-implementation-andmanagement-12th-edition-coronel-test-bank/
More informationThe Unified Modelling Language. Example Diagrams. Notation vs. Methodology. UML and Meta Modelling
UML and Meta ling Topics: UML as an example visual notation The UML meta model and the concept of meta modelling Driven Architecture and model engineering The AndroMDA open source project Applying cognitive
More informationData Vault Brisbane User Group
Data Vault Brisbane User Group 26-02-2013 Agenda Introductions A brief introduction to Data Vault Creating a Data Vault based Data Warehouse Comparisons with 3NF/Kimball When is it good for you? Examples
More informationJBoss DNA. Randall Hauch Principal Software Engineer JBoss Data Services
JBoss DNA Randall Hauch Principal Software Engineer JBoss Data Services 1 JBoss DNA New project A few months old http://labs.jboss.org/dna Prior repository experience and IP MetaMatrix Repository Drools
More informationCHAPTER 2: DATA MODELS
CHAPTER 2: DATA MODELS 1. A data model is usually graphical. PTS: 1 DIF: Difficulty: Easy REF: p.36 2. An implementation-ready data model needn't necessarily contain enforceable rules to guarantee the
More informationMetaBase Modeler User s Guide MetaMatrix Products, Release 4.2 SP2 (Second Service Pack for Release 4.2) Document Edition 1, June 10, 2005
MetaBase Modeler User s Guide MetaMatrix Products, Release 4.2 SP2 (Second Service Pack for Release 4.2) Document Edition 1, June 10, 2005 2001-2005 MetaMatrix, Inc. All rights reserved. You can obtain
More informationWebSphere Application Server, Version 5. What s New?
WebSphere Application Server, Version 5 What s New? 1 WebSphere Application Server, V5 represents a continuation of the evolution to a single, integrated, cost effective, Web services-enabled, J2EE server
More informationRich Hilliard 20 February 2011
Metamodels in 42010 Executive summary: The purpose of this note is to investigate the use of metamodels in IEEE 1471 ISO/IEC 42010. In the present draft, metamodels serve two roles: (1) to describe the
More informationTOPLink for WebLogic. Whitepaper. The Challenge: The Solution:
Whitepaper The Challenge: Enterprise JavaBeans (EJB) represents a new standard in enterprise computing: a component-based architecture for developing and deploying distributed object-oriented applications
More informationSolving the Enterprise Data Dilemma
Solving the Enterprise Data Dilemma Harmonizing Data Management and Data Governance to Accelerate Actionable Insights Learn More at erwin.com Is Our Company Realizing Value from Our Data? If your business
More informationDQpowersuite. Superior Architecture. A Complete Data Integration Package
DQpowersuite Superior Architecture Since its first release in 1995, DQpowersuite has made it easy to access and join distributed enterprise data. DQpowersuite provides an easy-toimplement architecture
More informationHow to Accelerate Merger and Acquisition Synergies
How to Accelerate Merger and Acquisition Synergies MERGER AND ACQUISITION CHALLENGES Mergers and acquisitions (M&A) occur frequently in today s business environment; $3 trillion in 2017 alone. 1 M&A enables
More informationIBM InfoSphere Information Analyzer
IBM InfoSphere Information Analyzer Understand, analyze and monitor your data Highlights Develop a greater understanding of data source structure, content and quality Leverage data quality rules continuously
More informationIBM B5280G - IBM COGNOS DATA MANAGER: BUILD DATA MARTS WITH ENTERPRISE DATA (V10.2)
IBM B5280G - IBM COGNOS DATA MANAGER: BUILD DATA MARTS WITH ENTERPRISE DATA (V10.2) Dauer: 5 Tage Durchführungsart: Präsenztraining Zielgruppe: This course is intended for Developers. Nr.: 35231 Preis:
More informationWhat s a BA to do with Data? Discover and define standard data elements in business terms
What s a BA to do with Data? Discover and define standard data elements in business terms Susan Block, Lead Business Systems Analyst The Vanguard Group Discussion Points Discovering Business Data The Data
More informationAn introduction to MOF MetaObject Facility.
An introduction to MOF MetaObject Facility pierre-alain.muller@irisa.fr About The MetaObject Facility Specification is the foundation of OMG's industry-standard standard environment where models can be
More informationRed Hat JBoss Data Virtualization 6.3 Glossary Guide
Red Hat JBoss Data Virtualization 6.3 Glossary Guide David Sage Nidhi Chaudhary Red Hat JBoss Data Virtualization 6.3 Glossary Guide David Sage dlesage@redhat.com Nidhi Chaudhary nchaudha@redhat.com Legal
More informationQuery-Time JOIN for Active Intelligence Engine (AIE)
Query-Time JOIN for Active Intelligence Engine (AIE) Ad hoc JOINing of Structured Data and Unstructured Content: An Attivio-Patented Breakthrough in Information- Centered Business Agility An Attivio Technology
More informationSupports 1-1, 1-many, and many to many relationships between objects
Author: Bill Ennis TOPLink provides container-managed persistence for BEA Weblogic. It has been available for Weblogic's application server since Weblogic version 4.5.1 released in December, 1999. TOPLink
More informationManaging Data Resources
Chapter 7 OBJECTIVES Describe basic file organization concepts and the problems of managing data resources in a traditional file environment Managing Data Resources Describe how a database management system
More informationAppDev StudioTM 3.2 SAS. Migration Guide
SAS Migration Guide AppDev StudioTM 3.2 The correct bibliographic citation for this manual is as follows: SAS Institute Inc. 2006. SAS AppDev TM Studio 3.2: Migration Guide. Cary, NC: SAS Institute Inc.
More informationComposite Data Virtualization Eight Ways Composite Data Virtualization Adds Value to Enterprise Data Warehousing
Composite Data Virtualization Eight Ways Composite Data Virtualization Adds Value to Enterprise Data Warehousing Composite Software January 2010 TABLE OF CONTENTS MAXIMIZING VALUE FROM ENTERPRISE DATA
More informationEMC Documentum xdb. High-performance native XML database optimized for storing and querying large volumes of XML content
DATA SHEET EMC Documentum xdb High-performance native XML database optimized for storing and querying large volumes of XML content The Big Picture Ideal for content-oriented applications like dynamic publishing
More informationAVOIDING SILOED DATA AND SILOED DATA MANAGEMENT
AVOIDING SILOED DATA AND SILOED DATA MANAGEMENT Dalton Cervo Author, Consultant, Data Management Expert March 2016 This presentation contains extracts from books that are: Copyright 2011 John Wiley & Sons,
More informationDatabase Management System. Fundamental Database Concepts
Database Management System Fundamental Database Concepts CONTENTS Basics of DBMS Purpose of DBMS Applications of DBMS Views of Data Instances and Schema Data Models Database Languages Responsibility of
More informationService-Oriented Architecture
Service-Oriented Architecture The Service Oriented Society Imagine if we had to do everything we need to get done by ourselves? From Craftsmen to Service Providers Our society has become what it is today
More informationModel Driven Architecture Targets Middleware Interoperability Challenges
Model Driven Architecture Targets Middleware Interoperability Challenges by Richard Soley Chairman and Chief Executive Officer Object Management Group and the OMG Staff Strategy Group "CORBA was a powerful
More information2. An implementation-ready data model needn't necessarily contain enforceable rules to guarantee the integrity of the data.
Test bank for Database Systems Design Implementation and Management 11th Edition by Carlos Coronel,Steven Morris Link full download test bank: http://testbankcollection.com/download/test-bank-for-database-systemsdesign-implementation-and-management-11th-edition-by-coronelmorris/
More informationA Guide to Using Cisco Data Virtualization
A Guide to Using Cisco Data Virtualization TONY YOUNG, SOLUTION ARCHITECT CISCO ADVANCED SERVICES SEPTEMBER 2015 2015 Cisco and/or its affiliates. All rights reserved. This document is Cisco Public. Page
More informationGet Started on SOA. People Entry Point Interaction and Collaboration Services. Case for an SOA Portal
Get Started on SOA People Entry Point Interaction and Collaboration Services Case for an SOA Our customers are our highest priorities; our employees are our highest cost We need to make our employees more
More informationQuickSpecs. ISG Navigator for Universal Data Access M ODELS OVERVIEW. Retired. ISG Navigator for Universal Data Access
M ODELS ISG Navigator from ISG International Software Group is a new-generation, standards-based middleware solution designed to access data from a full range of disparate data sources and formats.. OVERVIEW
More information5-1McGraw-Hill/Irwin. Copyright 2007 by The McGraw-Hill Companies, Inc. All rights reserved.
5-1McGraw-Hill/Irwin Copyright 2007 by The McGraw-Hill Companies, Inc. All rights reserved. 5 hapter Data Resource Management Data Concepts Database Management Types of Databases McGraw-Hill/Irwin Copyright
More informationThis tutorial has been prepared for computer science graduates to help them understand the basic-to-advanced concepts related to data mining.
About the Tutorial Data Mining is defined as the procedure of extracting information from huge sets of data. In other words, we can say that data mining is mining knowledge from data. The tutorial starts
More informationIntroduction to Dependable Systems: Meta-modeling and modeldriven
Introduction to Dependable Systems: Meta-modeling and modeldriven development http://d3s.mff.cuni.cz CHARLES UNIVERSITY IN PRAGUE faculty of mathematics and physics 3 Software development Automated software
More informationHybrid Data Platform
UniConnect-Powered Data Aggregation Across Enterprise Data Warehouses and Big Data Storage Platforms A Percipient Technology White Paper Author: Ai Meun Lim Chief Product Officer Updated Aug 2017 2017,
More informationwebmethods EntireX for ESB: Leveraging Platform and Application Flexibility While Optimizing Service Reuse
December 2008 webmethods EntireX for ESB: Leveraging Platform and Application Flexibility While Optimizing Service Reuse By Chris Pottinger, Sr. Manager Product Development, and Juergen Lind, Sr. Product
More informationIntroduction to Web Services & SOA
References: Web Services, A Technical Introduction, Deitel & Deitel Building Scalable and High Performance Java Web Applications, Barish Service-Oriented Programming (SOP) SOP A programming paradigm that
More informationCHAPTER 3 Implementation of Data warehouse in Data Mining
CHAPTER 3 Implementation of Data warehouse in Data Mining 3.1 Introduction to Data Warehousing A data warehouse is storage of convenient, consistent, complete and consolidated data, which is collected
More informationDiscover, Relate, Model, and Integrate Data Assets with Rational Data Architect
Discover, Relate, Model, and Integrate Data Assets with Rational Data Architect Niels C. Jacobsen (nielsj@dk.ibm.com) Associate IT Architect, IBM Software Group Rational IBM Software Group 2005 IBM Corporation
More informationThe Emerging Data Lake IT Strategy
The Emerging Data Lake IT Strategy An Evolving Approach for Dealing with Big Data & Changing Environments bit.ly/datalake SPEAKERS: Thomas Kelly, Practice Director Cognizant Technology Solutions Sean Martin,
More informationADT: Eclipse development tools for ATL
ADT: Eclipse development tools for ATL Freddy Allilaire (freddy.allilaire@laposte.net) Tarik Idrissi (tarik.idrissi@laposte.net) Université de Nantes Faculté de Sciences et Techniques LINA (Laboratoire
More informationFIVE BEST PRACTICES FOR ENSURING A SUCCESSFUL SQL SERVER MIGRATION
FIVE BEST PRACTICES FOR ENSURING A SUCCESSFUL SQL SERVER MIGRATION The process of planning and executing SQL Server migrations can be complex and risk-prone. This is a case where the right approach and
More informationDITA for Enterprise Business Documents Sub-committee Proposal Background Why an Enterprise Business Documents Sub committee
DITA for Enterprise Business Documents Sub-committee Proposal Background Why an Enterprise Business Documents Sub committee Documents initiate and record business change. It is easy to map some business
More informationLeverage SOA for increased business flexibility What, why, how, and when
Leverage SOA for increased business flexibility What, why, how, and when Dr. Bob Sutor Director, IBM WebSphere Product and Market Management sutor@us.ibm.com http://www.ibm.com/developerworks/blogs/dw_blog.jspa?blog=384
More informationObject Management Group Model Driven Architecture (MDA) MDA Guide rev. 2.0 OMG Document ormsc/
Executive Summary Object Management Group Model Driven Architecture (MDA) MDA Guide rev. 2.0 OMG Document ormsc/2014-06-01 This guide describes the Model Driven Architecture (MDA) approach as defined by
More informationSHARED DATA: THE ACHILLES HEEL OF SERVICE- ORIENTED ARCHITECTURES
SHARED DATA: THE ACHILLES HEEL OF SERVICE- ORIENTED ARCHITECTURES INTRODUCTION Service-oriented architectures (SOAs) are a significant advance in improving the flexibility of business logic. However, they
More informationModel-Based Social Networking Over Femtocell Environments
Proc. of World Cong. on Multimedia and Computer Science Model-Based Social Networking Over Femtocell Environments 1 Hajer Berhouma, 2 Kaouthar Sethom Ben Reguiga 1 ESPRIT, Institute of Engineering, Tunis,
More information3rd Lecture Languages for information modeling
3rd Lecture Languages for information modeling Agenda Languages for information modeling UML UML basic concepts Modeling by UML diagrams CASE tools: concepts, features and objectives CASE toolset architecture
More informationData Models: The Center of the Business Information Systems Universe
Data s: The Center of the Business Information Systems Universe Whitemarsh Information Systems Corporation 2008 Althea Lane Bowie, Maryland 20716 Tele: 301-249-1142 Email: Whitemarsh@wiscorp.com Web: www.wiscorp.com
More informationComposite Software Data Virtualization The Five Most Popular Uses of Data Virtualization
Composite Software Data Virtualization The Five Most Popular Uses of Data Virtualization Composite Software, Inc. June 2011 TABLE OF CONTENTS INTRODUCTION... 3 DATA FEDERATION... 4 PROBLEM DATA CONSOLIDATION
More informationMS-55045: Microsoft End to End Business Intelligence Boot Camp
MS-55045: Microsoft End to End Business Intelligence Boot Camp Description This five-day instructor-led course is a complete high-level tour of the Microsoft Business Intelligence stack. It introduces
More informationIBM Rational Software Architect
Unifying all aspects of software design and development IBM Rational Software Architect A complete design & development toolset Incorporates all the capabilities in IBM Rational Application Developer for
More informationOracle Warehouse Builder 10g Runtime Environment, an Update. An Oracle White Paper February 2004
Oracle Warehouse Builder 10g Runtime Environment, an Update An Oracle White Paper February 2004 Runtime Environment, an Update Executive Overview... 3 Introduction... 3 Runtime in warehouse builder 9.0.3...
More informationBreak Through Your Software Development Challenges with Microsoft Visual Studio 2008
Break Through Your Software Development Challenges with Microsoft Visual Studio 2008 White Paper November 2007 For the latest information, please see www.microsoft.com/vstudio This is a preliminary document
More information