Caché and Data Management in the Financial Services Industry

Similar documents
CONSOLIDATING RISK MANAGEMENT AND REGULATORY COMPLIANCE APPLICATIONS USING A UNIFIED DATA PLATFORM

IBM WebSphere Message Broker for z/os V6.1 delivers the enterprise service bus built for connectivity and transformation

Massive Scalability With InterSystems IRIS Data Platform

Full file at

Composite Software Data Virtualization The Five Most Popular Uses of Data Virtualization

COMMUNICATION PROTOCOLS

Achieving Horizontal Scalability. Alain Houf Sales Engineer

Abstract. The Challenges. ESG Lab Review InterSystems IRIS Data Platform: A Unified, Efficient Data Platform for Fast Business Insight

Incremental Updates VS Full Reload

Fast Innovation requires Fast IT

Model Driven Architecture and Rhapsody

BI Moves Operational - The Case for High-Performance Aggregation Infrastructure

The functions performed by a typical DBMS are the following:

Active Server Pages Architecture

Enterprise Data-warehouse (EDW) In Easy Steps

MetaMatrix Enterprise Data Services Platform

12 Minute Guide to Archival Search

Data Management Glossary

DQpowersuite. Superior Architecture. A Complete Data Integration Package

Partner Presentation Faster and Smarter Data Warehouses with Oracle OLAP 11g

Understanding the ESVA Architecture

Hospital System Lowers IT Costs After Epic Migration Flatirons Digital Innovations, Inc. All rights reserved.

1 Dulcian, Inc., 2001 All rights reserved. Oracle9i Data Warehouse Review. Agenda

Crystal Reports. Overview. Contents. How to report off a Teradata Database

OLAP Introduction and Overview

Capturing Your Changed Data

ISAM TO SQL MIGRATION

EMC Documentum xdb. High-performance native XML database optimized for storing and querying large volumes of XML content

JAVASCRIPT CHARTING. Scaling for the Enterprise with Metric Insights Copyright Metric insights, Inc.

QuickSpecs. Compaq NonStop Transaction Server for Java Solution. Models. Introduction. Creating a state-of-the-art transactional Java environment

Key Considerations for Improving Performance And Virtualization in Microsoft SQL Server Environments

<Insert Picture Here> Enterprise Data Management using Grid Technology

Technical Overview. Access control lists define the users, groups, and roles that can access content as well as the operations that can be performed.

An Oracle White Paper February Optimizing Storage for Oracle PeopleSoft Applications

Solace Message Routers and Cisco Ethernet Switches: Unified Infrastructure for Financial Services Middleware

Introduction to K2View Fabric

Frequently Asked Questions

Design Better. Reduce Risks. Ease Upgrades. Protect Your Software Investment

Shedding Tiers Creating a Simpler, More Manageable Storage Infrastructure

Advanced ODBC and JDBC Access to Salesforce Data

Best Practices for Deploying Web Services via Integration

CHAPTER 3 Implementation of Data warehouse in Data Mining

Optimizing Performance for Partitioned Mappings

Lenovo Database Configuration for Microsoft SQL Server TB

Lenovo Database Configuration

Composite Data Virtualization Maximizing Value from Enterprise Data Warehouse Investments

WHITE PAPER: BEST PRACTICES. Sizing and Scalability Recommendations for Symantec Endpoint Protection. Symantec Enterprise Security Solutions Group

IBM Europe Announcement ZP , dated November 6, 2007

Hitachi Adaptable Modular Storage and Hitachi Workgroup Modular Storage

Achieving Rapid Data Recovery for IBM AIX Environments An Executive Overview of EchoStream for AIX

Paper. Delivering Strong Security in a Hyperconverged Data Center Environment

Inside GigaSpaces XAP Technical Overview and Value Proposition

Low Friction Data Warehousing WITH PERSPECTIVE ILM DATA GOVERNOR

SGL FlashNet & Oracle s StorageTek Digital Tape for Media & Entertainment TECHNICAL ARTICLE

Xcelerated Business Insights (xbi): Going beyond business intelligence to drive information value

Moving From Reactive to Proactive Storage Management with an On-demand Cloud Solution

FLORIDA DEPARTMENT OF TRANSPORTATION PRODUCTION BIG DATA PLATFORM

Massively Parallel Processing. Big Data Really Fast. A Proven In-Memory Analytical Processing Platform for Big Data

Chapter 2 CommVault Data Management Concepts

HYCU and ExaGrid Hyper-converged Backup for Nutanix

Hitachi Adaptable Modular Storage and Workgroup Modular Storage

ScaleArc for SQL Server

ORACLE MESSAGEQ ORACLE DATA SHEET KEY FEATURES AND BENEFITS

SQL Server 2008 Consolidation

Capturing Your Changed Data With the CONNX Data Synchronization Tool. by Dann Corbit, Senior Software Engineer

CA ERwin Data Profiler

Data Model Considerations for Radar Systems

Whitepaper. Solving Complex Hierarchical Data Integration Issues. What is Complex Data? Types of Data

Was ist dran an einer spezialisierten Data Warehousing platform?

REPORTING AND QUERY TOOLS AND APPLICATIONS

Part II. Integration Use Cases

Provide Real-Time Data To Financial Applications

Moving Technology Infrastructure into the Future: Value and Performance through Consolidation

IT1105 Information Systems and Technology. BIT 1 ST YEAR SEMESTER 1 University of Colombo School of Computing. Student Manual

Archiving, Backup, and Recovery for Complete the Promise of Virtualisation Unified information management for enterprise Windows environments

MetaSuite : Advanced Data Integration And Extraction Software

Server software accepts requests for data from client software and returns the results to the client

TOPLink for WebLogic. Whitepaper. The Challenge: The Solution:

And OrACLE In A data MArT APPLICATIOn

CLOUD DISASTER RECOVERY. A Panel Discussion

Client/Server-Architecture

How to Accelerate Merger and Acquisition Synergies

Migration and Building of Data Centers in IBM SoftLayer

CA IDMS Server. Release Notes. r17

INTRODUCING VERITAS BACKUP EXEC SUITE

AcuConnect Versatile Remote COBOL Listener

On BigFix Performance: Disk is King. How to get your infrastructure right the first time! Case Study: IBM Cloud Development - WW IT Services

Seven Decision Points When Considering Containers

Several major software companies including IBM, Informix, Microsoft, Oracle, and Sybase have all released object-relational versions of their

Messaging Service Management and Analysis

CA Test Data Manager Key Scenarios

IBM TS4300 with IBM Spectrum Storage - The Perfect Match -

Using the Cisco ACE Application Control Engine Application Switches with the Cisco ACE XML Gateway

WHITE PAPER. How Deduplication Benefits Companies of All Sizes An Acronis White Paper

Scaling Data Center Application Infrastructure. Gary Orenstein, Gear6

IBM XIV Storage System

<Insert Picture Here> Introducing Oracle WebLogic Server on Oracle Database Appliance

a white paper from Corel Corporation

Configuration changes such as conversion from a single instance to RAC, ASM, etc.

Protecting Mission-Critical Workloads with VMware Fault Tolerance W H I T E P A P E R

Transcription:

Caché and Data Management in the Financial Services Industry Executive Overview One way financial services firms can improve their operational efficiency is to revamp their data management infrastructure. Creating a central repository for data that is used by multiple applications can ensure data consistency and quality across the enterprise, ease integration bottlenecks, and lower the number of failed trades. However, different applications have different database usage patterns. To satisfy them all, any central data repository must: n Support a large number of transactions with high performance n Provide fast response to queries, using up-to-date information n Provide reliable, scalable persistence for data n Be easily accessible by several disparate applications This paper discusses various database technologies and how well they meet the criteria for a central data repository. Traditional relational databases generally do not have the performance needed to support running queries against live transactional data. Also, it can be difficult to connect them with applications written in object-oriented technologies. In-memory databases are fast, but not very scalable. Nor do they provide reliable data persistence. InterSystems Caché, a high-performance, massively scalable, multidimensional database, is an ideal choice for financial services firms looking to streamline the way they manage critical data. Its transactional bit-map indexing technology supports simultaneous querying and transaction processing. Plus, Caché s wideopen connectivity lets it communicate with many disparate applications, including existing data stores, thus allowing organisations to make incremental changes to their data management infrastructure.

Caché and Data Management in the Financial Services Industry Introduction Within many financial services organisations, there is a growing trend towards internal data consolidation. Executives at buy-side and sell-side firms are realising that since many of their applications work off the same body of data, they can improve their operational efficiency by keeping common data in a central data repository. Such an approach will improve the consistency and quality of data used by critical applications. It will alleviate the integration bottlenecks caused by the need to synchronise data between disparate applications within the enterprise. It will enable real-time reporting and analysis using current data. The result will be fewer failed trades, faster performance and better decision-making. However, creating an internal central data repository can be a challenging task. Ideally, a central data repository should combine the attributes of both a transactional database and a data warehouse. It must be able to process a high volume of updates, while at the same time providing fast responses to queries involving large amounts of current and historical data. Another challenge is that different applications may be written in different languages, use different data access protocols, etc. Any central database must be compatible with them all. And finally, the ideal central data repository will allow a modular approach to data consolidation. It will work with existing messaging systems, and with existing data stores, allowing organisations to leverage the investment they have already made in these areas. This paper will discuss several database technologies commonly used within the financial services industry, and how well they meet the requirements of a central data repository. It will explain why Caché, the high-performance, multidimensional database from InterSystems, is the best choice for data consolidation, and give an example of how a financial services firm might use Caché to improve its operational efficiency by transforming its data management infrastructure. The Challenge of Data Consolidation The need to continuously improve operational efficiency, as well as mandates to provide straight-through processing and real-time compliance monitoring, is driving many financial services organisations to revamp their data management infrastructure. Since many applications within buy-side and sell-side firms access and update a common body of data, it makes sense to consolidate that common data in a central repository. Having one source for critical data ensures that the information used by disparate applications within the enterprise is both consistent, and up to date. This will result in fewer failed trades, better performance (because there will be less need to spend time reconciling data between systems), and decisions based on current data. 2

Although consolidation of data has great benefits, some firms are not ready to embark on a large data centralisation programme because of the perceived cost and complexity. Ideally, a central data repository would interface with many systems, including applications that handle risk management, performance measurement, accounting and trading, as well as reference data and corporate action services. In addition, the central repository should provide a fast response to queries from analysts, data integrity users and compliance users. To control costs and maintain smooth operations, many firms will want to take a gradual, modular approach to changing their data management infrastructure. All of the needs mentioned above place large demands on the database technology that powers a financial services firm s central data repository. The choice of database technology for the new data infrastructure becomes critical. It must enable: n Reliable persistence for large amounts of data Data is the lifeblood of financial services organisations. They must be able to store and retrieve huge volumes of data in such a way that it is always available, and recoverable in the case of system failure. n High-performance completion of large volumes of transactions Some of the applications accessing a central data repository need it to be a transactional database. It must allow very fast reading, inserting and updating of data. For these applications performance is key, because the faster the central database can process changes, the more transactions it can complete in a given time. n Quick response to complex queries of current and historical data Some of the applications accessing a central data repository will be running queries for the purposes of trend analysis, reporting and compliance monitoring. Many of these applications will benefit by being able to access up-to-date information. (In the case of compliance monitoring, this is a legally mandated requirement.) Typically, database systems use indexes to shorten response times to queries, so an important attribute of any central database will be the ability to build indexes on rapidly changing transactional data. n Seamless and high-speed interaction with a variety of different applications Different applications within an enterprise may use different technologies, protocols and standards. In order to interact with them, a central database may be called upon to communicate using Java, C++,.NET, XML, Web services, FTP, ODBC, RV, e-jms, or just about any other technology available. However, merely interacting is not enough. Interfaces to disparate applications must be efficient and fast, so as not to slow overall performance. n Easy adaptation and extensibility Since many financial services firms will take an evolutionary approach to revamping their data management infrastructure, it is important that their central repository be easily adaptable, in order to accommodate new data as applications are integrated into the system. Extensibility helps ensure that the central database will be compatible with emerging technologies. In the remainder of this paper, we will discuss how various database technologies stack up against the needs of a central data repository. 3

Relational Databases Relational database technology has been used to provide reliable data persistence for over 25 years. It has the advantage of using easily understood data structures all information is stored in tables using a simple rows-and-columns format. Plus, relational technology has a query language (SQL) that has become the de facto standard throughout the database world. However, relational technology falls short in situations when the very highest performance is needed. Its tabular data structures are not well-suited for storing complex, real-world information. Complex data is split into several interrelated tables, and must be joined before it can be accessed and used. The constant deconstructing and reconstructing of data increases processing overhead, and transactional performance tends to suffer. The same sort of problem occurs when a relational database must interact with object-oriented technologies like Java and C++. In such cases it is necessary to create a map that translates data from the objects used by target applications and the tables used by the relational database. Not only does mapping add to the processing overhead, it also lengthens the development cycle. The performance drawbacks of a relational database may not be apparent in simple applications. But they will become evident if relational technology is used to build a central data repository, especially when an attempt is made to build indexes in order to speed query response times. Relational databases are not designed to act as a transactional system and a data warehouse simultaneously. That being the case, some organisations take the approach of building two central data repositories. One is a transactional system that periodically feeds the data warehouse, against which queries are run. However, in order to maintain adequate performance of the transactional system, downloads to the query system cannot be done very often. Thus, the applications that access the query system are usually working with day-old, even week-old, data. That situation is undesirable at best. And in the face of the new mandate for real-time compliance monitoring, it is unacceptable. In-Memory Databases So-called in-memory or main memory databases are exactly what they sound like. Rather than writing data to disk, they hold it all in local memory. Because there is no reading from or writing to disk, transactions can be processed very quickly. High performance is the main advantage of using in-memory database technology. However, in-memory databases do not provide data persistence, and the mechanisms for data recovery after failure are not practical for most applications. In addition, because they require all data to be resident in memory, these systems cannot scale to handle extremely large volumes of data. In-memory databases may be useful for small applications where transaction speed is of utmost importance, but they are not suitable for building a central data repository. 4

The Caché Multidimensional Database InterSystems Caché is a high-performance multidimensional database. It stores data in flexible, efficient multidimensional arrays, rather than in a simplistic rows-andcolumns format. Caché s unique architecture and capabilities make it an ideal choice for financial services firms that are looking to build a central data repository. High Performance Because it does not constrain data to a rows-and-columns format, Caché can store complex information without deconstructing it. By eliminating the joins (and associated processing overhead) common to relational technology, Caché can provide reliable data persistence and transactional performance that rivals the speed of in-memory databases. Massive Scalability Caché s efficient multidimensional data structures are inherently more compact than relational representations of the same information. Because data is kept intact (that is, not split up into several interrelated tables), transactional performance does not suffer as the size of the database grows. In addition, Caché s Enterprise Cache Protocol (ECP) dramatically reduces network traffic between application servers and the data server in distributed architectures, making it possible to add application servers and support even more users. Caché-based applications can support thousands of concurrent users without sacrificing performance. Transactional Bit-Map Indexing Caché is the only data management technology fast enough to enable the building of indexes on dynamic data. That means it can function simultaneously as a transactional database supporting large volumes of updates, and as a data warehouse providing quick responses to complex queries. Seamless Connectivity Another advantage of Caché s multidimensional data structures is that they can be automatically projected in a variety of forms, providing connectivity to other technologies, without mapping and the processing overhead that goes with it. n Relational connectivity Multidimensional arrays can be projected as two-dimensional arrays (tables), thus allowing Caché to look like a relational database. The Caché database is ODBCand JDBC-compliant, and can be queried using SQL. In addition, Caché s Relational Gateway feature lets it access data stored in your existing relational databases. n Object connectivity Caché s multidimensional arrays are a good match for the complex data structures used by object-oriented technologies. Data can be stored as objects and projected in a variety of formats, including Java, C++, COM, EJB and XML. Plus, any Caché object method can be published as a Web service with just a few clicks of a mouse. n Connectivity with other technologies Caché can communicate via a wide variety of other protocols and standards, including TCP sockets, FTP and SMTP. It is also compatible with popular middleware and messaging technologies. Such wide-open connectivity will allow a Caché-based central data repository to be seamlessly accessed by any application within your organisation. 5

Rapid Development Caché s object model supports concepts such as multiple inheritance and polymorphism that enable the reuse of components and speed application development. Object-oriented development techniques also make any Caché-based system easily adaptable and extensible. A Sample Architecture Figure 1 gives an example of how Caché might be used as a central data repository for an asset management firm. It shows Caché acting as a transactional data warehouse that supports STP efforts, risk management, performance measurement, compliance and trading activities. A data infrastructure like the one shown will provide better data consistency and quality, and enhance operational efficiency. Data Integrity Analysts Analysts Compliance Users SQL Queries, Manual Overrides, etc Load Balancing of User Requests Application Server Application Server Application Server Shared Cache Shared Cache Shared Cache Market Data Trades Real-time Messages Data Server Validation Logic XML Messages Corporate Actions Provider Reference Data Service Trading System Access to Historical Trades and Positions Real-time Updates via TCP Caché Real-time Database (Trades, Positions, Security Master File, Prices) Caché Relational Gateway Java Binding C++ Binding Performance System Risk Management System Accounting System Data Warehouse Other Operational Databases FIGURE 1: Data Management Infrastructure with Caché 6

Figure 1 shows the end result of what will, in most cases, be an evolutionary process of linking applications to a central database. For this example, here is a list of the various steps taken by our hypothetical asset management firm to gradually revamp the firm s data infrastructure: 1. The risk management data model was implemented in Caché. This was accomplished by importing the DDL from the existing risk management system and using it to define data structures within Caché. That done, some modifications were made, taking advantage of Caché s object technology to improve the data model. Then historical reference and position data were migrated into Caché. Caché s C++ object binding was used to communicate natively with the risk management system. 2. The performance measurement system was connected to the central repository. This required some further modifications to the data model, a task made easy by using inheritance extending the classes created during step 1. In this hypothetical example, Java binding was required to allow the central database to communicate with the performance system. 3. Reference data services and corporate action services were directed to feed the Caché database directly. Thus the Caché repository became the master copy for all securities reference data, corporate actions and securities prices (which were fed from the trading system). Batch feeds to the accounting system could be turned off. (Caché s Relational Gateway was used to periodically update the accounting system. In addition, the Relational Gateway allowed the central database to pull information from the existing data warehouse and other operational systems, as needed.) 4. The trading system was connected to the central data repository in real time. This allowed the running of compliance queries for all activity in the fixed income and equities systems. Caché s ECP was used to distribute the processing load created by a large number of analysts, data integrity users and compliance users. Conclusion Financial services organisations can improve their data consistency and operational efficiency by changing their data management infrastructure so that critical applications access and update the same central data repository. To be effective, a central database must have certain capabilities, chiefly the ability to act simultaneously as a transactional database and a data warehouse. It must also provide high-performance connectivity to disparate applications that may be based on a wide variety of technologies, protocols and standards. Of the several database technologies commonly used in the financial industry, only Caché, the multidimensional database from InterSystems, has the performance, scalability, flexibility and reliability to meet the rigorous demands of a central data repository. 7

InterSystems InterSystems House 70 Tangier Lane, Eton, Windsor Berkshire, SL4 6BB, UK Tel: +44 (0)1753 855450 Fax: +44 (0)1753 855290 Email: info@intersystems.com InterSystems.co.uk/FS InterSystems Caché is a registered trademark of InterSystems Corporation. Other product names are trademarks of their respective vendors. Copyright 2012 InterSystems Corporation. All rights reserved. 100539