Strategic Briefing Paper Big Data

Size: px
Start display at page:

Download "Strategic Briefing Paper Big Data"

Transcription

1 Strategic Briefing Paper Big Data The promise of Big Data is improved competitiveness, reduced cost and minimized risk by taking better decisions. This requires affordable solution architectures which are able to react in real time and keep processing time constant while volumes are increasing. Fujitsu provides optimized Big Data infrastructure solutions based on a blend of various technologies, including hardware, software and services. Contents Some dreams of today s organizations 2 Business Intelligence for data-based decision support 2 Today s demands are different 2 What Big Data is about 2 Why traditional infrastructure approaches don t work 2 Distributed Parallel Processing 3 In-Memory Databases (IMDB) 3 Disk array with In-Memory Data Grid (IMDG) 3 NoSQL databases 4 Complex Event Processing (CEP) 4 Big Data solution architecture 5 Big Data is not just about infrastructure 6 How Fujitsu can help 6 Summary 6 Page 1 of 6

2 Some dreams of today s organizations Let us start with some dreams that today s organizations might dream when thinking about ways to improve their business: What if you could predict what will happen in terms of trends, customer behavior, or business opportunities? What if you could always take the best decisions? What if you could accelerate your decisions? What if you could have critical actions initiated automatically? What if you could fully understand the root cause of issues or costs? What if you could skip useless activities? What if you could quantify and minimize risks? What Big Data is about Big Data is characterized by large data volumes, of a variety of data types, coming from versatile data sources, which are generated at a high velocity and need to be analyzed at the same high velocity too, in order to create a high value for the organization and make the dreams we started with come true. Contemplating such questions, many managers immediately imagine the opportunities for their business. However, are these only nice dreams, or is there a chance that these dreams can come true? Business Intelligence for data-based decision support Remember the days when Business Intelligence emerged. People had recognized that utilizing data in an intelligent way could support decisions based on real facts rather than intuition, thus helping improve operations, minimize risk, increase revenue and gain competitive advantage. However, people also realized that data in its original form was usually of low value. Therefore structured data in the Gigabytes to Terabytes range was collected from few internal data sources, mainly relational databases, and then consolidated and transformed into a form convenient for analysis in order to discover relations, patterns and principles, to finally find the real value. Usually people analyzed periodically what had happened in the past. And it did not matter whether such an analytics job took hours or days. Today s demands are different Meanwhile the situation has changed tremendously. It is clearly understood that considering structured data from just a few relational databases means that most of the valuable data out there would remain untapped. Versatile internal and external data sources such as databases, text, audio and video files, IT logs, web logs, social media and the millions and billions of sensors which are built in everywhere can provide an enormous value, too. We are talking about large but ever increasing volumes of in the range of many Terabytes to Petabytes. Especially new sources provide data which rarely is structured, but rather unstructured, semi-structured and poly-structured. Very often this data is continuously generated at a very high velocity, and the same high velocity is also demanded when it comes to analyzing the data in real-time to create value for the organization. Getting the right insight and the desired analytics results in real-time opens up numerous new business opportunities for every organization, no matter to which industry it belongs. Why traditional infrastructure approaches don t work One of the biggest challenges with Big Data is to keep processing time constant while data volumes increase. This in turn has a strong impact on the infrastructure for Big Data. Big Data includes data at rest and data in motion. In a first step we will concentrate on data at rest. In traditional Business Intelligence solutions, the consolidated data for analytics is stored in a retrieval-optimized relational database, denoted as data warehouse. Relational databases are based on a rigid schema, which means that unstructured and semi-structured data would require some pre-processing. The tables resulting from pre-processing are often huge but sparse. This in turn means a high amount of metadata, high storage capacities and slow access to data. Furthermore, in a relational database, data is stored in rows. This is well-suited for Online Transaction Processing (OLTP), but when it comes to OLAP (Online Analytical Processing), much irrelevant data has to be read. This is because only certain information in certain columns is relevant for the queries. Why not use a powerful server and scale it up when data size increases? No matter how powerful your server might be, there will always be a hard limit for each resource type and therefore for the overall performance of the server. For sure, in the future the limits will move upwards, but the totality of all data involved in your analysis might grow much faster. Doing the right things right by predicting ad-hoc what is going to happen in the future is the weapon for organizations in today s competitive markets. This is exactly what Big Data is about. If scale up is not an option, the question remains about relational databases and server scale out. Sharing the database among several servers, the storage connections might prove to be a bottleneck when accessing the database. Likewise, the effort to coordinate the access to shared data will increase with the number of database servers. This will lead to decreasing server efficiency and limit parallelization. Page 2 of 6

3 Consequently, everything you try to improve the situation, be it scale up or scale out in conjunction with your relational database, would be time-consuming and expensive and far away from fulfilling real-time demands. Results would be returned too late and insights could be already obsolete when being presented to the user. For sure, you could think about building distinct databases and split up your analytics tasks. However doing so, you would create disconnected data silos and conduct distinct disconnected analytics which would not give you the comprehensive insight that you expect. Due to all limitations mentioned, new approaches are required, which can really keep processing time constant, while data volumes increase. Distributed Parallel Processing The idea of Distributed Parallel Processing is to distribute the data and thus the I/O load across the local disks of a server cluster and move the computing tasks exactly to the server nodes where the data resides. It is beyond all questions that the access to data on disk storage systems or even AFA (All-Flash-Arrays) providing highest I/O performance and low latency can never be as fast as if data are resident in main memory, and hence closer to the applications. That s why for real-time demands the distilled essence of the transformed data is consolidated into a fast responding in-memory platform. In-Memory Databases (IMDB) An In-Memory Database (IMDB) is characterized by loading its data entirely along with the database applications into the main memory of a server or server cluster to conduct rapid analysis. For analytics tasks, I/O is totally eliminated, because no disks are involved. Disks are just needed for snapshots of the resident data and change logs, to counteract the possible loss of memory content, e.g. after power outage. Fast recovery from server failure can be also achieved by mirroring the IMDB. A relational IMDB is the solution of choice, if several independent applications access the data from different viewpoints. The main limiting factor of an IMDB is the data size, because data size is limited by the overall memory capacity in the entire server cluster. Everything can then happen in parallel, nothing is shared; and therefore scalability is basically unlimited. You may start small, and whenever capacity or performance is not sufficient any more, you add more servers. Fault-tolerance can be achieved by replicating the same data to several nodes. Using cheap standard servers with Hadoop Open Source Software (Distributed File System, MapReduce and more) makes distributed parallel processing affordable. Distributed Parallel Processing is a great method to deal with large amounts of data fast and efficiently. But it does not support analysis in real-time. Therefore, in many cases, Distributed Parallel Processing is just used to transform data into a shape which is more appropriate for ad-hoc queries. The result of this transformation is typically much smaller than the initial data volume, but it contains all essential information for the respective use case. Thus, it may be seen as the distilled essence of the overall data. Disk array with In-Memory Data Grid (IMDG) If you organize the distilled essence of your transformed data on disk storage, an IMDG (In-Memory Data Grid) between disk storage and applications is worth considering. An IMDG is a memory-resident cache which can be distributed across several servers. By reducing the I/O load, it can be avoided that I/O becomes the bottleneck when a real-time analysis is intended. An IMDG can be applied, even if the size of the transformed data on disk exceeds the overall available main memory. Moreover, an IMDG can manage any data object which typically fits to the application semantics much better than a relational table. Combined with indexing and search, the IMDG becomes an in-memory object store, which has efficiency and performance advantages compared to IMDB. Data replication to multiple servers ensures data consistency and quick failover after a server crash. Moreover, using solid state disks in the individual servers as a persistence layer prevent data loss after power outage. Page 3 of 6

4 An IMDG is a great choice, in particular, if its data is not subject to a high update frequency, if the applications accessing the data can be used as they are, or if the skills are available to adapt the application in order to fully exploit the potential benefits of an IMDG. Of course it is also imaginable that Distributed Parallel Processing is used to transform the data, but the transformed data remain in the distributed Big Data store. An IMDG between the analytics applications and the distributed Big Data store will reduce I/O, accelerate analytics and can enable results of queries in real-time, too. Both concepts, IMDB and IMDG are able to scale-up and scale-out. Complex Event Processing (CEP) While the focus of the previous sections was on data at rest, we are now going to have a closer look at data in motion. Event streams are generated continuously at a very high velocity, for instance by sensors. These event streams have to be collected and analyzed on the fly in real-time. Depending on their relevance, events are filtered and then correlated. The analysis is based on a set of pre-defined rules that include a condition and an action. If the condition (which may be a correlation of logical and timely aspects) is fulfilled, the action will be triggered in real-time. This is how a CEP engine works. NoSQL databases Relational databases are best suited for structured data of a limited size. Big Data exceeds the size limits and includes lots of data which are not structured. NoSQL (Not only SQL) databases provide the advantages of database systems, for instance the easy query options, without the limitations of relational databases. They are based on a flexible data model, are designed to be distributed and for scale-out. There are various implementations of NoSQL databases for different use cases. The best-known examples are key value stores, document stores, graph databases and column-oriented databases. In a column-oriented database, data is stored in columns rather than rows. Thus, for queries only the columns are of interest have to be read. This reduces I/O and accelerates analytics. As a certain column contains only one type of data, and quite often there are only few distinct values per column, it is possible to store only these few distinct values and references to these values. Doing so, you achieve an extremely high compression. Typical compression factors are in the range from 10 to 50. This helps reduce storage capacities and consequently storage costs, too. In order to avoid time-consuming I/O, the collecting tank for the event streams is organized in-memory. In-memory data grids (IMDG) help meet higher main memory capacity demands. Depending on the intensity of the event streams, and/or the multitude of rules to be processed, it can be helpful to distribute CEP engines across several servers to enable parallel processing. The same applies for complex queries which cannot be executed by a single server. In this case, the complex queries will be split and distributed, too. Page 4 of 6

5 Big Data solution architecture With all the individual concepts, we are now in a position to present the overall Big Data solution architecture. Let us just give a brief summary of what is shown in the figure below. Data is extracted from versatile data sources. In addition to the mainly structured data from transactional databases, Data Warehouses and in-memory databases (IMDB), unstructured, semi-structured and poly-structured data is extracted and collected from versatile other sources, such as text, audio and video files, IT logs, web logs and sensors. Data at rest is loaded into the Distributed Data Store. Distributed Parallel Processing based on the MapReduce algorithm is used to transform data into a usable form and for analysis. The distilled essence of the transformed data can be exported to a SQL or NoSQL database management system. If a traditional disk storage array does not provide the required I/O performance when accessing the SQL database, a fast storage system, as for instance an AFA can be used to increase throughput and reduce latency, thus accelerating I/O. Real-time analysis is enabled by in-memory data platforms. An IMDB will be the fastest option, especially if new queries are created again and again. If the size of the distilled essence exceeds the totally available main memory capacity, a disk array with an IMDG in front can be used to accelerate the analytics applications, especially if data generated once is often read. Another advantage of IMDG is that is applicable to any data objects. Of course, an IMDG will also help reduce I/O, if the analysis is directly applied to the Distributed Data store. Event streams generated at a high velocity are collected in the data store of a CEP (Complex Event Processing) engine which in turn will look after analyzing the event streams and initiating the respective actions in real-time. As CEP is extremely time-critical, the data store is memory-resident, and often implemented as an IMDG. Depending on the use case, results of CEP can be forwarded into the Distributed Data Store or the distilled essence for further processing, in the same way as data from the Distributed Data Store is sometimes used for CEP. Page 5 of 6

6 Big Data is not just about infrastructure Infrastructure is a very important aspect of Big Data. However, it is not just about infrastructure, it is also about data and processes. This includes a lot of questions which need to be answered: Which data from which data sources to involve? How to transform data into high quality information? What to look for? Which questions to ask? What to do with the answers? Which analytic methods? How to visualize results? Which tools to use, and how to use them? What about data retention and deletion? To find the right answers, analytic skills and a deep knowledge of data, tools and intended results are required. How Fujitsu can help Regarding Big Data, Fujitsu covers the aspects of data, processes and infrastructure. In an end-to-end service approach, we consult our customers in terms of what can be achieved by Big Data and jointly define the roadmap to Big Data. We schedule and prioritize the use cases, design and implement the solution, we integrate it into the existing IT landscape, and we maintain the overall solution. What is more, Fujitsu stands for flexibility and choice in terms of sourcing options. If a customer wants to be relieved from routine tasks in order to focus his resources on other strategic projects, we will take over the operation. This is exactly what Fujitsu Managed Infrastructure Services are about. In addition, Big Data services are also delivered from the Fujitsu Cloud. Integrated Systems lower the entry barrier and enable a fast time-to-value. Especially worth mentioning are FUJITSU Integrated System PRIMEFLEX for Hadoop (a Hadoop cluster enabling even business users to conduct analytics in an easy manner) and FUJITSU Integrated System PRIMEFLEX for SAP HANA (SAP s IMDB solution). Attractive financing options enable the introduction of Big Data solutions without any pre-investment. Summary Big Data is more than large data sets. It is about large volumes of various types from versatile sources generated at a high velocity. The objective of analyzing Big Data in real-time is to predict things that are going to happen in future, to take the right decisions, and to initiate the respective actions, which will help to improve operations, reduce costs, minimize risks and create value for the organization, i.e. to gain real competitive advantage. The challenge regarding Big Data infrastructures is to keep processing time constant while data volumes increase and keeping it affordable at the same time. Therefore new approaches, such as Distributed Parallel Processing, IMDB, IMDG, AFA and CEP are needed. Fujitsu is a one-stop shop for Big Data where organizations can get optimized infrastructure solutions based on a blend of various technologies, including hardware, software and services all from a single source. Fujitsu supports all relevant Big Data infrastructure concepts, and will thus always elaborate the optimum combination of technologies according to the customer-specific situation and requirements. The well-designed solution stack includes proven servers and storage systems from Fujitsu. Software is either from Fujitsu itself, open source, or from leading independent software vendors with whom we have a close partnership. Contact FUJITSU Technology Solutions GmbH Address: Mies-van-der-Rohe-Strasse 8, Munich, Germany Website: WW EN ƒ Copyright 2016 Fujitsu, the Fujitsu logo are trademarks or registered trademarks of Fujitsu Limited in Japan and other countries. Other company, product and service names may be trademarks or registered trademarks of their respective owners. Technical data subject to modification and delivery subject to availability. Any liability that the data and illustrations are complete, actual or correct is excluded. Designations may be trademarks and/or copyrights of the respective manufacturer, the use of which by third parties for their own purposes may infringe the rights of such owner. Page 6 of 6

White paper Solution Approaches for Big Data

White paper Solution Approaches for Big Data White paper Solution Approaches for Big Data Big Data becomes relevant for more and more organizations. They move to new fields of applications where large volumes of data are automatically and continuously

More information

When, Where & Why to Use NoSQL?

When, Where & Why to Use NoSQL? When, Where & Why to Use NoSQL? 1 Big data is becoming a big challenge for enterprises. Many organizations have built environments for transactional data with Relational Database Management Systems (RDBMS),

More information

ATA DRIVEN GLOBAL VISION CLOUD PLATFORM STRATEG N POWERFUL RELEVANT PERFORMANCE SOLUTION CLO IRTUAL BIG DATA SOLUTION ROI FLEXIBLE DATA DRIVEN V

ATA DRIVEN GLOBAL VISION CLOUD PLATFORM STRATEG N POWERFUL RELEVANT PERFORMANCE SOLUTION CLO IRTUAL BIG DATA SOLUTION ROI FLEXIBLE DATA DRIVEN V ATA DRIVEN GLOBAL VISION CLOUD PLATFORM STRATEG N POWERFUL RELEVANT PERFORMANCE SOLUTION CLO IRTUAL BIG DATA SOLUTION ROI FLEXIBLE DATA DRIVEN V WHITE PAPER Create the Data Center of the Future Accelerate

More information

ACCELERATE YOUR ANALYTICS GAME WITH ORACLE SOLUTIONS ON PURE STORAGE

ACCELERATE YOUR ANALYTICS GAME WITH ORACLE SOLUTIONS ON PURE STORAGE ACCELERATE YOUR ANALYTICS GAME WITH ORACLE SOLUTIONS ON PURE STORAGE An innovative storage solution from Pure Storage can help you get the most business value from all of your data THE SINGLE MOST IMPORTANT

More information

Data Center Management and Automation Strategic Briefing

Data Center Management and Automation Strategic Briefing Data Center and Automation Strategic Briefing Contents Why is Data Center and Automation (DCMA) so important? 2 The Solution Pathway: Data Center and Automation 2 Identifying and Addressing the Challenges

More information

Combine Native SQL Flexibility with SAP HANA Platform Performance and Tools

Combine Native SQL Flexibility with SAP HANA Platform Performance and Tools SAP Technical Brief Data Warehousing SAP HANA Data Warehousing Combine Native SQL Flexibility with SAP HANA Platform Performance and Tools A data warehouse for the modern age Data warehouses have been

More information

UNLEASHING THE VALUE OF THE TERADATA UNIFIED DATA ARCHITECTURE WITH ALTERYX

UNLEASHING THE VALUE OF THE TERADATA UNIFIED DATA ARCHITECTURE WITH ALTERYX UNLEASHING THE VALUE OF THE TERADATA UNIFIED DATA ARCHITECTURE WITH ALTERYX 1 Successful companies know that analytics are key to winning customer loyalty, optimizing business processes and beating their

More information

Progress DataDirect For Business Intelligence And Analytics Vendors

Progress DataDirect For Business Intelligence And Analytics Vendors Progress DataDirect For Business Intelligence And Analytics Vendors DATA SHEET FEATURES: Direction connection to a variety of SaaS and on-premises data sources via Progress DataDirect Hybrid Data Pipeline

More information

Embedded Technosolutions

Embedded Technosolutions Hadoop Big Data An Important technology in IT Sector Hadoop - Big Data Oerie 90% of the worlds data was generated in the last few years. Due to the advent of new technologies, devices, and communication

More information

Four Steps to Unleashing The Full Potential of Your Database

Four Steps to Unleashing The Full Potential of Your Database Four Steps to Unleashing The Full Potential of Your Database This insightful technical guide offers recommendations on selecting a platform that helps unleash the performance of your database. What s the

More information

White paper Why PRIMEFLEX Attracts Service Providers

White paper Why PRIMEFLEX Attracts Service Providers White paper Why PRIMEFLEX Attracts Service Providers Building data center infrastructures is increasingly complex, error-prone, time-consuming, risky and expensive. FUJITSU Integrated System PRIMEFLEX

More information

Fast Innovation requires Fast IT

Fast Innovation requires Fast IT Fast Innovation requires Fast IT Cisco Data Virtualization Puneet Kumar Bhugra Business Solutions Manager 1 Challenge In Data, Big Data & Analytics Siloed, Multiple Sources Business Outcomes Business Opportunity:

More information

IN-MEMORY DATA FABRIC: Real-Time Streaming

IN-MEMORY DATA FABRIC: Real-Time Streaming WHITE PAPER IN-MEMORY DATA FABRIC: Real-Time Streaming COPYRIGHT AND TRADEMARK INFORMATION 2014 GridGain Systems. All rights reserved. This document is provided as is. Information and views expressed in

More information

White Paper Flash Forward a guide to find the right path to your storage solution

White Paper Flash Forward a guide to find the right path to your storage solution White Paper Flash Forward a guide to find the right path to your storage solution 2nd updated edition Content Digitization increases data traffic 2 Starting point: Choosing the scalability approach 3 Architectures

More information

DATACENTER SERVICES DATACENTER

DATACENTER SERVICES DATACENTER SERVICES SOLUTION SUMMARY ALL CHANGE React, grow and innovate faster with Computacenter s agile infrastructure services Customers expect an always-on, superfast response. Businesses need to release new

More information

5 Fundamental Strategies for Building a Data-centered Data Center

5 Fundamental Strategies for Building a Data-centered Data Center 5 Fundamental Strategies for Building a Data-centered Data Center June 3, 2014 Ken Krupa, Chief Field Architect Gary Vidal, Solutions Specialist Last generation Reference Data Unstructured OLTP Warehouse

More information

Introduction to the Active Everywhere Database

Introduction to the Active Everywhere Database Introduction to the Active Everywhere Database INTRODUCTION For almost half a century, the relational database management system (RDBMS) has been the dominant model for database management. This more than

More information

Shine a Light on Dark Data with Vertica Flex Tables

Shine a Light on Dark Data with Vertica Flex Tables White Paper Analytics and Big Data Shine a Light on Dark Data with Vertica Flex Tables Hidden within the dark recesses of your enterprise lurks dark data, information that exists but is forgotten, unused,

More information

Big Data Technology Ecosystem. Mark Burnette Pentaho Director Sales Engineering, Hitachi Vantara

Big Data Technology Ecosystem. Mark Burnette Pentaho Director Sales Engineering, Hitachi Vantara Big Data Technology Ecosystem Mark Burnette Pentaho Director Sales Engineering, Hitachi Vantara Agenda End-to-End Data Delivery Platform Ecosystem of Data Technologies Mapping an End-to-End Solution Case

More information

White Paper Hyper! Hyper? Worthwhile or Nonsense? Hyper-Convergence The End of Classical IT?

White Paper Hyper! Hyper? Worthwhile or Nonsense? Hyper-Convergence The End of Classical IT? White Paper Hyper! Hyper? Worthwhile or Nonsense? Hyper-Convergence The End of Classical IT? Hyper-converged infrastructures are the rising stars in data centers. What is behind this hype, what are its

More information

How to integrate data into Tableau

How to integrate data into Tableau 1 How to integrate data into Tableau a comparison of 3 approaches: ETL, Tableau self-service and WHITE PAPER WHITE PAPER 2 data How to integrate data into Tableau a comparison of 3 es: ETL, Tableau self-service

More information

Disaster Recovery Is A Business Strategy

Disaster Recovery Is A Business Strategy Disaster Recovery Is A Business Strategy A White Paper By Table of Contents Preface Disaster Recovery Is a Business Strategy Disaster Recovery Is a Business Strategy... 2 Disaster Recovery: The Facts...

More information

A Robust, Flexible Platform for Expanding Your Storage without Limits

A Robust, Flexible Platform for Expanding Your Storage without Limits White Paper SUSE Enterprise A Robust, Flexible Platform for Expanding Your without Limits White Paper A Robust, Flexible Platform for Expanding Your without Limits Unlimited Scalability That s Cost-Effective

More information

IBM DB2 BLU Acceleration vs. SAP HANA vs. Oracle Exadata

IBM DB2 BLU Acceleration vs. SAP HANA vs. Oracle Exadata Research Report IBM DB2 BLU Acceleration vs. SAP HANA vs. Oracle Exadata Executive Summary The problem: how to analyze vast amounts of data (Big Data) most efficiently. The solution: the solution is threefold:

More information

Business- Centric Storage

Business- Centric Storage Business- Centric Storage Providing reliable data services more efficiently with FUJITSU Storage ETERNUS ETERNUS Business-Centric Storage Digitization is fundamentally transforming businesses. This transformation

More information

QLogic/Lenovo 16Gb Gen 5 Fibre Channel for Database and Business Analytics

QLogic/Lenovo 16Gb Gen 5 Fibre Channel for Database and Business Analytics QLogic/ Gen 5 Fibre Channel for Database Assessment for Database and Business Analytics Using the information from databases and business analytics helps business-line managers to understand their customer

More information

White paper ETERNUS CS800 Data Deduplication Background

White paper ETERNUS CS800 Data Deduplication Background White paper ETERNUS CS800 - Data Deduplication Background This paper describes the process of Data Deduplication inside of ETERNUS CS800 in detail. The target group consists of presales, administrators,

More information

Distributed Databases: SQL vs NoSQL

Distributed Databases: SQL vs NoSQL Distributed Databases: SQL vs NoSQL Seda Unal, Yuchen Zheng April 23, 2017 1 Introduction Distributed databases have become increasingly popular in the era of big data because of their advantages over

More information

Hierarchy of knowledge BIG DATA 9/7/2017. Architecture

Hierarchy of knowledge BIG DATA 9/7/2017. Architecture BIG DATA Architecture Hierarchy of knowledge Data: Element (fact, figure, etc.) which is basic information that can be to be based on decisions, reasoning, research and which is treated by the human or

More information

SAP IQ Software16, Edge Edition. The Affordable High Performance Analytical Database Engine

SAP IQ Software16, Edge Edition. The Affordable High Performance Analytical Database Engine SAP IQ Software16, Edge Edition The Affordable High Performance Analytical Database Engine Agenda Agenda Introduction to Dobler Consulting Today s Data Challenges Overview of SAP IQ 16, Edge Edition SAP

More information

Was ist dran an einer spezialisierten Data Warehousing platform?

Was ist dran an einer spezialisierten Data Warehousing platform? Was ist dran an einer spezialisierten Data Warehousing platform? Hermann Bär Oracle USA Redwood Shores, CA Schlüsselworte Data warehousing, Exadata, specialized hardware proprietary hardware Introduction

More information

Composite Software Data Virtualization The Five Most Popular Uses of Data Virtualization

Composite Software Data Virtualization The Five Most Popular Uses of Data Virtualization Composite Software Data Virtualization The Five Most Popular Uses of Data Virtualization Composite Software, Inc. June 2011 TABLE OF CONTENTS INTRODUCTION... 3 DATA FEDERATION... 4 PROBLEM DATA CONSOLIDATION

More information

Automate Transform Analyze

Automate Transform Analyze Competitive Intelligence 2.0 Turning the Web s Big Data into Big Insights Automate Transform Analyze Introduction Today, the web continues to grow at a dizzying pace. There are more than 1 billion websites

More information

Oracle GoldenGate for Big Data

Oracle GoldenGate for Big Data Oracle GoldenGate for Big Data The Oracle GoldenGate for Big Data 12c product streams transactional data into big data systems in real time, without impacting the performance of source systems. It streamlines

More information

FUJITSU Server BS2000 SE Series. High-End Multi-OS Platform

FUJITSU Server BS2000 SE Series. High-End Multi-OS Platform FUJITSU Server SE Series High-End Multi-OS Platform Flexibility redefined: The mainframe platform for multiple scenarios «Outstanding performance scalability (both scale up and scale out) paired with extremely

More information

Evolving To The Big Data Warehouse

Evolving To The Big Data Warehouse Evolving To The Big Data Warehouse Kevin Lancaster 1 Copyright Director, 2012, Oracle and/or its Engineered affiliates. All rights Insert Systems, Information Protection Policy Oracle Classification from

More information

Abstract. The Challenges. ESG Lab Review InterSystems IRIS Data Platform: A Unified, Efficient Data Platform for Fast Business Insight

Abstract. The Challenges. ESG Lab Review InterSystems IRIS Data Platform: A Unified, Efficient Data Platform for Fast Business Insight ESG Lab Review InterSystems Data Platform: A Unified, Efficient Data Platform for Fast Business Insight Date: April 218 Author: Kerry Dolan, Senior IT Validation Analyst Abstract Enterprise Strategy Group

More information

Massive Scalability With InterSystems IRIS Data Platform

Massive Scalability With InterSystems IRIS Data Platform Massive Scalability With InterSystems IRIS Data Platform Introduction Faced with the enormous and ever-growing amounts of data being generated in the world today, software architects need to pay special

More information

White paper Why PRIMEFLEX Attracts Service Providers

White paper Why PRIMEFLEX Attracts Service Providers White paper Why PRIMEFLEX Attracts Service Providers Building data center infrastructures is increasingly complex, error-prone, time-consuming, risky and expensive. FUJITSU Integrated System PRIMEFLEX

More information

Oracle NoSQL Database Overview Marie-Anne Neimat, VP Development

Oracle NoSQL Database Overview Marie-Anne Neimat, VP Development Oracle NoSQL Database Overview Marie-Anne Neimat, VP Development June14, 2012 1 Copyright 2012, Oracle and/or its affiliates. All rights Agenda Big Data Overview Oracle NoSQL Database Architecture Technical

More information

Map-Reduce. Marco Mura 2010 March, 31th

Map-Reduce. Marco Mura 2010 March, 31th Map-Reduce Marco Mura (mura@di.unipi.it) 2010 March, 31th This paper is a note from the 2009-2010 course Strumenti di programmazione per sistemi paralleli e distribuiti and it s based by the lessons of

More information

2014 年 3 月 13 日星期四. From Big Data to Big Value Infrastructure Needs and Huawei Best Practice

2014 年 3 月 13 日星期四. From Big Data to Big Value Infrastructure Needs and Huawei Best Practice 2014 年 3 月 13 日星期四 From Big Data to Big Value Infrastructure Needs and Huawei Best Practice Data-driven insight Making better, more informed decisions, faster Raw Data Capture Store Process Insight 1 Data

More information

ScaleArc for SQL Server

ScaleArc for SQL Server Solution Brief ScaleArc for SQL Server Overview Organizations around the world depend on SQL Server for their revenuegenerating, customer-facing applications, running their most business-critical operations

More information

Big Data and Object Storage

Big Data and Object Storage Big Data and Object Storage or where to store the cold and small data? Sven Bauernfeind Computacenter AG & Co. ohg, Consultancy Germany 28.02.2018 Munich Volume, Variety & Velocity + Analytics Velocity

More information

QLogic 16Gb Gen 5 Fibre Channel for Database and Business Analytics

QLogic 16Gb Gen 5 Fibre Channel for Database and Business Analytics QLogic 16Gb Gen 5 Fibre Channel for Database Assessment for Database and Business Analytics Using the information from databases and business analytics helps business-line managers to understand their

More information

Data Protection for Cisco HyperFlex with Veeam Availability Suite. Solution Overview Cisco Public

Data Protection for Cisco HyperFlex with Veeam Availability Suite. Solution Overview Cisco Public Data Protection for Cisco HyperFlex with Veeam Availability Suite 1 2017 2017 Cisco Cisco and/or and/or its affiliates. its affiliates. All rights All rights reserved. reserved. Highlights Is Cisco compatible

More information

CONSOLIDATING RISK MANAGEMENT AND REGULATORY COMPLIANCE APPLICATIONS USING A UNIFIED DATA PLATFORM

CONSOLIDATING RISK MANAGEMENT AND REGULATORY COMPLIANCE APPLICATIONS USING A UNIFIED DATA PLATFORM CONSOLIDATING RISK MANAGEMENT AND REGULATORY COMPLIANCE APPLICATIONS USING A UNIFIED PLATFORM Executive Summary Financial institutions have implemented and continue to implement many disparate applications

More information

Symantec Data Center Transformation

Symantec Data Center Transformation Symantec Data Center Transformation A holistic framework for IT evolution As enterprises become increasingly dependent on information technology, the complexity, cost, and performance of IT environments

More information

Market Report. Scale-out 2.0: Simple, Scalable, Services- Oriented Storage. Scale-out Storage Meets the Enterprise. June 2010.

Market Report. Scale-out 2.0: Simple, Scalable, Services- Oriented Storage. Scale-out Storage Meets the Enterprise. June 2010. Market Report Scale-out 2.0: Simple, Scalable, Services- Oriented Storage Scale-out Storage Meets the Enterprise By Terri McClure June 2010 Market Report: Scale-out 2.0: Simple, Scalable, Services-Oriented

More information

What s New in SAP Sybase IQ 16 Tap Into Big Data at the Speed of Business

What s New in SAP Sybase IQ 16 Tap Into Big Data at the Speed of Business SAP White Paper SAP Database and Technology Solutions What s New in SAP Sybase IQ 16 Tap Into Big Data at the Speed of Business 2013 SAP AG or an SAP affiliate company. All rights reserved. The ability

More information

MODERNISE WITH ALL-FLASH. Intel Inside. Powerful Data Centre Outside.

MODERNISE WITH ALL-FLASH. Intel Inside. Powerful Data Centre Outside. MODERNISE WITH ALL-FLASH Intel Inside. Powerful Data Centre Outside. MODERNISE WITHOUT COMPROMISE In today s lightning-fast digital world, it s critical for businesses to make their move to the Modern

More information

An Introduction to Big Data Formats

An Introduction to Big Data Formats Introduction to Big Data Formats 1 An Introduction to Big Data Formats Understanding Avro, Parquet, and ORC WHITE PAPER Introduction to Big Data Formats 2 TABLE OF TABLE OF CONTENTS CONTENTS INTRODUCTION

More information

WHY BUILDING SECURITY SYSTEMS NEED CONTINUOUS AVAILABILITY

WHY BUILDING SECURITY SYSTEMS NEED CONTINUOUS AVAILABILITY WHY BUILDING SECURITY SYSTEMS NEED CONTINUOUS AVAILABILITY White Paper 2 Why Building Security Systems Need Continuous Availability Always On Is the Only Option. If All Systems Go Down, How Can You React

More information

THE FUTURE OF BUSINESS DEPENDS ON SOFTWARE DEFINED STORAGE (SDS)

THE FUTURE OF BUSINESS DEPENDS ON SOFTWARE DEFINED STORAGE (SDS) THE FUTURE OF BUSINESS DEPENDS ON SOFTWARE DEFINED STORAGE (SDS) How SSDs can fit into and accelerate an SDS strategy SPONSORED BY TABLE OF CONTENTS Introduction 3 An Overview of SDS 4 Achieving the Goals

More information

Big Data The end of Data Warehousing?

Big Data The end of Data Warehousing? Big Data The end of Data Warehousing? Hermann Bär Oracle USA Redwood Shores, CA Schlüsselworte Big data, data warehousing, advanced analytics, Hadoop, unstructured data Introduction If there was an Unwort

More information

The Future of Business Depends on Software Defined Storage (SDS) How SSDs can fit into and accelerate an SDS strategy

The Future of Business Depends on Software Defined Storage (SDS) How SSDs can fit into and accelerate an SDS strategy The Future of Business Depends on Software Defined Storage (SDS) Table of contents Introduction 2 An Overview of SDS 3 Achieving the Goals of SDS Hinges on Smart Hardware Decisions 5 Assessing the Role

More information

Processing Unstructured Data. Dinesh Priyankara Founder/Principal Architect dinesql Pvt Ltd.

Processing Unstructured Data. Dinesh Priyankara Founder/Principal Architect dinesql Pvt Ltd. Processing Unstructured Data Dinesh Priyankara Founder/Principal Architect dinesql Pvt Ltd. http://dinesql.com / Dinesh Priyankara @dinesh_priya Founder/Principal Architect dinesql Pvt Ltd. Microsoft Most

More information

Reasons to Deploy Oracle on EMC Symmetrix VMAX

Reasons to Deploy Oracle on EMC Symmetrix VMAX Enterprises are under growing urgency to optimize the efficiency of their Oracle databases. IT decision-makers and business leaders are constantly pushing the boundaries of their infrastructures and applications

More information

Fujitsu/Fujitsu Labs Technologies for Big Data in Cloud and Business Opportunities

Fujitsu/Fujitsu Labs Technologies for Big Data in Cloud and Business Opportunities Fujitsu/Fujitsu Labs Technologies for Big Data in Cloud and Business Opportunities Satoshi Tsuchiya Cloud Computing Research Center Fujitsu Laboratories Ltd. January, 2012 Overview: Fujitsu s Cloud and

More information

Taming Structured And Unstructured Data With SAP HANA Running On VCE Vblock Systems

Taming Structured And Unstructured Data With SAP HANA Running On VCE Vblock Systems 1 Taming Structured And Unstructured Data With SAP HANA Running On VCE Vblock Systems The Defacto Choice For Convergence 2 ABSTRACT & SPEAKER BIO Dealing with enormous data growth is a key challenge for

More information

Protecting Mission-Critical Application Environments The Top 5 Challenges and Solutions for Backup and Recovery

Protecting Mission-Critical Application Environments The Top 5 Challenges and Solutions for Backup and Recovery White Paper Business Continuity Protecting Mission-Critical Application Environments The Top 5 Challenges and Solutions for Backup and Recovery Table of Contents Executive Summary... 1 Key Facts About

More information

IBM Data Retrieval Technologies: RDBMS, BLU, IBM Netezza, and Hadoop

IBM Data Retrieval Technologies: RDBMS, BLU, IBM Netezza, and Hadoop #IDUG IBM Data Retrieval Technologies: RDBMS, BLU, IBM Netezza, and Hadoop Frank C. Fillmore, Jr. The Fillmore Group, Inc. The Baltimore/Washington DB2 Users Group December 11, 2014 Agenda The Fillmore

More information

Stages of Data Processing

Stages of Data Processing Data processing can be understood as the conversion of raw data into a meaningful and desired form. Basically, producing information that can be understood by the end user. So then, the question arises,

More information

Symantec NetBackup 7 for VMware

Symantec NetBackup 7 for VMware V-Ray visibility into virtual machine protection Overview There s little question that server virtualization is the single biggest game-changing trend in IT today. Budget-strapped IT departments are racing

More information

An Introduction to GPFS

An Introduction to GPFS IBM High Performance Computing July 2006 An Introduction to GPFS gpfsintro072506.doc Page 2 Contents Overview 2 What is GPFS? 3 The file system 3 Application interfaces 4 Performance and scalability 4

More information

BIG DATA TESTING: A UNIFIED VIEW

BIG DATA TESTING: A UNIFIED VIEW http://core.ecu.edu/strg BIG DATA TESTING: A UNIFIED VIEW BY NAM THAI ECU, Computer Science Department, March 16, 2016 2/30 PRESENTATION CONTENT 1. Overview of Big Data A. 5 V s of Big Data B. Data generation

More information

Hybrid Data Platform

Hybrid Data Platform UniConnect-Powered Data Aggregation Across Enterprise Data Warehouses and Big Data Storage Platforms A Percipient Technology White Paper Author: Ai Meun Lim Chief Product Officer Updated Aug 2017 2017,

More information

Downtime Prevention Buyer s Guide. 6 QUESTIONS to help you choose the right availability protection for your applications

Downtime Prevention Buyer s Guide. 6 QUESTIONS to help you choose the right availability protection for your applications Downtime Prevention Buyer s Guide 6 QUESTIONS to help you choose the right availability protection for your applications Question 6 questions to help you choose the right availability protection for your

More information

E X E C U T I V E B R I E F

E X E C U T I V E B R I E F Create a Better Way to Work: OpenText Suite 16 & OpenText Cloud 16 Over the next five years, executives expect digital disruption to displace four out of 10 incumbents or 40 percent of established market

More information

Best Practices in Securing a Multicloud World

Best Practices in Securing a Multicloud World Best Practices in Securing a Multicloud World Actions to take now to protect data, applications, and workloads We live in a multicloud world. A world where a multitude of offerings from Cloud Service Providers

More information

Dell EMC All-Flash solutions are powered by Intel Xeon processors. Learn more at DellEMC.com/All-Flash

Dell EMC All-Flash solutions are powered by Intel Xeon processors. Learn more at DellEMC.com/All-Flash N O I T A M R O F S N A R T T I L H E S FU FLA A IN Dell EMC All-Flash solutions are powered by Intel Xeon processors. MODERNIZE WITHOUT COMPROMISE I n today s lightning-fast digital world, your IT Transformation

More information

In-Memory Data Management

In-Memory Data Management In-Memory Data Management Martin Faust Research Assistant Research Group of Prof. Hasso Plattner Hasso Plattner Institute for Software Engineering University of Potsdam Agenda 2 1. Changed Hardware 2.

More information

Making hybrid IT simple with Capgemini and Microsoft Azure Stack

Making hybrid IT simple with Capgemini and Microsoft Azure Stack Making hybrid IT simple with Capgemini and Microsoft Azure Stack The significant evolution of cloud computing in the last few years has encouraged IT leaders to rethink their enterprise cloud strategy.

More information

Executive brief Create a Better Way to Work: OpenText Release 16

Executive brief Create a Better Way to Work: OpenText Release 16 Executive brief Create a Better Way to Work: OpenText Release 16 Over the next five years, executives expect digital disruption to displace four out of 10 incumbents or 40 percent of established market

More information

Oracle Database Exadata Cloud Service Exadata Performance, Cloud Simplicity DATABASE CLOUD SERVICE

Oracle Database Exadata Cloud Service Exadata Performance, Cloud Simplicity DATABASE CLOUD SERVICE Oracle Database Exadata Exadata Performance, Cloud Simplicity DATABASE CLOUD SERVICE Oracle Database Exadata combines the best database with the best cloud platform. Exadata is the culmination of more

More information

Practical Strategies For High Performance SQL Server High Availability

Practical Strategies For High Performance SQL Server High Availability Practical Strategies For High Performance SQL Server High Availability Jason Aw, Strategic Business Development SIOS Technology Join 3 question poll for lucky draw https://www.surveymonkey.com/r/8hmmg3n

More information

An Oracle White Paper April 2010

An Oracle White Paper April 2010 An Oracle White Paper April 2010 In October 2009, NEC Corporation ( NEC ) established development guidelines and a roadmap for IT platform products to realize a next-generation IT infrastructures suited

More information

Enterprise Data Architect

Enterprise Data Architect Enterprise Data Architect Position Summary Farmer Mac maintains a considerable repository of financial data that spans over two decades. Farmer Mac is looking for a hands-on technologist and data architect

More information

BUSTED! 5 COMMON MYTHS OF MODERN INFRASTRUCTURE. These Common Misconceptions Could Be Holding You Back

BUSTED! 5 COMMON MYTHS OF MODERN INFRASTRUCTURE. These Common Misconceptions Could Be Holding You Back BUSTED! 5 COMMON MYTHS OF MODERN INFRASTRUCTURE These Common Misconceptions Could Be Holding You Back 2 IT Is Facing a New Set of Challenges As technology continues to evolve, IT must adjust to changing

More information

Virtualizing SQL Server 2008 Using EMC VNX Series and VMware vsphere 4.1. Reference Architecture

Virtualizing SQL Server 2008 Using EMC VNX Series and VMware vsphere 4.1. Reference Architecture Virtualizing SQL Server 2008 Using EMC VNX Series and VMware vsphere 4.1 Copyright 2011, 2012 EMC Corporation. All rights reserved. Published March, 2012 EMC believes the information in this publication

More information

How to Protect SAP HANA Applications with the Data Protection Suite

How to Protect SAP HANA Applications with the Data Protection Suite White Paper Business Continuity How to Protect SAP HANA Applications with the Data Protection Suite As IT managers realize the benefits of in-memory database technology, they are accelerating their plans

More information

SIEM Solutions from McAfee

SIEM Solutions from McAfee SIEM Solutions from McAfee Monitor. Prioritize. Investigate. Respond. Today s security information and event management (SIEM) solutions need to be able to identify and defend against attacks within an

More information

ELTMaestro for Spark: Data integration on clusters

ELTMaestro for Spark: Data integration on clusters Introduction Spark represents an important milestone in the effort to make computing on clusters practical and generally available. Hadoop / MapReduce, introduced the early 2000s, allows clusters to be

More information

FUJITSU Backup as a Service Rapid Recovery Appliance

FUJITSU Backup as a Service Rapid Recovery Appliance FUJITSU Backup as a Service Rapid Recovery Appliance The unprecedented growth of business data The role that data plays in today s organisation is rapidly increasing in importance. It guides and supports

More information

FIVE BEST PRACTICES FOR ENSURING A SUCCESSFUL SQL SERVER MIGRATION

FIVE BEST PRACTICES FOR ENSURING A SUCCESSFUL SQL SERVER MIGRATION FIVE BEST PRACTICES FOR ENSURING A SUCCESSFUL SQL SERVER MIGRATION The process of planning and executing SQL Server migrations can be complex and risk-prone. This is a case where the right approach and

More information

Automated Netezza Migration to Big Data Open Source

Automated Netezza Migration to Big Data Open Source Automated Netezza Migration to Big Data Open Source CASE STUDY Client Overview Our client is one of the largest cable companies in the world*, offering a wide range of services including basic cable, digital

More information

For DBAs and LOB Managers: Using Flash Storage to Drive Performance and Efficiency in Oracle Databases

For DBAs and LOB Managers: Using Flash Storage to Drive Performance and Efficiency in Oracle Databases For DBAs and LOB Managers: Using Flash Storage to Drive Performance and Efficiency in Oracle Databases WHITE PAPER Table of Contents The Benefits of Flash Storage for Oracle Databases...2 What DBAs Need

More information

Data Warehouses Chapter 12. Class 10: Data Warehouses 1

Data Warehouses Chapter 12. Class 10: Data Warehouses 1 Data Warehouses Chapter 12 Class 10: Data Warehouses 1 OLTP vs OLAP Operational Database: a database designed to support the day today transactions of an organization Data Warehouse: historical data is

More information

Pervasive PSQL Summit v10 Highlights Performance and analytics

Pervasive PSQL Summit v10 Highlights Performance and analytics Pervasive PSQL Summit v10 Highlights Performance and analytics A Monash Information Services Bulletin by Curt A. Monash, PhD. September, 2007 Sponsored by: Pervasive PSQL Version 10 Highlights Page 2 PSQL

More information

The Hadoop Paradigm & the Need for Dataset Management

The Hadoop Paradigm & the Need for Dataset Management The Hadoop Paradigm & the Need for Dataset Management 1. Hadoop Adoption Hadoop is being adopted rapidly by many different types of enterprises and government entities and it is an extraordinarily complex

More information

CenturyLink for Microsoft

CenturyLink for Microsoft Strategic Partner Alliances CenturyLink for Microsoft EMPOWER REACH AGILITY 2017 CenturyLink. All Rights Reserved. The CenturyLink mark, pathways logo and certain CenturyLink product names are the property

More information

Solace JMS Broker Delivers Highest Throughput for Persistent and Non-Persistent Delivery

Solace JMS Broker Delivers Highest Throughput for Persistent and Non-Persistent Delivery Solace JMS Broker Delivers Highest Throughput for Persistent and Non-Persistent Delivery Java Message Service (JMS) is a standardized messaging interface that has become a pervasive part of the IT landscape

More information

Assignment 5. Georgia Koloniari

Assignment 5. Georgia Koloniari Assignment 5 Georgia Koloniari 2. "Peer-to-Peer Computing" 1. What is the definition of a p2p system given by the authors in sec 1? Compare it with at least one of the definitions surveyed in the last

More information

Investing in a Better Storage Environment:

Investing in a Better Storage Environment: Investing in a Better Storage Environment: Best Practices for the Public Sector Investing in a Better Storage Environment 2 EXECUTIVE SUMMARY The public sector faces numerous and known challenges that

More information

Enterprise Architectures The Pace Accelerates Camberley Bates Managing Partner & Analyst

Enterprise Architectures The Pace Accelerates Camberley Bates Managing Partner & Analyst Enterprise Architectures The Pace Accelerates Camberley Bates Managing Partner & Analyst Change is constant in IT.But some changes alter forever the way we do things Inflections & Architectures Solid State

More information

Experience server virtualization

Experience server virtualization Experience server virtualization Find the fast track to greater flexibility and lower costs Intel Xeon processor. As business priorities surge ahead, are you keeping pace? Today s IT infrastructures must

More information

Building a Data Strategy for a Digital World

Building a Data Strategy for a Digital World Building a Data Strategy for a Digital World Jason Hunter, CTO, APAC Data Challenge: Pushing the Limits of What's Possible The Art of the Possible Multiple Government Agencies Data Hub 100 s of Service

More information

BS2000/OSD DAB Disk Access Buffer Intelligent Caching with AutoDAB

BS2000/OSD DAB Disk Access Buffer Intelligent Caching with AutoDAB BS2000/OSD DAB Disk Access Buffer Intelligent Caching with AutoDAB Issue June 2009 Pages 7 To cache or not to cache? That is not the question! Business-critical computing is typified by high performance

More information

Veritas Backup Exec. Powerful, flexible and reliable data protection designed for cloud-ready organizations. Key Features and Benefits OVERVIEW

Veritas Backup Exec. Powerful, flexible and reliable data protection designed for cloud-ready organizations. Key Features and Benefits OVERVIEW Veritas Backup Exec Powerful, flexible and reliable data protection designed for cloud-ready organizations. OVERVIEW Veritas Backup Exec is the backup solution without barriers, delivered your way. You

More information

Instant evolution in the age of digitization. Turn technology into your competitive advantage

Instant evolution in the age of digitization. Turn technology into your competitive advantage Instant evolution in the age of digitization Turn technology into your competitive advantage It s easy to underestimate how far we ve come in such a relatively short space of time, and how much the world

More information