LOW LATENCY DATA DISTRIBUTION IN CAPITAL MARKETS: GETTING IT RIGHT

Similar documents
Top 4 considerations for choosing a converged infrastructure for private clouds

Dell EMC Hyper-Converged Infrastructure

Enterprise Case Study

Making hybrid IT simple with Capgemini and Microsoft Azure Stack

Massive Scalability With InterSystems IRIS Data Platform

Dell EMC Hyper-Converged Infrastructure

New Approach to Unstructured Data

How Microsoft IT Reduced Operating Expenses Using Virtualization

Oracle and Tangosol Acquisition Announcement

CORPORATE PERFORMANCE IMPROVEMENT DOES CLOUD MEAN THE PRIVATE DATA CENTER IS DEAD?

GET CLOUD EMPOWERED. SEE HOW THE CLOUD CAN TRANSFORM YOUR BUSINESS.

Perfect Balance of Public and Private Cloud

Group Ticker Plant Project Update

THE JOURNEY OVERVIEW THREE PHASES TO A SUCCESSFUL MIGRATION ADOPTION ACCENTURE IS 80% IN THE CLOUD

Extending the Benefits of GDDR Beyond Graphics

For personal use only

FROM A RIGID ECOSYSTEM TO A LOGICAL AND FLEXIBLE ENTITY: THE SOFTWARE- DEFINED DATA CENTRE

Why Enterprises Need to Optimize Their Data Centers

Cisco HyperFlex and the F5 BIG-IP Platform Accelerate Infrastructure and Application Deployments

AMD Disaggregates the Server, Defines New Hyperscale Building Block

Server Success Checklist: 20 Features Your New Hardware Must Have

Abstract. The Challenges. ESG Lab Review InterSystems IRIS Data Platform: A Unified, Efficient Data Platform for Fast Business Insight

Storage Networking Strategy for the Next Five Years

The Next Evolution of Enterprise Public Cloud. Bring the Oracle Cloud to Your Data Center

Cisco Cloud Services Router 1000V and Amazon Web Services CASE STUDY

Government IT Modernization and the Adoption of Hybrid Cloud

Cisco Crosswork Network Automation

FUJITSU Backup as a Service Rapid Recovery Appliance

THE RISE OF THE MODERN DATA CENTER

Solace Message Routers and Cisco Ethernet Switches: Unified Infrastructure for Financial Services Middleware

21ST century enterprise. HCL Technologies Presents. Roadmap for Data Center Transformation

Data center interconnect for the enterprise hybrid cloud

Future-proof managed hosting brings agility

Cisco Unified Computing System Delivering on Cisco's Unified Computing Vision

Lenovo Data Center Group. Define a different future

Preparing your network for the next wave of innovation

Toward a Memory-centric Architecture

INSIGHTS. FPGA - Beyond Market Data. Financial Markets

Business Benefits of Policy Based Data De-Duplication Data Footprint Reduction with Quality of Service (QoS) for Data Protection

Cisco 4000 Series Integrated Services Routers: Architecture for Branch-Office Agility

Powering Transformation With Cisco

RED HAT ENTERPRISE LINUX. STANDARDIZE & SAVE.

Predictive Insight, Automation and Expertise Drive Added Value for Managed Services

Live Broadcast: Video Services from AT&T

AWS & Intel: A Partnership Dedicated to fueling your Innovations. Thomas Kellerer BDM CSP, Intel Central Europe

IBM POWER SYSTEMS: YOUR UNFAIR ADVANTAGE

The Next Opportunity in the Data Centre

TREND REPORT. Ethernet, MPLS and SD-WAN: What Decision Makers Really Think

New Zealand Government IBM Infrastructure as a Service

MODERNISE WITH ALL-FLASH. Intel Inside. Powerful Data Centre Outside.

SIEM Solutions from McAfee

Your Data Center is Everywhere. Unified Computing System Data Center Campaign Overview Marketing Cheat Sheet

COMPUTACENTER AND CITRIX TOGETHER

Modernizing Government Storage for the Cloud Era

Cisco SAN Analytics and SAN Telemetry Streaming

Cisco APIC Enterprise Module Simplifies Network Operations

Taking Hyper-converged Infrastructure to a New Level of Performance, Efficiency and TCO

DELL POWERVAULT MD FAMILY MODULAR STORAGE THE DELL POWERVAULT MD STORAGE FAMILY

SD-WAN: A Simplified Network for Distributed Enterprises

Top 5 Reasons to Consider

It s Time to Move Your Critical Data to SSDs Introduction

Markus Kujala, Systems Engineering Manager

For Healthcare Providers: How All-Flash Storage in EHR and VDI Can Lower Costs and Improve Quality of Care

MANAGING THE COMPLEXITY.

THE FUTURE OF BUSINESS DEPENDS ON SOFTWARE DEFINED STORAGE (SDS)

The Future of Business Depends on Software Defined Storage (SDS) How SSDs can fit into and accelerate an SDS strategy

Remodel. New server deployment time is reduced from weeks to minutes

CONNECTING THE CLOUD WITH ON DEMAND INFRASTRUCTURE

HPE Nimble Storage Partner Sales Guide

BUYING SERVER HARDWARE FOR A SCALABLE VIRTUAL INFRASTRUCTURE

BUILD BETTER MICROSOFT SQL SERVER SOLUTIONS Sales Conversation Card

Composite Software Data Virtualization The Five Most Popular Uses of Data Virtualization

Accelerate Your Enterprise Private Cloud Initiative

Market Report. Scale-out 2.0: Simple, Scalable, Services- Oriented Storage. Scale-out Storage Meets the Enterprise. June 2010.

SOLUTION BRIEF NETWORK OPERATIONS AND ANALYTICS. How Can I Predict Network Behavior to Provide for an Exceptional Customer Experience?

SOLUTION BRIEF TOP 5 REASONS TO CHOOSE FLASHSTACK

Azul Disrupts the ROI Equation for High Performance Applications

A Low Latency Solution Stack for High Frequency Trading. High-Frequency Trading. Solution. White Paper

Upgrade Your MuleESB with Solace s Messaging Infrastructure

RISC-V: Enabling a New Era of Open Data-Centric Computing Architectures

Why Converged Infrastructure?

Public Cloud Adoption: 10 benefits your business can t ignore

Faster FAST Multicore Acceleration of

Lenovo Database Configuration for Microsoft SQL Server TB

DATA CENTRE SOLUTIONS

Modernizing Servers and Software

Virtualizing Networks:

Composable Infrastructure for Public Cloud Service Providers

Virtualizing the SAP Infrastructure through Grid Technology. WHITE PAPER March 2007

Overcoming the Challenges of Server Virtualisation

Why Converged Infrastructure?

Strong Consistency versus Weak Consistency

Choosing the Right Cloud. ebook

WHOLESALE HOSTED CENTREX

The Top Five Reasons to Deploy Software-Defined Networks and Network Functions Virtualization

For DBAs and LOB Managers: Using Flash Storage to Drive Performance and Efficiency in Oracle Databases

Oracle Exadata Statement of Direction NOVEMBER 2017

Paper. Delivering Strong Security in a Hyperconverged Data Center Environment

Iron Networks, Inc. Turnkey Converged Infrastructure-as-a-Service Platforms

Logicalis What we do

Transcription:

LOW LATENCY DATA DISTRIBUTION IN CAPITAL MARKETS: GETTING IT RIGHT PATRICK KUSTER Head of Business Development, Enterprise Capabilities, Thomson Reuters +358 (40) 840 7788; patrick.kuster@thomsonreuters.com

In seeking the most effective solution for low latency access to data from multiple venues, financial firms are challenged by the fact that technology has moved on, but the way they accumulate data has not. The platforms built in response to regulatory requirements to deliver best execution are now nearly a decade old and firms are looking for a faster way to execute on their clients behalf. A THIRD WAY During the past ten years, technology options have become broader and less costly single core processors have been replaced with multi-core processors, for example. At the same time, the amount of data with which firms are dealing has grown exponentially in terms of volume and complexity. Today, firms require high-performance access to real-time market data sourced directly from individual trading venues in a standardised, reliable and flexible format.recently, a debate has arisen about the virtues of embracing either hardware acceleration or software as a solution to these requirements. However, a third way a hybrid approach that harnesses the value of both is in our view the most effective way of delivering the required performance. 01

HARDWARE ACCELERATION VS SOFTWARE: FALSE DICHOTOMY? FPGA HAS GAINED MUCH ATTENTION IN FINANCIAL SERVICES Hardware acceleration the use of computer hardware to perform functions faster than is possible in software running on a general purpose CPU delivers low latency performance. Of the hardware acceleration techniques, field-programmable gate array (FPGA) technology has gained much attention in financial services. FPGA is an integrated circuit designed to be configured by a user or a designer after manufacturing. The circuits can undertake complex digital computations rapidly. FPGA is best at processing streams of data quickly and deterministically meaning users can accurately predict how long a process will take. The technology delivers parallel and concurrent resources so that users can deploy it, for example, for five or 500 tasks and the latency will be the same. The FPGA technology ensures that the latency from the venue to the receiving application is deterministic. You can be sure that the application is always a fixed amount of time behind real time, which is important since it affects the probability that a particular order/quote will have been filled by the time the app sees it. This allows trading applications to accurately predict the impact that a trade will have had on market conditions. This is important for the performance of an algorithm, as well as for risk management. For example Market Makers have quotes that can be exposed when the market moves, so they need to know how current their market data is to take appropriate action and set their spread appropriately. Both software and hardware approach the problem of increase in data rates through parallel processing, however software does allow a much more dynamic and flexible environment to work in. The flexibility and ease of programming of software are the main benefits. A challenge with the software approach is that the underlying server is designed for general purpose computing and therefore the software tool is competing with everything else that runs on that server. When the market is very busy and has high liquidity, latency tends to grow with a software solution. 02

IMPROVEMENTS IN FPGA One of the most important factors that is making FPGA a key technology in financial services is the growth in size of the FPGA chips. The FPGA is made up of a fixed set of logic cells, which the developer will configure into a certain arrangement to carry out the required task. The more logic cells that are available on an FPGA, the more complex operations are possible on a single chip. If we look at the past seven years, we can see a huge increase in the size of these chips available to the market. For example, the Xilinx FPGA cards on which the Thomson Reuters Elektron Direct Feed is based demonstrate huge increases in both computation power and storage capabilities. THEORETICAL MAXIMUM TIMES OPERATION CAN BE CARRIED OUT ON THE FPGA FAMILY DEVICE LOGIC CELLS (THOUSANDS) BRAM (KB) 16X16 MAC 16-BIT ADDITION 16:11-BIT MUX V4 XC4VLX60 60 360 64 3328 6656 V5 XC5VLX110T 110 666 64 4320 17280 V6 XC6VLX340T 240 1872 768 9420 37680 V7 XC7VX690T 690 5310 2880 21650 86600 The available Logic Cells on each generation of card. This is an approximate equivalent gate count on the FPGA itself. The available BRAM on each generation this is the directly accessible memory available to the FPGA device. The 16x16 MAC is a 16 bit x 16 bit multiply and accumulate function, implemented in special DSP blocks rather than in the generic FPGA fabric. 16 bit addition is a simple binary addition of two 16 bit numbers. 16:1 1-bit mux is a multiplexer that selects one bit from 16 values (this is a basic hardware function). XILINX FPGA GROWTH - 2008 TO 2015 V5 6656 V4 17280 V6 37680 Scale = Theoretical number of operations V7 86600 V5 4320 V4 3328 V6 9420 = V7 = V6 V7 21650 = V5 = V4 For basic operations, the available space on the FPGA has increased by more than tenfold during the past seven years. This will likely continue as the chips continue to reduce in size. In practice this means that there is a huge opportunity to use these cards for more complex operations and therefore bring the deterministic latency benefits of the hardware to more complex data sets. V5 64 V6 768 V7 2880 16:11-BIT MUX 16-BIT ADDITION 16X16 MAC V4 64 03

THE BENEFITS OF A HYBRID SYSTEM Hardware acceleration is good for some tasks but not others the same can be said of software tools. Therefore a hybrid solution that combines hardware acceleration to achieve low latency and software optimisation for recovery and static data mapping will help firms to achieve higher standards of best execution. This approach ensures that customers receive fast and robust real-time data feeds, which can be installed and integrated into their applications quickly and easily. HARDWARE FOR SIMPLE TASKS, SOFTWARE FOR MORE COMPLEX TASKS The determinism of FPGA is most useful for data feed handlers when providing real-time updates from data feeds. For other activities such as recovering logic, handling static data etc software is more appropriate. However, once a firm wants to run complex algorithms for trading strategies software programs come into their own. Simple, repetitive tasks can be handled by hardware accelerators, with software performing the more analytical, complex activities. As a general rule, the simpler a task is, the more appropriate it is to use hardware acceleration. Along a scale of complexity, the indicator will move towards software the more complex, dynamic or changeable a task becomes. If the market data feed is unbundled from other tasks and handled via hardware acceleration, firms will achieve deterministic latency. 04

HARDWARE AND SOFTWARE USE CASES Ed Rainer, Group Product Manager, Real-Time Information Services, London Stock Exchange Group, says that FPGA technology coupled with a high performance, scalable software chassis was a natural progression for LSEG s big data technology solutions. Innovation is at the core of everything we do at LSEG and through the research and development teams at MillenniumIT we ve got all of the tools to bring an exciting technology offering to market. The behaviour of the Group s varied customer base - Member Firms,Information Vendors and Network Service Providers - continues to be influenced by our product s performance and in particular the latency of our trading and information product suite. Feedback consistently indicates that constant performance and latency is what s important to our customers. This is a key benefit of an FPGA implementation deterministic, predictable performance and a scalable product build. Our Group Ticker Plant, a product designed to standardise real-time data publication across Group trading venues, irrespective of the trading technology employed, leverages the Group s FPGA research and promotes it to production environments for the very first time. The platform FPGAs disseminate latency-sensitive data from our matching engines such as order book changes, executions and trading sessions,directly to customers in IP multicast. This happens in parallel with more complex processing such as statistics, reference data management and TCP-based data recovery being powered by a traditional software chassis. Phase 2 Group Ticker Plant services, including full-depth order books powered by FPGA, are available across Group test environments now. Further detail on the programme can be found at: www.londonstockexchange.com/oneproductinfinitepossibilites. A KEY BENEFIT OF AN FPGA IMPLEMENTATION - DETERMINISTIC, PREDICTABLE PERFORMANCE... Software is best deployed in scenarios where large amounts of stored data are required, and the analysis of this data needs to change on a regular basis. For example, a firm may be using a solution that analyses the past 30 years of news headlines for specific pieces of information, and based on those results will adjust the program to look for new information. Due to the large amount of data needed, and the need for quick evolution of the search and analysis criteria, software would make a natural fit. On the other hand, FPGA and other hardware accelerators are best used for constant, predictable data. An example of this may be an equities or derivatives data feed from an exchange or other venue. The feed may change only twice or three times a year, so there is some degree of flexibility required to cope with these changes. Hardware accelerators can parse the information very efficiently, particularly for venues that provide a binary, fixed-length message. The predictability dovetails very well with the determinism that FPGA offers. Another scenario is that of an internal processing system, such as a large data set, database or capture system, the format of which is controlled by the trading firm. The requirement here is to move a large amount of data from one location to the other, using a standard translation. 05

DRIVERS OF FPGA A MASSIVE INCREASE IN DATA RATES One of the key drivers of FPGA use in recent years is the massive increase in data rates. This has been combined with an increase in the cost of data centre space (due largely to the cost of cooling). At a time when exchanges have moved to co-location strategies and trading firms pay a premium for being as close to the exchange servers as possible, space is costly. When you add up the rack space itself, the increased costs of power and cooling needed for modern data centres, there is a large cost of ownership for every server deployed. Putting a large number of servers running software in a location is no longer efficient. Hardware accelerators enable firms to do more with less kit. Configurable, multicast FPGA technology for real-time data processing means that multiple direct feeds can be supported on a single server, reducing the footprint of what is needed to process data. As a result, a firm s infrastructure and operating costs can be reduced. FPGA also efficiently handles microbursts huge spikes of data delivered by exchanges in very short periods of time. Determinism enables hardware accelerators to process data as it comes in, at the rate it comes in. Software cannot deal with microbursts as efficiently and will typically smooth the data out, buffering the information and thus delaying its delivery. When running software systems with multiple cores each one can be dedicated to a single process. A handoff is then required to the network interface, then on to a system that publishes the data out to multiple users on the network that want to consume the data. If distribution is done out of FPGA, these handoffs are removed as all steps are handled by the FPGA card. As already stated, FPGA is not always appropriate for all data feed activities. If an activity requires the storage of large volumes of data, software and storage on a server will deliver better results than FPGA. Also, lowvolume data feeds with low throughput of data are more efficiently processed via software than via hardware accelerators. Software is cheaper to develop than FPGA therefore feeds that are low-volume or are not contingent on having very low latency are better handled via software. Benjamin Stephens, Head of Latency Sensitive Electronic Trading at Instinet Europe, says there were two main reasons for choosing hardware acceleration: deterministic latency and a reduction in the firm s hardware footprint. Deterministic latency allows trading algorithms to perform more consistently and also reduces performance penalties during times of heavy trading and high data volumes. Additionally, the cost of co-location space is high and hardware accelerators enabled us to significantly reduce the hardware footprint we needed to run a large number of venues in co-location sites. For example, we could run a US equity smart order router on one server with hardware acceleration. This delivered large savings and a good performance benefit. HARDWARE ACCELERATION - DETERMINISTIC LATENCY AND REDUCTION IN HARDWARE FOOTPRINT As explained, hardware accelerators are not always the best fit. Stephens says for advanced calculations or complex data or when the logic required is constantly changing, software is more appropriate. 06

CONCLUSION The deployment of a hybrid approach to low latency data access can be complex; a thorough understanding of when FPGA can be used versus when software should be considered is key. In today s markets, electronic market makers and brokers are no longer achieving the price realisation they were ten years ago. In looking for faster ways of executing on their clients behalf, they realise that they must balance this need for speed with cost. Using hardware acceleration to strengthen software-based feed handlers is becoming more common, even among market vendors. The technology underpinning hardware accelerators is complex and firms that are not on the bleeding edge of latency or that don t have very specialised business models should seek out partners to help them develop hybrid approaches. Partnering with companies that are developing combined software and hardware direct feed solutions for multiple clients on a scale basis, will enable financial firms to cost effectively and reliably gain access to everincreasing volumes of market data. USING HARDWARE ACCELERATION TO STRENGTHEN SOFTWARE-BASED FEED HANDLERS IS BECOMING MORE COMMON PATRICK KUSTER Head of Business Development, Enterprise Capabilities, Thomson Reuters +358 (40) 840 7788; patrick.kuster@thomsonreuters.com Visit thomsonreuters.com 07 2015 Thomson Reuters. S022185 07/15. Thomson Reuters and the Kinesis logo are trademarks of Thomson Reuters.