Storage on the Lunatic Fringe. Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium

Similar documents
Storage on the Lunatic Fringe

Commodity Reliability And Practices or Building Reliable Systems with CRAP

Data Transfers Between LHC Grid Sites Dorian Kcira

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

Table 9. ASCI Data Storage Requirements

The CMS Computing Model

UW-ATLAS Experiences with Condor

Gigabyte Bandwidth Enables Global Co-Laboratories

Towards Network Awareness in LHC Computing

Insight: that s for NSA Decision making: that s for Google, Facebook. so they find the best way to push out adds and products

Distributed File Systems Part IV. Hierarchical Mass Storage Systems

Beyond Petascale. Roger Haskin Manager, Parallel File Systems IBM Almaden Research Center

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

NetApp: Solving I/O Challenges. Jeff Baxter February 2013

Data Movement & Storage Using the Data Capacitor Filesystem

High-Energy Physics Data-Storage Challenges

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Management Information Systems OUTLINE OBJECTIVES. Information Systems: Computer Hardware. Dr. Shankar Sundaresan

Data Movement & Tiering with DMF 7

Sun Lustre Storage System Simplifying and Accelerating Lustre Deployments

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

New Approach to Unstructured Data

Deduplication Storage System

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Terabytes, Petabytes and Beyond -- Data Storage Strategies

Coordinating Parallel HSM in Object-based Cluster Filesystems

Grid Computing at the IIHE

CSE 591/392: GPU Programming. Introduction. Klaus Mueller. Computer Science Department Stony Brook University

Virtualizing a Batch. University Grid Center

Deep Storage for Exponential Data. Nathan Thompson CEO, Spectra Logic

Today: Secondary Storage! Typical Disk Parameters!

Integrating Fibre Channel Storage Devices into the NCAR MSS

Netherlands Institute for Radio Astronomy. May 18th, 2009 Hanno Holties

Storage Industry Resource Domain Model

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

2 Databases for calibration and bookkeeping purposes

Storage for HPC, HPDA and Machine Learning (ML)

CHIPP Phoenix Cluster Inauguration

Clustering and Reclustering HEP Data in Object Databases

HPSS Treefrog Summary MARCH 1, 2018

CLOUDS OF JINR, UNIVERSITY OF SOFIA AND INRNE JOIN TOGETHER

The LCG 3D Project. Maria Girone, CERN. The 23rd Open Grid Forum - OGF23 4th June 2008, Barcelona. CERN IT Department CH-1211 Genève 23 Switzerland

Storage Resource Sharing with CASTOR.

LHC and LSST Use Cases

File Storage Management Systems (FSMS) and ANSI/AIIM MS66

IEPSAS-Kosice: experiences in running LCG site

Architecting Storage for Semiconductor Design: Manufacturing Preparation

High Performance Computing on MapReduce Programming Framework

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

Scale-out Object Store for PB/hr Backups and Long Term Archive April 24, 2014

A scalable storage element and its usage in HEP

1. Introduction. Outline

CSE 451: Operating Systems Spring Module 12 Secondary Storage. Steve Gribble

CERN s Business Computing

A GPFS Primer October 2005

Data services for LHC computing

CC-IN2P3: A High Performance Data Center for Research

Lustre2.5 Performance Evaluation: Performance Improvements with Large I/O Patches, Metadata Improvements, and Metadata Scaling with DNE

CERN Open Data and Data Analysis Knowledge Preservation

Data storage services at KEK/CRC -- status and plan

Communication has significant impact on application performance. Interconnection networks therefore have a vital role in cluster systems.

朱义普. Resolving High Performance Computing and Big Data Application Bottlenecks with Application-Defined Flash Acceleration. Director, North Asia, HPC

Data Intensive Science Impact on Networks

PetaSTAR A Real World Data Storage and Management Solution

Long Term Data Preservation for CDF at INFN-CNAF

Storage Systems Market Analysis Dec 04

WHITE PAPER QUANTUM S XCELLIS SCALE-OUT NAS. Industry-leading IP Performance for 4K, 8K and Beyond

BlueGene/L. Computer Science, University of Warwick. Source: IBM

IBM Storwize V7000 Unified

HPC Growing Pains. IT Lessons Learned from the Biomedical Data Deluge

Kinetic Open Storage Platform: Enabling Break-through Economics in Scale-out Object Storage PRESENTATION TITLE GOES HERE Ali Fenn & James Hughes

NetApp High-Performance Storage Solution for Lustre

Big Data Analytics and the LHC

Xcellis Technical Overview: A deep dive into the latest hardware designed for StorNext 5

Summary of the LHC Computing Review

Computing Model Tier-2 Plans for Germany Relations to GridKa/Tier-1

Data Analysis in Experimental Particle Physics

Next-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads

Building the Storage Internet. Dispersed Storage Overview March 2008

Survey: Users Share Their Storage Performance Needs. Jim Handy, Objective Analysis Thomas Coughlin, PhD, Coughlin Associates

Distributed File Systems II

EMC ISILON HARDWARE PLATFORM

Andrea Sciabà CERN, Switzerland

<Insert Picture Here> Tape Technologies April 4, 2011

Backup and archiving need not to create headaches new pain relievers are around

Scale-out Data Deduplication Architecture

COSC6376 Cloud Computing Lecture 17: Storage Systems

ALHAD G. APTE, BARC 2nd GARUDA PARTNERS MEET ON 15th & 16th SEPT. 2006

IBM Storage. Leading the 21st Century Growth. Freddy Lee Advanced Technical Support

New strategies of the LHC experiments to meet the computing requirements of the HL-LHC era

T e c h n o l o g y. LaserTAPE: The Future of Storage

Understanding Primary Storage Optimization Options Jered Floyd Permabit Technology Corp.

Shared Object-Based Storage and the HPC Data Center

Isilon: Raising The Bar On Performance & Archive Use Cases. John Har Solutions Product Manager Unstructured Data Storage Team

Introducing Panasas ActiveStor 14

A 101 Guide to Heterogeneous, Accelerated, Data Centric Computing Architectures

An Introduction to GPFS

Planning For Persistent Memory In The Data Center. Sarah Jelinek/Intel Corporation

I/O Challenges: Todays I/O Challenges for Big Data Analysis. Henry Newman CEO/CTO Instrumental, Inc. April 30, 2013

Study of the viability of a Green Storage for the ALICE-T1. Eduardo Murrieta Técnico Académico: ICN - UNAM

Transcription:

Storage on the Lunatic Fringe Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium tmruwart@dtc.umn.edu

Orientation Who are the lunatics? What are their requirements? Why is this interesting to the Storage Industry? What is SNIA doing about this? Conclusions

Who are the Lunatics? DoE Accelerated Strategic Computing Initiative (ASCI)! BIG data, locally and widely distributed, high bandwidth access, relatively few users, secure, short-term retention High Energy Physics (HEP) Fermilab, CERN, DESY! BIG data, locally distributed, widely available, moderate number of users, sparse access, long-term retention NASA Earth Observing System Data Information Systems (EOSDIS)! Moderately sized data, locally distributed, widely available, large number of users, very long-term retention DoD NSA! Lots of little data trillions of files, locally distributed, relatively few users, secure, long-term retention DoD Army High Performance Computing Centers and Naval Research Center! BIG data, locally and widely distributed, relatively few users, high bandwidth access, secure, very long term reliable retention

A bit of History 1990 Supercomputer Centers operating with HUGE disk farms of 50-100 GB! 1990 Laptop computers have 50MB internal disk drives! 1992 Fast/wide SCSI runs at break-necking speeds of 20 MB/sec! 1994 Built a 1+TB array of disks with a single SGI xfs file system and wrote a single 1TB file! Used 4GB disks in 7+1 RAID 5 disk arrays! 36 disk arrays mounted in 5 racks 1997 ASCI Mountain Blue - 75TB distributed 2002 ASCI Q 700TB online, high performance, pushing limits of traditional [legacy] block-based file systems

The not-too-distant Future 2004 ASCI Red Storm 240TB online, high bandwidth, massively parallel 2005 ASCI Purple 3000TB online, high performance, OSD/Lustre 2006 NASA RDS 6000TB online, global access, CAS,OSD, Data Grids, Lustre? 2007 DoE Fermi Lab / CERN 3 PB/year online / nearline, global sparse access 2010 Your laptop will have a 1TB internal disk that will still be barley adequate for MS Office

DoE ASCI 1998 Mountain Blue Los Alamos! 48 128-Processor SGI Origin 2000 systems! 75TB disk storage 2002 Q! 310 32-processor machines + 64 32-processor I/O nodes! 2048 2GB FC connections to 64 I/O nodes! 2048 2GB FC connections to disk storage subsystem! 692 TB disk storage, 20GB/sec bandwidth! 2 file systems of 346GB each! 4 file system layers between the application! and the disk media 2004 Red Storm! 10,000 processors, 10TB Main Memory! 240TB Disk, 50 GB/sec bandwidth

DoE ASCI Purple Requirements Parallel I/O Bandwidth - Multiple (up to 60,000) clients access one file at hundreds of GB/sec. Support for very large (multi-petabyte) file systems Single files of multi-terabyte size must be permitted. Scalable file creation & Metadata Operations! Tens of Millions of files in one directory! Thousands of file creates per second within the same directory Archive Driven Performance - The file system should support high bandwidth data movement to tertiary storage. Adaptive Pre-fetching - Sophisticated pre-fetch and write-behind schemes are encouraged, but a method to disable them must accompany them. Flow Control & Quality of I/O Service

HEP Fermilab and CMS The Compact Muon Solenoid (CMS)! $750M Experiment being built at CERN in Switzerland! Will be active in 2007! Data rate from the detectors is ~1 PB/sec! Data rate after filtering is ~hundreds of MB/sec The Data Problem! Dataset for a single experiment is ~1PB! Several experiments per year are run! Must be made available to 5000 scientists all over the planet (Earth primarily)! Dense dataset, sparse data access by any one user! Access patterns are not deterministic HEP experiments cost $US 1B, last 20 years, involve thousands of collaborators at hundreds of institutions worldwide, and collect and analyze several petabytes of data per year

LHC Data Grid Hierarchy CMS as example, Atlas is similar human=2m Tier 1 ~PByte/sec CMS detector: 15m X 15m X 22m 12,500 tons, $700M. Online System Tier 0 +1 ~2.5 Gbits/sec ~100 MBytes/sec event reconstruction event simulation French Regional Center German Regional Center Italian Center FermiLab, USA Regional Center Tier 3 Physics data cache Courtesy Harvey Newman, CalTech and CERN analysis Institute Institute ~0.25TIPS Institute 100-1000 Mbits/sec Tier 4 Workstations Institute ~0.6-2.5 Gbps Tier2 Center Tier2 Center Tier2 Center Tier2 Center Tier2 Center Tier 2 ~0.6-2.5 Gbps CERN/CMS data goes to 6-8 Tier 1 regional centers, and from each of these to 6-10 Tier 2 centers. Physicists work on analysis channels at 135 institutes. Each institute has ~10 physicists working on one or more channels. 2000 physicists in 31 countries are involved in this 20-year experiment in which DOE is a major player.

NASA EOSDIS Remote Data Store Project:! Build a 6PB Data archive with a life expectancy of at least 20 years, probably more! Make data and data products available to 2 million users What to use?! Online versus Nearline! SCSI vs ATA! Tape vs Optical! How much of each and when? Data Grids? Dealing with Technology Life Cycles continual migration

DoD NSA How to deal with a trillion files?! At 256 bytes of metadata per file -> 256TB just for the file system metadata for one trillion files! File System resiliency! Backups? Forget it. File Creation Rate is a challenge 32,000 file per second for 1 year will generate 1 trillion files How to search for any given file How to search for any given piece of information inside all the files

DoD MSRC 500TB per year data growth Longevity of data retention is critical! 100% reliable access of any piece of data for 20+ years Security is critical Reasonably quick access to any piece of data from anywhere at any time Heterogeneous computing and storage environment

History has shown The problems that the Lunatic Fringe is working on today are the problems that the main-stream storage industry will face in 5-10 years Legacy Block-based File Systems break at these scales Legacy Network File System protocols cannot scale to meet these extreme requirements

Looking Forward

What happens when. NEC Announces a 10Tbit Memory Chip Disk drives reach 1TByte and beyond MEMS devices become commercially viable Holographic Storage Devices become commercially viable Interface speeds reach 1Tbit/sec Intel develops the sub-space channel Vendors need better ways to exploit the capabilities of these technologies rather than react to them

Common thread Their data storage capacity, access, and retention requirements are continually increasing Some of the technologies and concepts the Lunatic Fringe are looking at include:! Object-based Storage Device! Intelligent Storage! Data Grid! Borg Assimilation Technologies, etc.

How does SNIA make a difference? Act as a point to achieve critical mass behind emerging technologies such as OSD, SMI, and Intelligent Storage Make sure that these emerging technologies come to market from the beginning as standards (not proprietary implementations that migrate to standards) Help to get beyond the potential barrier for emerging technologies OSD and Intelligent Storage Help to generate vendor and user awareness and education regarding future trends and emerging technologies

Conclusions Lunatic Fringe users will continue to push the limits of existing hardware and software technologies Lunatic Fringe is a moving target there will always be a Lunatic Fringe well beyond where you are The Storage Industry at large should pay more attention to! What they are doing! Why they are doing it! What they learn

References University of Minnesota Digital Technology Center www.dtc.umn.edu ASCI www.llnl.gov/asci/platforms Fermilab www.fnal.gov NASA EOSDIS www.nasa.gov NSA www.dod.mil

Contact Info tmruwart@dtc.umn.edu