ROBINHOOD POLICY ENGINE

Similar documents
RobinHood Policy Engine

MANAGING LUSTRE & ITS CEA

FROM RESEARCH TO INDUSTRY. RobinHood v3. Robinhood User Group Thomas Leibovici 16 septembre 2015

Lustre/HSM Binding. Aurélien Degrémont Aurélien Degrémont LUG April

RobinHood Project Update

Robin Hood 2.5 on Lustre 2.5 with DNE

LUG 2012 From Lustre 2.1 to Lustre HSM IFERC (Rokkasho, Japan)

RobinHood v3. Integrity check & thoughts on local VFS changelogs. Robinhood User Group Dominique Martinet

RobinHood Project Status

Nathan Rutman SC09 Portland, OR. Lustre HSM

TGCC OVERVIEW. 13 février 2014 CEA 10 AVRIL 2012 PAGE 1

Lustre usages and experiences

Data Movement & Tiering with DMF 7

Lustre overview and roadmap to Exascale computing

LCE: Lustre at CEA. Stéphane Thiell CEA/DAM

XtreemStore A SCALABLE STORAGE MANAGEMENT SOFTWARE WITHOUT LIMITS YOUR DATA. YOUR CONTROL

Robinhood v2. Temporary Filesystem Manager. Admin Guide

Implementing a Hierarchical Storage Management system in a large-scale Lustre and HPSS environment

Lustre* is designed to achieve the maximum performance and scalability for POSIX applications that need outstanding streamed I/O.

IBM Spectrum Scale Strategy Days

LustreFS and its ongoing Evolution for High Performance Computing and Data Analysis Solutions

UK LUG 10 th July Lustre at Exascale. Eric Barton. CTO Whamcloud, Inc Whamcloud, Inc.

Filesystems on SSCK's HP XC6000

Implementing a Hierarchical Storage Management system in a large-scale Lustre and HPSS environment

Coordinating Parallel HSM in Object-based Cluster Filesystems

Demonstration Milestone Completion for the LFSCK 2 Subproject 3.2 on the Lustre* File System FSCK Project of the SFS-DEV-001 contract.

Lustre HSM at Cambridge. Early user experience using Intel Lemur HSM agent

Using file systems at HC3

Dept. Of Computer Science, Colorado State University

Google File System. Arun Sundaram Operating Systems

COS 318: Operating Systems. NSF, Snapshot, Dedup and Review

Administering Lustre 2.0 at CEA

Experiences using a multi-tiered GPFS file system at Mount Sinai. Bhupender Thakur Patricia Kovatch Francesca Tartagliogne Dansha Jiang

Distributed Systems 16. Distributed File Systems II

<Insert Picture Here> Btrfs Filesystem

Lustre on ZFS. At The University of Wisconsin Space Science and Engineering Center. Scott Nolin September 17, 2013

Lustre * Features In Development Fan Yong High Performance Data Division, Intel CLUG

CSE 124: Networked Services Lecture-16

Next-Generation NVMe-Native Parallel Filesystem for Accelerating HPC Workloads

CSE 124: Networked Services Fall 2009 Lecture-19

PPC s SMART Practice Aids

Lustre Parallel Filesystem Best Practices

Daily, Weekly or Monthly Partitions? A discussion of several factors for this important decision

The Google File System

Triton file systems - an introduction. slide 1 of 28

European Lustre Workshop Paris, France September Hands on Lustre 2.x. Johann Lombardi. Principal Engineer Whamcloud, Inc Whamcloud, Inc.

NPTEL Course Jan K. Gopinath Indian Institute of Science

CA485 Ray Walshe Google File System

Extraordinary HPC file system solutions at KIT

Lustre 2.12 and Beyond. Andreas Dilger, Whamcloud

Computer Science Section. Computational and Information Systems Laboratory National Center for Atmospheric Research

DDN s Vision for the Future of Lustre LUG2015 Robert Triendl

HP Designing and Implementing HP Enterprise Backup Solutions. Download Full Version :

Shared Object-Based Storage and the HPC Data Center

GFS: The Google File System

Data Management. Parallel Filesystems. Dr David Henty HPC Training and Support

The Btrfs Filesystem. Chris Mason

CLOUD-SCALE FILE SYSTEMS

Lustre on ZFS. Andreas Dilger Software Architect High Performance Data Division September, Lustre Admin & Developer Workshop, Paris, 2012

Opendedupe & Veritas NetBackup ARCHITECTURE OVERVIEW AND USE CASES

Zhang, Hongchao

Oracle Zero Data Loss Recovery Appliance (ZDLRA)

Configuring Short RPO with Actifio StreamSnap and Dedup-Async Replication

Overcoming Obstacles to Petabyte Archives

IBM V7000 Unified R1.4.2 Asynchronous Replication Performance Reference Guide

PERSISTENCE: FSCK, JOURNALING. Shivaram Venkataraman CS 537, Spring 2019

A High Performance, High Reliability Perforce Server. Shiv Sikand, IC Manage Marc Lewert, Angela Thomas, TiVo

Lesson 9 Transcript: Backup and Recovery

Project Quota for Lustre

The HAMMER Filesystem DragonFlyBSD Project Matthew Dillon 11 October 2008

LFSCK High Performance Data Division

Andreas Dilger, Intel High Performance Data Division LAD 2017

Remote Directories High Level Design

Data life cycle monitoring using RoBinHood at scale. Gabriele Paciucci Solution Architect Bruno Faccini Senior Support Engineer September LAD

Storage Optimization with Oracle Database 11g

Database Management. Understanding Failure Resiliency CHAPTER

An Overview of Fujitsu s Lustre Based File System

Andreas Dilger, Intel High Performance Data Division Lustre User Group 2017

GFS: The Google File System. Dr. Yingwu Zhu

The Google File System

Insights into TSM/HSM for UNIX and Windows

NPTEL Course Jan K. Gopinath Indian Institute of Science

COS 318: Operating Systems. Journaling, NFS and WAFL

File Systems Management and Examples

6.5 Collective Open/Close & Epoch Distribution Demonstration

Challenges in making Lustre systems reliable

Enhancing Lustre Performance and Usability

Scale out Read Only Workload by sharing data files of InnoDB. Zhai weixiang Alibaba Cloud

Data Movement and Storage. 04/07/09 1

Lessons learned from Lustre file system operation

HP Data Protector 9.0 Deduplication

Still All on One Server: Perforce at Scale

Introduction to Lustre* Architecture

SFA12KX and Lustre Update

Blue Waters I/O Performance

Operating Systems. File Systems. Thomas Ropars.

5.4 - DAOS Demonstration and Benchmark Report

AFM Migration: The Road To Perdition

A. Deduplication rate is less than expected, accounting for the remaining GSAN capacity

Actifio Sky DB. Actifio s Solution for Oracle, Oracle EBS with standalone, RAC, ASM, EXADATA configurations

Transcription:

ROBINHOOD POLICY ENGINE Aurélien DEGREMONT Thomas LEIBOVICI CEA/DAM LUSTRE USER GROUP 2013 PAGE 1

ROBINHOOD: BIG PICTURE Admin rules & policies find and du clones Parallel scan (nighly, weekly, ) Lustre v2 ChangeLogs Robinhood database near real-time DB update Fine-grained statistics + web UI Mass action scheduling (policies) Scan sometimes, query at will or no scan at all (Lustre v2) Query at will for: Searches (rbh-find) Statistics (rbh-report, WebUI) Policies: old files cleaning, data archiving, HSM Lustre 1.8, posix FS: scanning cost is shared for all these purposes Lustre 2.x: only initial scan needed (Changelogs) PAGE 2

RBH-FIND, RBH-DU Fast find and du clones Query robinhood DB instead of performing POSIX namespace scan faster! > rbh-find [path] -user foo*" -size +1G ost 4 20sec for 40M entries Enhanced du : Detailed stats (by type ) Can filter by user > rbh-du -sh /fs/dir -u foo --details /fs/dir symlink count:30777, size:1.0m, spc_used:9.1m dir count:598024, size:2.4g, spc_used:2.4g file count:3093601, size:3.2t, spc_used:2.9t PAGE 3

ROBINHOOD: FINE-GRAINED STATISTICS Usage statistics Per user, per group, per type, Possibly split user usage by group > rbh-report u foo S user, group, type, count, spc_used, avg_size foo, proj001, dir, 2, 8.00 KB, 4.00 KB foo, gr1, dir, 74441, 291.64 MB, 4.01 KB foo, gr1, file, 422367, 71.01 GB, 335.54 KB foo, gr1, symlink, 1418, 1.35 MB, 46 foo, proj002, file, 2, 1.27 GB, 651.43 MB Total: 498230 entries, 77918785024 bytes used (72.57 GB) PAGE 4

ROBINHOOD: FINE-GRAINED STATISTICS Top user and groups Sorted by volume, object count, avg file size > rbh-report --top-users --by-count rank, user, spc_used, count, avg_size 1, john, 423.23 GB, 1599881, 275.30 KB 2, paul, 292.91 GB, 954153, 330.98 KB 3, mike, 65.37 GB, 543169, 130.98 KB Top directories Sorted by object count, avg file size > rbh-report --top-dirs --by-count rank, path, dircount, avgsize, user, group, last_mod 1, /hpss/foo1/dir1, 24832, 2.62 GB, foo1, gr59, 2013/03/11 17:13:45 2, /hpss/foo2/dir3, 20484, 339.88 MB, foo2, g03, 2013/02/03 06:59:05 3, /hpss/bar2/dir4, 19484, 543.82 MB, bar2, g03, 2012/05/28 12:45:26 PAGE 5

ROBINHOOD: FINE-GRAINED STATISTICS File size profile Global or for a given user/group: > rbh-report u foo -sz-prof user, type, count, spc_used, avg_size, 0, 1~31, 32~1K-, 1K~31K, 32K~1M-, 1M~31M, 32M~1G-, 1G~31G, 32G~1T-, +1T foo, file, 62091, 58.50 TB, 1.09 GB, 40, 44, 233, 548, 1879, 5820, 8004, 1835, 178, 0 User, group, directories can also be sorted by the ratio of file in a given range: > rbh-report --top-users --by-szratio=1..31m rank, user, ratio(1..31m) 1, john, 91.30% 2, perrez, 87.64% 3, matthiew, 85.76% 4, vladimir, 78.50% 5, gino, 77.02% PAGE 6

WEB INTERFACE Web UI Usage stats (per user, per group) File size profile (global / per user / per group) PAGE 7

ROBINHOOD POLICIES Mass action scheduling on filesystem entries Admin-defined rules Build-in policies: Purge Archiving Dir removal Deferred removal (undelete) HSM: schedule archive and release actions Policy definition: Flexible and highly customizable Attribute-based Using fileclass definitions Example: fileclass BigLogFiles { definition { type == file and size > 100MB and (path ( == /fs/ fs/logdir logdir/* or name == *.log) } } purge_policies { ignore_fileclass = my_fileclass; } policy purge_logs { target_fileclass = BigLogFiles; condition { last_mod > 15d } } PAGE 8

PURGE POLICIES Use cases Managing file lifetime, cleaning tmp data Cleanup after a code run (lifetime: few hours) Cleanup after a simulation (lifetime: several months) Clean old krb tickets, logs, core dumps, crash dumps, Avoid ENOSPC errors caused by full OSTs Admin defines high/low OST usage thresholds and purge policy rules Robinhood monitors free space per OST Applies purge policies on OSTs that exceed high threshold Disk space fair-share Admin defines max usage per user or group Robinhood monitors usage per user/group Applies purge policies for user/group over their limit ENOSPC OST PAGE 9

ARCHIVING POLICIES An alternative to rsync New data No need to scan each time you want to archive data Policy-driven: Archived data Can skip some kind of files Can control the delay before archiving a file Fine control of copy streams: number of simultaneous copies in parallel max files/hour, max volume/hour Robinhood detects file modification and maintains file status in its DB report available with current status of files (new/archived/modified ) Allows undelete Removed files in Lustre are not immediately removed in backup space Configurable delay before cleaning the backup copy Can undelete a file during this delay PAGE 10

ROBINHOOD AND LUSTRE-HSM Lustre-HSM support Archiving policies to schedule copy from Lustre to the HSM Purge policies to release disk space in Lustre OSTs when needed (file remains visible in Lustre for users) GC of deleted files Undelete Disaster recovery: to rebuild a Lustre filesystem from the archive Aware of HSM specific changelog records: To keep track of file status PAGE 11

WORK IN PROGRESS MDS disaster recovery Scenario: MDT changes disaster MDT snapshot (dd ) changelogs 1) Restore MDT snapshot 2) Replay changes between snapshot and disaster (rbh-diff / rbh-apply) New policies Rebalance files between OSTs / migrate files between OST pools Generic policies to schedule any kind of action, e.g. Datascrubbing (to detect silent corruption) MD scrubbing (FS consistency check) Run any command on FS entries PAGE 12

FUTURE WORK Distributed database For now: 1 single MySQL database Need to distribute the database: To handle higher MD rate (DNE) To manage xx billions of entries Turn it to a framework API to extract customized reports All components as dynamic customizable modules: Policy criteria Policies Stats Filesystem backend, DB backend Lustre User Group 2013 PAGE 13

ROBINHOOD WEBSITE & SUPPORT Last release: robinhood 2.4.2 Project home: http://robinhood.sourceforge.net Mailing lists: robinhood-news@lists.sourceforge.net robinhood-support@lists.sourceforge.net robinhood-devel@lists.sourceforge.net PAGE 14