EFF-IO M7.5 Demo. Semantic Migration of Multi-dimensional Arrays

Size: px
Start display at page:

Download "EFF-IO M7.5 Demo. Semantic Migration of Multi-dimensional Arrays"

Transcription

1 EFF-IO M7.5 Demo Semantic Migration of Multi-dimensional Arrays John Bent, Sorin Faibish, Xuezhao Liu, Harriet Qui, Haiying Tang, Jerry Tirrell, Jingwang Zhang, Kelly Zhang, Zhenhua Zhang NOTICE: THIS MANUSCRIPT HAS BEEN AUTHORED BY EMC UNDER INTEL S SUBCONTRACT WITH LAWRENCE LIVERMORE NATIONAL SECURITY, LLC WHO IS THE OPERATOR AND MANAGER OF LAWRENCE LIVERMORE NATIONAL LABORATORY UNDER CONTRACT NO. DE-AC52-07NA27344 WITH THE U.S. DEPARTMENT OF ENERGY. THE UNITED STATES GOVERNMENT RETAINS AND THE PUBLISHER, BY ACCEPTING THE ARTICLE OF PUBLICATION, ACKNOWLEDGES THAT THE UNITED STATES GOVERNMENT RETAINS A NON-EXCLUSIVE, PAID-UP, IRREVOCABLE, WORLD-WIDE LICENSE TO PUBLISH OR REPRODUCE THE PUBLISHED FORM OF THIS MANUSCRIPT, OR ALLOW OTHERS TO DO SO, FOR UNITED STATES GOVERNMENT PURPOSES. THE VIEWS AND OPINIONS OF AUTHORS EXPRESSED HEREIN DO NOT NECESSARILY REFLECT THOSE OF THE UNITED STATES GOVERNMENT OR LAWRENCE LIVERMORE NATIONAL SECURITY, LLC. 1

2 Demo System Configuration IONS: Four IONs lola-[12-15].hpdd.intel.com 8 SSD burst buffers on each DAOS: Global /mnt/daos mount on each DAOS-lustre MDS(lola-2);OSTS(lola-[16-19]/8xSSD) BB s: Local /mnt/lustre[1-4] on each ION All IONs have lustre cross-mounts to all BB s CN s: Four CNs available via Mercury lola-[20,25-27] Our demo runs directly on ION s though without HDF MPI job making calls directly to IOD API 2

3 Demo System Configuration CN1 lola-20 CN2 lola-25 CN3 lola-26 CN4 lola-27 M7.5 App is not using any App CN s. Running App test suite as App MPI HDF5 HDF5 HDF5 FShip-c job FShip-c across the IONs. FShip-c HDF5 FShip-c FShip-server FShip-server FShip-server FShip-server IOD-1 IOD-2 IOD-3 IOD-4 lola-12 lola-13 lola-14 lola-15 /lustre1 8xSSD /lustre2 8xSSD /lustre3 8xSSD /lustre4 8xSSD /DAOS /DAOS /DAOS /DAOS /lustre2,/lustre3,/lustre4 /lustre1,/lustre3,/lustre4 /lustre1,/lustre2,/lustre4 /lustre1,/lustre2,/lustre3 /lustre1/.plfs cannonical /lustre1/.plfs shadow /lustre1/.plfs cannonical /lustre2/.plfs shadow /lustre1/.plfs cannonical /lustre3/.plfs shadow /lustre1/.plfs cannonical /lustre4/.plfs shadow /DAOS Lustre/zfs MDS(lola-2);OSTS(lola-[16-19]/8xSSD) 3

4 Demo scope Semantic migration of arrays Persist (IOD -> DAOS) Fetch (DAOS -> IOD) Dimensional sequence transformation Changing the access dimension Reading from a persisted / purge array i.e. Reading directly from DAOS Query layout / sharding 4

5 Initialization 1/2 Script started on Fri 21 Mar :21:30 AM PDT M7.5_demo]#./M7.5_demo1.sh Cleaning the BB's and DAOS to return to initial state. M7.5 demonstration 1 -- Array Migration. will run M7.5_case1_array_migration_ops on 4 IODS: lola-12 lola-13 lola-14 lola-15 > mpirun -np 4 -hosts lola-12,lola-13,lola-14,lola-15 -c demo_config -t M7.5_case1_array_migration_ops Fri Mar 21 09:21:41 PDT 2014 ********* Test Configuration ********* inputfile = M7.5_case1_array_migration_ops nproc = 4 all_async = True 5

6 Initialization 2/2 ========================================================================= # Demonstration: Array Migration, Data Reorganization and Layout Discovery # # This demonstration will cover these elements: # Migration of array objects using structural descriptions between DAOS and IODs # - Will be shown in both directions # Reorganization of data layout during migration for array objects # - E.g. Migrate using a layout which specifies a different dimension sequence # Layout query of array objects # - E.g. Returning structure set to place data when migrating to DAOS # # Out of scope demonstration elements: # - Similar functionality for KV objects and blob objects will not be # demonstrated at this time # ========================================================================= 6

7 Step 1: Create and open container Running command: * contopen /containera create Call iod_container_open(), path: /containera, open_mode: 0x83... Finish to open container - passed (rc = 0) 7

8 Step : Skip Transaction 0 Running command: 0 transstart 0 write Call iod_trans_start(), tid: 0x0, num_ranks: 0, mode: 0x2... Finish to start transaction - passed (rc = 0) # ######## Step finish TID0 for write # Command format: tid transfinish flag # flag only meaningful for writing transaction. 'normal' for commonly finish > Press enter to continue... Running command: 0 transfinish normal Call iod_trans_finish(), tid: 0x0, num_ranks: 0, flag: 0 Finish to finish transaction - passed (rc = 0) 8

9 Step : Start TID=1, create array Running command: 1 transstart 0 write Call iod_trans_start(), tid: 0x1, num_ranks: 0, mode: 0x2... Finish to start transaction - passed (rc = 0) # ######## Step in TID1, create ARRAY object: cell_size=32 num_dims=3 (16 X 32 X 64) > Press enter to continue... Leader mode Running command: 1 create array array_obj ,32,64 16 Call iod_obj_create(), tid: 0x1, type: 0x3... Array struct: cell_size: 512, num_dims: 3, firstdim_max: 16, current_dims: 16,32,64 Finish to create object - passed (rc = 0) 16MB array (512x16x32x64) 9

10 16 x 32 x 64 Multi-dimensional Array

11 Step 2.5: Write into array # Call iod_array_write() to write the ARRAY object # start is [0,0,0], stride is [8,8,8], count is [2,4,8], block is [8,8,8] # > Press enter to continue... Running command: 1 write array_obj 0,0,0 8,8,8 2,4,8 8,8,8 Rank 0 ID 0: start[ 0,0,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 1 ID 0: start[ 0,8,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 2 ID 0: start[ 0,16,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 3 ID 0: start[ 0,24,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Call iod_array_write(), oh: 123, tid: 0x1 Rank 0 ID 0: iod_array_write() returns 0 Rank 1 ID 0: iod_array_write() returns 0 Rank 3 ID 0: iod_array_write() returns 0 Rank 2 ID 0: iod_array_write() returns 0 Finish to write array - passed (rc = 0) 11

12 Write Pattern Rank 3 Rank 2 Rank 1 Rank 0 12

13 Step 2.6: Prepare semantic persist # location "DAOS" (i.e. IOD_LOC_CENTRAL) # type currently only support stripped layout (i.e.iod_layout_striped). # target_start -- the offset in DAOS shard list, currently IOD ignores this value (chooses one internally for load balance). # target_num ---- number of storage targets (i.e. DAOS shards). If 0, it defaults to all shards. # stripe_size --- resharding granularity. The number of cells for contiguous layout array. # dims_seq For contigous layout, it is an array with num_dims size (i.e. a 4D array may have # [3][2][1][0] as its logical dimension sequence) Running command: 1 setlayout array_obj DAOS,striped,0,2,128,<1,2,0> Call iod_obj_set_layout(), tid: 0x1 Layout to set: location type target_start target_num stripe_size dims_seq DAOS striped , 2, 0 Finish to set object layout - passed (rc = 0) 13

14 Step 2.7: Commit TID=1 Running command: 1 transfinish normal Call iod_trans_finish(), tid: 0x1, num_ranks: 0, flag: 0 Finish to finish transaction - passed (rc = 0) 14

15 Step 3.2: Query Eventual Persist Layout Running command: 1 getlayout array_obj At tid 0x1: dump IOD returned layout: location type target_start target_num stripe_size dims_seq DAOS striped , 2, 0 Rank 0 ID 0: get_layout verification succeed Finish to get object layout - passed (rc = 0) 15

16 Step 3.3: Query Semantic Shards on IONs Running command: 1 querymap array_obj Call iod_obj_query_map(), tid: 0x1 Rank 0 in tid 0x1: dump IOD returned map: obj_id type n_range b6056d ARRAY 64 range: 0, start: 0, 0, 0, end: 0, 7, 63, n_cell: 512, loc: /mnt/lustre1/.plfs shadow range: 1, start: 0, 8, 0, end: 0, 15, 63, n_cell: 512, loc: /mnt/lustre2/.plfs shadow range: 2, start: 0, 16, 0, end: 0, 23, 63, n_cell: 512, loc: /mnt/lustre3/.plfs shadow range: 3, start: 0, 24, 0, end: 0, 31, 63, n_cell: 512, loc: /mnt/lustre4/.plfs shadow range: 4, start: 1, 0, 0, end: 1, 7, 63, n_cell: 512, loc: /mnt/lustre1/.plfs shadow range: 5, start: 1, 8, 0, end: 1, 15, 63, n_cell: 512, loc: /mnt/lustre2/.plfs shadow range: 6, start: 1, 16, 0, end: 1, 23, 63, n_cell: 512, loc: /mnt/lustre3/.plfs shadow range: 7, start: 1, 24, 0, end: 1, 31, 63, n_cell: 512, loc: /mnt/lustre4/.plfs shadow <snip> range: 56, start: 14, 0, 0, end: 14, 7, 63, n_cell: 512, loc: /mnt/lustre1/.plfs shadow range: 57, start: 14, 8, 0, end: 14, 15, 63, n_cell: 512, loc: /mnt/lustre2/.plfs shadow range: 58, start: 14, 16, 0, end: 14, 23, 63, n_cell: 512, loc: /mnt/lustre3/.plfs shadow range: 59, start: 14, 24, 0, end: 14, 31, 63, n_cell: 512, loc: /mnt/lustre4/.plfs shadow range: 60, start: 15, 0, 0, end: 15, 7, 63, n_cell: 512, loc: /mnt/lustre1/.plfs shadow range: 61, start: 15, 8, 0, end: 15, 15, 63, n_cell: 512, loc: /mnt/lustre2/.plfs shadow range: 62, start: 15, 16, 0, end: 15, 23, 63, n_cell: 512, loc: /mnt/lustre3/.plfs shadow range: 63, start: 15, 24, 0, end: 15, 31, 63, n_cell: 512, loc: /mnt/lustre4/.plfs shadow Result of write pattern 16

17 Semantic Sharding on IONs ION 4 ION 3 ION 2 ION 1 17

18 Step 3.3: Query Semantic Shards on IONs # Verify data in BB s: [root@lola-3 scripts]# plfs_query /tmp/iod_plfs/containera/ /1 \ grep dropping.data xargs du -h 4.0M /mnt/lustre7/.plfs shadow///containera/ /1/hostdir.1/droppin g.data lola-7.lola.whamcloud.com M /mnt/lustre6/.plfs shadow///containera/ /1/hostdir.2/droppin g.data lola-6.lola.whamcloud.com M /mnt/lustre5/.plfs shadow///containera/ /1/hostdir.3/droppin g.data lola-5.lola.whamcloud.com M /mnt/lustre3/.plfs shadow///containera/ /1/hostdir.5/droppin g.data lola-3.lola.whamcloud.com

19 Step 4.1: Persist container at TID=1 Running command: 1 persist Call iod_trans_persist(), coh: 25, tid : 0x1... Finish to persist container - passed (rc = 0) # Verify proper scatter-gather migration: > [root@lola-3 scripts]# egrep 'reqmsg ACK' /tmp/iod.debug.log head 2014/03/23-16:52:33.07 lola-3 plfs[107367] IOD INFO handling reqmsg, type: 1, msg_len: 63.(rank 0) 2014/03/23-16:52:33.07 lola-3 plfs[107367] IOD INFO ENTER: _container_ops_trans_start_reqmsg_handler (rank 0)(rank 0) 2014/03/23-16:52:33.07 lola-3 plfs[107367] IOD INFO EXIT : _container_ops_trans_start_reqmsg_handler, rc: 0.(rank 0)(rank 0) 2014/03/23-16:52:33.07 lola-3 plfs[107367] IOD INFO handling reqmsg, type: 1, msg_len: 63.(rank 0) 2014/03/23-16:52:33.07 lola-3 plfs[107367] IOD INFO ENTER: _container_ops_trans_start_reqmsg_handler (rank 0)(rank 0) 2014/03/23-16:52:33.07 lola-3 plfs[107367] IOD INFO EXIT : _container_ops_trans_start_reqmsg_handler, rc: 0.(rank 0)(rank 0) 2014/03/23-16:52:33.07 lola-3 plfs[107367] IOD INFO sending ACK req_id 2, result: 0, piggyback: ffffffffffffffff from 0 to 0... (rank 0) 2014/03/23-16:52:33.07 lola-3 plfs[107367] IOD INFO handling reqmsg, type: 2, msg_len: 32.(rank 0) 2014/03/23-16:52:33.07 lola-3 plfs[107367] IOD INFO ACK arrived, from IOD rank 0, req_id: 2, result: 0.(rank 0) 2014/03/23-16:52:33.07 lola-3 plfs[107367] IOD INFO ENTER: iod_reqmsg_ack_handler (rank 0)(rank 0) 19

20 Step 4.1: Persist container at TID=1 # Verify data on DAOS: > [root@lola-3 scripts]#./daos_query.sh /mnt/daos/containera [root@lola-3 scripts]# daos_query.sh /mnt/daos/containera /scratch/iod/tests/output/daos_query.sh/mar ### ls -l /mnt/daos/containera -rw-r--r-- 1 root root Mar 23 19:55 /mnt/daos/containera ### du -h /mnt/daos/containera 27M /mnt/daos/containera #### daos_ctl run -c /mnt/daos/containera Cor,Cq,Eq,Cc: 1 HCE is 1 #### daos_ctl run -c /mnt/daos/containera Cor,Cq,Eq,Cc: 8 8 shards #### daos_ctl run -c /mnt/daos/containera Cor,Sq1:0,Sl1:0,Cc: 141, Shard 0 has 141, objects and uses (15 MB) Objects in shard 0: [0: ] [0: ] [0: ] [0: ] [0: ] #### daos_ctl run -c /mnt/daos/containera Cor,Sq1:1,Sl1:1,Cc: Shard 1 has 135, objects and uses (5 MB) Objects in shard 1: [1: ] <snip> #### daos_ctl run -c /mnt/daos/containera Cor,Sq1:7,Sl1:7,Cc: Shard 7 has 135, objects and uses (5 MB) Objects in shard 7: [7: ] 20

21 Semantic Sharding on DAOS KEY Shard0 Shard1 Shard2 Shard3 Shard4 Shard5 Shard6 Shard7 21

22 Step 4.2: Read array (from IONs) Running command: 1 read array_obj 0,0,0 8,8,8 2,4,8 8,8,8 Rank 0 ID 0: start[ 0,8,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 1 ID 0: start[ 0,16,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 2 ID 0: start[ 0,24,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 3 ID 0: start[ 0,0,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Call iod_array_read(), oh: 234, tid: 0x1 Rank 2 ID 0: iod_array_read() returns 0 Rank 1 ID 0: iod_array_read() returns 0 Rank 0 ID 0: iod_array_read() returns 0 Finish to read array - passed (rc = 0) Rank 3 ID 0: iod_array_read() returns 0 22

23 Reading Array (from IONs) Rank 2 Rank 1 Rank 0 Rank 3 ION 4 ION 3 ION 2 ION 1 23

24 Step 4.3/4.4: Purge Array / Query map Running command: 1 purge array_obj Call iod_obj_purge(), oh: 242, tid: 0x1... Finish to purge object - passed (rc = 0) Running command: 1 querymap array_obj Call iod_obj_query_map(), tid: 0x1 Rank 0 in tid 0x1: dump IOD returned map: obj_id type n_range b6056d ARRAY 0 Finish to query object map - passed (rc = 0) No data is IOD resident # Verify data purged: [root@lola-3 scripts]# plfs_query /tmp/iod_plfs/containera/ /1 \ grep dropping.data [root@lola-3 scripts]# 24

25 Step 4.5: Read Array (from DAOS) Running command: 1 read array_obj 0,0,0 8,8,8 2,4,8 8,8,8 Rank 0 ID 0: start[ 0,8,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 1 ID 0: start[ 0,16,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 2 ID 0: start[ 0,24,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 3 ID 0: start[ 0,0,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Call iod_array_read(), oh: 258, tid: 0x1 Rank 0 ID 0: iod_array_read() returns 0 Finish to read array - passed (rc = 0) Rank 3 ID 0: iod_array_read() returns 0 Rank 1 ID 0: iod_array_read() returns 0 Rank 2 ID 0: iod_array_read() returns 0 25

26 Step 4.5: Read Array (from DAOS) 26

27 Step 5.1/5.2: Fetch Array into 2 replicas # layout contains: location,type,target_start,target_num,strip_size,<dims_seq> # location "BB" (i.e. IOD_LOC_BB) # type only support stripped layout(i.e.iod_layout_striped) # target_start --- the beginning IOD # target_num the number of IODs. If 0, it defaults to all IODs # stripe_size ---- resharding granularity. The number of cells for contiguous layout array # dims_seq For contiguous layout, it is an array with num_dims size # # ######## Step prefetch object 1st: using one IOD and default dims_seq > Press enter to continue... Running command: 1 fetch array_obj tag1 BB,striped,0,1,64,<> Call iod_obj_fetch(), tid: 0x1... Rank 0: check all information saved for fetch: fake_tag tag location type target_start target_num stripe_size dims_seq tag1 0x BB striped (nil) Finish to fetch object - passed (rc = 0) Running command: 1 fetch array_obj tag2 BB,striped,0,0,1024,<2,0,1> Call iod_obj_fetch(), tid: 0x1... Rank 0: check all information saved for fetch: fake_tag tag location type target_start target_num stripe_size dims_seq tag1 0x BB striped (nil) tag2 0x BB striped , 0, 1 Finish to fetch object - passed (rc = 0) 27

28 Step 5.3: Query Replica 1 Shards on IONs Running command: * querymap array_obj tag1 Call iod_obj_query_map(), tag: 0x Rank 0 in tid 0x : dump IOD returned map: obj_id type n_range b6056d ARRAY 1 range: 0, start: 0, 0, 0, end: 15, 31, 63, n_cell: 32768, loc: /mnt/lustre3/.plfs shadow Finish to query object map - passed (rc = 0) 28

29 Step 5.4: Query Replica 2 Shards on IONs # ######## Step Query object's map: 2nd prefetched copy (should striped on all IODs) > Press enter to continue... Running command: * querymap array_obj tag2 Call iod_obj_query_map(), tag: 0x Rank 0 in tid 0x : dump IOD returned map: obj_id type n_range b6056d ARRAY 32 range: 0, start: 0, 0, 0, end: 15, 31, 1, n_cell: 1024, loc: /mnt/lustre1/.plfs shadow range: 1, start: 0, 0, 2, end: 15, 31, 3, n_cell: 1024, loc: /mnt/lustre2/.plfs shadow range: 2, start: 0, 0, 4, end: 15, 31, 5, n_cell: 1024, loc: /mnt/lustre3/.plfs shadow range: 3, start: 0, 0, 6, end: 15, 31, 7, n_cell: 1024, loc: /mnt/lustre4/.plfs shadow range: 4, start: 0, 0, 8, end: 15, 31, 9, n_cell: 1024, loc: /mnt/lustre1/.plfs shadow range: 5, start: 0, 0, 10, end: 15, 31, 11, n_cell: 1024, loc: /mnt/lustre2/.plfs shadow range: 6, start: 0, 0, 12, end: 15, 31, 13, n_cell: 1024, loc: /mnt/lustre3/.plfs shadow range: 7, start: 0, 0, 14, end: 15, 31, 15, n_cell: 1024, loc: /mnt/lustre4/.plfs shadow <snip> range: 28, start: 0, 0, 56, end: 15, 31, 57, n_cell: 1024, loc: /mnt/lustre1/.plfs shadow range: 29, start: 0, 0, 58, end: 15, 31, 59, n_cell: 1024, loc: /mnt/lustre2/.plfs shadow range: 30, start: 0, 0, 60, end: 15, 31, 61, n_cell: 1024, loc: /mnt/lustre3/.plfs shadow range: 31, start: 0, 0, 62, end: 15, 31, 63, n_cell: 1024, loc: /mnt/lustre4/.plfs shadow Finish to query object map - passed (rc = 0) 29

30 Replica 1 Replica 2 KEY ION 1 ION 2 ION 3 ION 4 30

31 Step 5.5: Read Replica 1 # ######## Step 5.5 read data of the 1st prefetched copy with its tag Check prefetched copy of object array_obj(0x b6056d) at TAG = 0x Rank 0 ID 0: start[ 0,8,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 1 ID 0: start[ 0,16,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 2 ID 0: start[ 0,24,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 3 ID 0: start[ 0,0,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Call iod_array_read(), oh: 330, tid: 0x Rank 0 ID 0: iod_array_read() returns 0 Rank 0 ID 0: array IO data verification succeed Finish to read array - passed (rc = 0) Finish to readprefetched copy - passed (rc = 0) Rank 2 ID 0: iod_array_read() returns 0 Rank 1 ID 0: iod_array_read() returns 0 Rank 3 ID 0: iod_array_read() returns 0 31

32 Step 5.6: Read Replica 2 # ######## Step 5.6 read data of the 2nd prefetched copy with its tag Check prefetched copy of object array_obj(0x b6056d) at TAG = 0x Rank 0 ID 0: start[ 0,8,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 1 ID 0: start[ 0,16,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 2 ID 0: start[ 0,24,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Rank 3 ID 0: start[ 0,0,0 ], stride[ 8,8,8 ], count[ 2,1,8 ], block[ 8,8,8 ] Call iod_array_read(), oh: 338, tid: 0x Rank 1 ID 0: iod_array_read() returns 0 Rank 2 ID 0: iod_array_read() returns 0 Rank 0 ID 0: iod_array_read() returns 0 Finish to read array - passed (rc = 0) Finish to readprefetched copy - passed (rc = 0) 32

33 Reading Replica 1 KEY ION 1 ION 2 ION 3 ION 4 33

34 Reading Replica 2 KEY ION 1 ION 2 ION 3 ION 4 34

35 Step 5.7/6: Cleanup / Fini # ######## Step finish TID1 for read > Press enter to continue... Running command: 1 transfinish normal Call iod_trans_finish(), tid: 0x1, num_ranks: 0, flag: 0 Finish to finish transaction - passed (rc = 0) # # ######## Step 6: close the container and quit. > Press enter to continue... Running command: * contclose Call iod_container_close(), handle: Finish to close container - passed (rc = 0) > Press enter to continue... Running command: * quit M7.5_case1_array_migration_ops: overall_result============= 0(Success) M7.5_case1_array_migration_ops: overall_result============= 0(Success) M7.5_case1_array_migration_ops: overall_result============= 0(Success) M7.5_case1_array_migration_ops: overall_result============= 0(Success) 35

36 Semantic Migration Summary Original DAOS Replica 1 Replica 2 36

37

38 EFF-IO M7.5 Demo Semantic Migration of Multi-dimensional Arrays John Bent, Sorin Faibish, Xuezhao Liu, Harriet Qui, Haiying Tang, Jerry Tirrell, Jingwang Zhang, Kelly Zhang, Zhenhua Zhang 38

FastForward I/O and Storage: IOD M5 Demonstration (5.2, 5.3, 5.9, 5.10)

FastForward I/O and Storage: IOD M5 Demonstration (5.2, 5.3, 5.9, 5.10) FastForward I/O and Storage: IOD M5 Demonstration (5.2, 5.3, 5.9, 5.10) 1 EMC September, 2013 John Bent john.bent@emc.com Sorin Faibish faibish_sorin@emc.com Xuezhao Liu xuezhao.liu@emc.com Harriet Qiu

More information

High Level Design IOD KV Store FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O

High Level Design IOD KV Store FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O Date: January 10, 2013 High Level Design IOD KV Store FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O LLNS Subcontract No. Subcontractor Name Subcontractor Address B599860

More information

Milestone 6.3: Basic Analysis Shipping Demonstration

Milestone 6.3: Basic Analysis Shipping Demonstration The HDF Group Milestone 6.3: Basic Analysis Shipping Demonstration Ruth Aydt, Mohamad Chaarawi, Ivo Jimenez, Quincey Koziol, Jerome Soumagne 12/17/2013 NOTICE: THIS MANUSCRIPT HAS BEEN AUTHORED BY INTEL

More information

FastForward I/O and Storage: ACG 8.6 Demonstration

FastForward I/O and Storage: ACG 8.6 Demonstration FastForward I/O and Storage: ACG 8.6 Demonstration Kyle Ambert, Jaewook Yu, Arnab Paul Intel Labs June, 2014 NOTICE: THIS MANUSCRIPT HAS BEEN AUTHORED BY INTEL UNDER ITS SUBCONTRACT WITH LAWRENCE LIVERMORE

More information

8.5 End-to-End Demonstration Exascale Fast Forward Storage Team June 30 th, 2014

8.5 End-to-End Demonstration Exascale Fast Forward Storage Team June 30 th, 2014 8.5 End-to-End Demonstration Exascale Fast Forward Storage Team June 30 th, 2014 NOTICE: THIS MANUSCRIPT HAS BEEN AUTHORED BY INTEL, THE HDF GROUP, AND EMC UNDER INTEL S SUBCONTRACT WITH LAWRENCE LIVERMORE

More information

5.4 - DAOS Demonstration and Benchmark Report

5.4 - DAOS Demonstration and Benchmark Report 5.4 - DAOS Demonstration and Benchmark Report Johann LOMBARDI on behalf of the DAOS team September 25 th, 2013 Livermore (CA) NOTICE: THIS MANUSCRIPT HAS BEEN AUTHORED BY INTEL UNDER ITS SUBCONTRACT WITH

More information

The HDF Group Q5 Demo

The HDF Group Q5 Demo The HDF Group The HDF Group Q5 Demo 5.6 HDF5 Transaction API 5.7 Full HDF5 Dynamic Data Structure NOTICE: THIS MANUSCRIPT HAS BEEN AUTHORED BY INTEL UNDER ITS SUBCONTRACT WITH LAWRENCE LIVERMORE NATIONAL

More information

End-to-End Data Integrity in the Intel/EMC/HDF Group Exascale IO DOE Fast Forward Project

End-to-End Data Integrity in the Intel/EMC/HDF Group Exascale IO DOE Fast Forward Project End-to-End Data Integrity in the Intel/EMC/HDF Group Exascale IO DOE Fast Forward Project As presented by John Bent, EMC and Quincey Koziol, The HDF Group Truly End-to-End App provides checksum buffer

More information

Design Document (Historical) HDF5 Dynamic Data Structure Support FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O

Design Document (Historical) HDF5 Dynamic Data Structure Support FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O Date: July 24, 2013 Design Document (Historical) HDF5 Dynamic Data Structure Support FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O LLNS Subcontract No. Subcontractor

More information

Milestone 8.1: HDF5 Index Demonstration

Milestone 8.1: HDF5 Index Demonstration The HDF Group Milestone 8.1: HDF5 Index Demonstration Ruth Aydt, Mohamad Chaarawi, Quincey Koziol, Aleksandar Jelenak, Jerome Soumagne 06/30/2014 NOTICE: THIS MANUSCRIPT HAS BEEN AUTHORED BY THE HDF GROUP

More information

Milestone Burst Buffer & Data Integrity Demonstra>on Milestone End- to- End Epoch Recovery Demonstra>on

Milestone Burst Buffer & Data Integrity Demonstra>on Milestone End- to- End Epoch Recovery Demonstra>on he HF Group ilestone 7.2 - Burst Buffer & ata Integrity emonstra>on ilestone 7.3 - End- to- End Epoch Recovery emonstra>on NOICE: HIS ANUSCRIP HAS BEEN AUHORE BY HE HF GROUP UNER HE INEL SUBCONRAC WIH

More information

High Level Design Client Health and Global Eviction FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O MILESTONE: 4.

High Level Design Client Health and Global Eviction FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O MILESTONE: 4. Date: 2013-06-01 High Level Design Client Health and Global Eviction FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O MILESTONE: 4.1 LLNS Subcontract No. Subcontractor

More information

FastForward I/O and Storage: ACG 5.8 Demonstration

FastForward I/O and Storage: ACG 5.8 Demonstration FastForward I/O and Storage: ACG 5.8 Demonstration Jaewook Yu, Arnab Paul, Kyle Ambert Intel Labs September, 2013 NOTICE: THIS MANUSCRIPT HAS BEEN AUTHORED BY INTEL UNDER ITS SUBCONTRACT WITH LAWRENCE

More information

API and Usage of libhio on XC-40 Systems

API and Usage of libhio on XC-40 Systems API and Usage of libhio on XC-40 Systems May 24, 2018 Nathan Hjelm Cray Users Group May 24, 2018 Los Alamos National Laboratory LA-UR-18-24513 5/24/2018 1 Outline Background HIO Design HIO API HIO Configuration

More information

Lustre* - Fast Forward to Exascale High Performance Data Division. Eric Barton 18th April, 2013

Lustre* - Fast Forward to Exascale High Performance Data Division. Eric Barton 18th April, 2013 Lustre* - Fast Forward to Exascale High Performance Data Division Eric Barton 18th April, 2013 DOE Fast Forward IO and Storage Exascale R&D sponsored by 7 leading US national labs Solutions to currently

More information

libhio: Optimizing IO on Cray XC Systems With DataWarp

libhio: Optimizing IO on Cray XC Systems With DataWarp libhio: Optimizing IO on Cray XC Systems With DataWarp May 9, 2017 Nathan Hjelm Cray Users Group May 9, 2017 Los Alamos National Laboratory LA-UR-17-23841 5/8/2017 1 Outline Background HIO Design Functionality

More information

DAOS Epoch Recovery Design FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O

DAOS Epoch Recovery Design FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O Date: June 4, 2014 DAOS Epoch Recovery Design FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O LLNS Subcontract No. Subcontractor Name Subcontractor Address B599860 Intel

More information

Fast Forward I/O & Storage

Fast Forward I/O & Storage Fast Forward I/O & Storage Eric Barton Lead Architect 1 Department of Energy - Fast Forward Challenge FastForward RFP provided US Government funding for exascale research and development Sponsored by 7

More information

I/O in scientific applications

I/O in scientific applications COSC 4397 Parallel I/O (II) Access patterns Spring 2010 I/O in scientific applications Different classes of I/O operations Required I/O: reading input data and writing final results Checkpointing: data

More information

Versioning Object Storage Device (VOSD) Design Document FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O

Versioning Object Storage Device (VOSD) Design Document FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O Date: June 4, 2014 Versioning Object Storage Device (VOSD) Design Document FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O LLNS Subcontract No. Subcontractor Name Subcontractor

More information

Got Burst Buffer. Now What? Early experiences, exciting future possibilities, and what we need from the system to make it work

Got Burst Buffer. Now What? Early experiences, exciting future possibilities, and what we need from the system to make it work Got Burst Buffer. Now What? Early experiences, exciting future possibilities, and what we need from the system to make it work The Salishan Conference on High-Speed Computing April 26, 2016 Adam Moody

More information

DAOS API and DAOS POSIX DESIGN DOCUMENT FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O

DAOS API and DAOS POSIX DESIGN DOCUMENT FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O Date: December 13, 2012 DAOS API and DAOS POSIX DESIGN DOCUMENT FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O LLNS Subcontract No. Subcontractor Name Subcontractor

More information

EMC Lustre Contributions

EMC Lustre Contributions EMC Lustre Contributions It s all about speed. Tao Peng Xuezhao Liu as presented by John Bent Fast Data Group Office of the CTO 1 EMC Lustre activities Support Lustre bug fixes (LU-1126, LU-1322, etc.)

More information

RFC: HDF5 File Space Management: Paged Aggregation

RFC: HDF5 File Space Management: Paged Aggregation RFC: HDF5 File Space Management: Paged Aggregation Vailin Choi Quincey Koziol John Mainzer The current HDF5 file space allocation accumulates small pieces of metadata and raw data in aggregator blocks.

More information

Fast Forward Storage & I/O. Jeff Layton (Eric Barton)

Fast Forward Storage & I/O. Jeff Layton (Eric Barton) Fast Forward & I/O Jeff Layton (Eric Barton) DOE Fast Forward IO and Exascale R&D sponsored by 7 leading US national labs Solutions to currently intractable problems of Exascale required to meet the 2020

More information

Lustre and PLFS Parallel I/O Performance on a Cray XE6

Lustre and PLFS Parallel I/O Performance on a Cray XE6 Lustre and PLFS Parallel I/O Performance on a Cray XE6 Cray User Group 2014 Lugano, Switzerland May 4-8, 2014 April 2014 1 Many currently contributing to PLFS LANL: David Bonnie, Aaron Caldwell, Gary Grider,

More information

A Plugin for HDF5 using PLFS for Improved I/O Performance and Semantic Analysis

A Plugin for HDF5 using PLFS for Improved I/O Performance and Semantic Analysis 2012 SC Companion: High Performance Computing, Networking Storage and Analysis A for HDF5 using PLFS for Improved I/O Performance and Semantic Analysis Kshitij Mehta, John Bent, Aaron Torres, Gary Grider,

More information

High Level Design Server Collectives FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O

High Level Design Server Collectives FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O Date: June 05, 2013 High Level Design Server Collectives FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O LLNS Subcontract No. Subcontractor Name Subcontractor Address

More information

HDF5 User s Guide. HDF5 Release November

HDF5 User s Guide. HDF5 Release November HDF5 User s Guide HDF5 Release 1.8.8 November 2011 http://www.hdfgroup.org Copyright Notice and License Terms for HDF5 (Hierarchical Data Format 5) Software Library and Utilities HDF5 (Hierarchical Data

More information

What is a file system

What is a file system COSC 6397 Big Data Analytics Distributed File Systems Edgar Gabriel Spring 2017 What is a file system A clearly defined method that the OS uses to store, catalog and retrieve files Manage the bits that

More information

6.5 Collective Open/Close & Epoch Distribution Demonstration

6.5 Collective Open/Close & Epoch Distribution Demonstration 6.5 Collective Open/Close & Epoch Distribution Demonstration Johann LOMBARDI on behalf of the DAOS team December 17 th, 2013 Fast Forward Project - DAOS DAOS Development Update Major accomplishments of

More information

UK LUG 10 th July Lustre at Exascale. Eric Barton. CTO Whamcloud, Inc Whamcloud, Inc.

UK LUG 10 th July Lustre at Exascale. Eric Barton. CTO Whamcloud, Inc Whamcloud, Inc. UK LUG 10 th July 2012 Lustre at Exascale Eric Barton CTO Whamcloud, Inc. eeb@whamcloud.com Agenda Exascale I/O requirements Exascale I/O model 3 Lustre at Exascale - UK LUG 10th July 2012 Exascale I/O

More information

Structuring PLFS for Extensibility

Structuring PLFS for Extensibility Structuring PLFS for Extensibility Chuck Cranor, Milo Polte, Garth Gibson PARALLEL DATA LABORATORY Carnegie Mellon University What is PLFS? Parallel Log Structured File System Interposed filesystem b/w

More information

Vitess on Kubernetes. followed by a demo of VReplication. Jiten Vaidya

Vitess on Kubernetes. followed by a demo of VReplication. Jiten Vaidya Vitess on Kubernetes followed by a demo of VReplication Jiten Vaidya jiten@planetscale.com A word about me... Jiten Vaidya - Managed teams that operationalized Vitess at Youtube CEO at PlanetScale Founded

More information

Parallel I/O Libraries and Techniques

Parallel I/O Libraries and Techniques Parallel I/O Libraries and Techniques Mark Howison User Services & Support I/O for scientifc data I/O is commonly used by scientific applications to: Store numerical output from simulations Load initial

More information

Common Persistent Memory POSIX Runtime (CPPR) API Reference Manual. Reference Manual High Performance Data Division

Common Persistent Memory POSIX Runtime (CPPR) API Reference Manual. Reference Manual High Performance Data Division Common Persistent Memory POSIX Runtime (CPPR) Reference Manual High Performance Data Division INTEL FEDERAL, LLC PROPRIETARY October 2016 Generated under Argonne Contract number: B609815 DISTRIBUTION STATEMENT:

More information

SDS: A Framework for Scientific Data Services

SDS: A Framework for Scientific Data Services SDS: A Framework for Scientific Data Services Bin Dong, Suren Byna*, John Wu Scientific Data Management Group Lawrence Berkeley National Laboratory Finding Newspaper Articles of Interest Finding news articles

More information

DAOS Server Collectives Design FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O

DAOS Server Collectives Design FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O Date: June 05, 2013 DAOS Server Collectives Design FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O LLNS Subcontract No. Subcontractor Name Subcontractor Address B599860

More information

libhio: Optimizing IO on Cray XC Systems With DataWarp

libhio: Optimizing IO on Cray XC Systems With DataWarp libhio: Optimizing IO on Cray XC Systems With DataWarp Nathan T. Hjelm, Cornell Wright Los Alamos National Laboratory Los Alamos, NM {hjelmn, cornell}@lanl.gov Abstract High performance systems are rapidly

More information

PLFS and Lustre Performance Comparison

PLFS and Lustre Performance Comparison PLFS and Lustre Performance Comparison Lustre User Group 2014 Miami, FL April 8-10, 2014 April 2014 1 Many currently contributing to PLFS LANL: David Bonnie, Aaron Caldwell, Gary Grider, Brett Kettering,

More information

Reduction Network Discovery Design Document FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O

Reduction Network Discovery Design Document FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O Date: May 01, 2014 Reduction Network Discovery Design Document FOR EXTREME-SCALE COMPUTING RESEARCH AND DEVELOPMENT (FAST FORWARD) STORAGE AND I/O LLNS Subcontract No. Subcontractor Name Subcontractor

More information

DAOS and Friends: A Proposal for an Exascale Storage System

DAOS and Friends: A Proposal for an Exascale Storage System DAOS and Friends: A Proposal for an Exascale Storage System Jay Lofstead, Ivo Jimenez, Carlos Maltzahn, Quincey Koziol, John Bent, Eric Barton Sandia National Laboratories gflofst@sandia.gov University

More information

Designing a global repository using OceanStore

Designing a global  repository using OceanStore Designing a global EMail repository using OceanStore Steven Czerwinski, Anthony Joseph, John Kubiatowicz Summer Retreat June 11, 2002 UC Berkeley Goal Build an another interesting OceanStore app Further

More information

File Open, Close, and Flush Performance Issues in HDF5 Scot Breitenfeld John Mainzer Richard Warren 02/19/18

File Open, Close, and Flush Performance Issues in HDF5 Scot Breitenfeld John Mainzer Richard Warren 02/19/18 File Open, Close, and Flush Performance Issues in HDF5 Scot Breitenfeld John Mainzer Richard Warren 02/19/18 1 Introduction Historically, the parallel version of the HDF5 library has suffered from performance

More information

HPC Input/Output. I/O and Darshan. Cristian Simarro User Support Section

HPC Input/Output. I/O and Darshan. Cristian Simarro User Support Section HPC Input/Output I/O and Darshan Cristian Simarro Cristian.Simarro@ecmwf.int User Support Section Index Lustre summary HPC I/O Different I/O methods Darshan Introduction Goals Considerations How to use

More information

Power Bounds and Large Scale Computing

Power Bounds and Large Scale Computing 1 Power Bounds and Large Scale Computing Friday, March 1, 2013 Bronis R. de Supinski 1 Tapasya Patki 2, David K. Lowenthal 2, Barry L. Rountree 1 and Martin Schulz 1 2 University of Arizona This work has

More information

Move Exchange 2010 Database To Another Drive Powershell

Move Exchange 2010 Database To Another Drive Powershell Move Exchange 2010 Database To Another Drive Powershell Tip. How to move default database in Exchange 2010 / 2013. How to delete or move them to another drive automatically? Clear IIS logs:powershell script.

More information

Distributed File Systems II

Distributed File Systems II Distributed File Systems II To do q Very-large scale: Google FS, Hadoop FS, BigTable q Next time: Naming things GFS A radically new environment NFS, etc. Independence Small Scale Variety of workloads Cooperation

More information

DELL EMC ISILON F800 AND H600 I/O PERFORMANCE

DELL EMC ISILON F800 AND H600 I/O PERFORMANCE DELL EMC ISILON F800 AND H600 I/O PERFORMANCE ABSTRACT This white paper provides F800 and H600 performance data. It is intended for performance-minded administrators of large compute clusters that access

More information

OPERATING SYSTEM. Chapter 12: File System Implementation

OPERATING SYSTEM. Chapter 12: File System Implementation OPERATING SYSTEM Chapter 12: File System Implementation Chapter 12: File System Implementation File-System Structure File-System Implementation Directory Implementation Allocation Methods Free-Space Management

More information

Metadata Performance Evaluation LUG Sorin Faibish, EMC Branislav Radovanovic, NetApp and MD BWG April 8-10, 2014

Metadata Performance Evaluation LUG Sorin Faibish, EMC Branislav Radovanovic, NetApp and MD BWG April 8-10, 2014 Metadata Performance Evaluation Effort @ LUG 2014 Sorin Faibish, EMC Branislav Radovanovic, NetApp and MD BWG April 8-10, 2014 OpenBenchmark Metadata Performance Evaluation Effort (MPEE) Team Leader: Sorin

More information

Use of a new I/O stack for extreme-scale systems in scientific applications

Use of a new I/O stack for extreme-scale systems in scientific applications 1 Use of a new I/O stack for extreme-scale systems in scientific applications M. Scot Breitenfeld a, Quincey Koziol b, Neil Fortner a, Jerome Soumagne a, Mohamad Chaarawi a a The HDF Group, b Lawrence

More information

Megastore: Providing Scalable, Highly Available Storage for Interactive Services & Spanner: Google s Globally- Distributed Database.

Megastore: Providing Scalable, Highly Available Storage for Interactive Services & Spanner: Google s Globally- Distributed Database. Megastore: Providing Scalable, Highly Available Storage for Interactive Services & Spanner: Google s Globally- Distributed Database. Presented by Kewei Li The Problem db nosql complex legacy tuning expensive

More information

Adding a System Call to Plan 9

Adding a System Call to Plan 9 Adding a System Call to Plan 9 John Floren (john@csplan9.rit.edu) Sandia National Laboratories Livermore, CA 94551 DOE/NNSA Funding Statement Sandia is a multiprogram laboratory operated by Sandia Corporation,

More information

Common Persistent Memory POSIX* Runtime (CPPR) API Reference (MS21) API Reference High Performance Data Division

Common Persistent Memory POSIX* Runtime (CPPR) API Reference (MS21) API Reference High Performance Data Division Common Persistent Memory POSIX* Runtime (CPPR) API Reference High Performance Data Division INTEL FEDERAL, LLC PROPRIETARY December 2017 Generated under Argonne Contract number: B609815 DISTRIBUTION STATEMENT:

More information

Improving Per Processor Memory Use of ns-3 to Enable Large Scale Simulations

Improving Per Processor Memory Use of ns-3 to Enable Large Scale Simulations Improving Per Processor Memory Use of ns-3 to Enable Large Scale Simulations WNS3 2015, Castelldefels (Barcelona), Spain May 13, 2015 Steven Smith, David R. Jefferson Peter D. Barnes, Jr, Sergei Nikolaev

More information

Chapter 11: Implementing File Systems

Chapter 11: Implementing File Systems Chapter 11: Implementing File Systems Operating System Concepts 99h Edition DM510-14 Chapter 11: Implementing File Systems File-System Structure File-System Implementation Directory Implementation Allocation

More information

Teamcenter Volume Management Guide. Publication Number PLM00104 I

Teamcenter Volume Management Guide. Publication Number PLM00104 I Teamcenter 10.1 Volume Management Guide Publication Number PLM00104 I Proprietary and restricted rights notice This software and related documentation are proprietary to Siemens Product Lifecycle Management

More information

Chapter 12: File System Implementation

Chapter 12: File System Implementation Chapter 12: File System Implementation Silberschatz, Galvin and Gagne 2013 Chapter 12: File System Implementation File-System Structure File-System Implementation Allocation Methods Free-Space Management

More information

CS November 2017

CS November 2017 Bigtable Highly available distributed storage Distributed Systems 18. Bigtable Built with semi-structured data in mind URLs: content, metadata, links, anchors, page rank User data: preferences, account

More information

<Insert Picture Here> End-to-end Data Integrity for NFS

<Insert Picture Here> End-to-end Data Integrity for NFS End-to-end Data Integrity for NFS Chuck Lever Consulting Member of Technical Staff Today s Discussion What is end-to-end data integrity? T10 PI overview Adapting

More information

Sharding Introduction

Sharding Introduction search MongoDB Home Admin Zone Sharding Sharding Introduction Sharding Introduction MongoDB supports an automated sharding architecture, enabling horizontal scaling across multiple nodes. For applications

More information

The JANUS Computing Environment

The JANUS Computing Environment Research Computing UNIVERSITY OF COLORADO The JANUS Computing Environment Monte Lunacek monte.lunacek@colorado.edu rc-help@colorado.edu What is JANUS? November, 2011 1,368 Compute nodes 16,416 processors

More information

CHAPTER 11: IMPLEMENTING FILE SYSTEMS (COMPACT) By I-Chen Lin Textbook: Operating System Concepts 9th Ed.

CHAPTER 11: IMPLEMENTING FILE SYSTEMS (COMPACT) By I-Chen Lin Textbook: Operating System Concepts 9th Ed. CHAPTER 11: IMPLEMENTING FILE SYSTEMS (COMPACT) By I-Chen Lin Textbook: Operating System Concepts 9th Ed. File-System Structure File structure Logical storage unit Collection of related information File

More information

Parallel File Systems. John White Lawrence Berkeley National Lab

Parallel File Systems. John White Lawrence Berkeley National Lab Parallel File Systems John White Lawrence Berkeley National Lab Topics Defining a File System Our Specific Case for File Systems Parallel File Systems A Survey of Current Parallel File Systems Implementation

More information

MS 52 Distributed Persistent Memory Class Storage Model (DAOS-M) Solution Architecture Revision 1.2 September 28, 2015

MS 52 Distributed Persistent Memory Class Storage Model (DAOS-M) Solution Architecture Revision 1.2 September 28, 2015 MS 52 Distributed Persistent Memory Class Storage Model (DAOS-M) Solution Architecture Revision 1.2 September 28, 2015 Intel Federal, LLC Proprietary i Solution Architecture Generated under Argonne Contract

More information

Guidelines for Efficient Parallel I/O on the Cray XT3/XT4

Guidelines for Efficient Parallel I/O on the Cray XT3/XT4 Guidelines for Efficient Parallel I/O on the Cray XT3/XT4 Jeff Larkin, Cray Inc. and Mark Fahey, Oak Ridge National Laboratory ABSTRACT: This paper will present an overview of I/O methods on Cray XT3/XT4

More information

The State and Needs of IO Performance Tools

The State and Needs of IO Performance Tools The State and Needs of IO Performance Tools Scalable Tools Workshop Lake Tahoe, CA August 6 12, 2017 This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National

More information

Timely Safe Recording. Nero Copyright

Timely Safe Recording. Nero Copyright Timely Safe Recording ## Permission is granted to members of INCITS, its technical ## committees and their associated task groups to reproduce ## this document for the purposes of INCITS standardization

More information

MPI Runtime Error Detection with MUST

MPI Runtime Error Detection with MUST MPI Runtime Error Detection with MUST At the 27th VI-HPS Tuning Workshop Joachim Protze IT Center RWTH Aachen University April 2018 How many issues can you spot in this tiny example? #include #include

More information

Today CSCI Coda. Naming: Volumes. Coda GFS PAST. Instructor: Abhishek Chandra. Main Goals: Volume is a subtree in the naming space

Today CSCI Coda. Naming: Volumes. Coda GFS PAST. Instructor: Abhishek Chandra. Main Goals: Volume is a subtree in the naming space Today CSCI 5105 Coda GFS PAST Instructor: Abhishek Chandra 2 Coda Main Goals: Availability: Work in the presence of disconnection Scalability: Support large number of users Successor of Andrew File System

More information

Chapter 12: File System Implementation

Chapter 12: File System Implementation Chapter 12: File System Implementation Chapter 12: File System Implementation File-System Structure File-System Implementation Directory Implementation Allocation Methods Free-Space Management Efficiency

More information

State of OpenMP & Outlook on OpenMP 4.1

State of OpenMP & Outlook on OpenMP 4.1 State of OpenMP & Outlook on OpenMP 4.1 Thursday, October 11, 2015 Bronis R. de Supinski Chair, OpenMP Language Committee This work has been authored by Lawrence Livermore National Security, LLC under

More information

PNUTS and Weighted Voting. Vijay Chidambaram CS 380 D (Feb 8)

PNUTS and Weighted Voting. Vijay Chidambaram CS 380 D (Feb 8) PNUTS and Weighted Voting Vijay Chidambaram CS 380 D (Feb 8) PNUTS Distributed database built by Yahoo Paper describes a production system Goals: Scalability Low latency, predictable latency Must handle

More information

NIF ICCS Test Controller for Automated & Manual Testing

NIF ICCS Test Controller for Automated & Manual Testing UCRL-CONF-235325 NIF ICCS Test Controller for Automated & Manual Testing J. S. Zielinski October 5, 2007 International Conference on Accelerator and Large Experimental Physics Control Systems Knoxville,

More information

Lustre A Platform for Intelligent Scale-Out Storage

Lustre A Platform for Intelligent Scale-Out Storage Lustre A Platform for Intelligent Scale-Out Storage Rumi Zahir, rumi. May 2003 rumi.zahir@intel.com Agenda Problem Statement Trends & Current Data Center Storage Architectures The Lustre File System Project

More information

ECE 574 Cluster Computing Lecture 13

ECE 574 Cluster Computing Lecture 13 ECE 574 Cluster Computing Lecture 13 Vince Weaver http://web.eece.maine.edu/~vweaver vincent.weaver@maine.edu 21 March 2017 Announcements HW#5 Finally Graded Had right idea, but often result not an *exact*

More information

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme STO1926BU A Day in the Life of a VSAN I/O Diving in to the I/O Flow of vsan John Nicholson (@lost_signal) Pete Koehler (@vmpete) VMworld 2017 Content: Not for publication #VMworld #STO1926BU Disclaimer

More information

A More Realistic Way of Stressing the End-to-end I/O System

A More Realistic Way of Stressing the End-to-end I/O System A More Realistic Way of Stressing the End-to-end I/O System Verónica G. Vergara Larrea Sarp Oral Dustin Leverman Hai Ah Nam Feiyi Wang James Simmons CUG 2015 April 29, 2015 Chicago, IL ORNL is managed

More information

cstore_fdw Columnar store for analytic workloads Hadi Moshayedi & Ben Redman

cstore_fdw Columnar store for analytic workloads Hadi Moshayedi & Ben Redman cstore_fdw Columnar store for analytic workloads Hadi Moshayedi & Ben Redman What is CitusDB? CitusDB is a scalable analytics database that extends PostgreSQL Citus shards your data and automa/cally parallelizes

More information

Town Crier. Authenticated Data Feeds For Smart Contracts. CS5437 Lecture by Kyle Croman and Fan Zhang Mar 18, 2016

Town Crier. Authenticated Data Feeds For Smart Contracts. CS5437 Lecture by Kyle Croman and Fan Zhang Mar 18, 2016 Town Crier Authenticated Data Feeds For Smart Contracts CS5437 Lecture by Kyle Croman and Fan Zhang Mar 18, 2016 Smart Contract Decentralized App: Programs are executed by all miners who reach consensus

More information

MPI Runtime Error Detection with MUST

MPI Runtime Error Detection with MUST MPI Runtime Error Detection with MUST At the 25th VI-HPS Tuning Workshop Joachim Protze IT Center RWTH Aachen University March 2017 How many issues can you spot in this tiny example? #include #include

More information

METADATA REGISTRY, ISO/IEC 11179

METADATA REGISTRY, ISO/IEC 11179 LLNL-JRNL-400269 METADATA REGISTRY, ISO/IEC 11179 R. K. Pon, D. J. Buttler January 7, 2008 Encyclopedia of Database Systems Disclaimer This document was prepared as an account of work sponsored by an agency

More information

GLUSTER CAN DO THAT! Architecting and Performance Tuning Efficient Gluster Storage Pools

GLUSTER CAN DO THAT! Architecting and Performance Tuning Efficient Gluster Storage Pools GLUSTER CAN DO THAT! Architecting and Performance Tuning Efficient Gluster Storage Pools Dustin Black Senior Architect, Software-Defined Storage @dustinlblack 2017-05-02 Ben Turner Principal Quality Engineer

More information

Application Performance on IME

Application Performance on IME Application Performance on IME Toine Beckers, DDN Marco Grossi, ICHEC Burst Buffer Designs Introduce fast buffer layer Layer between memory and persistent storage Pre-stage application data Buffer writes

More information

Indexing. Jan Chomicki University at Buffalo. Jan Chomicki () Indexing 1 / 25

Indexing. Jan Chomicki University at Buffalo. Jan Chomicki () Indexing 1 / 25 Indexing Jan Chomicki University at Buffalo Jan Chomicki () Indexing 1 / 25 Storage hierarchy Cache Main memory Disk Tape Very fast Fast Slower Slow (nanosec) (10 nanosec) (millisec) (sec) Very small Small

More information

Update 9/16/16: Version published to the ServiceNow store now supports Helsinki, Istanbul and Jakarta.

Update 9/16/16: Version published to the ServiceNow store now supports Helsinki, Istanbul and Jakarta. Qualys CMDB Sync App The Qualys CMDB Sync App synchronizes Qualys IT asset discovery and classification with the ServiceNow Configuration Management Database (CMDB) system. The App automatically updates

More information

Reference Manual of Impulse System Calls. Abstract

Reference Manual of Impulse System Calls. Abstract Reference Manual of Impulse System Calls Lixin Zhang, Leigh Stoller UUCS-99-018 Department of Computer Science University of Utah Salt Lake City, UT 84112, USA January 20, 1999 Abstract This document describes

More information

Progress on Efficient Integration of Lustre* and Hadoop/YARN

Progress on Efficient Integration of Lustre* and Hadoop/YARN Progress on Efficient Integration of Lustre* and Hadoop/YARN Weikuan Yu Robin Goldstone Omkar Kulkarni Bryon Neitzel * Some name and brands may be claimed as the property of others. MapReduce l l l l A

More information

Leveraging Flash in HPC Systems

Leveraging Flash in HPC Systems Leveraging Flash in HPC Systems IEEE MSST June 3, 2015 This work was performed under the auspices of the U.S. Department of Energy by under Contract DE-AC52-07NA27344. Lawrence Livermore National Security,

More information

Checkpoint/Restart System Services Interface (SSI) Modules for LAM/MPI API Version / SSI Version 1.0.0

Checkpoint/Restart System Services Interface (SSI) Modules for LAM/MPI API Version / SSI Version 1.0.0 Checkpoint/Restart System Services Interface (SSI) Modules for LAM/MPI API Version 1.0.0 / SSI Version 1.0.0 Sriram Sankaran Jeffrey M. Squyres Brian Barrett Andrew Lumsdaine http://www.lam-mpi.org/ Open

More information

Distributed Memory Parallel Programming

Distributed Memory Parallel Programming COSC Big Data Analytics Parallel Programming using MPI Edgar Gabriel Spring 201 Distributed Memory Parallel Programming Vast majority of clusters are homogeneous Necessitated by the complexity of maintaining

More information

HP Data Protector Integration with Autonomy IDOL Server

HP Data Protector Integration with Autonomy IDOL Server Technical white paper HP Data Protector Integration with Autonomy IDOL Server Introducing e-discovery for HP Data Protector environments Table of contents Summary 2 Introduction 2 Integration concepts

More information

OpenZFS Performance Improvements

OpenZFS Performance Improvements OpenZFS Performance Improvements LUG Developer Day 2015 April 16, 2015 Brian, Behlendorf This work was performed under the auspices of the U.S. Department of Energy by under Contract DE-AC52-07NA27344.

More information

The UNIX Time- Sharing System

The UNIX Time- Sharing System The UNIX Time- Sharing System Dennis M. Ritchie and Ken Thompson Bell Laboratories Communications of the ACM July 1974, Volume 17, Number 7 UNIX overview Unix is a general-purpose, multi-user, interactive

More information

Introducing Cache Pseudo-Locking to reduce memory access latency. Reinette Chatre

Introducing Cache Pseudo-Locking to reduce memory access latency. Reinette Chatre Introducing Cache Pseudo-Locking to reduce memory access latency Reinette Chatre About me Software Engineer at Intel (~12 years) Open Source Technology Center (OTC) Currently Enabling Cache Pseudo-Locking

More information

Database infrastructure for electronic structure calculations

Database infrastructure for electronic structure calculations Database infrastructure for electronic structure calculations Fawzi Mohamed fawzi.mohamed@fhi-berlin.mpg.de 22.7.2015 Why should you be interested in databases? Can you find a calculation that you did

More information

Lustre * Features In Development Fan Yong High Performance Data Division, Intel CLUG

Lustre * Features In Development Fan Yong High Performance Data Division, Intel CLUG Lustre * Features In Development Fan Yong High Performance Data Division, Intel CLUG 2017 @Beijing Outline LNet reliability DNE improvements Small file performance File Level Redundancy Miscellaneous improvements

More information

DDN and Flash GRIDScaler, Flashscale Infinite Memory Engine

DDN and Flash GRIDScaler, Flashscale Infinite Memory Engine 1! DDN and Flash GRIDScaler, Flashscale Infinite Memory Engine T. Cecchi - September 21 st 2016 HPC Advisory Council 2! DDN END-TO-END DATA LIFECYCLE MANAGEMENT BURST & COMPUTE SSD, DISK & FILE SYSTEM

More information

Tutorial 4: Condor. John Watt, National e-science Centre

Tutorial 4: Condor. John Watt, National e-science Centre Tutorial 4: Condor John Watt, National e-science Centre Tutorials Timetable Week Day/Time Topic Staff 3 Fri 11am Introduction to Globus J.W. 4 Fri 11am Globus Development J.W. 5 Fri 11am Globus Development

More information