Overview of Enhancing Database Performance and I/O Techniques

Size: px
Start display at page:

Download "Overview of Enhancing Database Performance and I/O Techniques"

Transcription

1 Overview of Enhancing Database Performance and I/O Techniques By: Craig Borysowich Enterprise Architect Imagination Edge Inc.

2 Database Performance and I/O Overview Database Performance Components In addition to the disk subsystem there are various components affecting the performance of any database application. In most cases, a good part of the performance improvement can be achieved from database design, physical layout of tables and indexes, appropriate indexes and query optimization techniques. Detailed descriptions of these areas are beyond the scope of this document. Please refer to the respective vendor s performance tuning guide for more information. Database I/O Overview To correctly configure the disk subsystem of a database server a basic understanding of database access patterns and methods is required. This section provides a high level overview of the database I/O process and typical patterns of common application types. This information will be referenced throughout the remainder of this paper. Database I/O Schemes Database systems store data in a logical collection of disk files or disk partitions. The physically atomic unit of a database file is referred to as a database page or database block. The typical size of a database page is 2K-16K. Most databases don't allow the user to change this size. However, some databases allow the administrator to define the page size at database creation time. In general, smaller database page sizes are desired for OLTP applications, and larger page sizes are good for DSS applications. More about these application types will be discussed later in this section. Physical reads and writes of the data are done at the database page level. Therefore, most disk requests are performed using the 2K database page size. There are some exceptions to this on large table scans. If the database system detects that a long scan will be done it will often read in larger blocks, but this feature doesn't usually change the tuning process. The process of executing I/O routines is similar across most database systems. Reads of database pages are done by database process(es) or thread(s) servicing the application. Many read operations can usually be conducted in parallel and each is independent of the other. Writes of data pages, on the other hand, are not initiated by a user application. When an application updates or inserts new data, the change is actually put into the database cache by the application's database process. Some time later an independent database function will clean the dirty data pages from memory. These writes will usually occur as a large batch of asynchronous

3 writes. Hence write operations are not usually a steady stream of single page writes, but periodic, heavy bursts of write activities. In contrast to the data pages, writes of transaction log pages are a very steady stream of single page writes. These occur during insert, update and delete operations and again are controlled by background database processes, such as checkpoint or lazy write process. Database Application Types By examining different database applications and database functions, the fundamental I/O profile used to access the disk can be determined. This determines whether a particular data set will be accessed either sequentially or randomly. The key high-level application types include transaction processing, decision support and batch processing systems. In this section we will examine the various access patterns on the database files and later apply these patterns to how the drives should be configured. The distinct file types to be examined are data and index files, transaction logs, and import/export files. Online Transaction Processing (OLTP) Online transaction processing (OLTP) applications are high-throughput, insert/update intensive systems. These systems are usually characterized by constantly growing large volumes of data that large number of users access concurrently. Examples of typical OLTP applications are order processing applications, banking applications and reservation systems. The key goals of an OLTP system are availability, speed, concurrency, and recoverability. The I/O profile resulting from this load is heavy random reads and writes across the data and index files. The transaction logs, however, are hit with a heavy stream of sequential write operations of 2K-16K in size. Decision Support System (DSS) Decision support or data warehousing applications perform queries on the large amount of data, usually gathered from OLTP applications, and provide competitive information to the decisionmakers in an organization. Decision support systems must handle complex queries on data that is less frequently updated than OLTP systems. DSS is characteristic of multiple users executing complicated joins and aggregations on large sets of data. Even though many of the operations could use some sequential processing, contention with other users and join and indexing operations result in a fairly random access pattern to the data and index files. Usually, if the database is dedicated to DSS, no updates will be done to the database during heavy query load. In this case, no I/O will occur on the transaction logs. If online updates are applied to the database, it must be remembered when configuring the drive subsystem that log file response time may still be important to the overall performance of the system. Characteristics of OLTP and DSS are compared in the Table 1.

4 OLTP DSS Transaction Locking Performance Number of rows accessed OLTP Regular business transactions Heavy on retrievals and updates Short and quick queries Many locks of short duration Lock contention can be high User response time critical Complex queries Few DSS Ad hoc queries on historical data Heavy on retrievals, light on updates Complex and long running queries Few locks of long duration User response not critical Many Concurrent Users Many Few Disk I/O Frequent disk writes New records are continually added and existing records are updated or deleted Table 1: Characteristics of OLTP and DSS Relatively few disk writes Pages are moved in and out of cache Batch Processing Batch processing is the most likely application to produce significant sequential I/O on the data files. The types of activities referred to here often occur after hours and usually in isolation of other activities. Batch processing involves database dumps, database loads, detail report processing, and index creation. Overview of Database Vendor s I/O features The following are I/O features of the Oracle database. Both Oracle7 and Oracle8 server information is referenced. For further information regarding these features, please refer to the Oracle documentation set for your version of Oracle. Partitioning Partitioned tables and indexes provide the ability to divide large tables and indexes into smaller and more manageable pieces (partitions). The ability to partition tables and indexes allows for a more granular table/index layout, better management of table and disk storage and increased I/O performance. The partitioning feature is particularly beneficial for DSS systems.

5 Table Partitioning Oracle8 allows range partitioning, which uses a range of column values (partitioning key) to map rows or index entries to partitions. The partitioning key consists of an ordered list of up to sixteen columns and is based on the partitioning column values. The most common range partitioning is by date. A table cannot be partitioned if it is a cluster; contains LOBs, LONG or LONG RAW datatypes or objects; or is an index-organized table. Index Partitioning Is similar to table partitioning. There are two types of index partitioning: Local and Global. In a local partitioned index, all keys in a particular index partition refer only to rows stored in a single underlying table partition (i.e. All keys in a particular index partition map to all rows in a single underlying table partition). In a global partitioned index, the keys in a particular index partition may refer to rows stored in more than one underlying table partition (i.e. All indexes in one partition may map to rows in three different partitions.) An index cannot be partitioned if it is a cluster index, defined on a clustered table, or a bitmap index on a partitioned table. Oracle8 supports prefixed and non-prefixed local indexes and prefixed global indexes. A prefixed index is partitioned on a left prefix of the index columns, while a non-prefixed index is partitioned on something other than a left prefix of the index columns. SQL*Loader Partitioned Object Support Supports loading partitioned objects into the database. The direct path has been changed to accommodate mapping rows to partitions of tables and to support local and global indexes. Parallel direct-path includes concurrent loading of an individual partition as well as support for concurrent loading of a partitioned table. Parallelism The parallelism feature in Oracle allows multiple processes, called parallel query processes or parallel server processes, to work together simultaneously to execute a single SQL statement. When Oracle is not parallelizing the execution of SQL statements, each SQL statement is done sequentially by a single process. The number of parallel server processes associated with a single operation is known as the degree of parallelism. The degree of parallelism depends upon the number of disks in your system and optimizes the use of partitioning. Asynchronous I/O Oracle7 Server for Windows NT uses the asynchronous I/O capabilities of Windows NT. Asynchronous I/O allows the DBWR and/or the LGWR thread to avoid blocking on I/O. Using the asynchronous I/O capabilities of the operating system and Oracle, only one DBWR thread is necessary in the system. This cuts down on system idle time that might occur when the DBWR is blocked on I/O. RAID Technology Overview Redundant Array of Independent Disks A drive array is simply a collection of disk drives which are grouped together to create an array of physical drives. Drive array technology distributes data across a series of disk drives to unite these physical drives into one or more, higher performance logical drives. Distributing the data makes it possible to access data concurrently from multiple drives in the array, yielding I/O rates

6 faster than non-arrayed drives. This approach that makes a collection of disks appear to the host as a single virtual disk or logical volume is called disk striping. The storage capacity of a striped array will be equal to the sum of the individual capacities of the drives of which it is composed. The drawback to disk striping is its negative impact on array reliability. The failure of any drive in the array means you have lost the entire volume. The best way to protect against the possibility of disk drive failure is known as Redundant Array of Independent Disks (RAID). The RAID concept uses redundant information stored on different disks to ensure that the subsystem can survive the loss of any one in the array without affecting the availability of data to users. RAID storage technology was developed at University of California in Berkeley, is a decade old and expanding in acceptance. To define generally accepted RAID levels, as well as to educate users, a group of vendors and other interested parties formed the RAID Advisory Board (RAB, St. Peter, MN) in Compaq was among the first vendors to bring RAID technology to the world of Intel based servers. RAID provides many fault tolerance options to protect your data. However, each RAID level offers a different mix of performance, reliability, and cost to your environment. Out of the several RAID methods RAID 0, RAID 1 and RAID 5 are of interest in this White Paper. The RAID configuration you choose affects the amount of available disk storage capacity and performance of your drive array. The following section lists the supported RAID levels and illustrates how the RAID configuration you select affects the performance and capacity of the database server. RAID Levels RAID 0 (No Fault Tolerance): This RAID level is not a true fault tolerance method because it does not provide data redundancy; therefore, provides no fault protection against data loss. RAID 0 is known as stripe sets because data is simply striped across all of the drives in the array. This configuration provides high performance at a low cost. However, you incur a risk of possible data loss. You may consider assigning RAID level 0 to drives that require large capacity (some cases full capacity of the disks) and high speed, and can afford the loss of data in the event of a disk failure. RAID 1 (Disk Mirroring): This configuration of mirrored sets of data uses 50 percent of drive storage capacity to provide greater data reliability by storing a duplicate of all user data on a separate disk drive. Therefore, half of the drives in the array are duplicated or mirrored by the other half. This RAID level provides high level of fault tolerance, but your drive cost doubles because this level requires twice as many disk drives to store the same amount of data and therefore might not be cost-effective for your environment. RAID4 (Data Guarding): In RAID 4 one of the disks in the stripe set is used for drive parity. To calculate the parity, data from all the data drives in the stripe set are read. RAID 4 is not commonly used for database applications. RAID 5 (Distributed Data Guarding) RAID 5 is also called Stripe Sets with Parity. This level of RAID actually breaks data up into blocks, calculates parity, then writes the data blocks in stripes to the disk drives, saving one stripe on each drive for the parity data. This method is cost effective. The total amount of disk space used for redundancy is equivalent to the capacity of a single drive; therefore, the overall cost for this method of fault tolerance is lower than Disk Mirroring. In RAID 5 configuration, if a drive fails, the controller uses the data on the parity drive and the data on the remaining drives to reconstruct data from the failed drive. This allows the system to continue operating with slightly reduced performance until you replace the failed drive. Calculation of physical I/O operations for RAID 5 is given below. Although a read operation generates only one I/O, a write operation generates two physical reads and two physical writes. Read1: Read the old data

7 Read2: Read the old parity Write1: Write the new data Write2: Write the new parity For example, if you have a RAID 5 volume of five physical disk drives Disk1, Disk2, Disk3, Disk4 and Disk5. If you need to perform one physical write to Data0, the RAID controller reads the old data from Disk1 and reads the old parity from Disk5 (the data and parity are stored on different drives). The new data is written back to Disk1 and new parity if written back to Disk5. Table 2: A RAID5 volume of 5 disks Disk1 Disk2 Disk3 Disk4 Disk5 Disk1 Disk2 Disk3 Disk4 Disk5 Data0 Data1 Data2 Data3 Parity0 Data4 Data5 Data6 Parity1 Data7 Data8 Data9 Parity2 Data10 Data11 Data12 Parity3 Data13 Data14 Data15 Parity4 Data16 Data17 Data18 Data19 RAID 5 is suitable for read intensive volumes. Read operations can occur in parallel and independent write operations are possible because of the interleaved parity. For example, suppose the measured application I/O profile is 3 reads for every 1 write, and the average number of I/Os per seconds per drive in the RAID 0 configuration is between 30 and 35 I/Os. RAID 0 24 reads 8 writes = 32 I/Os per second per drive RAID 1 24 reads and 16 writes = 40 I/Os per second per drive RAID 5 40 reads and 16 writes = 56 I/Os per second per drive Empirical testing has shown that the 4.3GB drives support random I/Os per second without incurring latency delays, which means in all 3 configurations the drives are operating within specifications. Characteristics of RAID Levels are given in Table 3. Table 3: Characteristics of RAID levels RAID 0 RAID 1 RAID 5 RAID 0 RAID 1 RAID 5 Usable disk space 100% 50% 67% to 93 % Parity and Redundancy Minimum number of disks None Duplicate data Parity distributed over each drive I/Os per Read 1 Read 1 Read 1 Read I/Os per Write 1 Write 2 Writes 2 Reads + 2 Writes

8 Performance Best Good Worst for Writes Fault Tolerance Worst Best Good Cost Best Worst Good Characteristics Stripe Size on Compaq SMART controllers. Best over all performance, but data is lost if any drive in the logical drive fails. Uses no storage space for fault tolerance Tolerant of multiple, simultaneous drive failures. Higher write performance than RAID 5. Uses the most storage capacity for fault tolerance. Requires an even number of drives, Good for split reads Tolerant of single drive failures. Uses the least amount of storage capacity for fault tolerance 256 Sectors (128K) 256 Sectors (128K) 32 Sectors (16K) Fault-Tolerance Evaluation There are several fault-tolerance options to a System Administrator depending upon the integrity requirements, the availability requirements, and the recovery issues of the application being deployed. The Compaq SMART Array Controller RAID levels relevant to this discussion include 0, 1 and 5. RAID 0 simply refers to data striping multiple disks into a single logical volume and has no fault-tolerance. RAID 1 and RAID 5 are both fault-tolerant, but require a different amount of drives to accomplish the fault-tolerance. RAID 1 or drive mirroring uses 50% of the drives in a volume for the fault-tolerance. RAID 5 requires only a single drive worth of the volume's total capacity for the fault-tolerance. For example, if a RAID 5 volume is configured with 6 1GB drives, the usable capacity of the volume is 5GB. With RAID 1 this same 5GB's of usable space requires 10 1GB drives. Which level of RAID to use is completely dependent on the individual situation. No one alternative is best for every application. Some applications may need complete redundancy and the System Administrator is willing to pay for it; while others may be willing to risk some down time for a more cost-effective implementation. In addition, the I/O loads generated by a particular application may fit very well in the performance constraints of RAID 5, while other applications will suffer serious performance penalties with RAID 5. This section explains some of the trade-offs associated with each alternative and possible performance implications, but it is up to the implementers to evaluate each alternative's applicability to their environment. Data Protection Levels Data protection here refers to the degree the database integrity is protected from a physical disk failure. Which scheme is chosen is determined by the degree to which the organization can afford down time and/or data loss. Obviously the higher the degree of protection, the higher the cost of the system. However, the cost of a few thousand dollars of redundant disks may be much less than the cost of a system being down. This comparison does not compare fault-tolerant mode RAID 1 with RAID 5, only the trade-offs of using fault-tolerance or not. The next section discusses the issues around the choice of RAID level. The most secure option is to configure the entire drive subsystem to run in a fault-tolerant mode. In this case, no single disk failure will stop the system or risk any loss of data. If the drive subsystem supports hot-swap drives, such as the SMART Controller with a ProLiant system, the

9 system will not need to be brought down for repairs. If the system doesn't support this feature, down time will need to be scheduled. Most business critical applications, particularly OLTP systems, use this level of fault-tolerance. The second option uses fault-tolerance on some volumes, but leaves others unprotected from a single disk failure. This involves putting all database transaction logs, critical database system tables, the operating system, and boot partition on a volume with some form of fault-tolerance. The remainder of the system is configured with volumes using no fault-tolerance. These volumes contain general data tables, index structures, and other non-critical database objects such as temporary space. If a protected drive fails the system continues to run in the same manner as described in the previous configuration. If an unprotected drive fails, the system will eventually error and the current instance of the database will be lost. At this point the system is brought down and repaired, the database is restored from a previous backup, and the database restored to the instant of failure by rolling forward committed transactions from the protected transaction logs. The reason for protecting the operating system and system tables of the database is to ease and speed the recovery process. If the operating system or system catalogs are lost, the database can still be recovered. However, the process is more complicated and takes much longer. The final configuration option is to run with the disk subsystem completely unprotected. In this situation, any disk failure will usually result in the requirement to restore the full database from a previous backup or reload the database if loaded from an ASCII dump of another system. All updates done since the backup or original load are lost and are NOT recoverable. This may be acceptable for many types of operations, such as static weekly loads of a decision support system. RAID Level Considerations In order to protect a database from a drive failure, a SMART Array Controller volume may be configured with either RAID 1 or RAID 5. Compaq array controllers even allow RAID levels to be mixed on a single controller when system requirements need varying level of performance and protection. As is usually the case, no single configuration fits all application scenarios. This discussion gives a system administrator and DBA the information needed to make an educated decision regarding the appropriate choice of RAID configurations depending on an application's specific characteristics. Cost/Performance Trade-offs In general RAID 5 provides a more cost-effective solution compared to RAID 1, particularly when I/O performance requirements are well below the performance capabilities of the configured drives. The larger the capacity requirement the more pronounced the cost advantage becomes. For example, in a configuration which needs only 1GB a RAID 5 solution could be 3 550MB drives and the RAID 1 solution would be 2 1GB drives. There is very little difference between these two configurations' prices. However, if looking at a 14 GB system, RAID 5 could be done with 8 2GB drives as opposed to 14 2GB drives for the RAID 1 configuration. Even though RAID 5 often has a cost advantage, RAID 1 has its own advantages for price/performance and increased fault-tolerance. If the application(s) using a database server have high I/O requirements, RAID 1 will provide significantly better performance at the same user capacity due to the increased number of drives. To increase the performance of the RAID 5 volume requires increasing the number of drives on the volume and therefore the capacity over what the application really needs. As will become evident from the examples in the next section, RAID 1 and RAID 5 may have the same price/performance costs on high I/O systems.

10 Another advantage of RAID 1 over RAID 5 is its ability to tolerate more than a single drive failure. This can occur and the server continues running as long as the multiple failed drives are not the pair mirroring each other. For example, if a RAID 1 volume consists of 4 drives mirrored to 4 drives, the first drive of the first set of drives could fail and before that drive is replaced any other drive, except for the first drive in the second set, could also fail and the system will continue to operate. In contrast, on a RAID 5 volume only a single drive can fail. If a second drive fails, before the failed drive has been replaced and recovered, the data will be lost. Failed State Considerations Obviously, the only configuration where this is relevant is RAID 1 and RAID 5. The performance degradation for RAID 1 should be negligible since the only difference is that one less drive is available for read operations. However, under RAID 5 a more significant penalty may be incurred. The reason for the degradation is, that for every read or write request to the failed drive, an I/O operation against all other drives in the volume is required. In most database systems the sustainable load will drop until the drive is replaced and the rebuild process is complete. Recovery Time The failed drive recovery time is highly influenced by the following factors: Type and size of the drive RAID level Workload on the system Controller type Compaq SMART Array Accelerator setting Compaq SMART Array drive recovery priority level On an OLTP system configured with RAID 1 array of 9.1 GB disk drives on a SMART-2DH controller, Array Accelerator set for 100% write and drive recovery set for low priority; when the system was idle the approximate drive recovery time to recover a 9.1GB failed drive was 45 minutes. During OLTP processing the recovery time is approximately 20 to 30 minutes per GB. If the system is in use during the drive rebuild, recovery time may be very dependent on the level of activity. Most systems should recover in nearly the same time with moderate activity as with no load, particularly RAID 1. RAID 5 is much more sensitive to system load during the recovery period due to the considerably heavier I/O requirements of the failed system as described previously. Take this recovery time and failed-state performance into account when deciding between RAID 1 and RAID 5. Online Spare Drives If you are running a mission critical application that cannot tolerate any outages due to disk failures, consider using online spare disks supported by Compaq SMART Array controllers. This will further improve your system s fault tolerance. An online spare (sometimes called a hot-spare) is a drive the controller uses when a drive failure occurs. If a drive fails, the controller rebuilds the data that was on the failed drive onto the online spare. The controller also sends data that it

11 would normally store on the failed drive directly to the online spare. You can assign an online spare to an array containing a RAID 1 and RAID 5 logical drive. Database Transaction Log devices As the database transaction log files are used to recover from data loss, it is very important to fully protect these files. Testing done in our lab shows that dedicating a RAID1 volume is most suitable for log devices. You should isolate the transaction log device from any other database and system I/O activities during heavy update activity. In most of the insert/update intensive OLTP applications the ability of the log device to support the number of physical I/O often limits the overall throughput. Dedicating separate devices for data and log files will eliminate potential I/O device contention. Usually, if the database is dedicated to DSS no updates will be done to the database during heavy query load. In this case, very little I/O will occur on the transaction logs. However, a typical log file's space requirement may not need the space of an entire physical disk. You can usually share a single large disk between the log file and other files on the system, which are not active during heavy update activities. Examples of items which coexist well with the log file include the operating system, database executables, database dump files, etc. If you have multiple databases under the same database server, it may not be practical to have dedicated physical devices to each transaction log files. In such case, have dedicated log devices for insert/update intensive and performance critical databases. Compaq Array Controller Technology Overview SMART Array Controller Family The SMART Controller provides drive array technology, which distributes data across a series of hard drives, and links physical drives to unite the physical drives into a single, higher performance logical drive. Distributing the data makes it possible to access data concurrently from multiple drives in the array, yielding I/O rates many times faster than individual drives. With the increased I/O rate, the computer system can better meet the needs of file server and multi-user host environments. The SMART Controller offers the following fault management and data reliability features: Auto reliability monitoring Dynamic sector repairing Drive parameter tracking Drive failure features Interim data recovery Automatic data recovery Hot-pluggable drives Controller duplexing Online spares

12 The Array Accelerator Cache on the SMART Controller dramatically improves I/O performance. Depending on the model of the SMART Controller this cache varies from 4MB-16MB and can be configured for Read Cache, Write Cache or both. The ECC Memory mirroring along with rechargeable batteries located on the SMART Controller protect the cache. The Compaq SMART-2 Array Controller features advanced intelligent I/O subsystems. This advanced controller was designed specifically to meet today s network data storage requirements and deliver the power to respond to future needs. Hot-Pluggable Capacity Expansion - Server storage capacity can be expanded online without shutting the server down. This feature is a breakthrough in data storage management. It is quite useful for administrators expanding storage capacity and implementing fault tolerant array configurations. Data Availability - On-line administration of storage is now available thus reducing planned down time. Unplanned downtime is also reduced with fault tolerance features that guard against disk drive and controller component failures. Manageability - The new GUI (Graphic User Interface) based Compaq Array Configuration Utility simplifies array configuration and provides an intuitive, easy-touse method for making changes and initiating capacity expansion. High Performance - High levels of I/O throughput and data bandwidth are provided to support demanding workloads. The controller provides an advanced intelligent I/O architecture consists of specialized modules that significantly enhance performance. These include an enhanced Array Accelerator for write-back and read-ahead caching, a hardware-based RAID engine, and an optimized local processor. The Array Accelerator is quite useful for increasing performance in database and fault tolerance configurations. This will be discussed in details in the following section. Fault Management Features of SMART-2 Controllers 1. Auto reliability monitoring (ARM): is a background process that scans hard drives for bad sectors in fault tolerant logical drives. ARM also verifies the consistency of parity data in drives with data guarding or distributed data guarding. This process assures that you can recover all data successfully if a drive failure occurs in the future. ARM operates only when you select RAID 1, RAID 4, or RAID 5. ARM is only run when drives are not being accessed. 2. Dynamic sector remapping: Using the dynamic sector repairing process, the controller automatically remaps any sectors with media faults it detects either during normal operation or during auto reliability monitoring. 3. Drive parameter tracking: Drive parameters tracking monitors more than 15 drive operational parameters and functional tests. This includes parameters such as read, write, and seek errors, spin-up time, cable off, and functional tests such as track-to-track seek time, one-third stroke, and full stroke seek time. Drive parameter tracking allows the SMART-2 Controller to detect drive problems and predict drive failure before they actually occur. It also makes pre-failure warranty possible on Compaq disk drives. 4. Drive Failure notification: Drive failure features produce various drive alerts or error messages depending on the Compaq server model. Refer to the documentation included with your server to determine what drive failure features are included on your server model. Other Compaq options such as Compaq Insight Manager and Compaq Server Manager/R provide additional drive failure features. See your Authorized Compaq Reseller for more information on these products. 5. Controller level RAID support: In RAID 5, RAID 4, or RAID 1 fault tolerant configurations, if a drive fails, the system continues to operate in an interim data recovery mode. For example, if you

13 had selected RAID 5 for a logical drive with four physical drives and one of the drives fails, the system continues to process I/O requests, but at a reduced performance level. Replace the failed drive as soon as possible to restore performance and full fault tolerance for that logical drive. 6. On-line Spare capability: On-line spare capability allows users to allocate up to four disk drives as on-line spares. These drive remain up and running, but are not active. If an active drive fails during system operation, the controller automatically begins to rebuild the data from the failed drive onto a spare. Once the rebuild is complete, the system is again fully fault tolerant. 7. Automatic data recovery: After you replace a failed drive, automatic data recovery reconstructs the data and places it on the replaced drive. This allows a rapid recovery to full operating performance without interrupting normal system operations. You must specify RAID 5, RAID 4, or RAID 1 through the Array Configuration Utility to make the recovery feature available. The drive failure alert system and automatic data recovery are functions of the SMART-2 Controller; they operate independently of the operating system 8. Hot-pluggable drive support: The SMART-2 Controller, used with a Compaq ProLiant Storage System (such as the U1, U2, F1, F2, SCSI and Fast-Wide SCSI-2 Storage Systems), and a Compaq ProLiant Server or Rack-Mountable Compaq ProLiant Server, supports hotpluggable drives. You can install or remove these drives without turning off the system power. This feature is a function of the Compaq ProLiant Storage System and the Compaq ProLiant Server and operates independently of the operating system. 9. OS level fault tolerance features: Some operating systems support controller duplexing, a fault tolerance feature that requires two SMART-2 Controllers, and software level RAID support. With duplexing, the two controllers each have their own drives that contain identical data. In the unlikely event of a SMART-2 Controller failure, the remaining drives and SMART-2 Controller service all requests. Controller duplexing is a function of the operating system and takes the place of other fault tolerance methods. Refer to the documentation included with your operating system for implementation. Array Accelerator The Array Accelerator Cache on the SMART Controller dramatically improves I/O performance. Depending on the model of the SMART Controller this cache varies from 4MB-16MB and can be configured for Read Cache, Write Cache or both. The ECC Memory mirroring along with rechargeable batteries located on the SMART Controller protect the cache. Fully charged batteries will preserve data in the cache for up to 96 hours. This will ensure that the data temporarily stored is safe even with equipment failure or power outage. The battery cells used are Lithium Magnesium Dioxide and are recharged via a trickle charge applied while system power is present. Key features of the Compaq Array Accelerator are discussed below. Read-Ahead Caching The Array Accelerator uses an intelligent read-ahead algorithm that can anticipate data needs and reduce wait time. It can detect sequential read activity on single or multiple I/O threads and predict what sequential read requests will follow. It will then read-ahead, or pre-fetch data, from the disk drives before the data is actually requested. When the read request does occur, it can then be sent out of high-speed cache memory at microsecond speeds rather than from the disk drive at millisecond speed. This read-ahead algorithm provides excellent performance for sequential small block read requests. At the same time it avoids being penalized by random read patterns, because readahead is disabled when non-sequential read activity is detected. Thus, it overcomes the problem

14 with some array controllers in the industry today that use fixed read-ahead schemes that increase sequential read performance but degrade random read performance. Write-Back Caching The Array Accelerator also uses a write-back caching algorithm that allows applications to continue without waiting for write operations to complete. This technique is also referred to in the industry as a posted-write caching scheme. Without this type of caching, the controller would be forced to wait until the write data is actually written to disk before returning completion status to the operating system. With write-back caching, the controller can post write data to high-speed cache memory and immediately return back completion status to the operating system. The write operation is completed in microseconds rather than milliseconds. Data in the write cache is written to disk later, at the optimal time for the controller. The system then combines or coalesces the blocks of data into larger blocks and writes them to the hard disk. This results in fewer and larger sequential I/Os, thus improves performance. While the data is in cache, it is protected against memory chip failure by Error Checking and Correcting (ECC) and against system power loss by the integrated battery backup mechanism. Other array controllers on the market today do not have any type of backup; therefore, they run the risk of data loss during write-back caching if power to the controller should be disrupted or lost. Balanced Cache Size The enhanced Array Accelerator on the SMART-2/E Controller (the EISA version) has 4MB of usable cache for excellent performance with optimal cost effectiveness. The enhanced Array Accelerator on the SMART-2/DH Controller (the PCI version) has 16MB of usable cache. With the streamlined architecture of the SMART-2 Controller, these cache sizes provide excellent performance enhancements for sequential, small-size read requests and small-to-medium-size write requests. ECC Protection The Array Accelerator uses ECC technology to protect data in cache DRAM. This provides greater protection than mirroring and is much more cost effective. The ECC scheme generates 16 bits of check data for every 64 bits of user data. This check data can be used not only to detect errors but also to correct them. ECC can be used to detect and correct DRAM bit errors, DRAM chip failures, and memory bus errors on the Array Accelerator connector. Battery Backup The Array Accelerator has rechargeable batteries as one of its components. They provide backup power to the cache DRAM should the system power be interrupted or lost. When power is restored to the system, an initialization feature writes the preserved data to the disk drives. The controller recharges the batteries automatically under normal operating conditions. If the controller fails, you can move the Array Accelerator to another controller. The batteries can preserve data in the Array Accelerator for up to 4 days. Note that this depends on the present condition of the Array Accelerator batteries. If you feel that this risk is too great, you can disable the Array Accelerator.

15 Disk-Related Performance Characteristics The following terms are often used in the industry to describe the performance characteristics of disk performance. These general terms describe characteristics that can affect system performance, so it is important to understand how each term could impact your system. Seek Time - The time it takes for the disk head to move across the disk to find a particular track on a disk. Average Seek Time - The average time required for the disk head to move to the track that holds the data you want. This average time will generally be the time it takes to seek half way across the disk Latency - The time required for the disk to spin one complete revolution. Average Latency - The time required for the disk to spin half a revolution. Average Access time - The average length of time it takes the disk to seek to the required track plus the amount of time it takes for the disk to spin the data under the head. The average access time equals to the average seek time plus latency. Disk Transfer Rate - The speed at which the bits are being transferred to and from the disk media (actual disk platter) and is a function of the recording frequency. Typical units are bits per second (BPS) or bytes per second. RPM (Revolutions Per Minute) - The measurement of the rotational speed of a disk drive on a per minute basis. The slower the RPM, the higher the mechanical latencies. Disk RPM is a critical component of hard drive performance because it directly affects the rotational latency and the disk transfer rate. Transfer Rates A disk subsystem is made up of multiple hardware components such as hard disks, SCSI channel, disk controller, I/O bus, file system and disk controller cache. These subsystems communicate with each other using hardware interfaces such as SCSI channels and PCI buses. These communication highways, called channels and/or buses, communicate at different speeds known as transfer rates. Therefore, by knowing the transfer rate of each subsystem device, potential bottlenecks can be identified and corrected. The key to enhance system performance is focusing on how to maximize data throughput by minimizing the amount of time the disk subsystem has to wait to receive or send data. Remember that the slowest disk subsystem component determines the overall throughput of the system. Disk Transfer Rates - Normally, the disk transfer rate is lower than the bus speed. By adding more drives, the system can process more I/O requests, up to the saturation of the bus speed. Thus increasing overall all speed, which improves system performance. SCSI Channel Transfer Rates - The disk controllers being used today can transfer data up to 40MB/s to and from the hard disk controller by way of the SCSI bus. If the disk drive can transfer at only 5MB/s, then the disk drive is the bottleneck. Disk Controller Transfer Rates - Disk controllers are continuously improved to support wider data paths and faster transfer rates. Currently, Compaq supports three industry standard SCSI interfaces on their disk controllers, namely Fast-SCSI-2,

16 Fast-Wide SCSI-2, and Wide-Ultra SCSI with the transfer rate 10, 20 and 40 MB/s respectively. I/O Bus Transfer Rates - The PCI Bus has a transfer rate of 133 MB/s. Due to its fast transfer rate, it is the second least (file system cache being first) likely of all of the disk subsystem components to be performance bottleneck. File System and Disk Controller Caching Transfer Rates - File system and disk controller caching have significant impact on the system performance. Accessing data in memory, also know as Random Access Memory (RAM), is quite fast, while accessing data on disk is a slower process. Thus in order to improve system performance we should try to avoid disk accessing whenever possible and prefer to fetch data from memory. Concurrency - Concurrency is the process of eliminating the wait time involved to retrieve and return requested data. It occurs when multiple slow devices, e.g. disk drives, place I/O requests on a single faster device, e.g. SCSI bus. It allows multiple requests to be processed at the same time thus reducing idle time of the SCSI bus and utilize the fast transfer rate of the bus and improve performance. Compaq SMART Array Controllers for Database Servers The features offered by Compaq SMART Array Controllers are well suited for database servers where data availability and performance are of the utmost importance. The data stored on database servers is usually mission critical and must be available 24 hours a day and 7 days a week. Interruptions in the availability of that data can mean lost revenue. Compaq understands this type of scenario and has developed the Compaq SMART Array Controller for just such an environment. Compaq has loaded the SMART Array Controller with capabilities such as RAID (levels 1,4,5), online spares, automatic data recovery, and auto reliability monitoring to handle drive failures. However, Compaq also recognizes that early detection is also extremely critical. Drive parameter tracking allows the controller to identify problems early on before they cause drive failures. This early warning can be valuable to system administrators in preparing for drive replacement. Performance is also key for database servers, and Compaq SMART Array Controllers are optimized for database performance. Because the controller has a 32-bit RISC processor on it and RAID algorithms are handled by hardware based engines, the SMART Array Controller can offload processing cycles from the main server CPU s. This means high RAID5 performance and more server CPU cycles are available for the database software, which usually means better performance. A key part of database server performance is data access time. The read-ahead cache combined with intelligent read-ahead algorithms mean that database software can see microsecond read access times to data on disk when doing large sequential data access. These read-ahead capabilities can greatly speed up large decision support style queries. The write-back cache can allow microsecond write access times which allow system administrators to tune database page cleaning algorithms for optimal performance. The write-back cache can dramatically reduce the amount of time it takes to do a database checkpoint since that operation requires many write operations to be done in a limited amount of time. So the Compaq SMART Array Controllers offer a wide variety of features that enhance data availability and performance. This combination makes it a great choice for database server environments of all sizes.

17 Array Configuration Guidelines Disk Subsystem Monitoring on NT It is necessary to determine the I/O rates of your system to size your disk subsystem. This section provides information on tools to monitor your disk subsystem. Diskperf The diskperf option allows you to monitor the disk subsystem activity. It is very useful when monitoring performance of the drive subsystem. If this option is disabled, performance monitor will not be able to monitor low level disk-related activity, such as LogicalDisk and PhysicalDisk counters. Having this option enabled slightly degrades performance. Enable diskperf only when needed, then disable it to get maximum performance. You can enable or disable diskperf using Control Panel/Devices or by issuing the diskperf -y n command from the system prompt. When using the Control Panel, set diskperf to automatically start at boot time to enable. You must restart your system for the diskperf option to become effective. The diskperf option is described in a greater detail in the Microsoft Windows NT Resource Kit. Disk Performance Monitor Objects and their Meaning You can monitor LogicalDisk- and PhysicalDisk-related objects, such as Avg. Disk sec/transfer or Disk Transfers/sec. Monitoring LogicalDisk and PhysicalDisk-related objects requires you to have the diskperf option enabled. Object: Logical Disk Counter: Avg. Disk sec/read, Avg. Disk sec/write The Avg. Disk sec/read and Avg. Disk sec/write counters monitor the average number of seconds for read or write operations from or to a disk, respectively. Use the following guidelines when optimizing the disk subsystem. Do not have more average disk seconds per read or write operations than the values in the table below. Table 6: Recommended Average time for Disk Read and Write operations for database applications Transaction Log Access (with the ArrayAccelerator enabled) Database Access (with the Array Accelerator enabled) Avg. Disk sec/read N/a 20-40msec Avg. Disk sec/write <=10msec <=10msec If your values are much higher than those suggested in the above table, you may want to increase the speed of your disk subsystem by adding more drives or using faster drives. Object: Logical Disk

18 Counter: Disk Transfers/sec The Disk Transfers/sec counter monitors the rates of read and write operations on the disk. It is important to realize that if you have a Compaq SMART or SMART-2 Array controller and several drives allocated to a logical volume, this counter monitors the total number of disk transfers per logical volume. To calculate number of disk transfers per second per drive, you must divide the Disk Transfers/sec value by the number of drives in the logical volume. Volume Configuration The most confusing decision when configuring a system is how an array of physical drives should be partitioned into logical volumes. In reality this is a very easy step in the configuration process, once the methodology is understood. The key is to let an array do what it is intended to do, that is, distribute I/O requests among all available drives. Configuration and tuning of a disk subsystem for database applications should be focused on getting the most out of each drive and having an appropriate number of drives on a system. By carefully studying the applications accessing the database, the data can be distributed in such a manner that each drive receives an approximately equal number of requests. There are basic principles that are used in planning your disk layout, regardless of the type of database that you have. Spread randomly accessed data over as many disks drives as possible. Placing randomly accessed data on a single logical volume of many physical disk drives will help to balance the I/O load. Isolate purely sequential I/O to its own array. This will avoid excessive drive head movement. There are three basic database operations that are characteristically sequential: database transaction logging, bulk loading and unloading of the database, and batch reporting or updating requiring table scans. Group randomly accessed data Looking back at the environment definitions covered earlier, most data access on database servers is basically random. When the goal is to optimize these application classes, database data and index files should be grouped onto a single array volume for each disk array controller in a machine. For example, if an OLTP database contains a number of large tables and their associated indexes totaling 5GB, these should all be put on a single volume of six 1GB drives. This guarantees that all drives will have approximately equal I/O load during normal operation of the system. This load balancing is done by the disk controller, due to the low-level striping of files across all disks in the volume, with no additional effort required by the System Administrator or DBA. In contrast, techniques that have been carried over from traditional, non-array systems include the separation of data and index files to separate physical drives and separation of large data tables to their own dedicated drives. The reason for this was to allow simultaneous access to multiple drives. The problem with these techniques is that it puts the burden of load balancing upon the System Administrator. By carefully studying the applications accessing the database, the data can be distributed in such a manner so that each drive receives an approximately equal number of disk requests. Unfortunately, two assumptions are made that are difficult to achieve. Those are that the System Administrator and DBA can determine the actual I/O profile of an application and, that once determined, the I/O profile will remain constant throughout a day's processing and over the life of the application. For example, an application will often use different tables at different times of the day. Table A may have heavy access in the morning as a business gets started for the day. As the day wears on, most access may move over to Table B. If these

19 tables were separated on different physical disks, the application would overload the disk with Table A in the morning and overload the disk with Table B later. However, with an array controller these two tables would be put on the same logical volume spread across multiple drives allowing all drives to share equally in the I/O load throughout the day. Another example to further demonstrate this concept is data growth. Again using Table A and Table B, when the database is initially deployed Table A and B are approximately the same size and receive the same number of I/O requests. At that time, it makes sense for each table to be on its own drive. Over time, Table B grows by 10 percent per month in both size and I/O requests. After just 9 months Table B has doubled in access frequency and is now a bottleneck to system performance. If both these tables are put on a single array volume this table growth is shared among all drives on the volume, thus avoiding the disk bottleneck. The case covered thus far applies when all the disks are attached to a single controller. Obviously when the amount of data and indexes exceeds the capacity of a single controller you will have multiple volumes in the system. At this point you have three options for distributing the data across the controllers. 1. Use an operating system facility to stripe the volume across controllers. 2. Use the database facilities to stripe the data across the controllers. 3. Segment different data to each controller. There is no best solution to this situation. It is a matter of preference to the Administrator. General trade-offs of each are listed below to help you make your own decision. The degree of each of these may vary depending upon operating system and database software chosen. The pros and cons of each option are shown in Table 9. Table 9: Multiple Controller Options Use an operating system facility to stripe the volume across controllers. Pros Cons Ease of space management as database More difficult volume management grows Ease of backup with everything on one file system Very long tape recovery times of volume Performance penalty associated with file system striping Use the database facilities to stripe the data across the controllers. Pros Cons Little to no negative performance impact Most complicated database configuration process Ease of space management as database grows More complicated backup/recovery process Segment different data to each controller. Pros Cons

20 Best performance when balanced properly Fastest, easiest tape recovery of volume Fastest tape backup, allows multiple tape usage Requirement of administrator to understand access patterns Isolate sequentially accessed data Even though most data in a database server is read and/or written randomly, there are several cases of sequential access. During a long sequential read or write process, if the drive head can stay on the same physical disk track, performance will be dramatically better than if the head must continually seek between tracks. This principle is the same for traditional, non-array controllers and array controllers. Actually, the largest single time component in reading or writing a block of data on a disk is the time required to move the drive head between tracks. The best drives on the market today range from 8 to 10 milliseconds average seek time and many drives you may be using from just a few of years ago could be as high as 16 or 19 ms for an average seek time. The remaining activities in an I/O routine include: the database makes a request from the operating system, the operating system requests the data from the controller, the controller processes the request and makes the request from the drive, the block rotates under the drive head, the data is transferred to the controller, back to the operating system, and finally the application. All of this takes only 10 to 15 ms, depending upon the particular controller, operating system, and database. Obviously, minimizing seek time, where able, has a significant benefit. So much for low-level numbers, what does this mean to your application and how do you take advantage of it. There are three basic database operations that are characteristically sequential: transaction logging, bulk loading and unloading of the database, and batch reporting or updating requiring table scans. In many cases involving sequential I/O, the operations associated with it will not be the primary function of the system, but may be the most important aspect for your tuning due to nightly batch windows or backup time constraints. If the typical random access to the data will be relatively light or done by few users, you may want to give priority to tuning sequential operations over random I/O distribution. The transaction log is the most obvious sequential I/O routine. In any update environment the transaction log is sequentially written from beginning to end. You should isolate the log file from any other system I/O activities during heavy update activity. However, a typical log file's space requirement may not need the space of an entire physical disk. You can usually share a single large disk between the log file and other files on the system which are not active during heavy update activities. Examples of items which coexist well with the log file include the operating system, database executables, database dump files, etc. Bulk loads and batch updates are often associated with decision support systems. The typical scenario involves extracting large volumes of data from production transaction processing systems on a pre-defined schedule, such as daily or weekly. The data is then moved to the DSS server and loaded into a database. Performance of this load operation can be very critical due to nightly time constraints. Optimization of this process is very different from the typical system. In contrast to the random nature of most multi-user systems, a batch load or update can have up to four concurrent sequential operations. These operations include reading the input file, writing the database data file, writing the database index file, and writing the transaction log. To optimize the load time of the data each of these file types should be isolated to their own logical volumes. In contrast, pure optimization for daily multi-user decision support would put ALL of the above files on a single, large volume to allow concurrent access of all drives on the server by all read operations. The performance impact on daily access of optimization for load times should be less than 20%. However, the gain in load time can be over 200% which is often well worth the small random access penalty.

21 The final general application classes, which rely on sequential I/O performance include batch reporting and database dumps. These are fairly simple operations to optimize and there are only a couple of concepts to keep in mind. The profile of reporting and unloading is sequential table scans with some resulting output being written to a non-database file. In the case of a detailed analysis report, a report file of several megabytes may be spooled to a large disk file. The database dump sequentially reads tables writing the results to either a tape device or disk file. When the output is written to a disk file on the server, the target disk should be a different physical volume from the source of the data. If this is the only activity at the time the volume used for the log file is a good target; otherwise, another volume should be added to the system of sufficient size to accommodate the output file. Physical Disk Requirements Once you understand the target environment, the volume configuration guidelines, and the faulttolerance requirements the next step is to determine how many disks to distribute the data across and which type of disks are required. The goal is to use the lowest cost/megabyte drives, usually the largest drives available, while having enough drives to satisfy the physical I/O rate required by the application. The information presented here will help in meeting this goal. The number and type of disk drives required by a disk subsystem volume is dependent upon the throughput requirements of the environment and the total amount of data on the server. Throughput requirements are composed of two characteristics, the degree of concurrency among user requests and the request rate from the application. Concurrency and request rate are distinct, but related concepts. Concurrency refers to the number of physical requests that are in the process of being executed by the disk controller(s). The degree of concurrency is primarily influenced by the number of users on the system and the frequency of requests from those users. Request rate is the absolute number of reads and writes per unit of time, usually expressed as I/Os per second. Usually, the higher the concurrency level, the higher the request rate. Yet, fairly high request rates can be generated by a single batch process with no concurrent requests. Of course, the number of physical disk requests that actually occur will depend upon the size of the database relative to the amount of disk cache and the resulting cache hit rate. The worst case scenario would result in as many concurrent disk requests as simultaneous user requests. The reason for understanding the application I/O request rate is that multiple I/O requests can typically be processed in parallel if the data is distributed across multiple disk drives. Therefore, even though large drives are usually "faster" than smaller drives, this performance difference cannot offset the advantage of two drives ability to respond to multiple requests in parallel. The target number of drives in a server should be approximately equal to the sustained or average number of simultaneous I/O requests, with three drives being a typical minimum for the data and index files. The following evaluation process determines the number of drives required by a volume containing data accessed randomly by multiple users. Repeat this process for each volume to be used on a database server. In the case of high transaction rate systems, this process may yield too many drives to fit on a single controller, at which point the drives can be split into multiple, equal volumes across controllers. For systems or volumes using purely sequential I/O or having no concurrent disk requests, i.e. only one or two users on the system at a time, strict calculations for the number of drives is not required. For this case, use the fastest drives available and as many as required to meet the space requirements of the system. The steps used to determine the optimal number of drives are summarized here followed by a detailed explanation of each step and several case studies: 1. Determine typical I/O request rate by analyzing transaction I/O requirements and typical transaction rates of the target application(s).

22 2. Using Chart 1, determine the minimum number of drives required to fall within or below the I/Os per second rates for your disk drives. Please refer to Table 11 to find out the recommended number of I/Os per second rate for your disk drives. 3. Add at least an additional 20% space for growth and administration to the database size. 4. Divide drive requirements based on RAID usage into total space requirements to determine which drive size to use. The first step in calculating the number and size of drives to use is determining the request rate of the application. Without this information it is not possible to make an accurate estimation of the number and type of drives to use. The following equations provide rough estimates to the number of I/O requests generated by an application. Due to optimization routines in the controller to reduce the total I/O requests, these equations do not yield exact loads. They do, however, provide a tool for making conservative comparisons of the load the drives may be required to sustain under alternative RAID configurations. The RAID 5 equation component of (4*writes) implies that each database write incurs 4 physical I/Os, 2 reads and 2 writes; it is not meant to imply 4 physical write operations for each database write operation. Application I/O Requests per Second = RAID 0:(reads/transaction + writes/transaction) * transactions/second RAID 1:(reads/transaction + 2*writes/transaction) * transactions/second RAID 5:(reads/transaction + 4*writes/transaction) * transactions/second The next step in the process is determining how many drives are required to support the applications I/O demands. Advertised I/O rates of today's typical 500MB through 9GB drives can exceed 200 I/Os per second. However, these results are achieved with pure saturation testing without regard for request latency. Empirical testing has shown that these same disks sustain between 30 and 80 random database requests per second without incurring costly latency delays. Please refer to Table 11 to find out the recommended application I/Os per second for different disk drives. Using this range you can determine the approximate number of physical disks which should be configured on the server and what size those drives can be.

23 Chart 1: Smart Array Drive I/O Rate Capacities. Chart 1 is used in conjunction with the previously discussed equations to determine the number of physical disks required by an application. Locate on the Y-axis the number of I/Os per second required by the application. Move across the graph locating the intersection of the application I/O rate and the recommended application I/O per second for drives area. The number of drives is then determined by locating the drive count on the horizontal axis. The more conservative the implementation desired, the lower the I/O rates per drive that will be required. Also, if using RAID 1 an even number of drives is required. If the result is an odd number of drives, round up to the next even number. When throughput requirements are more than the chart indicates; continue adding controllers and drives until the required throughput is achieved. Chart 2 can be used for I/O subsystem capacity planning. If you know your application I/Os per second rates, you can determine the minimum number of SMART controllers needed. If your application I/O rate demands multiple controllers, you will have to distribute the disk drives evenly across the controllers for balancing the I/O load. Finally, the total space requirements for the database and support files, combined with the throughput requirements and fault-tolerance, determine the size of disks, which will provide an optimal solution. When calculating total space requirements, add an additional 20% free space above the actual data requirements for growth and miscellaneous overhead. Again, a set of equations provides the answer to the appropriate size drive. Drive size requirements based on total space needs and fault-tolerance: RAID 0: [total space / number of drives] Rounded up to next drive size RAID 1: [total space / (number of drives / 2)] Rounded up to next drive size RAID 5: [total space / (number of drives - number of controller volumes)] Rounded up to next drive size

24 Chart 2: Smart Array Controller I/O Capacities. Several examples are presented which illustrate the use of this process. All RAID levels are presented for each case. Some cases still require a judgment call by the implementers to choose the number of drives or drive sizes, but the alternatives are drastically narrowed. The alternatives chosen are conservative configurations and the reason for the choice, if not a single alternative, is explained. Case 1: Complex transaction processing system requiring multiple queries per update. Reads: Writes: Transaction rate: Space Requirements: 100/transaction 5/transaction 2/second 3GB Database + 20% Free space = 3.6GB RAID 0: ( ) * 2 = 205 I/Os / Second 205 I/Os / Second => 5 or 6 Drives of 35 to 45 I/Os per second 3.6 / 6 =.6 => Round up to next drive size => 1GB Drives A conservative implementation yields 6 1GB drives for the data volume. RAID 1: (100 + (2*5)) * 2 = 220 I/Os / Second 220 I/Os / Second => 6 or 8 Drives of 35 to 45 I/Os per second

RAID SEMINAR REPORT /09/2004 Asha.P.M NO: 612 S7 ECE

RAID SEMINAR REPORT /09/2004 Asha.P.M NO: 612 S7 ECE RAID SEMINAR REPORT 2004 Submitted on: Submitted by: 24/09/2004 Asha.P.M NO: 612 S7 ECE CONTENTS 1. Introduction 1 2. The array and RAID controller concept 2 2.1. Mirroring 3 2.2. Parity 5 2.3. Error correcting

More information

Definition of RAID Levels

Definition of RAID Levels RAID The basic idea of RAID (Redundant Array of Independent Disks) is to combine multiple inexpensive disk drives into an array of disk drives to obtain performance, capacity and reliability that exceeds

More information

hot plug RAID memory technology for fault tolerance and scalability

hot plug RAID memory technology for fault tolerance and scalability hp industry standard servers april 2003 technology brief TC030412TB hot plug RAID memory technology for fault tolerance and scalability table of contents abstract... 2 introduction... 2 memory reliability...

More information

Modern RAID Technology. RAID Primer A Configuration Guide

Modern RAID Technology. RAID Primer A Configuration Guide Modern RAID Technology RAID Primer A Configuration Guide E x c e l l e n c e i n C o n t r o l l e r s Modern RAID Technology RAID Primer A Configuration Guide 6th Edition Copyright 1997-2003 ICP vortex

More information

SYSTEM UPGRADE, INC Making Good Computers Better. System Upgrade Teaches RAID

SYSTEM UPGRADE, INC Making Good Computers Better. System Upgrade Teaches RAID System Upgrade Teaches RAID In the growing computer industry we often find it difficult to keep track of the everyday changes in technology. At System Upgrade, Inc it is our goal and mission to provide

More information

Storage. Hwansoo Han

Storage. Hwansoo Han Storage Hwansoo Han I/O Devices I/O devices can be characterized by Behavior: input, out, storage Partner: human or machine Data rate: bytes/sec, transfers/sec I/O bus connections 2 I/O System Characteristics

More information

OS and Hardware Tuning

OS and Hardware Tuning OS and Hardware Tuning Tuning Considerations OS Threads Thread Switching Priorities Virtual Memory DB buffer size File System Disk layout and access Hardware Storage subsystem Configuring the disk array

More information

OS and HW Tuning Considerations!

OS and HW Tuning Considerations! Administração e Optimização de Bases de Dados 2012/2013 Hardware and OS Tuning Bruno Martins DEI@Técnico e DMIR@INESC-ID OS and HW Tuning Considerations OS " Threads Thread Switching Priorities " Virtual

More information

I/O CANNOT BE IGNORED

I/O CANNOT BE IGNORED LECTURE 13 I/O I/O CANNOT BE IGNORED Assume a program requires 100 seconds, 90 seconds for main memory, 10 seconds for I/O. Assume main memory access improves by ~10% per year and I/O remains the same.

More information

Mladen Stefanov F48235 R.A.I.D

Mladen Stefanov F48235 R.A.I.D R.A.I.D Data is the most valuable asset of any business today. Lost data, in most cases, means lost business. Even if you backup regularly, you need a fail-safe way to ensure that your data is protected

More information

I/O CANNOT BE IGNORED

I/O CANNOT BE IGNORED LECTURE 13 I/O I/O CANNOT BE IGNORED Assume a program requires 100 seconds, 90 seconds for main memory, 10 seconds for I/O. Assume main memory access improves by ~10% per year and I/O remains the same.

More information

The term "physical drive" refers to a single hard disk module. Figure 1. Physical Drive

The term physical drive refers to a single hard disk module. Figure 1. Physical Drive HP NetRAID Tutorial RAID Overview HP NetRAID Series adapters let you link multiple hard disk drives together and write data across them as if they were one large drive. With the HP NetRAID Series adapter,

More information

Technical Note P/N REV A01 March 29, 2007

Technical Note P/N REV A01 March 29, 2007 EMC Symmetrix DMX-3 Best Practices Technical Note P/N 300-004-800 REV A01 March 29, 2007 This technical note contains information on these topics: Executive summary... 2 Introduction... 2 Tiered storage...

More information

Computer Organization and Structure. Bing-Yu Chen National Taiwan University

Computer Organization and Structure. Bing-Yu Chen National Taiwan University Computer Organization and Structure Bing-Yu Chen National Taiwan University Storage and Other I/O Topics I/O Performance Measures Types and Characteristics of I/O Devices Buses Interfacing I/O Devices

More information

QuickSpecs. Models. HP Smart Array 642 Controller. Overview. Retired

QuickSpecs. Models. HP Smart Array 642 Controller. Overview. Retired Overview The Smart Array 642 Controller (SA-642) is a 64-bit, 133-MHz PCI-X, dual channel, SCSI array controller for entry-level hardwarebased fault tolerance. Utilizing both SCSI channels of the SA-642

More information

Virtual Memory. Reading. Sections 5.4, 5.5, 5.6, 5.8, 5.10 (2) Lecture notes from MKP and S. Yalamanchili

Virtual Memory. Reading. Sections 5.4, 5.5, 5.6, 5.8, 5.10 (2) Lecture notes from MKP and S. Yalamanchili Virtual Memory Lecture notes from MKP and S. Yalamanchili Sections 5.4, 5.5, 5.6, 5.8, 5.10 Reading (2) 1 The Memory Hierarchy ALU registers Cache Memory Memory Memory Managed by the compiler Memory Managed

More information

Chapter 6. Storage and Other I/O Topics

Chapter 6. Storage and Other I/O Topics Chapter 6 Storage and Other I/O Topics Introduction I/O devices can be characterized by Behaviour: input, output, storage Partner: human or machine Data rate: bytes/sec, transfers/sec I/O bus connections

More information

Lenovo RAID Introduction Reference Information

Lenovo RAID Introduction Reference Information Lenovo RAID Introduction Reference Information Using a Redundant Array of Independent Disks (RAID) to store data remains one of the most common and cost-efficient methods to increase server's storage performance,

More information

B.H.GARDI COLLEGE OF ENGINEERING & TECHNOLOGY (MCA Dept.) Parallel Database Database Management System - 2

B.H.GARDI COLLEGE OF ENGINEERING & TECHNOLOGY (MCA Dept.) Parallel Database Database Management System - 2 Introduction :- Today single CPU based architecture is not capable enough for the modern database that are required to handle more demanding and complex requirements of the users, for example, high performance,

More information

Evaluation Report: Improving SQL Server Database Performance with Dot Hill AssuredSAN 4824 Flash Upgrades

Evaluation Report: Improving SQL Server Database Performance with Dot Hill AssuredSAN 4824 Flash Upgrades Evaluation Report: Improving SQL Server Database Performance with Dot Hill AssuredSAN 4824 Flash Upgrades Evaluation report prepared under contract with Dot Hill August 2015 Executive Summary Solid state

More information

Assessing performance in HP LeftHand SANs

Assessing performance in HP LeftHand SANs Assessing performance in HP LeftHand SANs HP LeftHand Starter, Virtualization, and Multi-Site SANs deliver reliable, scalable, and predictable performance White paper Introduction... 2 The advantages of

More information

Storage Systems. Storage Systems

Storage Systems. Storage Systems Storage Systems Storage Systems We already know about four levels of storage: Registers Cache Memory Disk But we've been a little vague on how these devices are interconnected In this unit, we study Input/output

More information

Models Smart Array 6402/128 Controller B21 Smart Array 6404/256 Controller B21

Models Smart Array 6402/128 Controller B21 Smart Array 6404/256 Controller B21 Overview The Smart Array 6400 high performance Ultra320, PCI-X controller family provides maximum performance, flexibility, and reliable data protection for HP ProLiant servers, through its unique modular

More information

Chapter 10: Mass-Storage Systems

Chapter 10: Mass-Storage Systems Chapter 10: Mass-Storage Systems Silberschatz, Galvin and Gagne 2013 Chapter 10: Mass-Storage Systems Overview of Mass Storage Structure Disk Structure Disk Attachment Disk Scheduling Disk Management Swap-Space

More information

1 of 6 4/8/2011 4:08 PM Electronic Hardware Information, Guides and Tools search newsletter subscribe Home Utilities Downloads Links Info Ads by Google Raid Hard Drives Raid Raid Data Recovery SSD in Raid

More information

Computer Architecture Computer Science & Engineering. Chapter 6. Storage and Other I/O Topics BK TP.HCM

Computer Architecture Computer Science & Engineering. Chapter 6. Storage and Other I/O Topics BK TP.HCM Computer Architecture Computer Science & Engineering Chapter 6 Storage and Other I/O Topics Introduction I/O devices can be characterized by Behaviour: input, output, storage Partner: human or machine

More information

Physical Storage Media

Physical Storage Media Physical Storage Media These slides are a modified version of the slides of the book Database System Concepts, 5th Ed., McGraw-Hill, by Silberschatz, Korth and Sudarshan. Original slides are available

More information

IBM i Version 7.3. Systems management Disk management IBM

IBM i Version 7.3. Systems management Disk management IBM IBM i Version 7.3 Systems management Disk management IBM IBM i Version 7.3 Systems management Disk management IBM Note Before using this information and the product it supports, read the information in

More information

Target Environments The Smart Array 6i Controller offers superior investment protection to the following environments: Non-RAID

Target Environments The Smart Array 6i Controller offers superior investment protection to the following environments: Non-RAID Overview The Smart Array 6i controller is an Ultra320 intelligent array controller for entry-level, hardware-based fault tolerance for protection of OS, applications, and logs. Most models have one internal-only

More information

CMSC 424 Database design Lecture 12 Storage. Mihai Pop

CMSC 424 Database design Lecture 12 Storage. Mihai Pop CMSC 424 Database design Lecture 12 Storage Mihai Pop Administrative Office hours tomorrow @ 10 Midterms are in solutions for part C will be posted later this week Project partners I have an odd number

More information

Chapter 10: Mass-Storage Systems. Operating System Concepts 9 th Edition

Chapter 10: Mass-Storage Systems. Operating System Concepts 9 th Edition Chapter 10: Mass-Storage Systems Silberschatz, Galvin and Gagne 2013 Chapter 10: Mass-Storage Systems Overview of Mass Storage Structure Disk Structure Disk Attachment Disk Scheduling Disk Management Swap-Space

More information

VERITAS Storage Foundation 4.0 TM for Databases

VERITAS Storage Foundation 4.0 TM for Databases VERITAS Storage Foundation 4.0 TM for Databases Powerful Manageability, High Availability and Superior Performance for Oracle, DB2 and Sybase Databases Enterprises today are experiencing tremendous growth

More information

Database Management Systems, 2nd edition, Raghu Ramakrishnan, Johannes Gehrke, McGraw-Hill

Database Management Systems, 2nd edition, Raghu Ramakrishnan, Johannes Gehrke, McGraw-Hill Lecture Handout Database Management System Lecture No. 34 Reading Material Database Management Systems, 2nd edition, Raghu Ramakrishnan, Johannes Gehrke, McGraw-Hill Modern Database Management, Fred McFadden,

More information

I/O Management and Disk Scheduling. Chapter 11

I/O Management and Disk Scheduling. Chapter 11 I/O Management and Disk Scheduling Chapter 11 Categories of I/O Devices Human readable used to communicate with the user video display terminals keyboard mouse printer Categories of I/O Devices Machine

More information

In the late 1980s, rapid adoption of computers

In the late 1980s, rapid adoption of computers hapter 3 ata Protection: RI In the late 1980s, rapid adoption of computers for business processes stimulated the KY ONPTS Hardware and Software RI growth of new applications and databases, significantly

More information

QuickSpecs. Models. Overview

QuickSpecs. Models. Overview Overview The HP Smart Array P400 is HP's first PCI-Express (PCIe) serial attached SCSI (SAS) RAID controller and provides new levels of performance and reliability for HP servers, through its support of

More information

Storage Optimization with Oracle Database 11g

Storage Optimization with Oracle Database 11g Storage Optimization with Oracle Database 11g Terabytes of Data Reduce Storage Costs by Factor of 10x Data Growth Continues to Outpace Budget Growth Rate of Database Growth 1000 800 600 400 200 1998 2000

More information

Introduction Disks RAID Tertiary storage. Mass Storage. CMSC 420, York College. November 21, 2006

Introduction Disks RAID Tertiary storage. Mass Storage. CMSC 420, York College. November 21, 2006 November 21, 2006 The memory hierarchy Red = Level Access time Capacity Features Registers nanoseconds 100s of bytes fixed Cache nanoseconds 1-2 MB fixed RAM nanoseconds MBs to GBs expandable Disk milliseconds

More information

EMC Tiered Storage for Microsoft SQL Server 2008 Enabled by EMC CLARiiON CX4 and Enterprise Flash Drives

EMC Tiered Storage for Microsoft SQL Server 2008 Enabled by EMC CLARiiON CX4 and Enterprise Flash Drives EMC Tiered Storage for Microsoft SQL Server 2008 Enabled by EMC CLARiiON CX4 and Enterprise Flash Drives A Detailed Review EMC Information Infrastructure Solutions Abstract This white paper demonstrates

More information

V. Mass Storage Systems

V. Mass Storage Systems TDIU25: Operating Systems V. Mass Storage Systems SGG9: chapter 12 o Mass storage: Hard disks, structure, scheduling, RAID Copyright Notice: The lecture notes are mainly based on modifications of the slides

More information

Performance of relational database management

Performance of relational database management Building a 3-D DRAM Architecture for Optimum Cost/Performance By Gene Bowles and Duke Lambert As systems increase in performance and power, magnetic disk storage speeds have lagged behind. But using solidstate

More information

Chapter 12: Mass-Storage

Chapter 12: Mass-Storage Chapter 12: Mass-Storage Systems Chapter 12: Mass-Storage Systems Revised 2010. Tao Yang Overview of Mass Storage Structure Disk Structure Disk Attachment Disk Scheduling Disk Management Swap-Space Management

More information

Storage Designed to Support an Oracle Database. White Paper

Storage Designed to Support an Oracle Database. White Paper Storage Designed to Support an Oracle Database White Paper Abstract Databases represent the backbone of most organizations. And Oracle databases in particular have become the mainstream data repository

More information

Java Without the Jitter

Java Without the Jitter TECHNOLOGY WHITE PAPER Achieving Ultra-Low Latency Table of Contents Executive Summary... 3 Introduction... 4 Why Java Pauses Can t Be Tuned Away.... 5 Modern Servers Have Huge Capacities Why Hasn t Latency

More information

CSE325 Principles of Operating Systems. Mass-Storage Systems. David P. Duggan. April 19, 2011

CSE325 Principles of Operating Systems. Mass-Storage Systems. David P. Duggan. April 19, 2011 CSE325 Principles of Operating Systems Mass-Storage Systems David P. Duggan dduggan@sandia.gov April 19, 2011 Outline Storage Devices Disk Scheduling FCFS SSTF SCAN, C-SCAN LOOK, C-LOOK Redundant Arrays

More information

COMP283-Lecture 3 Applied Database Management

COMP283-Lecture 3 Applied Database Management COMP283-Lecture 3 Applied Database Management Introduction DB Design Continued Disk Sizing Disk Types & Controllers DB Capacity 1 COMP283-Lecture 3 DB Storage: Linear Growth Disk space requirements increases

More information

IBM. Systems management Disk management. IBM i 7.1

IBM. Systems management Disk management. IBM i 7.1 IBM IBM i Systems management Disk management 7.1 IBM IBM i Systems management Disk management 7.1 Note Before using this information and the product it supports, read the information in Notices, on page

More information

Chapter 13: Mass-Storage Systems. Disk Scheduling. Disk Scheduling (Cont.) Disk Structure FCFS. Moving-Head Disk Mechanism

Chapter 13: Mass-Storage Systems. Disk Scheduling. Disk Scheduling (Cont.) Disk Structure FCFS. Moving-Head Disk Mechanism Chapter 13: Mass-Storage Systems Disk Scheduling Disk Structure Disk Scheduling Disk Management Swap-Space Management RAID Structure Disk Attachment Stable-Storage Implementation Tertiary Storage Devices

More information

Chapter 13: Mass-Storage Systems. Disk Structure

Chapter 13: Mass-Storage Systems. Disk Structure Chapter 13: Mass-Storage Systems Disk Structure Disk Scheduling Disk Management Swap-Space Management RAID Structure Disk Attachment Stable-Storage Implementation Tertiary Storage Devices Operating System

More information

HP SAS benchmark performance tests

HP SAS benchmark performance tests HP SAS benchmark performance tests technology brief Abstract... 2 Introduction... 2 Test hardware... 2 HP ProLiant DL585 server... 2 HP ProLiant DL380 G4 and G4 SAS servers... 3 HP Smart Array P600 SAS

More information

ZKLWHýSDSHU. 3UHð)DLOXUHý:DUUDQW\ý 0LQLPL]LQJý8QSODQQHGý'RZQWLPH. +3ý 1HW6HUYHUý 0DQDJHPHQW. Executive Summary. A Closer Look

ZKLWHýSDSHU. 3UHð)DLOXUHý:DUUDQW\ý 0LQLPL]LQJý8QSODQQHGý'RZQWLPH. +3ý 1HW6HUYHUý 0DQDJHPHQW. Executive Summary. A Closer Look 3UHð)DLOXUHý:DUUDQW\ý 0LQLPL]LQJý8QSODQQHGý'RZQWLPH ZKLWHýSDSHU Executive Summary The Hewlett-Packard Pre-Failure Warranty 1 helps you run your business with less downtime. It extends the advantage of

More information

High Availability through Warm-Standby Support in Sybase Replication Server A Whitepaper from Sybase, Inc.

High Availability through Warm-Standby Support in Sybase Replication Server A Whitepaper from Sybase, Inc. High Availability through Warm-Standby Support in Sybase Replication Server A Whitepaper from Sybase, Inc. Table of Contents Section I: The Need for Warm Standby...2 The Business Problem...2 Section II:

More information

Chapter 6. Storage and Other I/O Topics

Chapter 6. Storage and Other I/O Topics Chapter 6 Storage and Other I/O Topics Introduction I/O devices can be characterized by Behaviour: input, output, storage Partner: human or machine Data rate: bytes/sec, transfers/sec I/O bus connections

More information

CONFIGURING ftscalable STORAGE ARRAYS ON OpenVOS SYSTEMS

CONFIGURING ftscalable STORAGE ARRAYS ON OpenVOS SYSTEMS Best Practices CONFIGURING ftscalable STORAGE ARRAYS ON OpenVOS SYSTEMS Best Practices 2 Abstract ftscalable TM Storage G1, G2 and G3 arrays are highly flexible, scalable hardware storage subsystems that

More information

QuickSpecs. Models. Overview

QuickSpecs. Models. Overview Overview The HP Smart Array P800 is HP's first 16 port serial attached SCSI (SAS) RAID controller with PCI-Express (PCIe). It is the highest performing controller in the SAS portfolio and provides new

More information

Microsoft Office SharePoint Server 2007

Microsoft Office SharePoint Server 2007 Microsoft Office SharePoint Server 2007 Enabled by EMC Celerra Unified Storage and Microsoft Hyper-V Reference Architecture Copyright 2010 EMC Corporation. All rights reserved. Published May, 2010 EMC

More information

Administrivia. CMSC 411 Computer Systems Architecture Lecture 19 Storage Systems, cont. Disks (cont.) Disks - review

Administrivia. CMSC 411 Computer Systems Architecture Lecture 19 Storage Systems, cont. Disks (cont.) Disks - review Administrivia CMSC 411 Computer Systems Architecture Lecture 19 Storage Systems, cont. Homework #4 due Thursday answers posted soon after Exam #2 on Thursday, April 24 on memory hierarchy (Unit 4) and

More information

RAID: The Innovative Data Storage Manager

RAID: The Innovative Data Storage Manager RAID: The Innovative Data Storage Manager Amit Tyagi IIMT College of Engineering, Greater Noida, UP, India Abstract-RAID is a technology that is used to increase the performance and/or reliability of data

More information

RAID (Redundant Array of Inexpensive Disks)

RAID (Redundant Array of Inexpensive Disks) Magnetic Disk Characteristics I/O Connection Structure Types of Buses Cache & I/O I/O Performance Metrics I/O System Modeling Using Queuing Theory Designing an I/O System RAID (Redundant Array of Inexpensive

More information

QuickSpecs. Models. HP Smart Array P400i Controller. Overview

QuickSpecs. Models. HP Smart Array P400i Controller. Overview Overview The HP Smart Array P400 Serial Attached SCSI (SAS) controller (SA-P400) provides new levels of performance and reliability for HP servers, through its support of the latest SCSI technology and

More information

WHITEPAPER. Disk Configuration Tips for Ingres by Chip nickolett, Ingres Corporation

WHITEPAPER. Disk Configuration Tips for Ingres by Chip nickolett, Ingres Corporation WHITEPAPER Disk Configuration Tips for Ingres by Chip nickolett, Ingres Corporation table of contents: 3 Preface 3 Overview 4 How Many Disks Do I Need? 5 Should I Use RAID? 6 Ingres Configuration Recommendations

More information

1. Introduction. Traditionally, a high bandwidth file system comprises a supercomputer with disks connected

1. Introduction. Traditionally, a high bandwidth file system comprises a supercomputer with disks connected 1. Introduction Traditionally, a high bandwidth file system comprises a supercomputer with disks connected by a high speed backplane bus such as SCSI [3][4] or Fibre Channel [2][67][71]. These systems

More information

DELL EMC DATA DOMAIN SISL SCALING ARCHITECTURE

DELL EMC DATA DOMAIN SISL SCALING ARCHITECTURE WHITEPAPER DELL EMC DATA DOMAIN SISL SCALING ARCHITECTURE A Detailed Review ABSTRACT While tape has been the dominant storage medium for data protection for decades because of its low cost, it is steadily

More information

Storage Devices for Database Systems

Storage Devices for Database Systems Storage Devices for Database Systems 5DV120 Database System Principles Umeå University Department of Computing Science Stephen J. Hegner hegner@cs.umu.se http://www.cs.umu.se/~hegner Storage Devices for

More information

PowerVault MD3 Storage Array Enterprise % Availability

PowerVault MD3 Storage Array Enterprise % Availability PowerVault MD3 Storage Array Enterprise 99.999% Availability Dell Engineering June 2015 A Dell Technical White Paper THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS

More information

Demystifying Storage Area Networks. Michael Wells Microsoft Application Solutions Specialist EMC Corporation

Demystifying Storage Area Networks. Michael Wells Microsoft Application Solutions Specialist EMC Corporation Demystifying Storage Area Networks Michael Wells Microsoft Application Solutions Specialist EMC Corporation About Me DBA for 7+ years Developer for 10+ years MCSE: Data Platform MCSE: SQL Server 2012 MCITP:

More information

EMC CLARiiON Backup Storage Solutions

EMC CLARiiON Backup Storage Solutions Engineering White Paper Backup-to-Disk Guide with Computer Associates BrightStor ARCserve Backup Abstract This white paper describes how to configure EMC CLARiiON CX series storage systems with Computer

More information

Oracle Exadata: Strategy and Roadmap

Oracle Exadata: Strategy and Roadmap Oracle Exadata: Strategy and Roadmap - New Technologies, Cloud, and On-Premises Juan Loaiza Senior Vice President, Database Systems Technologies, Oracle Safe Harbor Statement The following is intended

More information

Recommendations for Aligning VMFS Partitions

Recommendations for Aligning VMFS Partitions VMWARE PERFORMANCE STUDY VMware ESX Server 3.0 Recommendations for Aligning VMFS Partitions Partition alignment is a known issue in physical file systems, and its remedy is well-documented. The goal of

More information

IBM Spectrum NAS. Easy-to-manage software-defined file storage for the enterprise. Overview. Highlights

IBM Spectrum NAS. Easy-to-manage software-defined file storage for the enterprise. Overview. Highlights IBM Spectrum NAS Easy-to-manage software-defined file storage for the enterprise Highlights Reduce capital expenditures with storage software on commodity servers Improve efficiency by consolidating all

More information

Chapter 11: File System Implementation. Objectives

Chapter 11: File System Implementation. Objectives Chapter 11: File System Implementation Objectives To describe the details of implementing local file systems and directory structures To describe the implementation of remote file systems To discuss block

More information

HP Smart Array Controller technology

HP Smart Array Controller technology HP Smart Array Controller technology technology brief, 3 rd edition Abstract... 3 Introduction... 3 Storage trends... 3 Meeting data storage requirements... 4 High performance... 4 PCI Express technology...

More information

Deploy a High-Performance Database Solution: Cisco UCS B420 M4 Blade Server with Fusion iomemory PX600 Using Oracle Database 12c

Deploy a High-Performance Database Solution: Cisco UCS B420 M4 Blade Server with Fusion iomemory PX600 Using Oracle Database 12c White Paper Deploy a High-Performance Database Solution: Cisco UCS B420 M4 Blade Server with Fusion iomemory PX600 Using Oracle Database 12c What You Will Learn This document demonstrates the benefits

More information

Introduction I/O 1. I/O devices can be characterized by Behavior: input, output, storage Partner: human or machine Data rate: bytes/sec, transfers/sec

Introduction I/O 1. I/O devices can be characterized by Behavior: input, output, storage Partner: human or machine Data rate: bytes/sec, transfers/sec Introduction I/O 1 I/O devices can be characterized by Behavior: input, output, storage Partner: human or machine Data rate: bytes/sec, transfers/sec I/O bus connections I/O Device Summary I/O 2 I/O System

More information

VERITAS Volume Replicator. Successful Replication and Disaster Recovery

VERITAS Volume Replicator. Successful Replication and Disaster Recovery VERITAS Volume Replicator Successful Replication and Disaster Recovery V E R I T A S W H I T E P A P E R Table of Contents Introduction.................................................................................1

More information

Today: Secondary Storage! Typical Disk Parameters!

Today: Secondary Storage! Typical Disk Parameters! Today: Secondary Storage! To read or write a disk block: Seek: (latency) position head over a track/cylinder. The seek time depends on how fast the hardware moves the arm. Rotational delay: (latency) time

More information

Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted from Hennessy & Patterson / 2003 Elsevier

Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted from Hennessy & Patterson / 2003 Elsevier Some material adapted from Mohamed Younis, UMBC CMSC 6 Spr 23 course slides Some material adapted from Hennessy & Patterson / 23 Elsevier Science Characteristics IBM 39 IBM UltraStar Integral 82 Disk diameter

More information

How to Speed up Database Applications with a Purpose-Built SSD Storage Solution

How to Speed up Database Applications with a Purpose-Built SSD Storage Solution How to Speed up Database Applications with a Purpose-Built SSD Storage Solution SAN Accessible Storage Array Speeds Applications by up to 25x Introduction Whether deployed in manufacturing, finance, web

More information

Configuring and Using Redundant Disks with Cisco MCS

Configuring and Using Redundant Disks with Cisco MCS Configuring and Using Redundant Disks with Cisco MCS Document ID: 9229 Contents Introduction Prerequisites Requirements Components Used Conventions Drive Mirroring (RAID 1) Recognize a Drive Failure Recover

More information

An Introduction to RAID

An Introduction to RAID Intro An Introduction to RAID Gursimtan Singh Dept. of CS & IT Doaba College RAID stands for Redundant Array of Inexpensive Disks. RAID is the organization of multiple disks into a large, high performance

More information

Storage systems. Computer Systems Architecture CMSC 411 Unit 6 Storage Systems. (Hard) Disks. Disk and Tape Technologies. Disks (cont.

Storage systems. Computer Systems Architecture CMSC 411 Unit 6 Storage Systems. (Hard) Disks. Disk and Tape Technologies. Disks (cont. Computer Systems Architecture CMSC 4 Unit 6 Storage Systems Alan Sussman November 23, 2004 Storage systems We already know about four levels of storage: registers cache memory disk but we've been a little

More information

Best Practices. Deploying Optim Performance Manager in large scale environments. IBM Optim Performance Manager Extended Edition V4.1.0.

Best Practices. Deploying Optim Performance Manager in large scale environments. IBM Optim Performance Manager Extended Edition V4.1.0. IBM Optim Performance Manager Extended Edition V4.1.0.1 Best Practices Deploying Optim Performance Manager in large scale environments Ute Baumbach (bmb@de.ibm.com) Optim Performance Manager Development

More information

Tape pictures. CSE 30341: Operating Systems Principles

Tape pictures. CSE 30341: Operating Systems Principles Tape pictures 4/11/07 CSE 30341: Operating Systems Principles page 1 Tape Drives The basic operations for a tape drive differ from those of a disk drive. locate positions the tape to a specific logical

More information

Che-Wei Chang Department of Computer Science and Information Engineering, Chang Gung University

Che-Wei Chang Department of Computer Science and Information Engineering, Chang Gung University Che-Wei Chang chewei@mail.cgu.edu.tw Department of Computer Science and Information Engineering, Chang Gung University l Chapter 10: File System l Chapter 11: Implementing File-Systems l Chapter 12: Mass-Storage

More information

vsan 6.6 Performance Improvements First Published On: Last Updated On:

vsan 6.6 Performance Improvements First Published On: Last Updated On: vsan 6.6 Performance Improvements First Published On: 07-24-2017 Last Updated On: 07-28-2017 1 Table of Contents 1. Overview 1.1.Executive Summary 1.2.Introduction 2. vsan Testing Configuration and Conditions

More information

Database Systems II. Secondary Storage

Database Systems II. Secondary Storage Database Systems II Secondary Storage CMPT 454, Simon Fraser University, Fall 2009, Martin Ester 29 The Memory Hierarchy Swapping, Main-memory DBMS s Tertiary Storage: Tape, Network Backup 3,200 MB/s (DDR-SDRAM

More information

Benefits of Multi-Node Scale-out Clusters running NetApp Clustered Data ONTAP. Silverton Consulting, Inc. StorInt Briefing

Benefits of Multi-Node Scale-out Clusters running NetApp Clustered Data ONTAP. Silverton Consulting, Inc. StorInt Briefing Benefits of Multi-Node Scale-out Clusters running NetApp Clustered Data ONTAP Silverton Consulting, Inc. StorInt Briefing BENEFITS OF MULTI- NODE SCALE- OUT CLUSTERS RUNNING NETAPP CDOT PAGE 2 OF 7 Introduction

More information

CS2410: Computer Architecture. Storage systems. Sangyeun Cho. Computer Science Department University of Pittsburgh

CS2410: Computer Architecture. Storage systems. Sangyeun Cho. Computer Science Department University of Pittsburgh CS24: Computer Architecture Storage systems Sangyeun Cho Computer Science Department (Some slides borrowed from D Patterson s lecture slides) Case for storage Shift in focus from computation to communication

More information

CS370: System Architecture & Software [Fall 2014] Dept. Of Computer Science, Colorado State University

CS370: System Architecture & Software [Fall 2014] Dept. Of Computer Science, Colorado State University CS 370: SYSTEM ARCHITECTURE & SOFTWARE [MASS STORAGE] Frequently asked questions from the previous class survey Shrideep Pallickara Computer Science Colorado State University L29.1 L29.2 Topics covered

More information

Storing Data: Disks and Files

Storing Data: Disks and Files Storing Data: Disks and Files Chapter 7 (2 nd edition) Chapter 9 (3 rd edition) Yea, from the table of my memory I ll wipe away all trivial fond records. -- Shakespeare, Hamlet Database Management Systems,

More information

Condusiv s V-locity VM Accelerates Exchange 2010 over 60% on Virtual Machines without Additional Hardware

Condusiv s V-locity VM Accelerates Exchange 2010 over 60% on Virtual Machines without Additional Hardware openbench Labs Executive Briefing: March 13, 2013 Condusiv s V-locity VM Accelerates Exchange 2010 over 60% on Virtual Machines without Additional Hardware Optimizing I/O for Increased Throughput and Reduced

More information

RocketRAID Intelli-VRM (Intelligent Virtual RAID Management) Early Warning System and Virtual System Rescue

RocketRAID Intelli-VRM (Intelligent Virtual RAID Management) Early Warning System and Virtual System Rescue RocketRAID Intelli-VRM (Intelligent Virtual RAID Management) Early Warning System and Virtual System Rescue Introduction The fast-paced, high-definition requirements of our modern, digital age, has increased

More information

PRESERVE DATABASE PERFORMANCE WHEN RUNNING MIXED WORKLOADS

PRESERVE DATABASE PERFORMANCE WHEN RUNNING MIXED WORKLOADS PRESERVE DATABASE PERFORMANCE WHEN RUNNING MIXED WORKLOADS Testing shows that a Pure Storage FlashArray//m storage array used for Microsoft SQL Server 2016 helps eliminate latency and preserve productivity.

More information

Database Systems. November 2, 2011 Lecture #7. topobo (mit)

Database Systems. November 2, 2011 Lecture #7. topobo (mit) Database Systems November 2, 2011 Lecture #7 1 topobo (mit) 1 Announcement Assignment #2 due today Assignment #3 out today & due on 11/16. Midterm exam in class next week. Cover Chapters 1, 2,

More information

IBM System Storage DS5020 Express

IBM System Storage DS5020 Express IBM DS5020 Express Manage growth, complexity, and risk with scalable, high-performance storage Highlights Mixed host interfaces support (FC/iSCSI) enables SAN tiering Balanced performance well-suited for

More information

RAID Technology White Paper

RAID Technology White Paper RAID Technology White Paper As specialists in data storage, LaCie recognizes that almost all computer users will need a storage or backup solution and that people use and store data in different ways.

More information

Appendix D: Storage Systems

Appendix D: Storage Systems Appendix D: Storage Systems Instructor: Josep Torrellas CS433 Copyright Josep Torrellas 1999, 2001, 2002, 2013 1 Storage Systems : Disks Used for long term storage of files temporarily store parts of pgm

More information

Chapter 10: Mass-Storage Systems. Operating System Concepts 9 th Edition

Chapter 10: Mass-Storage Systems. Operating System Concepts 9 th Edition Chapter 10: Mass-Storage Systems Silberschatz, Galvin and Gagne 2013 Objectives To describe the physical structure of secondary storage devices and its effects on the uses of the devices To explain the

More information

vsan Remote Office Deployment January 09, 2018

vsan Remote Office Deployment January 09, 2018 January 09, 2018 1 1. vsan Remote Office Deployment 1.1.Solution Overview Table of Contents 2 1. vsan Remote Office Deployment 3 1.1 Solution Overview Native vsphere Storage for Remote and Branch Offices

More information

Veritas InfoScale Enterprise for Oracle Real Application Clusters (RAC)

Veritas InfoScale Enterprise for Oracle Real Application Clusters (RAC) Veritas InfoScale Enterprise for Oracle Real Application Clusters (RAC) Manageability and availability for Oracle RAC databases Overview Veritas InfoScale Enterprise for Oracle Real Application Clusters

More information