A Remote Hot Standby System of Oracle

Similar documents
Dual-System Warm Standby of Remote Sensing Satellite Control System Technology

A Reverse Differential Archiving Method based on Zdelta

Research on Availability of Virtual Machine Hot Standby based on Double Shadow Page Tables

Construction Scheme for Cloud Platform of NSFC Information System

Design and Implementation of Dual-Mode Wireless Video Monitoring System

A Behavior Based File Checkpointing Strategy

Linux Software RAID Level 0 Technique for High Performance Computing by using PCI-Express based SSD

Efficient Path Finding Method Based Evaluation Function in Large Scene Online Games and Its Application

The Google File System

Design and Implementation of cache protecting from power failure in Disk Array

Replication Solutions with Open-E Data Storage Server (DSS) April 2009

Dell PowerVault MD3600f/MD3620f Remote Replication Functional Guide

Live Virtual Machine Migration with Efficient Working Set Prediction

Chapter 17: Recovery System

The Design and Implementation of Disaster Recovery in Dual-active Cloud Center

A Database Redo Log System Based on Virtual Memory Disk*

Current Topics in OS Research. So, what s hot?

The Right Choice for DR: Data Guard, Stretch Clusters, or Remote Mirroring. Ashish Ray Group Product Manager Oracle Corporation

Associate Professor Dr. Raed Ibraheem Hamed

GFS: The Google File System. Dr. Yingwu Zhu

Caching and reliability

Clipmap based Loading and Rendering of Large Terrain Texture

HUAWEI OceanStor Enterprise Unified Storage System. HyperReplication Technical White Paper. Issue 01. Date HUAWEI TECHNOLOGIES CO., LTD.

Implementation and Performance Evaluation of RAPID-Cache under Linux

M out of N Safety Computing System Based on General-Purpose Computers

SMD149 - Operating Systems - File systems

Picking up the First Arrivals in VSP Based on Edge Detection

Self-Adaptive Two-Dimensional RAID Arrays

HP Designing and Implementing HP Enterprise Backup Solutions. Download Full Version :

Design and Implement of CH376 Driver Based on FPGA

Research and Design of Crypto Card Virtualization Framework Lei SUN, Ze-wu WANG and Rui-chen SUN

Research on Load Balancing and Database Replication based on Linux

Research on Mass Image Storage Platform Based on Cloud Computing

High Availability and Disaster Recovery Solutions for Perforce

Application of RMAN Backup Technology in the Agricultural Products Wholesale Market System

Managing and Maintaining a Microsoft Windows Server 2003 Environment

In today s global business environment, companies must maintain

Australian Journal of Basic and Applied Sciences

Design and Implementation of unified Identity Authentication System Based on LDAP in Digital Campus

Synchronous. (Block Based)

Research Article Mobile Storage and Search Engine of Information Oriented to Food Cloud

HP2-T11. Industry Standard Architecture and Technology.

EIDE Disk Arrays and Its Implement

The Research of Collision Detection Algorithm Based on Separating axis Theorem

The Research and Improvement of HARQ in TD-LTE Systems

VERITAS Volume Replicator Successful Replication and Disaster Recovery

Control Theory and Informatics ISSN (print) ISSN (online) Vol 2, No.1, 2012

A Review on Backup-up Practices using Deduplication

The Google File System

Routing Journal Operations on Disks Using Striping With Parity 1

Delayed Partial Parity Scheme for Reliable and High-Performance Flash Memory SSD

The Google File System. Alexandru Costan

ECE 7650 Scalable and Secure Internet Services and Architecture ---- A Systems Perspective

Online Version Only. Book made by this file is ILLEGAL. Design and Implementation of Binary File Similarity Evaluation System. 1.

ApsaraDB for RDS. Product Introduction

GFS: The Google File System

Google File System. By Dinesh Amatya

C13: Files and Directories: System s Perspective

Resource Management on a Mixed Processor Linux Cluster. Haibo Wang. Mississippi Center for Supercomputing Research

RAID (Redundant Array of Inexpensive Disks)

RAID6L: A Log-Assisted RAID6 Storage Architecture with Improved Write Performance

CSE 451: Operating Systems. Section 10 Project 3 wrap-up, final exam review

Lecture 21: Reliable, High Performance Storage. CSC 469H1F Fall 2006 Angela Demke Brown

Chapter 17: Recovery System

Failure Classification. Chapter 17: Recovery System. Recovery Algorithms. Storage Structure

Huawei KunLun Mission Critical Server. KunLun 9008/9016/9032 Technical Specifications

The Design of Distributed File System Based on HDFS Yannan Wang 1, a, Shudong Zhang 2, b, Hui Liu 3, c

Design and Implementation of Various File Deduplication Schemes on Storage Devices

The Google File System (GFS)

HP AutoRAID (Lecture 5, cs262a)

Research on Software Scheduling Technology Based on Multi-Buffered Parallel Encryption

Engineering Goals. Scalability Availability. Transactional behavior Security EAI... CS530 S05

Huawei OceanStor ReplicationDirector Software Technical White Paper HUAWEI TECHNOLOGIES CO., LTD. Issue 01. Date

VERITAS Volume Replicator. Successful Replication and Disaster Recovery

Oracle Database 11g Data Guard

Basics of SQL Transactions

Survey on MapReduce Scheduling Algorithms

The Design and Implementation of Electronic Program Guide In Digital TV Set-Top-Box

BusinessObjects XI Release 2

Consistency and Scalability

Design and Implementation of a Random Access File System for NVRAM

Chapter 11: File System Implementation. Objectives

Outline. Failure Types

The Construction of Open Source Cloud Storage System for Digital Resources

Open Access Research on the Prediction Model of Material Cost Based on Data Mining

Alfresco 2.1. Backup and High Availability Guide

Overcoming Obstacles to Petabyte Archives

A Multi-party Decision Hot-standby Model

Outline. INF3190:Distributed Systems - Examples. Last week: Definitions Transparencies Challenges&pitfalls Architecturalstyles

iscsi Target Usage Guide December 15, 2017

The Design of Supermarket Electronic Shopping Guide System Based on ZigBee Communication

A Design and Implementation of Custom Communication Protocol Based on Aurora

Disaster Recovery (DR) Planning with the Cloud Desktop

Moment-preserving Based Watermarking for Color Image Authentication and Recovery

CHAPTER 11: IMPLEMENTING FILE SYSTEMS (COMPACT) By I-Chen Lin Textbook: Operating System Concepts 9th Ed.

A Step-by-Step Guide to Synchronous Volume Replication (Block Based) over a LAN with Open-E DSS. DSS ver up60 February 12, 2009

A Quantum Cryptography Communication Network Based on Software Defined Network

Oracle E-Business Availability Options. Solution Series for Oracle: 2 of 5

Research on Full-text Retrieval based on Lucene in Enterprise Content Management System Lixin Xu 1, a, XiaoLin Fu 2, b, Chunhua Zhang 1, c

Discover CephFS TECHNICAL REPORT SPONSORED BY. image vlastas, 123RF.com

Transcription:

2012 International Conference on Image, Vision and Computing (ICIVC 2012) IPCSIT vol. 50 (2012) (2012) IACSIT Press, Singapore DOI: 10.7763/IPCSIT.2012.V50.51 A Remote Hot Standby System of Oracle Qiu Long-jin and Gao Guang-qiang College of Computer Science, Sichuan University, Chengdu, China Abstract. For the problem of database hot standby system of oracle based on shared storage is difficult to ensure disaster recovery capacity, this paper proposed a remote host standby system of oracle. The system has features of remote hot backup, disaster recovery. It ensures the service continuity and disaster recovery capability of the server of oracle database. Experiments show that this system has high stability and capabilities of disaster survivability. Key words: hot standby system; disaste rtolerant solution; remote backup; service continuity; 1. Introduction The security and disaster recovery capability of oracle database are very important aspects for enterprise applications [1], many applications currently using hot standby system to ensure continuity of service. The mainstream oracle hot standby scheme is based on shared storage and two servers in the same local area network generally. When the master server can not provide services due to failure, it can switch to standby server quickly. The apparent drawback is that such scheme does not have disaster recovery capability, once the disaster occurs and the storage devices of oracle server are damaged, the data will not be recovered. The remote oracle hot standby system, referred to ORDB, is a network-based remote backup and recovery system [2]. Data changes of the backup disk on oracle master server (OMS) can be capture in real-time, then packaged into the Backup-Record and sent to remote control center (RCC) through the internet. RCC storage thread will write the back up records cached in buffer queue to the storage logical volume according to the taskid.the synchronization thread will send Backup-Records to the oracle standby server (OSS) to archive the data synchronization, storage or synchronization thread will plus 1 to the Backup-Record s Using-Count after the completion of a manipulation of the Backup-Record, the buffer queue will remove the Backup-Record while it s Using-Count equal to 2. This system can recovery the oracle database even if the OMS and the OSS data corruption at the same time, having a high capabilities for disaster recovery. Meanwhile, the real-time synchronization of OMS and OSS ensures the service connectivity.it is inexpensive but reliable and safe. 2. System Architecture 2.1. Architecture OSDB is consisted of oracle master server (OMS), oracle standby server (OSS) and remote control center (RCC).The architecture is shown in Figure 1. Corresponding author. Tel.: +13666272102. E-mail address: jinlong2204@yahoo.cn.

Internet Internet OMS RCC OSS Storage Server Fig. 1 OSDB system structure The responsibility of OMS is handing normal business of the database. We installed OSDB client on the OMS for the task of synchronization, recovery and monitoring data changes on local disk volume. OSS needs to install oracle with the same configuration as OMS, even the storage path of database files. The service of OSS is stopped when the service of OMS is normal. After installed the same version of the OSDB client, the RCC will send backup data of the OMS to the OSS constantly to achieve real-time synchronization of two servers in different places. When OMS be in fault, start oracle services and database instance on OSS, then set the IP of the database server to OSSIP, OSS will become the main server to provide services, and OMS will become standby server after data recovery by the control center. RCC maintain the task information, create storage logical volume corresponding taskid and control the synchronization, recovery, services switch, thus ensuring the service connectivity and capabilities of disaster survivability. 2.2. Modules OSDB client is consisted of disk management module, security control module and network communication modules. The remote control center is consisted of security control module, network communication module, storage management module. The module association is shown in Figure 2. Disk Manage Security Control Storage Manager Security Control Task Control Disk Manage Security Control Network Communication OMS Network Communication RCC Network Communication OMS Fig. 2 Functional Block Diagram The client's disk management module is made up of the massive cache and volume filter driver. Massive cache [3] is a memory-mapped file and can cache Backup-Records that needed to synchronize to the remote control center. Using memory mapped file as a large cache is mainly to improve the efficiency of the system IO. Volume filter driver [4] monitor the OMS data files, log files, control files and configuration files on the disk partition by intercepted writing IRP package, extract the data and package into Backup-Record which will be stored in massive cache and then sent to RCC use communication module to achieve real-time synchronization. When conducting data recovery, disk management module is responsible for extract the valid date from Backup-Records and write them into the disk volume according to the disk offset. Safety control module checks the network transmission errors of Backup-Record to ensure the accuracy of data transmission. Security control module computes the MD5 checksum of Backup-Records using MD5 algorithm, the Backup-Record format is: Struct record{ BYTE taskid; DWORD offset; BYTE * data; DWORD len; MD5 h;} taskid indicated the task id, a pair of OMS and OSS can only belong to a taskid, and RCC can control multiple tasks; offset corresponds to offset of valid data in the OMS disk partition or OSS storage logical

volume; data is a pointer which point to valid data; len is the size of record; h record the MD5 checksum of the Backup-Record, MD5 checksum calculation formula as: h MD5( taskid, offset, data, len) (1) When the RCC or the OSS receives the Backup-Records, MD5 checksum calculated using the formula 1 is compared with h in the Backup-Records, if the checking passed the data in the Backup-Records is to be written to the appropriate location according to the offset, or request a retransmission. RCC Mission Control module is used to control the creation of new task, modify task configuration, initializing synchronization, real-time synchronization and recovery, the main components of the module is APACHE, PHP,MYSQL and WEB management system, when users access the system needs to install a ActiveX control provided by the system which used to send control message to OMS, OSS or RCC, the corresponding control message service thread parses the message received and does the corresponding operation, the task structure as follows: Struct task{ PTSTR taskid; TCHAR vol; //disk volume PTSTR OMSIP; PTSTR RCCIP; PTSTR OSSIP; } RCC storage management module get OMS partition size when the new task is creating, the storage server then create the storage logical volume with the same size for storing backup data, which will be recycled when the task is deleting. Storage server can be disk array, NAS and other storage technology, this paper uses the disk array [5] as a physical storage server. 3. EY Processes for Achieving ORDB 3.1. Initializing Synchrony When creating a new task, you need to configure the parameters of structure task in the WEB management system, RCC create the corresponding storage logical volume, then the Initializing synchrony Must be started, which get the data of local disk volume where the database files storage then capture it into Backup-Record, OMS client network communication module will send it to RCC, RCC will write the Backup-Records to the storage logical volume and sent it to the OSS, thus the RCC and the OSS will have the same database files as OMS. The OMS client synchronization algorithm pseudo code is described as follows: Procedure synchro: SET totallength=getdisklength(task.vol) SET blocksize=64kb //the length of backup block SET offset=0; //current volume offset While(offset<totalLength) { record r=new record; r.taskid=task.taskid r.data=getrecord(task.vol,offset) r.offset=offset r.h=md5(r.taskid,r.offset,r.data) sendqueue.push(r) } While(!sendQueue.isempt()) { record r=sendqueue.pop(); Send r TO RCC; } RCC put the received Backup-Record into receive buffer queue, then the Backup-Records will be processed by storage threads and synchronization thread, storage thread find the corresponding logical volume

by the taskid and write data to it according to the offset, synchronize thread forward Backup-Records to OSS, OSS will check the value of MD5 after receiving the record, if the checking passed, the valid data in Backup- Records will be write to local disk volume according task s vol and offset. 3.2. Real-Time Synchrony Real-time synchrony [6] is the core of the system, real-time synchrony guarantees the consistence of database file between OMS, RCC and OSS real-time, making the remote hot standby achieved. the core module of real-time synchrony is the volume filter driver, the driver is located in block device layer, intercepted all packages which type are IRP_MJ_WRITE IRP, extract valid data and package into the Backup-Records, OMS real-time synchrony client flow chart shown in Figure 3 instructions. START Load Volume Filter Driver Not success Start the driver success YES Is the buffer queue empty? NO Send backup record Whether to stop monitoring NO YES Stop monitoring END Fig. 3 Real-time synchrony flow chart 3.3. Service Switch When the primary server failure or completely damaged by the disaster, switching to the standby server is the main purpose of a two-machine standby system. But the two-machine hot standby system is not disaster tolerant, if the storage device has been devastating damage, we can not recover data and services in the short term. This is the main issues addressed by ORDB, after failure of OMS, we can switch to the OSS to achieve service continuity immediately, does not require any data recovery. Service switch depends on the ORDB failure detection algorithm; OMS periodically send a status packet S to the RCC, the structure of S as follow: Struct S{ TIME ts //send time TIME tr //receive time } TYPE sta //statue of OMS service sta represent the statue of oracle service on OMS,which can be USABLE or STOPED. RCC defined a time window T for each task, the size of T is 100 times of the state packet transmission period, the collection of state package received S is stored in the time window, defined as follows: n S s n [1,100] (2) i i 1

RCC start a judgment once every 10 cycles for whether to run failure detection algorithm, the determine conditions as follows: n t T / 2 t { T ' s null} (3) i i i i 1 If the formula (3) hold, then start the failure detection algorithm, first calculate the average transmission time difference tv, calculated as follows: 100 tv ( s. ts s. tr) /100 s S (4) i 1 i i i If the average time span tv is more than 10 cycles, then we concede network failure and continue to wait, or empty the time window T, run the failure detection after 100 cycles, formula t s null s. sta USABLE s S, t T (5) i i i i i If Equation 5 is not met, we determine the OMS service is not available, so start the OSS database service and database instance, set the oracle server IP to OSSIP, the standby server becomes the primary server, origin server after the resumption of a standby server. 3.4. Data Recovery When the OMS, OSS damage occurs, you need to start the data recovery. First, RCC finds the corresponding logical volume by taskid, the data block in logic volume is packaged into Backup-Record and sent to the OMS or OSS, after completion of data file recovery we can start the service of oracle, if there are some inconsistencies generally, because all the files needed to recover database, so we can resolve the inconsistencies by oracle s mechanisms, such as rollback or media recovery. 4. Experiment To test ORDB time required for service switching, the author deployed the experimental environment of OMS, OSS and RCC, they are not in the same LAN, with independent IP, we install the oracle 10g and ORDB client on OMS and OSS, the ORDB server and web manage system on RCC, the operation system of OMS and OSS is Windows Server 2003, and Linux Enterprise 5 on RCC. Experimental steps are as follows: 1) Install oracle on the D drive In the OMS, built the database instance TESTDB, stored configuration files, control files, log files and data files in the E drive, create the test table, and then install the client, stop the all services of oracle; 2) Install oracle on OSS with the same configuration as OMS, but do not build the table; 3) Built management environment for RCC control center, logon WEB management system, then establish a new task on the OMS, OSS's E drives, set the member of struct task such as OMSIP, OSSIP and other information, and then delete all the files in the E drive on OSS, start full synchronization of the OMS, when the E disk data synchronized to the RCC the OSS and RCC will have complete copies of the database file on OMS ; 4) Start data monitoring on the E disk in OMS for real time synchronization, then run the stored procedure to simulation DML, DDL operations on database, some times after make OMS power-down; 5) Because of RCC detected that OMS has not sent status packets in 100 cycles, so determine OMS failure,then send the start command to OSS, OSS start the service of oracle. 6) OSS send a message to inform RCC that standby database startup completed, RCC set the primary database server IP to OSSIP, the connection access to the OMS will be switch to OSS, then start the recovery of the OMS. The author carried out 10 repeated experiments, recorded time required to complete the switching from the OMS power-down to the OSS can provide oracle service, and the data is shown in Table 1. Table 1 Times needed to switch services counts 1 759 2 541 time(s)

3 499 4 587 5 647 6 698 7 867 8 602 9 751 10 549 Shown in Table 1, ORDB system can be completed in about 10 minutes service switch, the system can not only prove ORDB complete database synchronization, data backup and disaster recovery, but also a very good warranty service connectivity. 5. Conclutions We have presented an architecture for remote hot standby system of oracle, which can backup all the oracle database file needed to recover a database on the remote server in real-time. When the Oracle Master Server collapsed, we can switch the oracle service to the Oracle Standby Server in about 10 minutes. Because we have three copies of the database file, even the OMS and OSS are damaged at the same time, we can also recover the database using the backup data in RCC. ORDB integrated the advantage of the two-machine hot standby system and the disaster recovers system, ensure the disaster tolerance and services continuity of oracle services. 6. References [1] Velpuri, Rama. Oracle backup & recovery handbook [M], 1997. [2] LAWLER C M, SZYGENDA S A, THORNTON M A. Techniques for disaster tolerant information technology systems. Proceedings of the 1st Annual 2007 IEEE Systems Conference[C]. Honolulu, HI, United States, 2007. 333-338. [3] GE Liang, LU Zheng-tian, YI Gu-wu, ZHOU Yu. Mass buffer in net work backup system [J]. Applicati on Research of Computers, China, 2009, 26(1):9-12. [4] WANG Y L, LI Z H, LIN W. A fast disaster recovery mechanism for volume replication systems [A]. The 3rd International Conference on High Performance Computing and Communications[C]. Houston, TX, United States, 2007. 732-743. [5] Peter M. Chen, Edward K. Lee, Garth A. Gibson, Randy H. Katz, David A. Patterson, RAID: highperformance, reliable secondary storage, ACM Computing Surveys (CSUR), v.26 n.2, p.145-185, June 1994. [6] Lloyd S J, Joan P, Jian L, et al. RORIB: An Economic and Efficient Solution for Real-time Online Remote Information Backup[J]. Journal of Database Management, 2003, 14(3): 56-73.