HDFS Fair Scheduler Preemption for Pending Request
|
|
- Marshall Watts
- 5 years ago
- Views:
Transcription
1 HDFS Fair Scheduler Preemption for Pending Request B.PURNACHANDRA RAO Research Scholor, Dept of Computer Science & Engg, ANU College of Engg & Technology,Guntur,India. Dr. N.NAGAMALLESWARA RAO Prof. Dept of Information Technology, R.V.R. & J.C. College of Engg& Technology,Guntur, India. ABSTRACT Hadoop is a popular open-source framework software framework for storing and processing large sets of data on a platform consisiting of commodity hardware. The main idea behind Hadoop is to move computation to the data, instead of the traditional way of moving data to computation. Hadoop file sizes are usually very large, ranging from gigabytes to terabytes,and large Hadoop clusters store millions of these files. Resource allocation among competing users, and scheduling jobs in Hadoop cluster is one of the key task in Hadoop internal oprations.hadoop is having three built in resource schedulers, such as the Fair scheduler, FIFO scheduler and the Capacity scheduler. First In First Out is a simple and early scheduler, and it uses a single queue for all jobs. There is no concept of priority in chosing a job for execution, with the oldest jobs getting chosen first from the head of the queue. The Fair scheduler enables multiple queue placement policies, which dictate where the scheduler places new applications. You can submit applications to a nonexistent queue by setting the create flag, which creates a new queue. The capacity scheduler chooses jobs with the highest gap between currently used and granted capacity, that is the most undeserved queues are offerred resources before other queues. The Fair scheduler selects jobs on the basis of the highest time deficit. The Fair scheduler uses preemption to support fairness among the queues and assigns priorities to users through weights. The capacity scheduler uses preemption to return guaranteed capacity back to the queues. Fair Scheduler is not having the capability of preemption for pending request. That means it doesn t consider what s the actual resource request, this causes when an under-utilized queue needs one single container with 60GB, scheduler could preemption 60 * 1GB containers on different nodes. Such preempted resources cannot be used by target resource request. This paper shows how to consider the actual metrics of the request before preempting resources. KEYWORDS Hadoop Distributed File System (HDFS), Schedulers, FIFO Scheduler, Fair Scheduler, Capacity Scheduler, Preemption, Preemption for Pending Request, NameNode, DataNode. 1 INTRODUCTION Apache Hadoop [1] is explicitly designed to handle large amounts of data, which can easily run into many petabytes and even exabytes. Hadoop file sizes are very usually very large,ranging from gigabytes to terabytes, and large Hadoop clusters store millions of these files. Everything in Hadoop is based on the assumption that hardware fails. Hadoop depends on large numbers of servers so it can parallelize work across them. Server and storage failures are to be expected, and the system is not affected by nonfunctioning storage units or even failed server. Traditional databases are geared mostly for fast access to data and not for both batch processing. Hadoop was originally designed for batch processing such as the indexing of millions of web pages and provides streaming access to data sets. Unlike traditional databases, Hadoop data files employ a write-once-read-many access model. Data consistency issues that may arise in an updatable database are not anissue with Hadoop file systems, because only a single writer can write to a file at any time. HDFS [2] [3] is designed for write-oncereadmany access model for files. Hadoop, MapReduce [5], Dryad [10] and HPCC (High- Performance Computing Cluster) [11] frameworks are Data-intensive and they are depending on the disk based file systems to meet their exponential storage demands. Hadoop is having the master nodes called NameNode and it is having namespace. Data will be stored in DataNodes where these are connected to Namenode and periodically sends status report to NameNode. Namenode maintains the blocks info where the data is available in datanode and empty blocks info. Each block data is stored in couple of locations, it means it is replicated to number of 119 B.PURNACHANDRA RAO, Dr. N.NAGAMALLESWARA RAO
2 blocks and the replication factor we have to configure in the hadoop configuration file. Hadoop distributed file system (HDFS) [6] has the capacity to store large amounts of data. Whenever client wants to store data to HDFS file system, client will talk to Namenode. Namenode will find out the metadata available in namespace to get the number of data blocks info which are vacant. The list of data blocks will be returned to HDFS client. Based on the list client will talk to DataNode to write the data[9]. Resource allocation refers to the allocation of scarce, finite computing resources, such as CPU time, memory, storage space and network bandwith, among the users that utikize a Hadoop cluster.the two more important resources that you have control over are processing power (CPU) and memory (RAM). Hadoop resource schedukers are components that are responsible for assigning tasks to available YARN containers on various DataNodes. The scheduler is a plugin with in Resource Manager. Hadoop is having three types of schedulers FIFO, Capacity scheduler and Fair Scheduler. FIFO is a simple early Hadoop scheduler, and it uses a single queue for all jobs. Capacity scheduler submits jobs into queues, each of which is guranteed a minimum amount of resources such as RAM and CPU. Fair scheduler assigns jobs to queues with guaranteed minimum resources. Preemption "When the cluster doesn t have enough idle resource, one queue is over-utilized and another queue is under-utilized, scheduler can preemption resource from the overutilized queue" is supported by both Capacity and Fair scheduler. But preemption for pending request is not supported by Fair scheduler. In this paper we will show that how we can utilize the concept of preemption for pending request in Fair scheduler. 2 LITERATURE REVIEW 2.1 HDFS SCHEDULERS Resource allocation is a crucial part of Hadoop Administration[4]. Hadoop resource schedulers are components that are responsible for assigning tasks to available YARN containers on various DataNodes. The scheduler is a plugin within the ResourceManager. First In First Out scheduler is a simple early Hadoop scheduler, and it uses a single queue for all jobs. There is no concept of priority in choosing a job for execution, with the oldest jobs first from the head of the queue. Capacity Scheduler submits the jobs to queues, each of which is guranteed a minimum amount of resources such as RAM and CPU. The queues with a greater gap between their used capacity and their granted resources are offerred priority in the allocation of new resources are released by completing jobs. If it has excess capacity, the scheduler shares it among the cluster users, just as the Fair scheduler does. It uses the concept of reservation and preemption to return the guaranteed capacity to the queus. Fair scheduler asigns jobs to queues with guranteed minimum resources. The scheduler picks up the jobs with the greatest time defecit for allocatiing resources that are freed by other applications. This scheduler can also allocates excess capacity from a pool to other pools. The Fair scheduler uses the concept of priority to support the importance of an application within a pool. It uses the concept of preemption to support fairness among different resource pools. Please refer Fig 1 for configuring Fair Scheduler. Need to use the value as org.apacahe. hadoop.yarn.server.resource.resourcemanager.schedu ler.capacity.capacityscheduler to utilize CapacityScheduler. Fig 1: Fair Scheduler Configuration The essential concept of CapacityScheduler is it uses dedicated queues to which you assign jobs. Each queue has a predetermined amount of resources allocated to it. However you pay in terms of the clusters resource utilization, since you are reserving and guaranteeing queue resource capacities. The goal of the capacity scheduler is to enable multiple tenants (users) of an organization to share the resources of a Hadoop cluster in a predictable fashion. Hadoop achieves this goal by using job queues. The scheduler provides guaranteed capacity of the job queues, while providing elasticity for the utilization of the cluster by the queues. Elasticity in this context means that the assignment of the resources is not set in concrete. As the queues wend their way through the cluster, it is common for some queues to be overloaded and for some others to be relatively idle. The capacity scheduler realizes this and automatically transfers the unused capacity of the lighlty used queues to the overloaded queues. A queue is an ordered list of jobs. A queue is allocated a certain portion of your clusters resources. When you create a queue, you allocate it a certain portion 120 B.PURNACHANDRA RAO, Dr. N.NAGAMALLESWARA RAO
3 of your clusters resources. User applications will then be submitted to this queue to access the queue's allotted resources. Applications submitted to a queue will run in a FIFO order. Once the apllications submitted to queue start running, they cant be preempted, but as the tasks complete, any free resources will be assigned to queues running below the capacity allowed to them. Please refer Fig 3 for creation of queues. Fig 3: Create queues NAMENODE HDFS is having master node called Namenode having the namespace which maintains metadata regarding datanode blocks info. It will manage the user access to file system. It will process the block reports sent by the DataNodes and maintaining the location where data blocks live. The NameNode also updates the fsimage file with the updated HDFS state information[12]. Once client receives set of datanodes the data will be written to datanodes in pipeline fashion [7,8] DATANODE HDFS is having slave nodes called DataNode. It provides the block storage by storing blocks on the local file system. Fulfilling the read/write requests from the clients who want to work with the data stored on the DataNodes. Create and delete of the data DataNodes will keep in touch with the NameNode by sending periodic block reports and heartbeats. A heartbeat confirms the DataNode is alive and healthy, and a block report shows the blocks being managed by the DataNode. Please refer Fig 4 for HDFS architecture. Fig 4: HDFS Architecture FAIR SCHEDULER The Fair Scheduler is a built-in Hadoop resource scheduler whose goal is to let smaller jobs finish fast (short response times) and provide a guaranteed service level for production jobs. The jobs are not usually of same type, some are production jobs that involve data imports and hourly reports. Some other jobs are run by the data analysis who are running adhoc Hive queries and Pig jobs. Usually there are some long-running data analyses or machine learning jobs running at the same time. The main task here is how to allocate the resources of the cluster in an efficient manner among these competing jobs. You dont need to reserve a predefined amount of capacity to groups or queues. The scheduler dynamically distributes the available resources among all the running jobs in a cluster. Please refer Fig 5 for FairScheduler. 121 B.PURNACHANDRA RAO, Dr. N.NAGAMALLESWARA RAO
4 Fig 5: FairScheduler when a large job starts first, and it happens to be the only job running, it starts using all the clusters resources by default (unless you specify maximum resource limits). Subsequently when a second job starts up, it is allocated roughly half of the total cluster resources(by default), now both jobs share the cluster resources on an equal basis. This is the concept of fairness that led to naming this scheduler the Fair Scheduler. The Fair scheduler ensures the resource allocation for application is fair meaning that all applications get roughly equal amounts of resources over time. When we talk about resources in the context of the Fair Scheduler, I am referring to memory only. However, you can also employ a variation of the Fair Scheduler called the Dominant Resource Fairness DRF scheduler, which uses both memory and CPU as resources. Dominant Resource Fairness is a concept wherein the YARN schedulers examine each user's dominant resource and use it as a measure of the resource usage by that user PREEMPTING APPLICATIONS Preempting an application means containers from other applications may need to be killed if necessary, to make room for the new applications. If you dont want late arriving applications to a specific leaf queue to wait because the running applications in other leaf queues are taking up all the allotted resources, you can use preemption. Under these situations although you have guaranteed a set capacity for a queue, there are no free resources available to be allotted to this leaf queue. The application Master container is killed only as a last resort, with preference being given to killing containers that have not been executed yet. Minimum Sharing preemption is when a pool is operating below its configured minimum share and Fair share preemption is when a pool is operating below its fair share. Of these two minimum share preemption is stricter and kicks in immediately when a pool, starts operating below its minimum allocated share for a specific period, before job preemption begins. Once preemption starts a pool thats currently below its minimum allocated share can go up to its minimum share, where as a pool that is now below 50 percent of its fair share will go all the way up until it hits full fair share. You can configure task preemption to ensure that key jobs are processed on time. However, preemption is not arbitrary - it's used to kill containers for queues that are using more than their fair share of resources. If you enable preemption in the cluster the fair scheduler will preempt applications in other queues if a queue's minimum share is not met for some period of time. Preemption ensures that your key production jobs are not delayed because other less important jobs are already running in the cluster. The Fair scheduler kills the most recently launched application to minimize the waste of resources in the cluster. To enable preemption, set the yarn.scheduler.fair.preemption property to true in the yarn-site.xml file. Both the Fair scheduler and the Capacity Scheduler have an identical goal : Allow long running jobs to complete in a decent time while simultaneously enabling users running queries to get their results back quickly. Both schedulers support hierarchical queues. All queues descend from a root or default queue. You can submit applications to leaf queues. Both queues support minimum and maximum capacities. Both queues support maximum application limits on a per-queue basis. Both schedulers let you move applications across queues. The Fair Scheduler contains scheduling policies that determine which jobs get resources each time the scheduler allocates resources. You can use the three types of sheduling policies -fifo, fair and drf by specifying the policy with the defaultqueueschedulingpolicy top-level element. The Capacity Scheduler on the other hand always schedules jobs within each queue with the FIFO principle. The Fair scheduker enables multiple queue placement policies., which dictate where the scheduker places new applications among the queues based on users, groups, or the queue requests made by the applications. You can submit applications to 122 B.PURNACHANDRA RAO, Dr. N.NAGAMALLESWARA RAO
5 non existent queue by setting the create flag so that it will create new queue. Preemption is a way to balance resource usage between queues: When the cluster doesn t have enough idle resource, one queue is over-utilized and another queue is under-utilized, scheduler can preemption resource from the overutilized queue. Preemption for pending request means it doesn t consider what s the actual resource request, this causes when an under-utilized queue needs one single container with 60GB, scheduler could preemption 60 * 1GB containers on different nodes. Such preempted resources cannot be used by target resource request. Capacity Scheduler is having the capability of preemption for pending request. But Fair scheduler is not having this capability. If it requests for resources using preemption policy for the pending request, it is not guarantee that It will utilize the allotted resource. This is the problem in the existing architecture. 3 PROPOSED METHOD 3.1 PROBLEM STATEMENT HDFS Fair scheduler is not supporting Preemption for pending request. It doesn t consider what s the actual resource request, this causes when an underutilized queue needs one single container with 60GB, scheduler could preemption 60 * 1GB containers on different nodes. Such preempted resources cannot be used by target resource request. If it requests for resources using preemption policy for the pending request, it is not guarantee that It will utilize all the allotted resources. This is the problem in the existing architecture. 3.2 PROPOSAL When there is a resource request Fair Scheduler needs to consider the resource request by any means so that It can request/preempt for exact need. The solution is Fair Scheduler needs to look for the 60GB space at only one location instead of getting at different locations/nodes. So we need to implement the Best fit memory allocation algorithm in Fair Scheduling process. Input memory blocks and processes with sizes. Initialize all memory blocks as free. Start by picking each process and find the minimum block size that can be assigned to current process i.e, find min(blocksize[1], blocksize[2],...blocksize[n]) > processsize[current], if found then assign it to the current process. If not then leave that process and keep checking the further processes. As per this allocation The best fit deals with allocating the smallest free partition which meets the requirement of the requesting process. This algorithm first searches the entire list of free partitions and considers the smallest hole that is adequate. It then tries to find a hole which is close to actual process size needed. As we discussed it is allocating 1GB resources at 60 different nodes, but if we implement this it will look for smallest free partition which meets the requirement. If it is not there it will preempt for expected resource. Please Refer Fig 6 for Best fit memory allocation, where P3 is placed at smallest affordable place in best fit, where as in worstfit it will be placed at highest capacity block, where as in fitrst fit it will look for first capable slot. Fig 6: Best Fit Memory allocation 4 IMPLEMENTATION Refer Fig 7 for the Implementation architecture using Best Fit allocation algorithm. Whenever pre emption is required process will look for exact fit and allocate the same. Fair scheduler uses best fit allocation process as shown in the figure before opting for the block to write the data. The extra complexity includes searching time required to get the exact fit. 123 B.PURNACHANDRA RAO, Dr. N.NAGAMALLESWARA RAO
6 entire list of free partitions and considers the smallest hole that is adequate. It then tries to find a hole which is close to actual process size needed. As we discussed it is allocating 1GB resources at 60 different nodes, but if we implement this it will look for smallest free partition which meets the requirement. If it is not there it will not allocate small resources, so that we can utilize these small resources for another applications. This will result in least wasted space. Internal fragmentation reduced but not eliminated. Only disadvantage in best fit algorithm is searching time increases for finding the exact fit. The future work includes finding the procedure for reducing the time to get the exact fit. Fig 7: Fair Scheduler with Best fit allocation 5 CONCLUSION Preempting an application means containers from other applications may need to be killed if necessary, to make room for the new applications. If you dont want late arriving applications to a specific leaf queue to wait because the running applications in other leaf queues are taking up all the allotted resources, you can use preemption. Preemption is a way to balance resource usage between queues: When the cluster doesn t have enough idle resource, one queue is over-utilized and another queue is under-utilized, scheduler can preemption resource from the over-utilized queue. Preemption for pending request means it doesn t consider what s the actual resource request, this causes when an under-utilized queue needs one single container with 60GB, scheduler could preemption 60 * 1GB containers on different nodes. Such preempted resources cannot be used by target resource request. Capacity Scheduler is having the capability of preemption for pending request. But Fair scheduler is not having this capability. The solution is Fair Scheduler needs to look for the 60GB sapce at only one location instead of getting at different locations/nodes. So we need to implement the Best fit memory allocation algorithm in Fair Scheduling process. As per this allocation The best fit deals with allocating the smallest free partition which meets the requirement of the requesting process. This algorithm first searches the REFERENCES [1] Apache Hadoop. Available at Hadoop Apache. [2] Tom White, "Hadoop:The Definitive Guide", Storage and Analysis at Internet Scale, Second ed., Yahoo Press, [3] Konstantin V. Shvachko, Scalability of Hadoop Distributed File System. [4] George Porter. Decoupling storage and computation in Hadoop with SuperDataNodes, ACM SIGOPS Operating System Review, 44, [5] Hadoop Distributed File System with Cache system - a paradigm for performnace improvement by Archana Kakade and Dr. SuhasRaut, International journal of scientific research and management (IJSRM), Vol.2,Issue.1: Pp, /Aug [6] J. Dean and S. Ghemawat (2004), Mapreduce: Simplified Data Processing on Large Clusters. In Proceeding of the 6th Conference on Symposium on operating Systems Design and Implementation (OSDI 04), Berkeley, CA, USA, 2004, pp [7] Shafer J, Rixner S, Cox AL. The Hadoop Distributed Filesystem: Balancing Portability and Performance, in Proceedings of IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS 2010), White Plains, NY, [8] Feng Wang, Jie Qiu, Jie Yang, Bo Dong, Xinhui Li, Ying Li, " Hadoop High Availability through Metadata Replication", IBM China Research Laboratory, ACM, pp 37-44,2009. [9] Derek Tankel. Scalability of Hadoop Distributed File System, Yahoo developer work, [10] The Case for RAMClouds: Scalable High- Performance Storage Entirely in DRAM Department of ComputerScience Stanford University. [11] J. Shafer and S Rixner (2010), "The Hadoop distributed file system: balancing portability and performance, In 2010 IEEE International Symposium on Performance Analysis of System and Software (ISPASS2010), White Plains, NY, March Pp B.PURNACHANDRA RAO, Dr. N.NAGAMALLESWARA RAO
7 [12] SAM R. ALAPATI, Expert Hadoop Administration, Managing, Tuning and Securing Spark, YARN and HDFS, Addison wesley data & analytics series, B.PURNACHANDRA RAO, Dr. N.NAGAMALLESWARA RAO
HDFS Pipeline Reconstruction to Avoid Data Loss
HDFS Pipeline Reconstruction to Avoid Data Loss Purnachandra Rao. B Department of Computer Science & Engineering ANU College of Engg & Technology, Guntur, India. pcr.bobbepalli@gmail.com Nagamalleswara
More informationA SURVEY ON SCHEDULING IN HADOOP FOR BIGDATA PROCESSING
Journal homepage: www.mjret.in ISSN:2348-6953 A SURVEY ON SCHEDULING IN HADOOP FOR BIGDATA PROCESSING Bhavsar Nikhil, Bhavsar Riddhikesh,Patil Balu,Tad Mukesh Department of Computer Engineering JSPM s
More informationCCA-410. Cloudera. Cloudera Certified Administrator for Apache Hadoop (CCAH)
Cloudera CCA-410 Cloudera Certified Administrator for Apache Hadoop (CCAH) Download Full Version : http://killexams.com/pass4sure/exam-detail/cca-410 Reference: CONFIGURATION PARAMETERS DFS.BLOCK.SIZE
More informationHDFS CACHE PERFORMANCE USING SET ASSOCIATIVE CACHE MEMORY
HDFS CACHE PERFORMANCE USING SET ASSOCIATIVE CACHE MEMORY 1 B.PURNACHANDRA RAO, 2 Dr.N.NAGAMALLESWARA RAO 1 Research Scholor, Dept of Computer Science & Engg,ANU College of Engg & Technology,Guntur,India.
More informationDistributed Systems 16. Distributed File Systems II
Distributed Systems 16. Distributed File Systems II Paul Krzyzanowski pxk@cs.rutgers.edu 1 Review NFS RPC-based access AFS Long-term caching CODA Read/write replication & disconnected operation DFS AFS
More informationLocality Aware Fair Scheduling for Hammr
Locality Aware Fair Scheduling for Hammr Li Jin January 12, 2012 Abstract Hammr is a distributed execution engine for data parallel applications modeled after Dryad. In this report, we present a locality
More informationHadoop File System S L I D E S M O D I F I E D F R O M P R E S E N T A T I O N B Y B. R A M A M U R T H Y 11/15/2017
Hadoop File System 1 S L I D E S M O D I F I E D F R O M P R E S E N T A T I O N B Y B. R A M A M U R T H Y Moving Computation is Cheaper than Moving Data Motivation: Big Data! What is BigData? - Google
More informationMI-PDB, MIE-PDB: Advanced Database Systems
MI-PDB, MIE-PDB: Advanced Database Systems http://www.ksi.mff.cuni.cz/~svoboda/courses/2015-2-mie-pdb/ Lecture 10: MapReduce, Hadoop 26. 4. 2016 Lecturer: Martin Svoboda svoboda@ksi.mff.cuni.cz Author:
More informationGoogle File System (GFS) and Hadoop Distributed File System (HDFS)
Google File System (GFS) and Hadoop Distributed File System (HDFS) 1 Hadoop: Architectural Design Principles Linear scalability More nodes can do more work within the same time Linear on data size, linear
More informationThe Google File System. Alexandru Costan
1 The Google File System Alexandru Costan Actions on Big Data 2 Storage Analysis Acquisition Handling the data stream Data structured unstructured semi-structured Results Transactions Outline File systems
More informationHadoop Virtualization Extensions on VMware vsphere 5 T E C H N I C A L W H I T E P A P E R
Hadoop Virtualization Extensions on VMware vsphere 5 T E C H N I C A L W H I T E P A P E R Table of Contents Introduction... 3 Topology Awareness in Hadoop... 3 Virtual Hadoop... 4 HVE Solution... 5 Architecture...
More informationExam Questions CCA-500
Exam Questions CCA-500 Cloudera Certified Administrator for Apache Hadoop (CCAH) https://www.2passeasy.com/dumps/cca-500/ Question No : 1 Your cluster s mapred-start.xml includes the following parameters
More informationCloud Computing and Hadoop Distributed File System. UCSB CS170, Spring 2018
Cloud Computing and Hadoop Distributed File System UCSB CS70, Spring 08 Cluster Computing Motivations Large-scale data processing on clusters Scan 000 TB on node @ 00 MB/s = days Scan on 000-node cluster
More informationTop 25 Big Data Interview Questions And Answers
Top 25 Big Data Interview Questions And Answers By: Neeru Jain - Big Data The era of big data has just begun. With more companies inclined towards big data to run their operations, the demand for talent
More informationLecture 11 Hadoop & Spark
Lecture 11 Hadoop & Spark Dr. Wilson Rivera ICOM 6025: High Performance Computing Electrical and Computer Engineering Department University of Puerto Rico Outline Distributed File Systems Hadoop Ecosystem
More informationInternational Journal of Advance Engineering and Research Development. A Study: Hadoop Framework
Scientific Journal of Impact Factor (SJIF): e-issn (O): 2348- International Journal of Advance Engineering and Research Development Volume 3, Issue 2, February -2016 A Study: Hadoop Framework Devateja
More informationBig Data 7. Resource Management
Ghislain Fourny Big Data 7. Resource Management artjazz / 123RF Stock Photo Data Technology Stack User interfaces Querying Data stores Indexing Processing Validation Data models Syntax Encoding Storage
More informationL5-6:Runtime Platforms Hadoop and HDFS
Indian Institute of Science Bangalore, India भ रत य व ज ञ न स स थ न ब गल र, भ रत Department of Computational and Data Sciences SE256:Jan16 (2:1) L5-6:Runtime Platforms Hadoop and HDFS Yogesh Simmhan 03/
More informationHadoop and HDFS Overview. Madhu Ankam
Hadoop and HDFS Overview Madhu Ankam Why Hadoop We are gathering more data than ever Examples of data : Server logs Web logs Financial transactions Analytics Emails and text messages Social media like
More informationCCA Administrator Exam (CCA131)
CCA Administrator Exam (CCA131) Cloudera CCA-500 Dumps Available Here at: /cloudera-exam/cca-500-dumps.html Enrolling now you will get access to 60 questions in a unique set of CCA- 500 dumps Question
More informationCLIENT DATA NODE NAME NODE
Volume 6, Issue 12, December 2016 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Efficiency
More informationTITLE: PRE-REQUISITE THEORY. 1. Introduction to Hadoop. 2. Cluster. Implement sort algorithm and run it using HADOOP
TITLE: Implement sort algorithm and run it using HADOOP PRE-REQUISITE Preliminary knowledge of clusters and overview of Hadoop and its basic functionality. THEORY 1. Introduction to Hadoop The Apache Hadoop
More informationPaaS and Hadoop. Dr. Laiping Zhao ( 赵来平 ) School of Computer Software, Tianjin University
PaaS and Hadoop Dr. Laiping Zhao ( 赵来平 ) School of Computer Software, Tianjin University laiping@tju.edu.cn 1 Outline PaaS Hadoop: HDFS and Mapreduce YARN Single-Processor Scheduling Hadoop Scheduling
More informationYuval Carmel Tel-Aviv University "Advanced Topics in Storage Systems" - Spring 2013
Yuval Carmel Tel-Aviv University "Advanced Topics in About & Keywords Motivation & Purpose Assumptions Architecture overview & Comparison Measurements How does it fit in? The Future 2 About & Keywords
More informationIntroduction to Hadoop. Owen O Malley Yahoo!, Grid Team
Introduction to Hadoop Owen O Malley Yahoo!, Grid Team owen@yahoo-inc.com Who Am I? Yahoo! Architect on Hadoop Map/Reduce Design, review, and implement features in Hadoop Working on Hadoop full time since
More informationDynamic processing slots scheduling for I/O intensive jobs of Hadoop MapReduce
Dynamic processing slots scheduling for I/O intensive jobs of Hadoop MapReduce Shiori KURAZUMI, Tomoaki TSUMURA, Shoichi SAITO and Hiroshi MATSUO Nagoya Institute of Technology Gokiso, Showa, Nagoya, Aichi,
More informationMapReduce. U of Toronto, 2014
MapReduce U of Toronto, 2014 http://www.google.org/flutrends/ca/ (2012) Average Searches Per Day: 5,134,000,000 2 Motivation Process lots of data Google processed about 24 petabytes of data per day in
More information18-hdfs-gfs.txt Thu Oct 27 10:05: Notes on Parallel File Systems: HDFS & GFS , Fall 2011 Carnegie Mellon University Randal E.
18-hdfs-gfs.txt Thu Oct 27 10:05:07 2011 1 Notes on Parallel File Systems: HDFS & GFS 15-440, Fall 2011 Carnegie Mellon University Randal E. Bryant References: Ghemawat, Gobioff, Leung, "The Google File
More informationBig Data for Engineers Spring Resource Management
Ghislain Fourny Big Data for Engineers Spring 2018 7. Resource Management artjazz / 123RF Stock Photo Data Technology Stack User interfaces Querying Data stores Indexing Processing Validation Data models
More informationHADOOP FRAMEWORK FOR BIG DATA
HADOOP FRAMEWORK FOR BIG DATA Mr K. Srinivas Babu 1,Dr K. Rameshwaraiah 2 1 Research Scholar S V University, Tirupathi 2 Professor and Head NNRESGI, Hyderabad Abstract - Data has to be stored for further
More informationKey aspects of cloud computing. Towards fuller utilization. Two main sources of resource demand. Cluster Scheduling
Key aspects of cloud computing Cluster Scheduling 1. Illusion of infinite computing resources available on demand, eliminating need for up-front provisioning. The elimination of an up-front commitment
More informationHDFS Architecture. Gregory Kesden, CSE-291 (Storage Systems) Fall 2017
HDFS Architecture Gregory Kesden, CSE-291 (Storage Systems) Fall 2017 Based Upon: http://hadoop.apache.org/docs/r3.0.0-alpha1/hadoopproject-dist/hadoop-hdfs/hdfsdesign.html Assumptions At scale, hardware
More informationKonstantin Shvachko, Hairong Kuang, Sanjay Radia, Robert Chansler Yahoo! Sunnyvale, California USA {Shv, Hairong, SRadia,
Konstantin Shvachko, Hairong Kuang, Sanjay Radia, Robert Chansler Yahoo! Sunnyvale, California USA {Shv, Hairong, SRadia, Chansler}@Yahoo-Inc.com Presenter: Alex Hu } Introduction } Architecture } File
More informationHadoop 2.x Core: YARN, Tez, and Spark. Hortonworks Inc All Rights Reserved
Hadoop 2.x Core: YARN, Tez, and Spark YARN Hadoop Machine Types top-of-rack switches core switch client machines have client-side software used to access a cluster to process data master nodes run Hadoop
More informationHDFS Federation. Sanjay Radia Founder and Hortonworks. Page 1
HDFS Federation Sanjay Radia Founder and Architect @ Hortonworks Page 1 About Me Apache Hadoop Committer and Member of Hadoop PMC Architect of core-hadoop @ Yahoo - Focusing on HDFS, MapReduce scheduler,
More informationEfficient Map Reduce Model with Hadoop Framework for Data Processing
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 4, April 2015,
More informationKey aspects of cloud computing. Towards fuller utilization. Two main sources of resource demand. Cluster Scheduling
Key aspects of cloud computing Cluster Scheduling 1. Illusion of infinite computing resources available on demand, eliminating need for up-front provisioning. The elimination of an up-front commitment
More informationSurvey on MapReduce Scheduling Algorithms
Survey on MapReduce Scheduling Algorithms Liya Thomas, Mtech Student, Department of CSE, SCTCE,TVM Syama R, Assistant Professor Department of CSE, SCTCE,TVM ABSTRACT MapReduce is a programming model used
More informationHadoop MapReduce Framework
Hadoop MapReduce Framework Contents Hadoop MapReduce Framework Architecture Interaction Diagram of MapReduce Framework (Hadoop 1.0) Interaction Diagram of MapReduce Framework (Hadoop 2.0) Hadoop MapReduce
More informationA BigData Tour HDFS, Ceph and MapReduce
A BigData Tour HDFS, Ceph and MapReduce These slides are possible thanks to these sources Jonathan Drusi - SCInet Toronto Hadoop Tutorial, Amir Payberah - Course in Data Intensive Computing SICS; Yahoo!
More informationA priority based dynamic bandwidth scheduling in SDN networks 1
Acta Technica 62 No. 2A/2017, 445 454 c 2017 Institute of Thermomechanics CAS, v.v.i. A priority based dynamic bandwidth scheduling in SDN networks 1 Zun Wang 2 Abstract. In order to solve the problems
More informationIntroduction to MapReduce
Basics of Cloud Computing Lecture 4 Introduction to MapReduce Satish Srirama Some material adapted from slides by Jimmy Lin, Christophe Bisciglia, Aaron Kimball, & Sierra Michels-Slettvet, Google Distributed
More informationThe Google File System
October 13, 2010 Based on: S. Ghemawat, H. Gobioff, and S.-T. Leung: The Google file system, in Proceedings ACM SOSP 2003, Lake George, NY, USA, October 2003. 1 Assumptions Interface Architecture Single
More informationCLOUD-SCALE FILE SYSTEMS
Data Management in the Cloud CLOUD-SCALE FILE SYSTEMS 92 Google File System (GFS) Designing a file system for the Cloud design assumptions design choices Architecture GFS Master GFS Chunkservers GFS Clients
More information50 Must Read Hadoop Interview Questions & Answers
50 Must Read Hadoop Interview Questions & Answers Whizlabs Dec 29th, 2017 Big Data Are you planning to land a job with big data and data analytics? Are you worried about cracking the Hadoop job interview?
More informationHigh Performance Computing on MapReduce Programming Framework
International Journal of Private Cloud Computing Environment and Management Vol. 2, No. 1, (2015), pp. 27-32 http://dx.doi.org/10.21742/ijpccem.2015.2.1.04 High Performance Computing on MapReduce Programming
More informationBatch Inherence of Map Reduce Framework
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 6, June 2015, pg.287
More informationIntroduction to MapReduce
Basics of Cloud Computing Lecture 4 Introduction to MapReduce Satish Srirama Some material adapted from slides by Jimmy Lin, Christophe Bisciglia, Aaron Kimball, & Sierra Michels-Slettvet, Google Distributed
More informationBigData and Map Reduce VITMAC03
BigData and Map Reduce VITMAC03 1 Motivation Process lots of data Google processed about 24 petabytes of data per day in 2009. A single machine cannot serve all the data You need a distributed system to
More informationOptimizing Hadoop Block Placement Policy & Cluster Blocks Distribution
Vol:6, No:1, 212 Optimizing Hadoop Block Placement Policy & Cluster Blocks Distribution Nchimbi Edward Pius, Liu Qin, Fion Yang, Zhu Hong Ming International Science Index, Computer and Information Engineering
More informationHadoop Job Scheduling with Dynamic Task Splitting
Hadoop Job Scheduling with Dynamic Task Splitting Xu YongLiang School of Computer Engineering 2015 Hadoop Job Scheduling with Dynamic Task Splitting Xu YongLiang (G1002570K) Supervisor Professor Cai Wentong
More informationA Micro Partitioning Technique in MapReduce for Massive Data Analysis
A Micro Partitioning Technique in MapReduce for Massive Data Analysis Nandhini.C, Premadevi.P PG Scholar, Dept. of CSE, Angel College of Engg and Tech, Tiruppur, Tamil Nadu Assistant Professor, Dept. of
More information18-hdfs-gfs.txt Thu Nov 01 09:53: Notes on Parallel File Systems: HDFS & GFS , Fall 2012 Carnegie Mellon University Randal E.
18-hdfs-gfs.txt Thu Nov 01 09:53:32 2012 1 Notes on Parallel File Systems: HDFS & GFS 15-440, Fall 2012 Carnegie Mellon University Randal E. Bryant References: Ghemawat, Gobioff, Leung, "The Google File
More informationPerformance evaluation of job schedulers on Hadoop YARN
Performance evaluation of job schedulers on Hadoop YARN Jia-Chun Lin Department of Informatics, University of Oslo Gaustadallèen 23 B, Oslo, N-0373, Norway kellylin@ifi.uio.no Ming-Chang Lee Department
More informationA brief history on Hadoop
Hadoop Basics A brief history on Hadoop 2003 - Google launches project Nutch to handle billions of searches and indexing millions of web pages. Oct 2003 - Google releases papers with GFS (Google File System)
More informationData Clustering on the Parallel Hadoop MapReduce Model. Dimitrios Verraros
Data Clustering on the Parallel Hadoop MapReduce Model Dimitrios Verraros Overview The purpose of this thesis is to implement and benchmark the performance of a parallel K- means clustering algorithm on
More informationCS60021: Scalable Data Mining. Sourangshu Bhattacharya
CS60021: Scalable Data Mining Sourangshu Bhattacharya In this Lecture: Outline: HDFS Motivation HDFS User commands HDFS System architecture HDFS Implementation details Sourangshu Bhattacharya Computer
More informationEnhanced Hadoop with Search and MapReduce Concurrency Optimization
Volume 114 No. 12 2017, 323-331 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu ijpam.eu Enhanced Hadoop with Search and MapReduce Concurrency Optimization
More informationEFFICIENT ALLOCATION OF DYNAMIC RESOURCES IN A CLOUD
EFFICIENT ALLOCATION OF DYNAMIC RESOURCES IN A CLOUD S.THIRUNAVUKKARASU 1, DR.K.P.KALIYAMURTHIE 2 Assistant Professor, Dept of IT, Bharath University, Chennai-73 1 Professor& Head, Dept of IT, Bharath
More informationAutomation of Rolling Upgrade for Hadoop Cluster without Data Loss and Job Failures. Hiroshi Yamaguchi & Hiroyuki Adachi
Automation of Rolling Upgrade for Hadoop Cluster without Data Loss and Job Failures Hiroshi Yamaguchi & Hiroyuki Adachi About Us 2 Hiroshi Yamaguchi Hiroyuki Adachi Hadoop DevOps Engineer Hadoop Engineer
More informationA Review Paper on Big data & Hadoop
A Review Paper on Big data & Hadoop Rupali Jagadale MCA Department, Modern College of Engg. Modern College of Engginering Pune,India rupalijagadale02@gmail.com Pratibha Adkar MCA Department, Modern College
More informationDynamic Data Placement Strategy in MapReduce-styled Data Processing Platform Hua-Ci WANG 1,a,*, Cai CHEN 2,b,*, Yi LIANG 3,c
2016 Joint International Conference on Service Science, Management and Engineering (SSME 2016) and International Conference on Information Science and Technology (IST 2016) ISBN: 978-1-60595-379-3 Dynamic
More informationReal-time Scheduling of Skewed MapReduce Jobs in Heterogeneous Environments
Real-time Scheduling of Skewed MapReduce Jobs in Heterogeneous Environments Nikos Zacheilas, Vana Kalogeraki Department of Informatics Athens University of Economics and Business 1 Big Data era has arrived!
More informationHadoop-PR Hortonworks Certified Apache Hadoop 2.0 Developer (Pig and Hive Developer)
Hortonworks Hadoop-PR000007 Hortonworks Certified Apache Hadoop 2.0 Developer (Pig and Hive Developer) http://killexams.com/pass4sure/exam-detail/hadoop-pr000007 QUESTION: 99 Which one of the following
More informationBig Data Programming: an Introduction. Spring 2015, X. Zhang Fordham Univ.
Big Data Programming: an Introduction Spring 2015, X. Zhang Fordham Univ. Outline What the course is about? scope Introduction to big data programming Opportunity and challenge of big data Origin of Hadoop
More informationThe Analysis Research of Hierarchical Storage System Based on Hadoop Framework Yan LIU 1, a, Tianjian ZHENG 1, Mingjiang LI 1, Jinpeng YUAN 1
International Conference on Intelligent Systems Research and Mechatronics Engineering (ISRME 2015) The Analysis Research of Hierarchical Storage System Based on Hadoop Framework Yan LIU 1, a, Tianjian
More informationEXTRACT DATA IN LARGE DATABASE WITH HADOOP
International Journal of Advances in Engineering & Scientific Research (IJAESR) ISSN: 2349 3607 (Online), ISSN: 2349 4824 (Print) Download Full paper from : http://www.arseam.com/content/volume-1-issue-7-nov-2014-0
More informationMixApart: Decoupled Analytics for Shared Storage Systems. Madalin Mihailescu, Gokul Soundararajan, Cristiana Amza University of Toronto and NetApp
MixApart: Decoupled Analytics for Shared Storage Systems Madalin Mihailescu, Gokul Soundararajan, Cristiana Amza University of Toronto and NetApp Hadoop Pig, Hive Hadoop + Enterprise storage?! Shared storage
More informationCorrelation based File Prefetching Approach for Hadoop
IEEE 2nd International Conference on Cloud Computing Technology and Science Correlation based File Prefetching Approach for Hadoop Bo Dong 1, Xiao Zhong 2, Qinghua Zheng 1, Lirong Jian 2, Jian Liu 1, Jie
More informationA Survey on Big Data
A Survey on Big Data D.Prudhvi 1, D.Jaswitha 2, B. Mounika 3, Monika Bagal 4 1 2 3 4 B.Tech Final Year, CSE, Dadi Institute of Engineering & Technology,Andhra Pradesh,INDIA ---------------------------------------------------------------------***---------------------------------------------------------------------
More informationHigh Throughput WAN Data Transfer with Hadoop-based Storage
High Throughput WAN Data Transfer with Hadoop-based Storage A Amin 2, B Bockelman 4, J Letts 1, T Levshina 3, T Martin 1, H Pi 1, I Sfiligoi 1, M Thomas 2, F Wuerthwein 1 1 University of California, San
More informationPLATFORM AND SOFTWARE AS A SERVICE THE MAPREDUCE PROGRAMMING MODEL AND IMPLEMENTATIONS
PLATFORM AND SOFTWARE AS A SERVICE THE MAPREDUCE PROGRAMMING MODEL AND IMPLEMENTATIONS By HAI JIN, SHADI IBRAHIM, LI QI, HAIJUN CAO, SONG WU and XUANHUA SHI Prepared by: Dr. Faramarz Safi Islamic Azad
More informationCS370 Operating Systems
CS370 Operating Systems Colorado State University Yashwant K Malaiya Fall 2017 Lecture 26 File Systems Slides based on Text by Silberschatz, Galvin, Gagne Various sources 1 1 FAQ Cylinders: all the platters?
More informationImproved MapReduce k-means Clustering Algorithm with Combiner
2014 UKSim-AMSS 16th International Conference on Computer Modelling and Simulation Improved MapReduce k-means Clustering Algorithm with Combiner Prajesh P Anchalia Department Of Computer Science and Engineering
More informationAn Enhanced Approach for Resource Management Optimization in Hadoop
An Enhanced Approach for Resource Management Optimization in Hadoop R. Sandeep Raj 1, G. Prabhakar Raju 2 1 MTech Student, Department of CSE, Anurag Group of Institutions, India 2 Associate Professor,
More informationParallel HITS Algorithm Implemented Using HADOOP GIRAPH Framework to resolve Big Data Problem
I J C T A, 9(41) 2016, pp. 1235-1239 International Science Press Parallel HITS Algorithm Implemented Using HADOOP GIRAPH Framework to resolve Big Data Problem Hema Dubey *, Nilay Khare *, Alind Khare **
More informationTopics. Big Data Analytics What is and Why Hadoop? Comparison to other technologies Hadoop architecture Hadoop ecosystem Hadoop usage examples
Hadoop Introduction 1 Topics Big Data Analytics What is and Why Hadoop? Comparison to other technologies Hadoop architecture Hadoop ecosystem Hadoop usage examples 2 Big Data Analytics What is Big Data?
More informationDistributed Systems. 15. Distributed File Systems. Paul Krzyzanowski. Rutgers University. Fall 2016
Distributed Systems 15. Distributed File Systems Paul Krzyzanowski Rutgers University Fall 2016 1 Google Chubby 2 Chubby Distributed lock service + simple fault-tolerant file system Interfaces File access
More informationDatabase Applications (15-415)
Database Applications (15-415) Hadoop Lecture 24, April 23, 2014 Mohammad Hammoud Today Last Session: NoSQL databases Today s Session: Hadoop = HDFS + MapReduce Announcements: Final Exam is on Sunday April
More informationHuge Data Analysis and Processing Platform based on Hadoop Yuanbin LI1, a, Rong CHEN2
2nd International Conference on Materials Science, Machinery and Energy Engineering (MSMEE 2017) Huge Data Analysis and Processing Platform based on Hadoop Yuanbin LI1, a, Rong CHEN2 1 Information Engineering
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung SOSP 2003 presented by Kun Suo Outline GFS Background, Concepts and Key words Example of GFS Operations Some optimizations in
More informationUNIT-IV HDFS. Ms. Selva Mary. G
UNIT-IV HDFS HDFS ARCHITECTURE Dataset partition across a number of separate machines Hadoop Distributed File system The Design of HDFS HDFS is a file system designed for storing very large files with
More informationThe Google File System
The Google File System Sanjay Ghemawat, Howard Gobioff and Shun Tak Leung Google* Shivesh Kumar Sharma fl4164@wayne.edu Fall 2015 004395771 Overview Google file system is a scalable distributed file system
More informationCPSC 426/526. Cloud Computing. Ennan Zhai. Computer Science Department Yale University
CPSC 426/526 Cloud Computing Ennan Zhai Computer Science Department Yale University Recall: Lec-7 In the lec-7, I talked about: - P2P vs Enterprise control - Firewall - NATs - Software defined network
More informationDept. Of Computer Science, Colorado State University
CS 455: INTRODUCTION TO DISTRIBUTED SYSTEMS [HADOOP/HDFS] Trying to have your cake and eat it too Each phase pines for tasks with locality and their numbers on a tether Alas within a phase, you get one,
More informationSurvey Paper on Traditional Hadoop and Pipelined Map Reduce
International Journal of Computational Engineering Research Vol, 03 Issue, 12 Survey Paper on Traditional Hadoop and Pipelined Map Reduce Dhole Poonam B 1, Gunjal Baisa L 2 1 M.E.ComputerAVCOE, Sangamner,
More informationGoogle File System. Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google fall DIP Heerak lim, Donghun Koo
Google File System Sanjay Ghemawat, Howard Gobioff, and Shun-Tak Leung Google 2017 fall DIP Heerak lim, Donghun Koo 1 Agenda Introduction Design overview Systems interactions Master operation Fault tolerance
More informationMitigating Data Skew Using Map Reduce Application
Ms. Archana P.M Mitigating Data Skew Using Map Reduce Application Mr. Malathesh S.H 4 th sem, M.Tech (C.S.E) Associate Professor C.S.E Dept. M.S.E.C, V.T.U Bangalore, India archanaanil062@gmail.com M.S.E.C,
More informationMounica B, Aditya Srivastava, Md. Faisal Alam
International Journal of Scientific Research in Computer Science, Engineering and Information Technology 2017 IJSRCSEIT Volume 2 Issue 3 ISSN : 2456-3307 Clustering of large datasets using Hadoop Ecosystem
More informationDistributed Systems. 15. Distributed File Systems. Paul Krzyzanowski. Rutgers University. Fall 2017
Distributed Systems 15. Distributed File Systems Paul Krzyzanowski Rutgers University Fall 2017 1 Google Chubby ( Apache Zookeeper) 2 Chubby Distributed lock service + simple fault-tolerant file system
More informationChapter 5. The MapReduce Programming Model and Implementation
Chapter 5. The MapReduce Programming Model and Implementation - Traditional computing: data-to-computing (send data to computing) * Data stored in separate repository * Data brought into system for computing
More informationPage 1. Goals for Today" Background of Cloud Computing" Sources Driving Big Data" CS162 Operating Systems and Systems Programming Lecture 24
Goals for Today" CS162 Operating Systems and Systems Programming Lecture 24 Capstone: Cloud Computing" Distributed systems Cloud Computing programming paradigms Cloud Computing OS December 2, 2013 Anthony
More informationPerformance Enhancement of Data Processing using Multiple Intelligent Cache in Hadoop
Performance Enhancement of Data Processing using Multiple Intelligent Cache in Hadoop K. Senthilkumar PG Scholar Department of Computer Science and Engineering SRM University, Chennai, Tamilnadu, India
More informationHadoop/MapReduce Computing Paradigm
Hadoop/Reduce Computing Paradigm 1 Large-Scale Data Analytics Reduce computing paradigm (E.g., Hadoop) vs. Traditional database systems vs. Database Many enterprises are turning to Hadoop Especially applications
More informationIndexing Strategies of MapReduce for Information Retrieval in Big Data
International Journal of Advances in Computer Science and Technology (IJACST), Vol.5, No.3, Pages : 01-06 (2016) Indexing Strategies of MapReduce for Information Retrieval in Big Data Mazen Farid, Rohaya
More informationA Survey on Job Scheduling in Big Data
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 16, No 3 Sofia 2016 Print ISSN: 1311-9702; Online ISSN: 1314-4081 DOI: 10.1515/cait-2016-0033 A Survey on Job Scheduling in
More informationAn Adaptive Scheduling Technique for Improving the Efficiency of Hadoop
An Adaptive Scheduling Technique for Improving the Efficiency of Hadoop Ms Punitha R Computer Science Engineering M.S Engineering College, Bangalore, Karnataka, India. Mr Malatesh S H Computer Science
More informationDistributed Face Recognition Using Hadoop
Distributed Face Recognition Using Hadoop A. Thorat, V. Malhotra, S. Narvekar and A. Joshi Dept. of Computer Engineering and IT College of Engineering, Pune {abhishekthorat02@gmail.com, vinayak.malhotra20@gmail.com,
More informationDistributed Filesystem
Distributed Filesystem 1 How do we get data to the workers? NAS Compute Nodes SAN 2 Distributing Code! Don t move data to workers move workers to the data! - Store data on the local disks of nodes in the
More informationDistributed File Systems II
Distributed File Systems II To do q Very-large scale: Google FS, Hadoop FS, BigTable q Next time: Naming things GFS A radically new environment NFS, etc. Independence Small Scale Variety of workloads Cooperation
More information