Application of Virtualization Technologies & CernVM. Benedikt Hegner CERN

Similar documents
CernVM-FS beyond LHC computing

Using CernVM-FS to deploy Euclid processing S/W on Science Data Centres

CernVM a virtual software appliance for LHC applications

The Business Value of Virtualizing Oracle ebusiness Suite. Haroon Qureshi QSolve, Inc.

CernVM-FS. Catalin Condurache STFC RAL UK

<Insert Picture Here> Linux: The Journey, Milestones, and What s Ahead Edward Screven, Chief Corporate Architect, Oracle

RADU POPESCU IMPROVING THE WRITE SCALABILITY OF THE CERNVM FILE SYSTEM WITH ERLANG/OTP

Ivane Javakhishvili Tbilisi State University High Energy Physics Institute HEPI TSU

Private Cloud Database Consolidation Name, Title

Virtualization. A very short summary by Owen Synge

STATUS OF PLANS TO USE CONTAINERS IN THE WORLDWIDE LHC COMPUTING GRID

Docker 101 Workshop. Eric Smalling - Solution Architect, Docker

1/10/2011. Topics. What is the Cloud? Cloud Computing

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Global Software Distribution with CernVM-FS

Linux Automation.

<Insert Picture Here> Enterprise Data Management using Grid Technology

OpenStack Magnum Pike and the CERN cloud. Spyros

Switch to Parallels Remote Application Server and Save 60% Compared to Citrix XenApp

EGEE and Interoperation

Exploring cloud storage for scien3fic research

I Tier-3 di CMS-Italia: stato e prospettive. Hassen Riahi Claudio Grandi Workshop CCR GRID 2011

VMs at a Tier-1 site. EGEE 09, Sander Klous, Nikhef

LHCb Distributed Conditions Database

The INFN Tier1. 1. INFN-CNAF, Italy

Enterprise-class desktop virtualization with NComputing. Clear the hurdles that block you from getting ahead. Whitepaper

LHCb experience running jobs in virtual machines

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

Evolution of Cloud Computing in ATLAS

Software installation and condition data distribution via CernVM File System in ATLAS

Database Consolidation onto Private Cloud. Piotr Kołodziej, Oracle Polska

10 Steps to Virtualization

High Performance Computing Course Notes Grid Computing I

Online data storage service strategy for the CERN computer Centre G. Cancio, D. Duellmann, M. Lamanna, A. Pace CERN, Geneva, Switzerland

Taming your heterogeneous cloud with Red Hat OpenShift Container Platform.

The LHC Computing Grid

2011 IBM Research Strategic Initiative: Workload Optimized Systems

[Docker] Containerization

Geant4 on Azure using Docker containers

IBM Bluemix compute capabilities IBM Corporation

An Introduction to Virtualization and Cloud Technologies to Support Grid Computing

Grid Computing: dealing with GB/s dataflows

Large Scale Computing Infrastructures

Lenovo Database Configuration

CouchDB-based system for data management in a Grid environment Implementation and Experience

Genomics on Cisco Metacloud + SwiftStack

Connected vehicle cloud Commercial presentation

Clouds at other sites T2-type computing

EXAM Pro: Windows Server 2008 R2, Virtualization Administrator. Buy Full Product.

Test Methodology We conducted tests by adding load and measuring the performance of the environment components:

Copyright 2011, Oracle and/or its affiliates. All rights reserved.

Multi-threaded, discrete event simulation of distributed computing systems

Configuration changes such as conversion from a single instance to RAC, ASM, etc.

The Challenge of Managing WebSphere Farm Configuration. Rational Automation Framework for WebSphere

Using Kollective with Citrix Virtual Desktop Infrastructure (VDI)

1 Copyright 2011, Oracle and/or its affiliates. All rights reserved. reserved. Insert Information Protection Policy Classification from Slide 8

MONTE CARLO SIMULATION FOR RADIOTHERAPY IN A DISTRIBUTED COMPUTING ENVIRONMENT

How Microsoft Built MySQL, PostgreSQL and MariaDB for the Cloud. Santa Clara, California April 23th 25th, 2018

ELFms industrialisation plans

where the Web was born Experience of Adding New Architectures to the LCG Production Environment

Security in the CernVM File System and the Frontier Distributed Database Caching System

Beyond 1001 Dedicated Data Service Instances

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

Clouds in High Energy Physics

TEN LAYERS OF CONTAINER SECURITY

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

Get more out of technology starting day one. ProDeploy Enterprise Suite

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

VMware vcloud Architecture Toolkit Hybrid VMware vcloud Use Case

Your cloud solution for EO Data access and processing

Virtualizing a Batch. University Grid Center

Virtualization Strategies on Oracle x86. Hwanki Lee Hardware Solution Specialist, Local Product Server Sales

Some thoughts on the evolution of Grid and Cloud computing

Evaluation of the Huawei UDS cloud storage system for CERN specific data

Assistant Professor, School of Computer Applications,Career Point University,Kota, Rajasthan, India Id

Oracle Real Application Clusters One Node

Distributing storage of LHC data - in the nordic countries

Was ist dran an einer spezialisierten Data Warehousing platform?

Identifying Workloads for the Cloud

Connected vehicle cloud

Scalability Testing with Login VSI v16.2. White Paper Parallels Remote Application Server 2018

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

vsan 6.6 Performance Improvements First Published On: Last Updated On:

ATLAS Nightly Build System Upgrade

Virtualization of the ATLAS Tier-2/3 environment on the HPC cluster NEMO

Cisco Expo 2009 Bratislava. Chief Technology Officer VMware, Inc.

Flex Tenancy :48:27 UTC Citrix Systems, Inc. All rights reserved. Terms of Use Trademarks Privacy Statement

On Demand Cryptographic Resources for Your Virtual Data Center and the Cloud: Introducing SafeNet s Crypto Hypervisor

Systematic Cooperation in P2P Grids

Clustering and Reclustering HEP Data in Object Databases

Andrea Sciabà CERN, Switzerland

vbranch Introduction and Demo

Disclaimer This presentation may contain product features that are currently under development. This overview of new technology represents no commitme

5 Reasons to Choose Parallels RAS Over Citrix Solutions

NFV Infrastructure for Media Data Center Applications

Architecture: Consolidated Platform. Eddie Augustine Major Accounts Manager: Federal

Docker and HPE Accelerate Digital Transformation to Enable Hybrid IT. Steven Follis Solutions Engineer Docker Inc.

Optimize Your Heterogeneous SOA Infrastructure

Lightweight scheduling of elastic analysis containers in a competitive cloud environment: a Docked Analysis Facility for ALICE

Distributed Systems CS6421

Transcription:

Application of Virtualization Technologies & CernVM Benedikt Hegner CERN

Virtualization Use Cases Worker Node Virtualization Software Testing Training Platform Software Deployment }Covered today Server consolidation Cloud Computing Volunteer Computing Data Preserveration

Worker Node Virtualization Helps in the management of computing centres Decoupling of jobs and physical resources Eases management of batch farm resources Enables computing centres to move to new computing models more easily Examples: CERN virtual batch system, CNAF worker nodes on demand

Software Testing Virtual machines can cut time and money out of the software development and testing process Great opportunity to test software in a large variety of platforms Each platform can be realized by a differently configured virtual machines Easy to duplicate same environment in several virtual machines Testing installation procedures from well defined state Etc. Example: Execution Infrastructure in ETICS (spin-off of the EGEE project) Set of virtual machines that run a variety of platforms attached to an Execution Engine where Build and Test Jobs are executed on behalf of the submitting users

Training Platform Similar as for software testing infrastructure, virtualization helps to deploy rapidly dedicated software and workstations/servers for training Need for many nodes rapidly and typically for a rather short period of time Isolation with respect production servers Disposable workstations/servers The resources for today s hands-on are set up following this idea

Software Deployment Problem Software @ LHC Millions of lines of code Different packaging and software distribution models Complicated software installation/update/configuration procedure Long and slow validation and certification process Very difficult to roll out major OS upgrade (SLC4 -> SLC5) Additional constraints imposed by the grid middleware development Effectively locked on one Linux flavour Whole process is focused on middleware and not on applications How to effectively harvest multi and many core CPU power of user laptops/desktops if LHC applications cannot run in such environment? Good news: We are not the only one with such problems

Horizontal Integration Application Libraries Traditional model Horizontal layers Independently developed Maintained by the different groups Different lifecycle Tools Databases OS Application is deployed on top of the stack Breaks if any layer changes Needs to be certified every time when something changes Results in deployment and support nightmare Hardware

Vertical Integration Virtual Machine Application Libraries Tools Databases OS Application driven approach Analyzing application requirements and dependencies Adding required tools and libraries Building minimal OS Bundling all this into Virtual Machine image Virtual Machine images should be versioned just like the applications Assuring accountability to mitigate possible negative aspects of newly acquired application freedom Problem in HEP The Application is rather huge (10GB) and changes once per week per experiment Overloads the deployment infrastructure at the sites

The CernVM project CernVM is a R&D project started 3 years ago on Virtualization at CERN The CernVM image is an attempt to mitigate the standard difficulties of VMs (performance, image distribution, trust, contextualization, etc.) Tuned for best performance of HEP applications Single image fits all [LHC] experiments Very small is size (only 250MB) with just-enough OS Experiment software is factorized out (dedicated File System) Flexible configuration and contextualization mechanisms

The CernVM filesystem } Experiment software is changing frequently and we want to avoid need to frequently update, certify and redistribute VM images with every release } Only a small fraction of software release is really used } Demonstrated scalability and reliability } Now being deployed on across all Grid sites as the channel for software distributions

Application Software Delivery CernVM comes with the read-only file system (CernVM-FS) optimized for software distribution Very little fraction of the experiment software is actually used (~10%) Very aggressive local caching, web proxy cache (squids) Transparent file compression Integrity checks using checksums, signed file catalog Operational in off-line mode No need to install any experiment software Virtually all versions of all applications are already installed The user just needs to start using it to trigger the download CernVM-FS can be used outside the CernVM context You can use it as well on a standard Scientific Linux 5/6 installation Removes the load of software deployment at your site

Stratum Model + Fast and Scalable + No single point of failure - Complex hierarchy Content Distribution

Client-Side Fail-Over Proxies SL5 Squid, load-balancing + fail-over e. g. CVMFS_HTTP_PROXY="A B C" Mirrors Fail-over mirrors at CERN, RAL, BNL For roaming users automatic ordering based on RTT

Client-Side Fail-Over Proxies SL5 Squid, load-balancing + fail-over e. g. CVMFS_HTTP_PROXY="A B C" Mirrors Fail-over mirrors at CERN, RAL, BNL For roaming users automatic ordering based on RTT

Publishing Releases 2. Publishing is an atomic operation 1. Each experiment is given a VM to install and test their software using own installation tools

That s all :-)

Hands-On Session Install Squid to serve the software on one of the provided VMs see exercise for storage Install the CernVM-FS fuse-module on another of the provided VMs Chapter 2 of https://cernvm.cern.ch/project/trac/downloads/cernvm/cvmfstech-2.0-1.pdf Let the installed fuse module point to the already set up proxy If time allows, try out of the box CernVM images on your local machine http://cernvm.cern.ch/portal/ For more hints for a local installation at your institute http://cernvm.cern.ch/portal/cvmfs/examples