Analysis & Tier 3s. Amir Farbin University of Texas at Arlington

Size: px
Start display at page:

Download "Analysis & Tier 3s. Amir Farbin University of Texas at Arlington"

Transcription

1 Analysis & Tier 3s Amir Farbin University of Texas at Arlington

2 Introduction Tug of war between analysis and computing requirements: Analysis Model Physics Requirements/User preferences (whims) Organization of data/steps of analysis Software Technologies/Design Computing Model Disk, CPU, Network, Cluster Architecture Production, DDM, DA Everyone would love to have the data everywhere, instant processing... Simulated data (eg FDR) is far short of the LHC data volume Today users practice analysis models which won t scale to real data... We all systematically underestimate LHC computing challenges... reflected in our AM. Important that this community presses back on the Analysis Model... it should be an iterative process.

3 Overview I will not talk about production activity at tier 3. Overview relevant parts computing model/analysis model. Talks specifically about what are the Analysis Model expectations from tier 3s. Tier 3s are still not defined. One ideas about an easy way to build a tier 3. There are lots of different types of tier 3. Apologies for being provocative... please point out my mistakes.

4 Analysis Model

5 ATLAS Computing Full Simulation Fast Simulation Generation Simulation KHz Digitization mhz Hz Data Store New KHz Generation Hz Fast Simulation ATLAS will only simulate 20% of data High-level Trigger 200 Hz Reconstruction Data chz KHz Algorithmic Hz Analysis MHz Interactive Analysis 10 9 events/year Base Data Analysis & Calibration Statistical Analysis

6 The Event Data Model Raw Channels. 1.6 MB/event. Reconstruction Output. Intended for calibration. 500 KB/event. Cells,Hits, Tracks, Clusters,Electrons, Jets,... TAG Analysis Object Data Summary of Event. Intended for selection. 1 KB/event. Trigger decision, pt of 4 best electrons, jets... Derived Physics Data Event Summary Raw Data Objects Data Data refinement Intended for Analysis. 100 KB/event. Light-weight Tracks, Clusters,Electrons, Jets, Electron Cells, Muon HitOnTrack,... Intended for interactive Analysis. ~10-20 KB/event. What-ever is necessary for a specific analysis/ calibration/study.

7 The Computing Model Resources Spread Around the GRID Derive 1st pass calibrations within 24 hours. Reconstruct rest of the data keeping up with data taking. RAW Reprocessing of full data with improved calibrations 2 months after data taking. Managed Tape Access: RAW, ESD Disk Access: AOD, fraction of ESD Tier 0 CERN Analysis Facility RAW/ AOD/ ESD Tier 1 10 Sites Worldwide AOD Interactive Analysis Plots, Fits, Toy MC, Studies,... Tier 2 30 Sites Worldwide Primary purpose: calibrations Small subset of collaboration will have access to full ESD. Limited Access to RAW Data. Tier 3 Tier 3 DPD Production of simulated events. User Analysis: 12 CPU/ Analyzer Disk Store: AOD

8 Analysis Activity Re-reconstruction/re-calibration- CPU intensive... often necessary. Algorithmic Analysis: Data Manipulations ESD AOD DPD DPD Skimming- Keep interesting events Thinning- Keep interesting objects in events Slimming- Keep interesting info in objects Reduction- Build higher-level data which encapsulates results of algorithms Basic principle: Data Optimization + CPU intensive algs more portable input & less CPU in later stages. Interactive Analysis: Making plots/performing studies on highly reduced data. Statistical Analysis: Perform fits, produce toy Monte Carlos, calculate significance. Tier 1/2 Activity Framework (ie Athena) based Resource intensive Large scale (lots of data) Organized Batch access only Tier 3 Activity Often exoframework Interactive Primary difference

9 Analysis Activity Re-reconstruction/re-calibration- CPU intensive... often Tier 1/2 Activity necessary. To the user, the defining difference between tier 2 and 3 is Framework interactive (ie Algorithmic Analysis: Data Manipulations access Athena) based ESD AOD DPD DPD Skimming- Users Keep need interesting a place to events login, in order to Resource intensive Thinning- Develop Keep interesting code, run objects test jobs in events Large scale Slimming- Submit Keep interesting large scale info analysis in objects to GRID (lots of data) Reduction- Gather Build higher-level results from data GRID which and encapsulates perform final stages of Organized analysis results of algorithms But tier 2s cannot manage user accounts and interactive usage Batch access Basic principle: Data Optimization + CPU intensive algs only patterns. more portable input & less CPU in later stages. Some tier 2s allow access to local users... unfair? Interactive Analysis: Making plots/performing studies on highly reduced data. Statistical Analysis: Perform fits, produce toy Monte Carlos, calculate significance. Tier 3 Activity Often exoframework Interactive Primary difference

10 Evolution of the EDM I AOD/ESD Merger- Experience from other experiments indicate that calibration/ reconstruction are often performed in final steps of analysis The structure of the AOD and ESD have been unified so that AOD is essentially the ESD with less detail Code (eg analysis or reconstruction) can run on either AOD/ESD. Re-reconstruct/Re-calibrate on AOD, when sufficient information has been retained. For Example: AOD now contains calo cell / hit info for all leptons. All containers in the ESD Data Data Data Electron Cell Cell Cell Cell Data Data CaloCluster CaloCluster Data Data Data Data Data Data Data Data Data Data Links to constituents TrackParticle TrackParticle Hit Hit Hit Hit Available in AOD Benefits: 1. Move data between ESD/AOD/DPD w/o schema change. 2. Read on Demand

11 Evolution of the EDM I AOD/ESD Merger- Experience from other experiments indicate that calibration/ reconstruction are often performed in final steps of analysis The structure of the AOD and ESD have been unified so that Enhancing the AOD with ESD-like capabilities: AOD is essentially the ESD with less detail The AOD is largely w/in budget! The reason the AOD is larger than Code (eg analysis or reconstruction) can run on either AOD/ESD. model is not due to the inclusion of ESD quantities, rather inefficiencies that will hopefully be resolved (trigger info, jets, etc...) Re-reconstruct/Re-calibrate on AOD, when sufficient information has been retained. AOD is supposed to be accessible at all Tier 2s. For Example: AOD now contains calo cell / hit info for all leptons. If some ESD-like tasks can performed w/ AOD, less ESD replicas needed. Electron All containers in the ESD Data Data Data Data Data CaloCluster CaloCluster But the AOD/ESD merger is new, and no one has moved from using Data ESD Data to using Data AOD... and Data very few Data are seriously considering it. TrackParticle TrackParticle There is hope for less Data need for Data ESD, but this isn t reflected in CM. Data Data Data Links to constituents Cell Cell Cell Cell Hit Hit Hit Hit Available in AOD Benefits: 1. Move data between ESD/AOD/DPD w/o schema change. 2. Read on Demand

12 Evolution of the EDM II Size and speed are obviously very important Realization that these go hand-in-hand ATLAS Data-structures are complex Transient/Persistent splitting 2 different representations of each data object: optimized for manipulation in memory or storage. We have an AOD which is store much more info than previously thought possible. ie more detail available on all tier 2s. Access times near optimal- less incentive for users to produce flat data structures in order to maximize speed.

13 Evolution of the EDM III AthenaROOTAccess: Reading framework data (ie POOL) w/o framework (athena) Many users have an aversion to framework and prefer ROOT. Excellent means of quickly inspecting a file for validation. Some stages of analysis must be outside of framework anyway. Specifically the DPD must be accessible outside framework... now DPD can be in the same format as framework data. Use POOL to write DPDs. Provides a unique DPD format for everyone... ie same format as AOD/ESD. DPD can be analyzed in framework.

14 Streams The online will split the data into O(10) streams based on triggers. Write events into different RAW files... different processing priority. Inclusive: one event may end up in more than one file. The stream boundaries will be obeyed through ESD and AOD. DPDs may process multiple streams. Overlaps must be resolved if > 1 stream is used in an analysis.

15 AMF Report Analysis Model Centrally produced Primary DPDs (aka D 1 PD) with total volume = AOD. D 1 PD: AOD volume reduction based on simple requirements. No strict decisions like electron ID or overlap-removal. Slim: Throw out containers (and info) which will be not used down-stream (eg tracks, clusters, or cells) Thin: eg keep useful truth particles or tracks. Skim: Keep events after loose preselection. In practice, this is the main source of reduction. May be analyzed with AthenaROOTAccess in ROOT. In response to: Strict selections made before DPD making. Inability to quickly remake & download DPD. Lack of familiarity with athena-based analysis tools.

16 AMF Report Analysis Model Why create D 1 PD? If you are on a tier 2, the D 1 PD provides little added value: All AOD is present on tier 2s. Skimming can be achieved using TAG... don t read any events you don t want. No per event speed improvements: athena only reads in containers you request. But some argue for faster access at Tier 2s because of smaller files. Is this really true? But since the D 1 PD volume is times smaller than the AOD, you can transfer it to a tier 3 and do all your analysis using ROOT + AthenaROOTAccess. This is a dream-come-true for a lot of people. AOD is too big for analysis at tier 3... but perhaps D 1 PD is small enough. Essentially using the D 1 PD as a means of moving what we imagined was AOD analysis to tier 3s. Users imagine this allows them to escape using Athena and GRID... Huge implications on Tier 3.

17 AMF Report Analysis Model Secondary DPDs (aka D 2 PD): POOL format, so it is re-readable into athena for further processing. Output of analysis on AOD or D 1 PD. D 1 PD is encouraged (why?). Perhaps D 2 PD output will include stricter object and event selection. Results of analysis can be saved into new containers, composite particles, UserData, EventView, ParticleView Tertiary DPDs (aka D 3 PD) Flat-ntuple. For quickly making final plots. Output of analysis on AOD, D 1 PD, or D 2 PD. AOD is discouraged (this would be like the CSC model). Very similar to D 2 PD in terms of potential content... but hopefully much reduced. If it s the 3rd step in a analysis, likely to be smaller than D 2 PD.

18 AMF Report Analysis Model The AMF Report presents a stringent analysis model AOD D 1 PD other analysis steps with D 2 PD and D 3 PD optional. Doesn t present other potential models (ie w/ no D 1 PD), though some are skeptical and plan to explore alternatives for FDR. Very specific recommendations on the nature of the operations allowed at each step and the software used. Original recommendation: prioritizing D 1 PD disk at expense of AOD disk at tier 2s... a change to CM. This has been removed as requested by TOB. Some recommendation of the TOB (Fabiola and Dave C.) have not been incorporated into newest version (4) due to disagreement. Very little accounting of access speed, disk space, etc requirements. These must be tested during FDR. Tier 3 TF must assess the feasibility of the AMF Report model for tier 3s. Either: The AMF Report sets the scale of tier 3... or Must change AM to be consistent with what is practical at a tier 3.

19 Evolution of DPDs (AMF) POOL-Based: Analyze in Athena or ROOT via ARA D 1 PD (primary) Subset of the AOD through skimming, thinning, slimming Centrally produced Defined by physics/ combined-reco groups different D1PDs w/ total volume = AOD ESD AOD D 2 PD (2ndary) Similar to D1PD More analysis specific Store UserDataresults of complex analysis algorithms: CompositeParticles, EventViews, ParticleViews, etc Flat : Analyze in plain ROOT D 3 PD (tertiary) Maybe similar in content to D1PD or D2PD But most likely highly reduced Just the few quantities necessary to quickly make the final plots for you analysis Complex analysis on D3PD is discouraged Format is proprietary Note: AOD D1PD D2PD D3PD chain is not necessary. Steps may be skipped.

20 AMF and Tier 3

21 Tier 3 The original analysis motivations for tier 3s are interactive environment and instant access to data. Eventually users need to make plots from a prompt. Not clear what sort of data (D 1 PD, D 2 PD, D 3 PD) is practical for analysis at tier 3. Lots of recent focus on D 1 PD... so my primary goal here is to evaluate performing D 1 PD analysis at tier 3s. Some believe that tier 2 resources will be insufficient for analysis, so tier 3s will need to pick up the slack. This is a departure from the original motivation. The questions are therefore: What are the required resources for D 1 PD analysis? Can we shift the tier 2 analysis to tier 3? (ie analysis on D 1 PDs) Can the analysis resources at tier 3 > tier 2? We should not forget about tasks that cannot be performed at the tier 2 so they must be provided on tier 3.

22 The Factors Obvious, I suppose: Disk: Does the data fit? CPU: Is there enough CPU to process it within reasonable time frame? DDM/Network: Can the D n PD be brought to tier 3 within appropriate time frame? Can the DDM handle the traffic?

23 D 1 PD: Numbers (Size) AMF Report: 1/10-1/20 of AOD... so 5-10 TB per year, per DPD. Studies by A. Shibata indicated 1/3 of the per event AOD size ( at best ). Suggests skim rates of ~ 15-30%. Produced monthly... needs downloading monthly. D 2 PD: D 2 PD may be larger per event than D 1 PD due to addition of user data. But D 2 PD per event size may be compensated by more skimming. Guess: % of D 1 PD. D 3 PD can be as large as D 2 PD... but hopefully will be highly reduced. Guess: 1-33% of D 1 PD. D 2 PD/D 3 PD might be downloaded as frequently as daily... unless already produced at tier 3.

24 Numbers (CPU) D1 PD will be produced at tier 1 (or 2s). So not really a tier 3 issue. Recent studies (Balint Radics) indicate simple D1 PD making jobs run at 40 ms/event total (including IO, I calculate 15 ms w/o IO). If there is any re-reconstruction/re-calibration (likely), this can easily be 250 ms/event or more. Lots of CPU required for D2 PD/D 3 PD production. This is where analysis really occurs. Time consumers: obj selection, overlap removal, combinatorics, observable calculation. TopView: 275 ms/event (including IO) My profiling: need at least 20 ms/event for top analysis (w/o IO). Realistic number ms.

25 D2PD Production In/out 33/40 KB/event CPU:100 ms/event D3PD Production In/out 33/10 KB/ event CPU: 50 ms/event Plotting In: 1 KB/ event CPU: 0 D 1 PD Processing Guesses of IO/CPU need for different analysis stages. Calculate fraction of a year s worth of data processed: ->D2PD Laptop Tier 3 Tier 2 Tier 1/2 Cores Hour >0.01% 0.08% 0.34% 3.36% Overnight 0.04% 1.01% 4.03% 40.26% 1 Week 0.56% 14.09% 56.37% All 1 Month 2.42% 60.39% All All ->D3PD Laptop Tier 3 Tier 2 Tier 1/2 Cores Hour 0.01% 0.17% 0.66% 6.63% Overnight 0.08% 1.99% 7.96% 79.56% 1 Week 1.11% 27.85% All All 1 Month 4.77% All All All Plotting Laptop Tier 3 Tier 2 Tier 1/2 Cores Hour 3.56% 89.02% All All Overnight 42.73% All All All 1 Week All All All All 1 Month All All All All Assuming perfect hardware/software

26 Tier 3 Disk for D 1 PD Processing How much data do you need to feed the CPUs? D 1 PD D 2 PD jobs at tier 3 (25 cores) Overnight: ~1% of 1 year s data ~ 0.33 TB of D 1 PD TB of D 2 PD. A week: ~ 15% of 1 year s data ~ 5 TB of D 1 PD. 6 TB of D 2 PD. D 1 PD D 3 PD jobs at tier 3 (25 cores) Overnight: ~2% of 1 year s data 0.66 TB of D 1 PD. 0.1 TB of D 3 PD. A week: ~30% of 1 year s data 10 TB of D 1 PD. 3.3 TB of D 1 PD. Estimate need 3x more space for 2 concurrent versions + simulation. If you are willing to wait a week for a single iteration: Can process 15-30% of a year s worth of D 1 PDs. Need ~30-40 TB per D 1 PD stream... you can process a whole D 1 PD stream. If you are willing to only wait overnight for a single iteration: Can process 1-2% of a year s worth of D 1 PDs. Need < 3 TB per D 1 PD stream... but you need to go to tier 2 to get full D 1 PDs. Unknown: how long will wait on a tier 3 before giving up and moving to tier 2?

27 Data Transfer AMF calls for D 1 PDs produced at tier 1/2 s, once a month. Full D 1 PD stream is 5-10 TB. Assuming each site will download 1 D 1 PD and O(30) Tier 3s, can DDM/network transfer TB to tier 3s once a month?

28 Summary D 1 PDs and Tier 3s If you want to process D 1 PDs at tier 3s: Need TB of disk per year. 25 cores = 1 analysis iteration per week. Multiply for faster response. Will need to wait a long time for the D 1 PD to arrive at tier 3. We probably can t afford this! If you agree, we should make a strong statement that large-scale D 1 PD analysis cannot be supported at tier 3. But then we should worry about having enough analysis resources at tier 2s? This brings into question the whole motivation for the D 1 PDs. Remember the D 1 PDs are a mechanism for moving AOD analysis to tier 3s. We should also evaluate the benefits of D 1 PDs on tier 2s. You can just use AOD + Tag instead and have access to more info for your analysis. Are are performance gains (if any) worth the extra disk space? My opinion: we should make D 1 PDs as long as we can afford them...

29 D 2 PD/D 3 PD at Tier 3 D 2 PD or D 3 PD must be at tier 3s for final stages of analysis... there is no other place to work interactively on these samples. If the CPU intensive parts of the analysis have been done on tier 2 when producing D 2 PD/D 3 PD, Analysis of D 2 PD/D 3 PD is much faster... like plotting Example. IO limited. But need more iterations of D 2 PD/D 3 PD production at tier 2... and copy to tier 3. (eg Once a week?) The actual volume of D 2 PD or D 3 PD will be defined by what is practical for tier 3s. People will fine tune D 2 PD and D 3 PD so they can analyze them on their tier 3. So I believe tier 3 parameters is defined by what you can afford and manage, not by the analysis model. In other words, tier 3 set the Analysis Model.

30 Path forward If we want to support D 1 PD analysis at tier 3, then I provided the relevant numbers. If not, We should provide feed-back to Common Analysis meeting. We should evaluate what tier 3 size we can afford and manage. Set the D 2 PD/D 3 PD requirements. Adjust the Analysis Model.

31 Building a Tier3

32 Tier 3 Hardware/Software There isn t just one type of tier 3 A tier 3 may be a laptop... or a few desktops. Today I can buy 3 eight-core machines, with 2 GB/core, and 12TB of disk for ~$12K. May be a departmental cluster or other leveraged (shared) resource. There may or may not be a sys-admin who may or may not be willing/capable of ATLAS specific support. Analysis is IO intensive, will be difficult to be optimal. Lots of software components must be setup... should be simple.

33 Tier 3 Software On one or few machines: On few machines: Very Minimal: dq2 client tools (to get DPDs) and ROOT. Today, this already means linux for dq2 client tools. Minimal: above + Athena kit for ARA analysis. Default: above + data aggregation via xrootd (or nfs), batch queues, and/or PROOF. On a cluster: Super: above + OSG middleware + ATLAS services (perhaps via a tier 2).

34 Virtual Machines? Intel Core Duo/Core 2 Duo natively support virtualization => virtual machines run nearly as fast as host OS running on the machine. CERN group now dedicated to providing pre-configured VMs. Allows running ATLAS software (Athena, ARA, etc) on ANY machine (any OS) with minimal effort. No sysadmin required. Linux (especially SLC4) isn t always the ideal OS. ATLAS/GRID software is easiest to setup in SLC4. Still, some software components are hard to setup... once it is setup, I don t want to have to repeat this for every machine I want to use. Virtual Machine allows us to encapsulate all the software necessary for ATLAS into a file that a user can run on any machine with no additional effort.

35 Example Setup Hardware: Intel Core Duo MacBook (2GB/120GB), Intel 2 Core MacBook Pro(4GB/160GB), Intel 2x 4-core Core 2 Duo (Penryn) Mac Pro(14GB/3.5TB). Should work the same (for the most part) in other OSs. VM Software: VMWare fusion ($80). VMWare player is free for Windows/Linux (can t create a VM, but can run one). VM: SLC4, with recompiled kernel. 30 GB disk image. I ve copied this VM many times between VMs (or ran multiple instances on the same host). Environment: Do not start X in VM (no graphics performance hit). Simply ssh to VM and use X-client in Mac OS (or linux or Windows). Gives unified environment (Ex: copy code from VM to Mac Mail). Data Disks (Kits/Home/AOD/Ntuples): Usually I keep these on the host machine, mount them from the VM. Allows me to edit code and do ROOT analysis w/o running VM.

36 User Perspective Download/install VM software (eg the free VMWare player) Download/copy VM from CERN VM group. Download ATLAS specific VM additions. VM files total O(20-30 Gigs) Maybe some local setup for giving VM access to data on host. Double-click on VM. ssh to the VM. Maybe download a kit into VM or local disk. Immediately use all GRID and ATLAS software.

37 Analysis Workflow Develop analysis, test on a small sample, submit jobs to GRID, collect results. Work cycle: 1. Install recent ATLAS kit and download small amount of data. 2. Setup test release, check out packages, setup environment. 3. Modify/develop analysis code (Athena, ARA, etc...) 4. Build software 5. Run small test job, check results in ROOT (on host machine, if you like). Iterate. 6. Submit jobs to GRID... wait... download DPDs. 7. Analyze DPDs in VM (ARA, Athena, ROOT) or Host (ROOT) Ideal for analysis development on Laptop... again very nice because you can keep sessions open for weeks and take it with you anywhere. But laptop disk typically too small for all DPDs, so step 7 is only good for small samples... same model can be deployed on desktop(s).

38 VMs and Tier 3s Easiest way to get ATLAS software running on your laptop or desktop. Currently investigating various issues: VM software, kernel optimization, data access, VM configuration, VM distribution, xrootd, batch, PROOF... measuring performance. many of these will be resolved in collaboration with CERN VM group. Currently, running one machine (desktop/laptop) has been very successful. Plan to provide a VM/recipe for ATLAS SW on any machine in the next few weeks. Next step is to scale this up to a cluster. Clearly an option worth exploring... lots of possibilities. VMs which deploy various ATLAS services. Ask me more about this if you are interested.

39 Final Remarks The ATLAS Analysis Model is continuously evolving. The AMF recommendations should be evaluated, not just accepted... the computing facilities should provide feed-back before deployment. AMF recommendations have huge implications on Tier 3s... my worry is that it won t work. Tier 3s should focus on what the tier 2s don t provide. One critical issue for tier 3s is how to easily build one. VMs could be one way of making things easier. Should we provide a tier 3 support team?

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko

Considerations for a grid-based Physics Analysis Facility. Dietrich Liko Considerations for a grid-based Physics Analysis Facility Dietrich Liko Introduction Aim of our grid activities is to enable physicists to do their work Latest GANGA developments PANDA Tier-3 Taskforce

More information

Data Analysis in ATLAS. Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser

Data Analysis in ATLAS. Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser Data Analysis in ATLAS Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser 1 ATLAS Data Flow into Analysis RAW detector data and simulated RDO data are reconstructed into our xaod

More information

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN Software and computing evolution: the HL-LHC challenge Simone Campana, CERN Higgs discovery in Run-1 The Large Hadron Collider at CERN We are here: Run-2 (Fernando s talk) High Luminosity: the HL-LHC challenge

More information

The ATLAS EventIndex: Full chain deployment and first operation

The ATLAS EventIndex: Full chain deployment and first operation The ATLAS EventIndex: Full chain deployment and first operation Álvaro Fernández Casaní Instituto de Física Corpuscular () Universitat de València CSIC On behalf of the ATLAS Collaboration 1 Outline ATLAS

More information

Virtualizing a Batch. University Grid Center

Virtualizing a Batch. University Grid Center Virtualizing a Batch Queuing System at a University Grid Center Volker Büge (1,2), Yves Kemp (1), Günter Quast (1), Oliver Oberst (1), Marcel Kunze (2) (1) University of Karlsruhe (2) Forschungszentrum

More information

Early experience with the Run 2 ATLAS analysis model

Early experience with the Run 2 ATLAS analysis model Early experience with the Run 2 ATLAS analysis model Argonne National Laboratory E-mail: cranshaw@anl.gov During the long shutdown of the LHC, the ATLAS collaboration redesigned its analysis model based

More information

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk Challenges and Evolution of the LHC Production Grid April 13, 2011 Ian Fisk 1 Evolution Uni x ALICE Remote Access PD2P/ Popularity Tier-2 Tier-2 Uni u Open Lab m Tier-2 Science Uni x Grid Uni z USA Tier-2

More information

The ATLAS Distributed Analysis System

The ATLAS Distributed Analysis System The ATLAS Distributed Analysis System F. Legger (LMU) on behalf of the ATLAS collaboration October 17th, 2013 20th International Conference on Computing in High Energy and Nuclear Physics (CHEP), Amsterdam

More information

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005 Compact Muon Solenoid: Cyberinfrastructure Solutions Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005 Computing Demands CMS must provide computing to handle huge data rates and sizes, and

More information

The ATLAS Tier-3 in Geneva and the Trigger Development Facility

The ATLAS Tier-3 in Geneva and the Trigger Development Facility Journal of Physics: Conference Series The ATLAS Tier-3 in Geneva and the Trigger Development Facility To cite this article: S Gadomski et al 2011 J. Phys.: Conf. Ser. 331 052026 View the article online

More information

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator

Computing. DOE Program Review SLAC. Rainer Bartoldus. Breakout Session 3 June BaBar Deputy Computing Coordinator Computing DOE Program Review SLAC Breakout Session 3 June 2004 Rainer Bartoldus BaBar Deputy Computing Coordinator 1 Outline The New Computing Model (CM2) New Kanga/ROOT event store, new Analysis Model,

More information

Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S)

Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S) Stephen J. Gowdy (CERN) 12 th September 2012 XLDB Conference FINDING THE HIGGS IN THE HAYSTACK(S) Overview Large Hadron Collider (LHC) Compact Muon Solenoid (CMS) experiment The Challenge Worldwide LHC

More information

Overview of ATLAS PanDA Workload Management

Overview of ATLAS PanDA Workload Management Overview of ATLAS PanDA Workload Management T. Maeno 1, K. De 2, T. Wenaus 1, P. Nilsson 2, G. A. Stewart 3, R. Walker 4, A. Stradling 2, J. Caballero 1, M. Potekhin 1, D. Smith 5, for The ATLAS Collaboration

More information

Installation of CMSSW in the Grid DESY Computing Seminar May 17th, 2010 Wolf Behrenhoff, Christoph Wissing

Installation of CMSSW in the Grid DESY Computing Seminar May 17th, 2010 Wolf Behrenhoff, Christoph Wissing Installation of CMSSW in the Grid DESY Computing Seminar May 17th, 2010 Wolf Behrenhoff, Christoph Wissing Wolf Behrenhoff, Christoph Wissing DESY Computing Seminar May 17th, 2010 Page 1 Installation of

More information

PROOF-Condor integration for ATLAS

PROOF-Condor integration for ATLAS PROOF-Condor integration for ATLAS G. Ganis,, J. Iwaszkiewicz, F. Rademakers CERN / PH-SFT M. Livny, B. Mellado, Neng Xu,, Sau Lan Wu University Of Wisconsin Condor Week, Madison, 29 Apr 2 May 2008 Outline

More information

ALICE ANALYSIS PRESERVATION. Mihaela Gheata DASPOS/DPHEP7 workshop

ALICE ANALYSIS PRESERVATION. Mihaela Gheata DASPOS/DPHEP7 workshop 1 ALICE ANALYSIS PRESERVATION Mihaela Gheata DASPOS/DPHEP7 workshop 2 Outline ALICE data flow ALICE analysis Data & software preservation Open access and sharing analysis tools Conclusions 3 ALICE data

More information

Data handling and processing at the LHC experiments

Data handling and processing at the LHC experiments 1 Data handling and processing at the LHC experiments Astronomy and Bio-informatic Farida Fassi CC-IN2P3/CNRS EPAM 2011, Taza, Morocco 2 The presentation will be LHC centric, which is very relevant for

More information

The Run 2 ATLAS Analysis Event Data Model

The Run 2 ATLAS Analysis Event Data Model The Run 2 ATLAS Analysis Event Data Model Marcin Nowak, BNL On behalf of the ATLAS Analysis Software Group and Event Store Group 16 th International workshop on Advanced Computing and Analysis Techniques

More information

HEP replica management

HEP replica management Primary actor Goal in context Scope Level Stakeholders and interests Precondition Minimal guarantees Success guarantees Trigger Technology and data variations Priority Releases Response time Frequency

More information

The Global Grid and the Local Analysis

The Global Grid and the Local Analysis The Global Grid and the Local Analysis Yves Kemp DESY IT GridKA School, 11.9.2008 Overview Global and globalization : Some thoughts Anatomy of an analysis and the computing resources needed Boundary between

More information

CMS Analysis Workflow

CMS Analysis Workflow CMS Analysis Workflow Sudhir Malik Fermilab/University of Nebraska- Lincoln, U.S.A. 1 CMS Software CMS so)ware (CMSSW) based on Event Data Model (EDM) - as event data is processed, products stored in the

More information

Prompt data reconstruction at the ATLAS experiment

Prompt data reconstruction at the ATLAS experiment Prompt data reconstruction at the ATLAS experiment Graeme Andrew Stewart 1, Jamie Boyd 1, João Firmino da Costa 2, Joseph Tuggle 3 and Guillaume Unal 1, on behalf of the ATLAS Collaboration 1 European

More information

Clustering and Reclustering HEP Data in Object Databases

Clustering and Reclustering HEP Data in Object Databases Clustering and Reclustering HEP Data in Object Databases Koen Holtman CERN EP division CH - Geneva 3, Switzerland We formulate principles for the clustering of data, applicable to both sequential HEP applications

More information

Big Data Analytics Tools. Applied to ATLAS Event Data

Big Data Analytics Tools. Applied to ATLAS Event Data Big Data Analytics Tools Applied to ATLAS Event Data Ilija Vukotic University of Chicago CHEP 2016, San Francisco Idea Big Data technologies have proven to be very useful for storage, visualization and

More information

Machine Learning in Data Quality Monitoring

Machine Learning in Data Quality Monitoring CERN openlab workshop on Machine Learning and Data Analytics April 27 th, 2017 Machine Learning in Data Quality Monitoring a point of view Goal Maximize the best Quality Data for physics analysis Data

More information

Monitoring of Computing Resource Use of Active Software Releases at ATLAS

Monitoring of Computing Resource Use of Active Software Releases at ATLAS 1 2 3 4 5 6 Monitoring of Computing Resource Use of Active Software Releases at ATLAS Antonio Limosani on behalf of the ATLAS Collaboration CERN CH-1211 Geneva 23 Switzerland and University of Sydney,

More information

PARALLEL PROCESSING OF LARGE DATA SETS IN PARTICLE PHYSICS

PARALLEL PROCESSING OF LARGE DATA SETS IN PARTICLE PHYSICS PARALLEL PROCESSING OF LARGE DATA SETS IN PARTICLE PHYSICS MARINA ROTARU 1, MIHAI CIUBĂNCAN 1, GABRIEL STOICEA 1 1 Horia Hulubei National Institute for Physics and Nuclear Engineering, Reactorului 30,

More information

Introduction to Distributed HTC and overlay systems

Introduction to Distributed HTC and overlay systems Introduction to Distributed HTC and overlay systems Tuesday morning session Igor Sfiligoi University of California San Diego Logistical reminder It is OK to ask questions - During

More information

STREAMLINING THE DELIVERY, PROTECTION AND MANAGEMENT OF VIRTUAL DESKTOPS. VMware Workstation and Fusion. A White Paper for IT Professionals

STREAMLINING THE DELIVERY, PROTECTION AND MANAGEMENT OF VIRTUAL DESKTOPS. VMware Workstation and Fusion. A White Paper for IT Professionals WHITE PAPER NOVEMBER 2016 STREAMLINING THE DELIVERY, PROTECTION AND MANAGEMENT OF VIRTUAL DESKTOPS VMware Workstation and Fusion A White Paper for IT Professionals Table of Contents Overview 3 The Changing

More information

Computing Model Tier-2 Plans for Germany Relations to GridKa/Tier-1

Computing Model Tier-2 Plans for Germany Relations to GridKa/Tier-1 ATLAS Tier-2 Computing in D GridKa-TAB, Karlsruhe, 30.9.2005 München Computing Model Tier-2 Plans for Germany Relations to GridKa/Tier-1 GridKa-TAB, 30.9.05 1 ATLAS Offline Computing ~Pb/sec PC (2004)

More information

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy Texas A&M Big Data Workshop October 2011 January 2015, Texas A&M University Research Topics Seminar 1 Outline Overview of

More information

UK Tier-2 site evolution for ATLAS. Alastair Dewhurst

UK Tier-2 site evolution for ATLAS. Alastair Dewhurst UK Tier-2 site evolution for ATLAS Alastair Dewhurst Introduction My understanding is that GridPP funding is only part of the story when it comes to paying for a Tier 2 site. Each site is unique. Aim to

More information

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science T. Maeno, K. De, A. Klimentov, P. Nilsson, D. Oleynik, S. Panitkin, A. Petrosyan, J. Schovancova, A. Vaniachine,

More information

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez Scientific data processing at global scale The LHC Computing Grid Chengdu (China), July 5th 2011 Who I am 2 Computing science background Working in the field of computing for high-energy physics since

More information

Summary of the LHC Computing Review

Summary of the LHC Computing Review Summary of the LHC Computing Review http://lhc-computing-review-public.web.cern.ch John Harvey CERN/EP May 10 th, 2001 LHCb Collaboration Meeting The Scale Data taking rate : 50,100, 200 Hz (ALICE, ATLAS-CMS,

More information

Monte Carlo programs

Monte Carlo programs Monte Carlo programs Alexander Khanov PHYS6260: Experimental Methods is HEP Oklahoma State University November 15, 2017 Simulation steps: event generator Input = data cards (program options) this is the

More information

Belle & Belle II. Takanori Hara (KEK) 9 June, 2015 DPHEP Collaboration CERN

Belle & Belle II. Takanori Hara (KEK) 9 June, 2015 DPHEP Collaboration CERN 1 Belle & Belle II Takanori Hara (KEK) takanori.hara@kek.jp 9 June, 2015 DPHEP Collaboration Workshop @ CERN 2 Belle Data Belle : started in 1999, data-taking completed in 2010 still keep analysing the

More information

Data services for LHC computing

Data services for LHC computing Data services for LHC computing SLAC 1 Xavier Espinal on behalf of IT/ST DAQ to CC 8GB/s+4xReco Hot files Reliable Fast Processing DAQ Feedback loop WAN aware Tier-1/2 replica, multi-site High throughout

More information

ATLAS operations in the GridKa T1/T2 Cloud

ATLAS operations in the GridKa T1/T2 Cloud Journal of Physics: Conference Series ATLAS operations in the GridKa T1/T2 Cloud To cite this article: G Duckeck et al 2011 J. Phys.: Conf. Ser. 331 072047 View the article online for updates and enhancements.

More information

Geant4 on Azure using Docker containers

Geant4 on Azure using Docker containers http://www.geant4.org Geant4 on Azure using Docker containers Andrea Dotti (adotti@slac.stanford.edu) ; SD/EPP/Computing 1 Outlook Motivation/overview Docker + G4 Azure + G4 Conclusions 2 Motivation/overview

More information

Popularity Prediction Tool for ATLAS Distributed Data Management

Popularity Prediction Tool for ATLAS Distributed Data Management Popularity Prediction Tool for ATLAS Distributed Data Management T Beermann 1,2, P Maettig 1, G Stewart 2, 3, M Lassnig 2, V Garonne 2, M Barisits 2, R Vigne 2, C Serfon 2, L Goossens 2, A Nairz 2 and

More information

LHC Computing Models

LHC Computing Models LHC Computing Models Commissione I 31/1/2005 Francesco Forti,, Pisa Gruppo di referaggio Forti (chair), Belforte, Menasce, Simone, Taiuti, Ferrari, Morandin, Zoccoli Outline Comparative analysis of the

More information

pyframe Ryan Reece A light-weight Python framework for analyzing ROOT ntuples in ATLAS University of Pennsylvania

pyframe Ryan Reece A light-weight Python framework for analyzing ROOT ntuples in ATLAS University of Pennsylvania pyframe A light-weight Python framework for analyzing ROOT ntuples in ATLAS Ryan Reece University of Pennsylvania ryan.reece@cern.ch Physics Analysis Tools Meeting, CERN September 21, 2011 What is this

More information

Considerations for Mobilizing your Lotus Notes Applications

Considerations for Mobilizing your Lotus Notes Applications Considerations for Mobilizing your Lotus Notes Applications John Kingsley Teamstudio Technical Director Why Mobilize? It all started with email. Not any one email in particular, just the fact that you

More information

ATLAS Analysis Workshop Summary

ATLAS Analysis Workshop Summary ATLAS Analysis Workshop Summary Matthew Feickert 1 1 Southern Methodist University March 29th, 2016 Matthew Feickert (SMU) ATLAS Analysis Workshop Summary March 29th, 2016 1 Outline 1 ATLAS Analysis with

More information

UW-ATLAS Experiences with Condor

UW-ATLAS Experiences with Condor UW-ATLAS Experiences with Condor M.Chen, A. Leung, B.Mellado Sau Lan Wu and N.Xu Paradyn / Condor Week, Madison, 05/01/08 Outline Our first success story with Condor - ATLAS production in 2004~2005. CRONUS

More information

ATLAS NOTE. December 4, ATLAS offline reconstruction timing improvements for run-2. The ATLAS Collaboration. Abstract

ATLAS NOTE. December 4, ATLAS offline reconstruction timing improvements for run-2. The ATLAS Collaboration. Abstract ATLAS NOTE December 4, 2014 ATLAS offline reconstruction timing improvements for run-2 The ATLAS Collaboration Abstract ATL-SOFT-PUB-2014-004 04/12/2014 From 2013 to 2014 the LHC underwent an upgrade to

More information

Enterprise print management in VMware Horizon

Enterprise print management in VMware Horizon Enterprise print management in VMware Horizon Introduction: Embracing and Extending VMware Horizon Tricerat Simplify Printing enhances the capabilities of VMware Horizon environments by enabling reliable

More information

Configuring Short RPO with Actifio StreamSnap and Dedup-Async Replication

Configuring Short RPO with Actifio StreamSnap and Dedup-Async Replication CDS and Sky Tech Brief Configuring Short RPO with Actifio StreamSnap and Dedup-Async Replication Actifio recommends using Dedup-Async Replication (DAR) for RPO of 4 hours or more and using StreamSnap for

More information

Optimizing Parallel Access to the BaBar Database System Using CORBA Servers

Optimizing Parallel Access to the BaBar Database System Using CORBA Servers SLAC-PUB-9176 September 2001 Optimizing Parallel Access to the BaBar Database System Using CORBA Servers Jacek Becla 1, Igor Gaponenko 2 1 Stanford Linear Accelerator Center Stanford University, Stanford,

More information

The CMS Computing Model

The CMS Computing Model The CMS Computing Model Dorian Kcira California Institute of Technology SuperComputing 2009 November 14-20 2009, Portland, OR CERN s Large Hadron Collider 5000+ Physicists/Engineers 300+ Institutes 70+

More information

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF Conference 2017 The Data Challenges of the LHC Reda Tafirout, TRIUMF Outline LHC Science goals, tools and data Worldwide LHC Computing Grid Collaboration & Scale Key challenges Networking ATLAS experiment

More information

TAG Based Skimming In ATLAS

TAG Based Skimming In ATLAS Journal of Physics: Conference Series TAG Based Skimming In ATLAS To cite this article: T Doherty et al 2012 J. Phys.: Conf. Ser. 396 052028 View the article online for updates and enhancements. Related

More information

Veeam with Cohesity Data Platform

Veeam with Cohesity Data Platform Veeam with Cohesity Data Platform Table of Contents About This Guide: 2 Data Protection for VMware Environments: 2 Benefits of using the Cohesity Data Platform with Veeam Backup & Replication: 4 Appendix

More information

Deploying and Using SLIM Suite in a Global Environment

Deploying and Using SLIM Suite in a Global Environment Deploying and Using SLIM Suite in a Global Environment Introduction The SLIM Suite of applications includes SLIM Estimate, SLIM Control, SLIM Metrics, SLIM DataManager and SLIM MasterPlan. Our clients

More information

The INFN Tier1. 1. INFN-CNAF, Italy

The INFN Tier1. 1. INFN-CNAF, Italy IV WORKSHOP ITALIANO SULLA FISICA DI ATLAS E CMS BOLOGNA, 23-25/11/2006 The INFN Tier1 L. dell Agnello 1), D. Bonacorsi 1), A. Chierici 1), M. Donatelli 1), A. Italiano 1), G. Lo Re 1), B. Martelli 1),

More information

ANSE: Advanced Network Services for [LHC] Experiments

ANSE: Advanced Network Services for [LHC] Experiments ANSE: Advanced Network Services for [LHC] Experiments Artur Barczyk California Institute of Technology Joint Techs 2013 Honolulu, January 16, 2013 Introduction ANSE is a project funded by NSF s CC-NIE

More information

Still All on One Server: Perforce at Scale

Still All on One Server: Perforce at Scale Still All on One Server: Perforce at Scale Dan Bloch Senior Site Reliability Engineer Google Inc. June 3, 2011 GOOGLE Google's mission: Organize the world's information and make it universally accessible

More information

One Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool. D. Mason for CMS Software & Computing

One Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool. D. Mason for CMS Software & Computing One Pool To Rule Them All The CMS HTCondor/glideinWMS Global Pool D. Mason for CMS Software & Computing 1 Going to try to give you a picture of the CMS HTCondor/ glideinwms global pool What s the use case

More information

Benchmarking the ATLAS software through the Kit Validation engine

Benchmarking the ATLAS software through the Kit Validation engine Benchmarking the ATLAS software through the Kit Validation engine Alessandro De Salvo (1), Franco Brasolin (2) (1) Istituto Nazionale di Fisica Nucleare, Sezione di Roma, (2) Istituto Nazionale di Fisica

More information

N. Marusov, I. Semenov

N. Marusov, I. Semenov GRID TECHNOLOGY FOR CONTROLLED FUSION: CONCEPTION OF THE UNIFIED CYBERSPACE AND ITER DATA MANAGEMENT N. Marusov, I. Semenov Project Center ITER (ITER Russian Domestic Agency N.Marusov@ITERRF.RU) Challenges

More information

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model

The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model Journal of Physics: Conference Series The evolving role of Tier2s in ATLAS with the new Computing and Data Distribution model To cite this article: S González de la Hoz 2012 J. Phys.: Conf. Ser. 396 032050

More information

The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data

The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data D. Barberis 1*, J. Cranshaw 2, G. Dimitrov 3, A. Favareto 1, Á. Fernández Casaní 4, S. González de la Hoz 4, J.

More information

AUTOMATED RESTORE TESTING FOR TIVOLI STORAGE MANAGER

AUTOMATED RESTORE TESTING FOR TIVOLI STORAGE MANAGER AUTOMATED RESTORE TESTING FOR TIVOLI STORAGE MANAGER TSMworks, Inc. Based in Research Triangle area, NC, USA IBM Advanced Business Partner Big fans of Tivoli Storage Manager Broad experience with Fortune

More information

SUCCESS STORY THE POLYCLINIC THE POLYCLINIC SPEEDS UP ITS VDI ENVIRONMENT WITH NVIDIA GRID

SUCCESS STORY THE POLYCLINIC THE POLYCLINIC SPEEDS UP ITS VDI ENVIRONMENT WITH NVIDIA GRID THE POLYCLINIC SPEEDS UP ITS VDI ENVIRONMENT WITH NVIDIA GRID Updated VDI with NVIDIA GRID wins over staff and increases user density across the organization. was founded in 1917 by a handful of doctors.

More information

Edinburgh (ECDF) Update

Edinburgh (ECDF) Update Edinburgh (ECDF) Update Wahid Bhimji On behalf of the ECDF Team HepSysMan,10 th June 2010 Edinburgh Setup Hardware upgrades Progress in last year Current Issues June-10 Hepsysman Wahid Bhimji - ECDF 1

More information

COMPUTER HARDWARE SOFTWARE. B1 ORIENTATION WORKSHOP Emily Wilson (M5)

COMPUTER HARDWARE SOFTWARE. B1 ORIENTATION WORKSHOP Emily Wilson (M5) COMPUTER HARDWARE SOFTWARE B1 ORIENTATION WORKSHOP Emily Wilson (M5) etwilson@dal.ca COMPUTER HARDWARE THE BIG QUESTION: MAC OR PC? SHORT ANSWER... Whatever you feel most comfortable with LONG ANSWER...

More information

Rok: Decentralized storage for the cloud native world

Rok: Decentralized storage for the cloud native world Whitepaper Rok: Decentralized storage for the cloud native world Cloud native applications and containers are becoming more and more popular, as enterprises realize their benefits: containers are great

More information

LHCb Computing Strategy

LHCb Computing Strategy LHCb Computing Strategy Nick Brook Computing Model 2008 needs Physics software Harnessing the Grid DIRC GNG Experience & Readiness HCP, Elba May 07 1 Dataflow RW data is reconstructed: e.g. Calo. Energy

More information

Embedded Systems Dr. Santanu Chaudhury Department of Electrical Engineering Indian Institute of Technology, Delhi

Embedded Systems Dr. Santanu Chaudhury Department of Electrical Engineering Indian Institute of Technology, Delhi Embedded Systems Dr. Santanu Chaudhury Department of Electrical Engineering Indian Institute of Technology, Delhi Lecture - 13 Virtual memory and memory management unit In the last class, we had discussed

More information

Data Analysis in Experimental Particle Physics

Data Analysis in Experimental Particle Physics Data Analysis in Experimental Particle Physics C. Javier Solano S. Grupo de Física Fundamental Facultad de Ciencias Universidad Nacional de Ingeniería Data Analysis in Particle Physics Outline of Lecture

More information

DQ2 - Data distribution with DQ2 in Atlas

DQ2 - Data distribution with DQ2 in Atlas DQ2 - Data distribution with DQ2 in Atlas DQ2 - A data handling tool Kai Leffhalm DESY March 19, 2008 Technisches Seminar Zeuthen Kai Leffhalm (DESY) DQ2 - Data distribution with DQ2 in Atlas March 19,

More information

DIGITALGLOBE ENHANCES PRODUCTIVITY

DIGITALGLOBE ENHANCES PRODUCTIVITY DIGITALGLOBE ENHANCES PRODUCTIVITY WITH NVIDIA GRID High-performance virtualized desktops transform daily tasks and drastically improve staff efficiency. ABOUT DIGITALGLOBE FIVE REASONS FOR NVIDIA GRID

More information

Author: Steve Gorman Title: Programming with the Intel architecture in the flat memory model

Author: Steve Gorman Title: Programming with the Intel architecture in the flat memory model Author: Steve Gorman Title: Programming with the Intel architecture in the flat memory model Abstract: As the Intel architecture moves off the desktop into a variety of other computing applications, developers

More information

Multi-threaded, discrete event simulation of distributed computing systems

Multi-threaded, discrete event simulation of distributed computing systems Multi-threaded, discrete event simulation of distributed computing systems Iosif C. Legrand California Institute of Technology, Pasadena, CA, U.S.A Abstract The LHC experiments have envisaged computing

More information

StorageCraft OneXafe and Veeam 9.5

StorageCraft OneXafe and Veeam 9.5 TECHNICAL DEPLOYMENT GUIDE NOV 2018 StorageCraft OneXafe and Veeam 9.5 Expert Deployment Guide Overview StorageCraft, with its scale-out storage solution OneXafe, compliments Veeam to create a differentiated

More information

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

LHCb Computing Status. Andrei Tsaregorodtsev CPPM LHCb Computing Status Andrei Tsaregorodtsev CPPM Plan Run II Computing Model Results of the 2015 data processing 2016-2017 outlook Preparing for Run III Conclusions 2 HLT Output Stream Splitting 12.5 khz

More information

(Refer Slide Time: 01:25)

(Refer Slide Time: 01:25) Computer Architecture Prof. Anshul Kumar Department of Computer Science and Engineering Indian Institute of Technology, Delhi Lecture - 32 Memory Hierarchy: Virtual Memory (contd.) We have discussed virtual

More information

CyberStore DSS. Multi Award Winning. Broadberry. CyberStore DSS. Open-E DSS v7 based Storage Appliances. Powering these organisations

CyberStore DSS. Multi Award Winning. Broadberry. CyberStore DSS. Open-E DSS v7 based Storage Appliances. Powering these organisations Multi Award Winning CyberStore DSS CyberStore DSS CyberStore DSS Storage Appliances: Open-E DSS v7 based Storage Appliances The CyberStore DSS range of iscsi SAN / NAS storage appliances include 1U-4U

More information

Hall D and IT. at Internal Review of IT in the 12 GeV Era. Mark M. Ito. May 20, Hall D. Hall D and IT. M. Ito. Introduction.

Hall D and IT. at Internal Review of IT in the 12 GeV Era. Mark M. Ito. May 20, Hall D. Hall D and IT. M. Ito. Introduction. at Internal Review of IT in the 12 GeV Era Mark Hall D May 20, 2011 Hall D in a Nutshell search for exotic mesons in the 1.5 to 2.0 GeV region 12 GeV electron beam coherent bremsstrahlung photon beam coherent

More information

Linux Automation.

Linux Automation. Linux Automation Using Red Hat Enterprise Linux to extract maximum value from IT infrastructure www.redhat.com Table of contents Summary statement Page 3 Background Page 4 Creating a more efficient infrastructure:

More information

DIAL: Distributed Interactive Analysis of Large Datasets

DIAL: Distributed Interactive Analysis of Large Datasets DIAL: Distributed Interactive Analysis of Large Datasets D. L. Adams Brookhaven National Laboratory, Upton NY 11973, USA DIAL will enable users to analyze very large, event-based datasets using an application

More information

Status of KISTI Tier2 Center for ALICE

Status of KISTI Tier2 Center for ALICE APCTP 2009 LHC Physics Workshop at Korea Status of KISTI Tier2 Center for ALICE August 27, 2009 Soonwook Hwang KISTI e-science Division 1 Outline ALICE Computing Model KISTI ALICE Tier2 Center Future Plan

More information

Virtualization. Q&A with an industry leader. Virtualization is rapidly becoming a fact of life for agency executives,

Virtualization. Q&A with an industry leader. Virtualization is rapidly becoming a fact of life for agency executives, Virtualization Q&A with an industry leader Virtualization is rapidly becoming a fact of life for agency executives, as the basis for data center consolidation and cloud computing and, increasingly, as

More information

The LHC Computing Grid

The LHC Computing Grid The LHC Computing Grid Visit of Finnish IT Centre for Science CSC Board Members Finland Tuesday 19 th May 2009 Frédéric Hemmer IT Department Head The LHC and Detectors Outline Computing Challenges Current

More information

An SQL-based approach to physics analysis

An SQL-based approach to physics analysis Journal of Physics: Conference Series OPEN ACCESS An SQL-based approach to physics analysis To cite this article: Dr Maaike Limper 2014 J. Phys.: Conf. Ser. 513 022022 View the article online for updates

More information

Memory Hierarchy. Mehran Rezaei

Memory Hierarchy. Mehran Rezaei Memory Hierarchy Mehran Rezaei What types of memory do we have? Registers Cache (Static RAM) Main Memory (Dynamic RAM) Disk (Magnetic Disk) Option : Build It Out of Fast SRAM About 5- ns access Decoders

More information

Using Automated Network Management at Fiserv. June 2012

Using Automated Network Management at Fiserv. June 2012 Using Automated Network Management at Fiserv June 2012 Brought to you by Join Group Vivit Network Automation Special Interest Group (SIG) Leaders: Chris Powers & Wendy Wheeler Your input is welcomed on

More information

Best Practices for Alert Tuning. This white paper will provide best practices for alert tuning to ensure two related outcomes:

Best Practices for Alert Tuning. This white paper will provide best practices for alert tuning to ensure two related outcomes: This white paper will provide best practices for alert tuning to ensure two related outcomes: 1. Monitoring is in place to catch critical conditions and alert the right people 2. Noise is reduced and people

More information

CMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster

CMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster CMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster Vaikunth Thukral Department of Physics and Astronomy Texas A&M University 1 Outline Grid Computing with CMS:

More information

Event Displays and LArg

Event Displays and LArg Event Displays and LArg Columbia U. / Nevis Labs Slide 1 Introduction Displays are needed in various ways at different stages of the experiment: Software development: understanding offline & trigger algorithms

More information

Full Website Audit. Conducted by Mathew McCorry. Digimush.co.uk

Full Website Audit. Conducted by Mathew McCorry. Digimush.co.uk Full Website Audit Conducted by Mathew McCorry Digimush.co.uk 1 Table of Contents Full Website Audit 1 Conducted by Mathew McCorry... 1 1. Overview... 3 2. Technical Issues... 4 2.1 URL Structure... 4

More information

The MOSIX Scalable Cluster Computing for Linux. mosix.org

The MOSIX Scalable Cluster Computing for Linux.  mosix.org The MOSIX Scalable Cluster Computing for Linux Prof. Amnon Barak Computer Science Hebrew University http://www. mosix.org 1 Presentation overview Part I : Why computing clusters (slide 3-7) Part II : What

More information

Web Hosting. Important features to consider

Web Hosting. Important features to consider Web Hosting Important features to consider Amount of Storage When choosing your web hosting, one of your primary concerns will obviously be How much data can I store? For most small and medium web sites,

More information

Map-Reduce. Marco Mura 2010 March, 31th

Map-Reduce. Marco Mura 2010 March, 31th Map-Reduce Marco Mura (mura@di.unipi.it) 2010 March, 31th This paper is a note from the 2009-2010 course Strumenti di programmazione per sistemi paralleli e distribuiti and it s based by the lessons of

More information

Session 7: Configuration Manager

Session 7: Configuration Manager Session 7: Configuration Manager Mark Aslett Consultant Adam Shepherd Consultant MCS Talks Infrastructure Architecture Agenda Introduction Gathering requirements Core Concepts Hierarchy Design Scaling

More information

VMWARE EBOOK. Easily Deployed Software-Defined Storage: A Customer Love Story

VMWARE EBOOK. Easily Deployed Software-Defined Storage: A Customer Love Story VMWARE EBOOK Easily Deployed Software-Defined Storage: A Customer Love Story TABLE OF CONTENTS The Software-Defined Data Center... 1 VMware Virtual SAN... 3 A Proven Enterprise Platform... 4 Proven Results:

More information

The ATLAS Production System

The ATLAS Production System The ATLAS MC and Data Rodney Walker Ludwig Maximilians Universität Munich 2nd Feb, 2009 / DESY Computing Seminar Outline 1 Monte Carlo Production Data 2 3 MC Production Data MC Production Data Group and

More information

Lecture#16: VM, thrashing, Replacement, Cache state

Lecture#16: VM, thrashing, Replacement, Cache state Lecture#16: VM, thrashing, Replacement, Cache state Review -- 1 min Multi-level translation tree: multi-level page table, paged paging, paged segmentation, hash table: inverted page table combination:

More information

Monitoring Tool Made to Measure for SharePoint Admins. By Stacy Simpkins

Monitoring Tool Made to Measure for SharePoint Admins. By Stacy Simpkins Monitoring Tool Made to Measure for SharePoint Admins By Stacy Simpkins Contents About the Author... 3 Introduction... 4 Who s it for and what all can it do?... 4 SysKit Insights Features... 6 Drillable

More information