Data Analysis in Experimental Particle Physics

Size: px
Start display at page:

Download "Data Analysis in Experimental Particle Physics"

Transcription

1 Data Analysis in Experimental Particle Physics C. Javier Solano S. Grupo de Física Fundamental Facultad de Ciencias Universidad Nacional de Ingeniería

2 Data Analysis in Particle Physics Outline of Lecture Characteristics of data from particle experiments From DAQ data to Event Records: Event Building From hits to tracks and clusters From tracks and clusters to particles : Correlating sub-detector information Uncertainties and resolution Data reconstruction and production : Data Summary Tapes Personal data analysis: n-tuples

3 Data Analysis in Particle Physics Outline of Lecture (cont.) Monte Carlo simulation Statistics and error analysis Hypothesis testing Simulation of particle production and interactions with the detector Digital representations of event data Monitoring and Calibration Why physicists don t (yet) use Excel and Oracle for their daily analysis. The challenge of analysis for the LHC experiments The challenge of computing for the LHC Solving the LHC computing challenge

4 Characteristics of data from particle experiments

5 Characteristics of data from particle experiments Most data comes from digitized information from sensors activated by particles crossing them. We call the data resulting from the observation of a particle collision an event. During hours, days, weeks, months, years or even decades, we observe many events.. We group them according to the time- varying experimental conditions into runs. Calibration and environmental information is also stored, usually in a periodic fashion. For practical reasons, this data is stored in data files of many events. Almost always, events are independent from each other.

6 Characteristics of data from particle experiments The Experimental Particle Physics Data Worm Data file 418 Data file 419 Run 137 Run 138 Run 139 Run 140 Calibration records Event number 31896

7 From DAQ data to Event Records Event Building

8 From hits to tracks and clusters

9 From hits to tracks and clusters Occupancy and point resolution are related to ambiguities in track finding

10 From hits to tracks and clusters Calibration, monitoring and software are needed to resolve these ambiguities

11 From hits to tracks and clusters What you see is not always what there was! Nuclear interaction

12 Monitoring and Calibration Particles deposit energy in sensors Sensors give Voltages, Currents, Charges Space position of sensor is known On-detector Analog-to-Digital Converters change these into numbers representing these or other quantities (for example clock-ticks between V pulses) Calibration establishes the relationship between the ADC units and the physical units (ev, {x,y,z}, ns)

13 From tracks and clusters to particles Correlating sub-detector information

14 Uncertainties and resolution Each measurement or hit has some uncertainty,, due to alignment and the characteristic of the sensor. These uncertainties get propagated, often in a non- linear manner, to resolution functions for the physics quantities used in analysis. Resolution has various consequences: Direct on measurements Signal-Background confusion Combinatorics

15 Data reconstruction and production : Data Summary Tapes Reconstruction turns hits+calibration+geometry into particle hypothesis Reconstruction is time consuming and must be made coherently Centrally organized production Output is one or more levels of so-called Data Summary Tapes (DST( DST) ) which are used as input to Personal Analysis In practice, there is a lot of utility software to organize these data for easy analysis (bookkeeping( bookkeeping) Programming of complicated event structures Old: FORTRAN with home-made memory managers Today: Object-Oriented design using C++ or Java

16 Personal data analysis Most modern detectors can address multiple physics topics. Hundreds or thousands of professors and students distributed around the world. Modern experimental collaborations are early example of virtual communities. Historical enablers for virtual communities: Fellowship and exchange programmes Telegraph, telex, telephone and telefax National and International Laboratories Reasonably priced airline tickets Computer inter-networking, and ftp The World Wide Web Multi-media applications on the Internet

17 Personal data analysis Today, physics analysis topics are increasingly tackled by virtual teams within these virtual communities. Must maintain coherency of data and algorithms within the virtual team. Production for a modern detector is very complex and consumes many resources. DST contains all imagined reconstruction objects for all foreseen analysis,, so they are big. Handling a DST often requires installation of special software libraries and writing code in reconstruction dialect.

18 Personal data analysis Solution: Each virtual team develops a code to extract a common analysis dataset for a given topic which is written and manipulated using a lingua franca : n-tuples and the Physics Analysis Workstation (PAW)/ROOT Physicist s version of business data mining with Excel Iterative process (time-scale of weeks or months): Team agrees on complex algorithms to be coded in the extraction program. Algorithms coded and tested, extraction from DST. n-tuple file is rapidly distributed via computer network. n-tuple is analyzed using non-compiled platform- independent code (PAW/ROOT macros today, Java in future?) that are easily modified and shared by . Eventually limitations are reached, go back to step 1.

19 Personal data analysis PAW was the killer application for physics in the 90s Interactive, just as powerful workstations became available Platform independent, in a very diverse workstation world Graphical, just as X-windows gave graphics over network Simple to write analysis macros, just as the complexity of FORTRAN programming required in experiments decoupled most of the collaborators from the experiment s code. In summary, PAW was like going from DOS to Macintosh. One major limitation of PAW is the lack of variable length structures or more generally data objects. ROOT overcomes these limitations keeping a similar philosophy as PAW. Java Analysis Studio tries to go further with agents.

20 Personal data analysis Which will be the killer application for LHC analysis? Is a Mac Classic on Appletalk enough or do we need the conceptual leap equivalent of Web + Java-enabled browser? Will the personal n-tuple model work for LHC? Do we need and can we afford to support our own interactive data analysis tool? Will one of the newer tools, such as Java Analysis Studio, go exponential in the open source world? Many questions, one simple answer: It will be young people like you who will make the next step happen.

21 Monte Carlo simulation Monte Carlo simulation uses random numbers ( mathematics textbooks) Try the following: Find a source of random numbers in the interval [0,1] (calculator, Excel, etc.) Take a function that you want to simulate (e.g. y=x 2 ) and normalize it to fit in the interval [0,1] for both x and y. Find graph paper to histogram values of x Repeat this at least 20 times: Throw two random numbers. Use first as value for x Evaluate the function y and compare its value to 2 nd random number If function value is less than random number, add a count to histogram in the correct bin for x If function value is more than random number, forget it If function value is more than random number, forget it Compare your histogram to the shape of the function

22 Monte Carlo simulation If you don t know how to program, you can pick up an Excel file from Here is the result for 100 trials: Note there are entries so the 8 efficiency is 30% 7 6 Note the statistical 5 fluctuations 4 Homework: How is the 3 2 normalization done? y 1 Example of Monte Carlo simulation of y=x*x x Y SAMPLE

23 Statistics and error analysis Analysis involves selecting, counting and normalizing. Things are easier when you actually have a signal. Understand underlying statistics: Poisson, Binomial,Multinomial, etc. If measuring a differential distribution, understand relation between normalization of binned counts vs. total counts. Understand selection biases and their impact on observed distributions. Things are a lot harder when you place limits. Two observations: If you cannot make an analytical estimate of the uncertainties, I won t believe your result. The expression n-sigma effect should be banned.

24 Hypothesis testing You must understand Bayes theorem. And every time you think you understand it, you must make a big effort to understand it better! Compare differential distributions of data with predictions of theory or model Different theories Different parameters for same model Setting up the statistical test is often straight-forward, which is why it is surprising most people do it wrong Taking account of resolution and systematic uncertainties is hard Make simulation look like data to get your answers Even if graphics looks better the other way around!!!

25 Simulation of particle production and interactions with the detector For particle production, combine Monte Carlo with Detailed particle properties Detailed cross-sections predicted by theory of phenomenology Computation of phase-space Output consists of event records containing simulated particles (often called 4-vectors by experimentalists) For simulating the detector, combine MC with Detailed description of the detector Detailed cross-sections for interaction with detector materials Detailed phenomenology of mechanism producing signal Transport (Ray-tracing) algorithms including B fields Digitization model mapping of {x,y,z} to read-out channel

26 Simulation of particle production and interactions with the detector Example: Small part of design of GEANT4 Reference to Jackson s textboook in documentation!

27 Digital representations of event data In principle, representing event data digitally should be very simple, except: everything comes in variable numbers: hits, tracks, clusters ambiguities lead to multiple relations particle identification may depend on analysis hypothesis etc. In simple terms, events don t look like bank account data, they look like collections of objects. You can do a reasonable representation using relational tables, but actually using the data structures from Fortran/ROOT programs is still cumbersome Object Oriented Programming is a better match, but C++ does not resolve all problems Frameworks

28 Why physicists don t (yet) use Excel and Oracle for their daily analysis. Spreadsheets like Excel and relational databases like Oracle have a very square view of data. This is not a good match to the Data Worm. Normal people (banks and insurance companies) can define a priori the quantities that they will select on (the( keys of the database). We usually derive selection criteria a posteriori using quantities calculated from the stored data. We like (need?) to express queries as individualistic detailed low-level computer codes.. Difficult to support in database. But this is changing very rapidly due to Data Mining: Businesses are interested in analyzing their raw data in unpredictable ways. Example: Cash register tickets to choose sale items Support for this requires a more organic view of data, for example object-relational databases.

29 Why physicists don t (yet) use Excel and Oracle for their daily analysis. Idealized Cluster Calorimeter hit Particle hypothesis Mass Charge Momentum Origin Simple relation One to Many One to Many Position Width Depth Energy Number of hits Track Origin Curvature Extrapolation Number of hits One to Many One to Many Position Response Tracker hit Position Response

30 Why physicists don t (yet) use Excel and Oracle for their daily analysis. Reality Cluster Calorimeter hit Particle hypothesis Mass Charge Momentum Origin Complicated algorithmic relation Many to Many Many to Many Position Width Depth Energy Number of hits Track Origin Curvature Extrapolation Number of hits Many to Many Many to Many Position Response Tracker hit Position Response

31 The challenge of analysis for the LHC experiments

32 The challenge of analysis for the LHC experiments 1:10 12 Online 1:10 7 Analysis 1:10 5

33 The challenge of analysis for the LHC experiments

34 The challenge of analysis for the LHC experiments Detector 0.1 to 1 GB/sec Raw data 35K SI95 Event Filter (selection & reconstruction) 1 PB / year 500 TB One Experiment Event Summary Data ~200 MB/sec 350K SI95 64 GB/sec Batch Physics Analysis ~100 MB/sec Event analysis objects Reconstruction 250K SI95 Event Simulation Thousands of scientists distributed around the planet

35 The challenge of computing for the LHC TeraBytes Long Term Tape Storage Estimates 14'000 12'000 10'000 8'000 6'000 4'000 2'000 0 Current Experiments COMPASS LHC Year

36 The challenge of computing for the LHC TeraBytes 14'000 12'000 10'000 8'000 6'000 4'000 2' Long Term Tape Storage Estimates Current Experiments Accumulation: 10 PB/year Signal/Background up to 1: COMPASS LHC Year

37 The challenge of computing for the LHC K SI95 5,000 4,000 3,000 2,000 1,000 Estimated CPU Capacity required at CERN LHC Moore s law some measure of the capacity technology advances provide for a constant number of processors or investment Jan 2000: 3.5K SI95

38 Thousands of CERN Units The challenge of computing for the LHC CERN Centre Physics Computing Capacity 40 Mainframes decomissioned First PC services CERN RD47 project RISC decomissioning agreed Moore's law (based on 1988) year

39 Thousands of CERN Units The challenge of computing for the LHC CERN Centre Physics Computing Capacity 40 Mainframes decomissioned First PC services CERN RD47 project RISC decomissioning agreed Continued innovation Moore's law (based on 1988) year

40 Solving the LHC Computing Challenge: Technology Development Domains DEVELOPER VIEW FABRIC GRID APPLICATION USER VIEW

41 Solving the LHC Computing Challenge Storage Network * Thousand dual-cpu boxes Hundreds of tape drives 24 * Farm Network * Data Rate in Gbps 960 * LAN-WAN Routers Storage Network 0.8 Real-time detector data Grid Interface 10 Thousand disk units Computing fabric at CERN (2006)

42 Internet Protocol Architecture Solving the LHC Computing Challenge: Data-Intensive Grid Research Grid Protocol Architecture Specialized services : user- or appln-specific distributed services Application User Managing multiple resources : ubiquitous infrastructure services Sharing single resources : negotiating access, controlling use Talking to things : communication (Internet protocols) & security Controlling things locally : Access to, & control of, resources Collective Resource Connectivity Fabric Application Transport Internet Link

43 Acknowledgements Many of the figures in this talk are from the Web sites of ATLAS, CMS, Aleph and Delphi.

Monte Carlo programs

Monte Carlo programs Monte Carlo programs Alexander Khanov PHYS6260: Experimental Methods is HEP Oklahoma State University November 15, 2017 Simulation steps: event generator Input = data cards (program options) this is the

More information

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez

Scientific data processing at global scale The LHC Computing Grid. fabio hernandez Scientific data processing at global scale The LHC Computing Grid Chengdu (China), July 5th 2011 Who I am 2 Computing science background Working in the field of computing for high-energy physics since

More information

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF

Conference The Data Challenges of the LHC. Reda Tafirout, TRIUMF Conference 2017 The Data Challenges of the LHC Reda Tafirout, TRIUMF Outline LHC Science goals, tools and data Worldwide LHC Computing Grid Collaboration & Scale Key challenges Networking ATLAS experiment

More information

8.882 LHC Physics. Track Reconstruction and Fitting. [Lecture 8, March 2, 2009] Experimental Methods and Measurements

8.882 LHC Physics. Track Reconstruction and Fitting. [Lecture 8, March 2, 2009] Experimental Methods and Measurements 8.882 LHC Physics Experimental Methods and Measurements Track Reconstruction and Fitting [Lecture 8, March 2, 2009] Organizational Issues Due days for the documented analyses project 1 is due March 12

More information

Lecture 1: January 23

Lecture 1: January 23 CMPSCI 677 Distributed and Operating Systems Spring 2019 Lecture 1: January 23 Lecturer: Prashant Shenoy Scribe: Jonathan Westin (2019), Bin Wang (2018) 1.1 Introduction to the course The lecture started

More information

Andrea Sciabà CERN, Switzerland

Andrea Sciabà CERN, Switzerland Frascati Physics Series Vol. VVVVVV (xxxx), pp. 000-000 XX Conference Location, Date-start - Date-end, Year THE LHC COMPUTING GRID Andrea Sciabà CERN, Switzerland Abstract The LHC experiments will start

More information

Summary of the LHC Computing Review

Summary of the LHC Computing Review Summary of the LHC Computing Review http://lhc-computing-review-public.web.cern.ch John Harvey CERN/EP May 10 th, 2001 LHCb Collaboration Meeting The Scale Data taking rate : 50,100, 200 Hz (ALICE, ATLAS-CMS,

More information

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities

DESY at the LHC. Klaus Mőnig. On behalf of the ATLAS, CMS and the Grid/Tier2 communities DESY at the LHC Klaus Mőnig On behalf of the ATLAS, CMS and the Grid/Tier2 communities A bit of History In Spring 2005 DESY decided to participate in the LHC experimental program During summer 2005 a group

More information

HPS Data Analysis Group Summary. Matt Graham HPS Collaboration Meeting June 6, 2013

HPS Data Analysis Group Summary. Matt Graham HPS Collaboration Meeting June 6, 2013 HPS Data Analysis Group Summary Matt Graham HPS Collaboration Meeting June 6, 2013 Data Analysis 101 define what you want to measure get data to disk (by magic or whatever) select subset of data to optimize

More information

Data Analysis in ATLAS. Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser

Data Analysis in ATLAS. Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser Data Analysis in ATLAS Graeme Stewart with thanks to Attila Krasznahorkay and Johannes Elmsheuser 1 ATLAS Data Flow into Analysis RAW detector data and simulated RDO data are reconstructed into our xaod

More information

Lecture 1: January 22

Lecture 1: January 22 CMPSCI 677 Distributed and Operating Systems Spring 2018 Lecture 1: January 22 Lecturer: Prashant Shenoy Scribe: Bin Wang 1.1 Introduction to the course The lecture started by outlining the administrative

More information

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback

Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy. David Toback Big Computing and the Mitchell Institute for Fundamental Physics and Astronomy Texas A&M Big Data Workshop October 2011 January 2015, Texas A&M University Research Topics Seminar 1 Outline Overview of

More information

Starting a Data Analysis

Starting a Data Analysis 03/20/07 PHY310: Statistical Data Analysis 1 PHY310: Lecture 17 Starting a Data Analysis Road Map Your Analysis Log Exploring the Data Reading the input file (and making sure it's right) Taking a first

More information

Atlantis: Visualization Tool in Particle Physics

Atlantis: Visualization Tool in Particle Physics Atlantis: Visualization Tool in Particle Physics F.J.G.H. Crijns 2, H. Drevermann 1, J.G. Drohan 3, E. Jansen 2, P.F. Klok 2, N. Konstantinidis 3, Z. Maxa 3, D. Petrusca 1, G. Taylor 4, C. Timmermans 2

More information

8.882 LHC Physics. Analysis Tips. [Lecture 9, March 4, 2009] Experimental Methods and Measurements

8.882 LHC Physics. Analysis Tips. [Lecture 9, March 4, 2009] Experimental Methods and Measurements 8.882 LHC Physics Experimental Methods and Measurements Analysis Tips [Lecture 9, March 4, 2009] Physics Colloquium Series 09 The Physics Colloquium Series Thursday, March 5 at 4:15 pm in room 10-250 Spring

More information

CSCS CERN videoconference CFD applications

CSCS CERN videoconference CFD applications CSCS CERN videoconference CFD applications TS/CV/Detector Cooling - CFD Team CERN June 13 th 2006 Michele Battistin June 2006 CERN & CFD Presentation 1 TOPICS - Some feedback about already existing collaboration

More information

The CMS Computing Model

The CMS Computing Model The CMS Computing Model Dorian Kcira California Institute of Technology SuperComputing 2009 November 14-20 2009, Portland, OR CERN s Large Hadron Collider 5000+ Physicists/Engineers 300+ Institutes 70+

More information

Virtualizing a Batch. University Grid Center

Virtualizing a Batch. University Grid Center Virtualizing a Batch Queuing System at a University Grid Center Volker Büge (1,2), Yves Kemp (1), Günter Quast (1), Oliver Oberst (1), Marcel Kunze (2) (1) University of Karlsruhe (2) Forschungszentrum

More information

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008

CERN openlab II. CERN openlab and. Sverre Jarp CERN openlab CTO 16 September 2008 CERN openlab II CERN openlab and Intel: Today and Tomorrow Sverre Jarp CERN openlab CTO 16 September 2008 Overview of CERN 2 CERN is the world's largest particle physics centre What is CERN? Particle physics

More information

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010

Worldwide Production Distributed Data Management at the LHC. Brian Bockelman MSST 2010, 4 May 2010 Worldwide Production Distributed Data Management at the LHC Brian Bockelman MSST 2010, 4 May 2010 At the LHC http://op-webtools.web.cern.ch/opwebtools/vistar/vistars.php?usr=lhc1 Gratuitous detector pictures:

More information

Deep Learning Photon Identification in a SuperGranular Calorimeter

Deep Learning Photon Identification in a SuperGranular Calorimeter Deep Learning Photon Identification in a SuperGranular Calorimeter Nikolaus Howe Maurizio Pierini Jean-Roch Vlimant @ Williams College @ CERN @ Caltech 1 Outline Introduction to the problem What is Machine

More information

Early experience with the Run 2 ATLAS analysis model

Early experience with the Run 2 ATLAS analysis model Early experience with the Run 2 ATLAS analysis model Argonne National Laboratory E-mail: cranshaw@anl.gov During the long shutdown of the LHC, the ATLAS collaboration redesigned its analysis model based

More information

N. Marusov, I. Semenov

N. Marusov, I. Semenov GRID TECHNOLOGY FOR CONTROLLED FUSION: CONCEPTION OF THE UNIFIED CYBERSPACE AND ITER DATA MANAGEMENT N. Marusov, I. Semenov Project Center ITER (ITER Russian Domestic Agency N.Marusov@ITERRF.RU) Challenges

More information

Multi-threaded, discrete event simulation of distributed computing systems

Multi-threaded, discrete event simulation of distributed computing systems Multi-threaded, discrete event simulation of distributed computing systems Iosif C. Legrand California Institute of Technology, Pasadena, CA, U.S.A Abstract The LHC experiments have envisaged computing

More information

The TDAQ Analytics Dashboard: a real-time web application for the ATLAS TDAQ control infrastructure

The TDAQ Analytics Dashboard: a real-time web application for the ATLAS TDAQ control infrastructure The TDAQ Analytics Dashboard: a real-time web application for the ATLAS TDAQ control infrastructure Giovanna Lehmann Miotto, Luca Magnoni, John Erik Sloper European Laboratory for Particle Physics (CERN),

More information

The CMS data quality monitoring software: experience and future prospects

The CMS data quality monitoring software: experience and future prospects The CMS data quality monitoring software: experience and future prospects Federico De Guio on behalf of the CMS Collaboration CERN, Geneva, Switzerland E-mail: federico.de.guio@cern.ch Abstract. The Data

More information

Insight: that s for NSA Decision making: that s for Google, Facebook. so they find the best way to push out adds and products

Insight: that s for NSA Decision making: that s for Google, Facebook. so they find the best way to push out adds and products What is big data? Big data is high-volume, high-velocity and high-variety information assets that demand cost-effective, innovative forms of information processing for enhanced insight and decision making.

More information

IEPSAS-Kosice: experiences in running LCG site

IEPSAS-Kosice: experiences in running LCG site IEPSAS-Kosice: experiences in running LCG site Marian Babik 1, Dusan Bruncko 2, Tomas Daranyi 1, Ladislav Hluchy 1 and Pavol Strizenec 2 1 Department of Parallel and Distributed Computing, Institute of

More information

Klaus Dehmelt EIC Detector R&D Weekly Meeting November 28, 2011 GEM SIMULATION FRAMEWORK

Klaus Dehmelt EIC Detector R&D Weekly Meeting November 28, 2011 GEM SIMULATION FRAMEWORK Klaus Dehmelt EIC Detector R&D Weekly Meeting November 28, 2011 GEM SIMULATION FRAMEWORK Overview GEM Simulation Framework in the context of Simulation Studies for a High Resolution Time Projection Chamber

More information

Reliability Engineering Analysis of ATLAS Data Reprocessing Campaigns

Reliability Engineering Analysis of ATLAS Data Reprocessing Campaigns Journal of Physics: Conference Series OPEN ACCESS Reliability Engineering Analysis of ATLAS Data Reprocessing Campaigns To cite this article: A Vaniachine et al 2014 J. Phys.: Conf. Ser. 513 032101 View

More information

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN

Software and computing evolution: the HL-LHC challenge. Simone Campana, CERN Software and computing evolution: the HL-LHC challenge Simone Campana, CERN Higgs discovery in Run-1 The Large Hadron Collider at CERN We are here: Run-2 (Fernando s talk) High Luminosity: the HL-LHC challenge

More information

v.m.g.rajasekaran ramani sri sarada sakthi mat. Hr. sec. school

v.m.g.rajasekaran ramani sri sarada sakthi mat. Hr. sec. school v.m.g.rajasekaran ramani sri sarada sakthi mat. Hr. sec. school STD: XI COMPUTER SCIENCE L.1. Introduction to computers 2 marks 1. What is a computer? Computer: A computer is an electronic machine, capable

More information

Storage on the Lunatic Fringe. Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium

Storage on the Lunatic Fringe. Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium Storage on the Lunatic Fringe Thomas M. Ruwart University of Minnesota Digital Technology Center Intelligent Storage Consortium tmruwart@dtc.umn.edu Orientation Who are the lunatics? What are their requirements?

More information

Upgraded Swimmer for Computationally Efficient Particle Tracking for Jefferson Lab s CLAS12 Spectrometer

Upgraded Swimmer for Computationally Efficient Particle Tracking for Jefferson Lab s CLAS12 Spectrometer Upgraded Swimmer for Computationally Efficient Particle Tracking for Jefferson Lab s CLAS12 Spectrometer Lydia Lorenti Advisor: David Heddle April 29, 2018 Abstract The CLAS12 spectrometer at Jefferson

More information

INTRODUCTION TO THE ANAPHE/LHC++ SOFTWARE SUITE

INTRODUCTION TO THE ANAPHE/LHC++ SOFTWARE SUITE INTRODUCTION TO THE ANAPHE/LHC++ SOFTWARE SUITE Andreas Pfeiffer CERN, Geneva, Switzerland Abstract The Anaphe/LHC++ project is an ongoing effort to provide an Object-Oriented software environment for

More information

UW-ATLAS Experiences with Condor

UW-ATLAS Experiences with Condor UW-ATLAS Experiences with Condor M.Chen, A. Leung, B.Mellado Sau Lan Wu and N.Xu Paradyn / Condor Week, Madison, 05/01/08 Outline Our first success story with Condor - ATLAS production in 2004~2005. CRONUS

More information

The LHC Computing Grid Project in Spain (LCG-ES) Presentation to RECFA

The LHC Computing Grid Project in Spain (LCG-ES) Presentation to RECFA * Collaboration of the Generalitat de Catalunya, IFAE and Universitat Autònoma de Barcelona Port d The LHC Computing Grid Project in Spain (LCG-ES) Presentation to RECFA Prof. Manuel Delfino Coordinating

More information

A New Segment Building Algorithm for the Cathode Strip Chambers in the CMS Experiment

A New Segment Building Algorithm for the Cathode Strip Chambers in the CMS Experiment EPJ Web of Conferences 108, 02023 (2016) DOI: 10.1051/ epjconf/ 201610802023 C Owned by the authors, published by EDP Sciences, 2016 A New Segment Building Algorithm for the Cathode Strip Chambers in the

More information

CERN and Scientific Computing

CERN and Scientific Computing CERN and Scientific Computing Massimo Lamanna CERN Information Technology Department Experiment Support Group 1960: 26 GeV proton in the 32 cm CERN hydrogen bubble chamber 1960: IBM 709 at the Geneva airport

More information

CMS Simulation Software

CMS Simulation Software CMS Simulation Software Dmitry Onoprienko Kansas State University on behalf of the CMS collaboration 10th Topical Seminar on Innovative Particle and Radiation Detectors 1-5 October 2006. Siena, Italy Simulation

More information

Hall D and IT. at Internal Review of IT in the 12 GeV Era. Mark M. Ito. May 20, Hall D. Hall D and IT. M. Ito. Introduction.

Hall D and IT. at Internal Review of IT in the 12 GeV Era. Mark M. Ito. May 20, Hall D. Hall D and IT. M. Ito. Introduction. at Internal Review of IT in the 12 GeV Era Mark Hall D May 20, 2011 Hall D in a Nutshell search for exotic mesons in the 1.5 to 2.0 GeV region 12 GeV electron beam coherent bremsstrahlung photon beam coherent

More information

Detector Control LHC

Detector Control LHC Detector Control Systems @ LHC Matthias Richter Department of Physics, University of Oslo IRTG Lecture week Autumn 2012 Oct 18 2012 M. Richter (UiO) DCS @ LHC Oct 09 2012 1 / 39 Detectors in High Energy

More information

Electron and Photon Reconstruction and Identification with the ATLAS Detector

Electron and Photon Reconstruction and Identification with the ATLAS Detector Electron and Photon Reconstruction and Identification with the ATLAS Detector IPRD10 S12 Calorimetry 7th-10th June 2010 Siena, Italy Marine Kuna (CPPM/IN2P3 Univ. de la Méditerranée) on behalf of the ATLAS

More information

ISTITUTO NAZIONALE DI FISICA NUCLEARE

ISTITUTO NAZIONALE DI FISICA NUCLEARE ISTITUTO NAZIONALE DI FISICA NUCLEARE Sezione di Perugia INFN/TC-05/10 July 4, 2005 DESIGN, IMPLEMENTATION AND CONFIGURATION OF A GRID SITE WITH A PRIVATE NETWORK ARCHITECTURE Leonello Servoli 1,2!, Mirko

More information

Storage and I/O requirements of the LHC experiments

Storage and I/O requirements of the LHC experiments Storage and I/O requirements of the LHC experiments Sverre Jarp CERN openlab, IT Dept where the Web was born 22 June 2006 OpenFabrics Workshop, Paris 1 Briefly about CERN 22 June 2006 OpenFabrics Workshop,

More information

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010

Physics Computing at CERN. Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010 Physics Computing at CERN Helge Meinhard CERN, IT Department OpenLab Student Lecture 27 July 2010 Location Building 513 (opposite of restaurant no. 2) Building Large building with 2700 m 2 surface for

More information

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk

Challenges and Evolution of the LHC Production Grid. April 13, 2011 Ian Fisk Challenges and Evolution of the LHC Production Grid April 13, 2011 Ian Fisk 1 Evolution Uni x ALICE Remote Access PD2P/ Popularity Tier-2 Tier-2 Uni u Open Lab m Tier-2 Science Uni x Grid Uni z USA Tier-2

More information

First LHCb measurement with data from the LHC Run 2

First LHCb measurement with data from the LHC Run 2 IL NUOVO CIMENTO 40 C (2017) 35 DOI 10.1393/ncc/i2017-17035-4 Colloquia: IFAE 2016 First LHCb measurement with data from the LHC Run 2 L. Anderlini( 1 )ands. Amerio( 2 ) ( 1 ) INFN, Sezione di Firenze

More information

Machine Learning in Data Quality Monitoring

Machine Learning in Data Quality Monitoring CERN openlab workshop on Machine Learning and Data Analytics April 27 th, 2017 Machine Learning in Data Quality Monitoring a point of view Goal Maximize the best Quality Data for physics analysis Data

More information

CERN s Business Computing

CERN s Business Computing CERN s Business Computing Where Accelerated the infinitely by Large Pentaho Meets the Infinitely small Jan Janke Deputy Group Leader CERN Administrative Information Systems Group CERN World s Leading Particle

More information

DIAL: Distributed Interactive Analysis of Large Datasets

DIAL: Distributed Interactive Analysis of Large Datasets DIAL: Distributed Interactive Analysis of Large Datasets D. L. Adams Brookhaven National Laboratory, Upton NY 11973, USA DIAL will enable users to analyze very large, event-based datasets using an application

More information

Machine Learning for (fast) simulation

Machine Learning for (fast) simulation Machine Learning for (fast) simulation Sofia Vallecorsa for the GeantV team CERN, April 2017 1 Monte Carlo Simulation: Why Detailed simulation of subatomic particles is essential for data analysis, detector

More information

Data Intensive Science Impact on Networks

Data Intensive Science Impact on Networks Data Intensive Science Impact on Networks Eli Dart, Network Engineer ESnet Network Engineering g Group IEEE Bandwidth Assessment Ad Hoc December 13, 2011 Outline Data intensive science examples Collaboration

More information

Topics for the TKR Software Review Tracy Usher, Leon Rochester

Topics for the TKR Software Review Tracy Usher, Leon Rochester Topics for the TKR Software Review Tracy Usher, Leon Rochester Progress in reconstruction Reconstruction short-term plans Simulation Calibration issues Balloon-specific support Personnel and Schedule TKR

More information

Fast pattern recognition with the ATLAS L1Track trigger for the HL-LHC

Fast pattern recognition with the ATLAS L1Track trigger for the HL-LHC Fast pattern recognition with the ATLAS L1Track trigger for the HL-LHC On behalf of the ATLAS Collaboration Uppsala Universitet E-mail: mikael.martensson@cern.ch ATL-DAQ-PROC-2016-034 09/01/2017 A fast

More information

Practical 2: Using Minitab (not assessed, for practice only!)

Practical 2: Using Minitab (not assessed, for practice only!) Practical 2: Using Minitab (not assessed, for practice only!) Instructions 1. Read through the instructions below for Accessing Minitab. 2. Work through all of the exercises on this handout. If you need

More information

Work in Tbilisi. David Mchedlishvili (SMART EDM_lab of TSU) GGSWBS , Tbilisi. Shota Rustaveli National Science Foundation

Work in Tbilisi. David Mchedlishvili (SMART EDM_lab of TSU) GGSWBS , Tbilisi. Shota Rustaveli National Science Foundation Mitglied der Helmholtz-Gemeinschaft David Mchedlishvili (SMART EDM_lab of TSU) Work in Tbilisi GGSWBS 18 23.08.2018, Tbilisi JEDI: Charged-Particle EDM Search Main principle: Inject polarized particles

More information

High-Energy Physics Data-Storage Challenges

High-Energy Physics Data-Storage Challenges High-Energy Physics Data-Storage Challenges Richard P. Mount SLAC SC2003 Experimental HENP Understanding the quantum world requires: Repeated measurement billions of collisions Large (500 2000 physicist)

More information

An introduction to plotting data

An introduction to plotting data An introduction to plotting data Eric D. Black California Institute of Technology February 25, 2014 1 Introduction Plotting data is one of the essential skills every scientist must have. We use it on a

More information

Optimizing Parallel Access to the BaBar Database System Using CORBA Servers

Optimizing Parallel Access to the BaBar Database System Using CORBA Servers SLAC-PUB-9176 September 2001 Optimizing Parallel Access to the BaBar Database System Using CORBA Servers Jacek Becla 1, Igor Gaponenko 2 1 Stanford Linear Accelerator Center Stanford University, Stanford,

More information

Event reconstruction in STAR

Event reconstruction in STAR Chapter 4 Event reconstruction in STAR 4.1 Data aquisition and trigger The STAR data aquisition system (DAQ) [54] receives the input from multiple detectors at different readout rates. The typical recorded

More information

ATLAS PILE-UP AND OVERLAY SIMULATION

ATLAS PILE-UP AND OVERLAY SIMULATION ATLAS PILE-UP AND OVERLAY SIMULATION LPCC Detector Simulation Workshop, June 26-27, 2017 ATL-SOFT-SLIDE-2017-375 22/06/2017 Tadej Novak on behalf of the ATLAS Collaboration INTRODUCTION In addition to

More information

Adding timing to the VELO

Adding timing to the VELO Summer student project report: Adding timing to the VELO supervisor: Mark Williams Biljana Mitreska Cern Summer Student Internship from June 12 to August 4, 2017 Acknowledgements I would like to thank

More information

ATLAS NOTE. December 4, ATLAS offline reconstruction timing improvements for run-2. The ATLAS Collaboration. Abstract

ATLAS NOTE. December 4, ATLAS offline reconstruction timing improvements for run-2. The ATLAS Collaboration. Abstract ATLAS NOTE December 4, 2014 ATLAS offline reconstruction timing improvements for run-2 The ATLAS Collaboration Abstract ATL-SOFT-PUB-2014-004 04/12/2014 From 2013 to 2014 the LHC underwent an upgrade to

More information

A Geometrical Modeller for HEP

A Geometrical Modeller for HEP A Geometrical Modeller for HEP R. Brun, A. Gheata CERN, CH 1211, Geneva 23, Switzerland M. Gheata ISS, RO 76900, Bucharest MG23, Romania For ALICE off-line collaboration Geometrical modelling generally

More information

GPFS Experiences from the Argonne Leadership Computing Facility (ALCF) William (Bill) E. Allcock ALCF Director of Operations

GPFS Experiences from the Argonne Leadership Computing Facility (ALCF) William (Bill) E. Allcock ALCF Director of Operations GPFS Experiences from the Argonne Leadership Computing Facility (ALCF) William (Bill) E. Allcock ALCF Director of Operations Argonne National Laboratory Argonne National Laboratory is located on 1,500

More information

LSGI 521: Principles of GIS. Lecture 5: Spatial Data Management in GIS. Dr. Bo Wu

LSGI 521: Principles of GIS. Lecture 5: Spatial Data Management in GIS. Dr. Bo Wu Lecture 5: Spatial Data Management in GIS Dr. Bo Wu lsbowu@polyu.edu.hk Department of Land Surveying & Geo-Informatics The Hong Kong Polytechnic University Contents 1. Learning outcomes 2. From files to

More information

Tracking POG Update. Tracking POG Meeting March 17, 2009

Tracking POG Update. Tracking POG Meeting March 17, 2009 Tracking POG Update Tracking POG Meeting March 17, 2009 Outline Recent accomplishments in Tracking POG - Reconstruction improvements for collisions - Analysis of CRAFT Data Upcoming Tasks Announcements

More information

Data Transfers Between LHC Grid Sites Dorian Kcira

Data Transfers Between LHC Grid Sites Dorian Kcira Data Transfers Between LHC Grid Sites Dorian Kcira dkcira@caltech.edu Caltech High Energy Physics Group hep.caltech.edu/cms CERN Site: LHC and the Experiments Large Hadron Collider 27 km circumference

More information

Grid Computing Activities at KIT

Grid Computing Activities at KIT Grid Computing Activities at KIT Meeting between NCP and KIT, 21.09.2015 Manuel Giffels Karlsruhe Institute of Technology Institute of Experimental Nuclear Physics & Steinbuch Center for Computing Courtesy

More information

The JINR Tier1 Site Simulation for Research and Development Purposes

The JINR Tier1 Site Simulation for Research and Development Purposes EPJ Web of Conferences 108, 02033 (2016) DOI: 10.1051/ epjconf/ 201610802033 C Owned by the authors, published by EDP Sciences, 2016 The JINR Tier1 Site Simulation for Research and Development Purposes

More information

Belle & Belle II. Takanori Hara (KEK) 9 June, 2015 DPHEP Collaboration CERN

Belle & Belle II. Takanori Hara (KEK) 9 June, 2015 DPHEP Collaboration CERN 1 Belle & Belle II Takanori Hara (KEK) takanori.hara@kek.jp 9 June, 2015 DPHEP Collaboration Workshop @ CERN 2 Belle Data Belle : started in 1999, data-taking completed in 2010 still keep analysing the

More information

LHCb Computing Status. Andrei Tsaregorodtsev CPPM

LHCb Computing Status. Andrei Tsaregorodtsev CPPM LHCb Computing Status Andrei Tsaregorodtsev CPPM Plan Run II Computing Model Results of the 2015 data processing 2016-2017 outlook Preparing for Run III Conclusions 2 HLT Output Stream Splitting 12.5 khz

More information

Big Data Analytics and the LHC

Big Data Analytics and the LHC Big Data Analytics and the LHC Maria Girone CERN openlab CTO Computing Frontiers 2016, Como, May 2016 DOI: 10.5281/zenodo.45449, CC-BY-SA, images courtesy of CERN 2 3 xx 4 Big bang in the laboratory We

More information

The performance of the ATLAS Inner Detector Trigger Algorithms in pp collisions at the LHC

The performance of the ATLAS Inner Detector Trigger Algorithms in pp collisions at the LHC X11 opical Seminar IPRD, Siena - 7- th June 20 he performance of the ALAS Inner Detector rigger Algorithms in pp collisions at the LHC Mark Sutton University of Sheffield on behalf of the ALAS Collaboration

More information

High Throughput WAN Data Transfer with Hadoop-based Storage

High Throughput WAN Data Transfer with Hadoop-based Storage High Throughput WAN Data Transfer with Hadoop-based Storage A Amin 2, B Bockelman 4, J Letts 1, T Levshina 3, T Martin 1, H Pi 1, I Sfiligoi 1, M Thomas 2, F Wuerthwein 1 1 University of California, San

More information

Scattering/Wave Terminology A few terms show up throughout the discussion of electron microscopy:

Scattering/Wave Terminology A few terms show up throughout the discussion of electron microscopy: 1. Scattering and Diffraction Scattering/Wave Terology A few terms show up throughout the discussion of electron microscopy: First, what do we mean by the terms elastic and inelastic? These are both related

More information

Lies, Damn Lies and Performance Metrics. PRESENTATION TITLE GOES HERE Barry Cooks Virtual Instruments

Lies, Damn Lies and Performance Metrics. PRESENTATION TITLE GOES HERE Barry Cooks Virtual Instruments Lies, Damn Lies and Performance Metrics PRESENTATION TITLE GOES HERE Barry Cooks Virtual Instruments Goal for This Talk Take away a sense of how to make the move from: Improving your mean time to innocence

More information

CC-IN2P3: A High Performance Data Center for Research

CC-IN2P3: A High Performance Data Center for Research April 15 th, 2011 CC-IN2P3: A High Performance Data Center for Research Toward a partnership with DELL Dominique Boutigny Agenda Welcome Introduction to CC-IN2P3 Visit of the computer room Lunch Discussion

More information

Introducing Robotics Vision System to a Manufacturing Robotics Course

Introducing Robotics Vision System to a Manufacturing Robotics Course Paper ID #16241 Introducing Robotics Vision System to a Manufacturing Robotics Course Dr. Yuqiu You, Ohio University c American Society for Engineering Education, 2016 Introducing Robotics Vision System

More information

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider

From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider From raw data to new fundamental particles: The data management lifecycle at the Large Hadron Collider Andrew Washbrook School of Physics and Astronomy University of Edinburgh Dealing with Data Conference

More information

Post Silicon Electrical Validation

Post Silicon Electrical Validation Post Silicon Electrical Validation Tony Muilenburg 1 1/21/2014 Homework 4 Review 2 1/21/2014 Architecture / Integration History 3 1/21/2014 4 1/21/2014 Brief History Of Microprocessors 5 1/21/2014 6 1/21/2014

More information

The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data

The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data The ATLAS EventIndex: an event catalogue for experiments collecting large amounts of data D. Barberis 1*, J. Cranshaw 2, G. Dimitrov 3, A. Favareto 1, Á. Fernández Casaní 4, S. González de la Hoz 4, J.

More information

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005

Compact Muon Solenoid: Cyberinfrastructure Solutions. Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005 Compact Muon Solenoid: Cyberinfrastructure Solutions Ken Bloom UNL Cyberinfrastructure Workshop -- August 15, 2005 Computing Demands CMS must provide computing to handle huge data rates and sizes, and

More information

Travelling securely on the Grid to the origin of the Universe

Travelling securely on the Grid to the origin of the Universe 1 Travelling securely on the Grid to the origin of the Universe F-Secure SPECIES 2007 conference Wolfgang von Rüden 1 Head, IT Department, CERN, Geneva 24 January 2007 2 CERN stands for over 50 years of

More information

ATLAS DCS Overview SCADA Front-End I/O Applications

ATLAS DCS Overview SCADA Front-End I/O Applications Overview SCADA Front-End I/O Applications ATLAS TDAQ week, July 6th 2001, H.J.Burckhart 1 SCX1 USA15 Common Infrastructure PC Local Area Network Server Operation Local Control Stations (LCS) Expert Subdetector

More information

Simulation. Variance reduction Example

Simulation. Variance reduction Example Simulation Variance reduction Example M C Example Consider scattering of laser beam from a material layer MSc thesis of Jukka Räbinä 2005 Goal is to simulate different statistics of the scattered image

More information

Expressing Parallelism with ROOT

Expressing Parallelism with ROOT Expressing Parallelism with ROOT https://root.cern D. Piparo (CERN) for the ROOT team CHEP 2016 2 This Talk ROOT helps scientists to express parallelism Adopting multi-threading (MT) and multi-processing

More information

CMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster

CMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster CMS Grid Computing at TAMU Performance, Monitoring and Current Status of the Brazos Cluster Vaikunth Thukral Department of Physics and Astronomy Texas A&M University 1 Outline Grid Computing with CMS:

More information

Oracle and Tangosol Acquisition Announcement

Oracle and Tangosol Acquisition Announcement Oracle and Tangosol Acquisition Announcement March 23, 2007 The following is intended to outline our general product direction. It is intended for information purposes only, and may

More information

Monitoring of Computing Resource Use of Active Software Releases at ATLAS

Monitoring of Computing Resource Use of Active Software Releases at ATLAS 1 2 3 4 5 6 Monitoring of Computing Resource Use of Active Software Releases at ATLAS Antonio Limosani on behalf of the ATLAS Collaboration CERN CH-1211 Geneva 23 Switzerland and University of Sydney,

More information

Charged Particle Reconstruction in HIC Detectors

Charged Particle Reconstruction in HIC Detectors Charged Particle Reconstruction in HIC Detectors Ralf-Arno Tripolt, Qiyan Li [http://de.wikipedia.org/wiki/marienburg_(mosel)] H-QM Lecture Week on Introduction to Heavy Ion Physics Kloster Marienburg/Mosel,

More information

CS252 S05. CMSC 411 Computer Systems Architecture Lecture 18 Storage Systems 2. I/O performance measures. I/O performance measures

CS252 S05. CMSC 411 Computer Systems Architecture Lecture 18 Storage Systems 2. I/O performance measures. I/O performance measures CMSC 411 Computer Systems Architecture Lecture 18 Storage Systems 2 I/O performance measures I/O performance measures diversity: which I/O devices can connect to the system? capacity: how many I/O devices

More information

CLAS 12 Reconstruction Software

CLAS 12 Reconstruction Software CLAS 12 Reconstruction Software G.Gavalian (ODU) Outline CLAS-12 computing Requirements CLARA framework (SOA architecture) CLARA based application examples (DataMining) CLAS 12 Reconstruction Where do

More information

Linux Installation Planning

Linux Installation Planning Linux Installation Planning Mark Post Novell, Inc. March 4, 2011 Session 8986 Agenda More Questions Than Answers First Things First Pick the Right Architecture Disk Storage Selection Application Selection

More information

Quad Module Hybrid Development for the ATLAS Pixel Layer Upgrade

Quad Module Hybrid Development for the ATLAS Pixel Layer Upgrade Quad Module Hybrid Development for the ATLAS Pixel Layer Upgrade Lawrence Berkeley National Lab E-mail: kedunne@lbl.gov Maurice Garcia-Sciveres, Timon Heim Lawrence Berkeley National Lab, Berkeley, USA

More information

CMS Conference Report

CMS Conference Report Available on CMS information server CMS CR 2005/021 CMS Conference Report 29 Septemebr 2005 Track and Vertex Reconstruction with the CMS Detector at LHC S. Cucciarelli CERN, Geneva, Switzerland Abstract

More information

ALICE ANALYSIS PRESERVATION. Mihaela Gheata DASPOS/DPHEP7 workshop

ALICE ANALYSIS PRESERVATION. Mihaela Gheata DASPOS/DPHEP7 workshop 1 ALICE ANALYSIS PRESERVATION Mihaela Gheata DASPOS/DPHEP7 workshop 2 Outline ALICE data flow ALICE analysis Data & software preservation Open access and sharing analysis tools Conclusions 3 ALICE data

More information

Diploma Of Computing

Diploma Of Computing Diploma Of Computing Course Outline Campus Intake CRICOS Course Duration Teaching Methods Assessment Course Structure Units Melbourne Burwood Campus / Jakarta Campus, Indonesia March, June, October 022638B

More information

Eight units must be completed and passed to be awarded the Diploma.

Eight units must be completed and passed to be awarded the Diploma. Diploma of Computing Course Outline Campus Intake CRICOS Course Duration Teaching Methods Assessment Course Structure Units Melbourne Burwood Campus / Jakarta Campus, Indonesia March, June, October 022638B

More information