Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL1 An incomplete view of the future of HEP Computing Matthias Kasemann Fermilab A view of the future.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
Distributed IT Infrastructure for U.S. ATLAS Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 15 th April 2009 Visit of Spanish Royal Academy.
ATLAS users and SLAC Jason Nielsen Santa Cruz Institute for Particle Physics University of California, Santa Cruz SLAC Users Organization Meeting 7 February.
DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
1 Towards an Upgrade TDR: LHCb Computing workshop May 2015 Introduction to Upgrade Computing Session Peter Clarke Many peoples ideas Vava, Conor,
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
International collaboration in high energy physics experiments  All large high energy physics experiments today are strongly international.  A necessary.
GridPP Steve Lloyd, Chair of the GridPP Collaboration Board.
1. 2 CERN European Organization for Nuclear Research Founded in 1954 by 12 countries – Norway one of them Today: 20 member states, around 2500 staff –
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
Requirements Review – July 21, Requirements for CMS Patricia McBride July 21, 2005.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Fermilab June 29, 2001Data collection and handling for HEP1 Matthias Kasemann Fermilab Overview of Data collection and handling for High Energy Physics.
5 May 98 1 Jürgen Knobloch Computing Planning for ATLAS ATLAS Software Week 5 May 1998 Jürgen Knobloch Slides also on:
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
…building the next IT revolution From Web to Grid…
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
CERN – IT Department CH-1211 Genève 23 Switzerland t Working with Large Data Sets Tim Smith CERN/IT Open Access and Research Data Session.
GridPP Collaboration Meeting 5 th November 2001 Dan Tovey, University of Sheffield Non-LHC and Non-US-Collider Experiments’ Requirements Dan Tovey, University.
1 LHCC RRB SG 16 Sep P. Vande Vyvre CERN-PH On-line Computing M&O LHCC RRB SG 16 Sep 2004 P. Vande Vyvre CERN/PH for 4 LHC DAQ project leaders.
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
23.March 2004Bernd Panzer-Steindel, CERN/IT1 LCG Workshop Computing Fabric.
Computing R&D and Milestones LHCb Plenary June 18th, 1998 These slides are on WWW at:
Feb. 14, 2002DØRAM Proposal DØ IB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) Introduction Partial Workshop Results DØRAM Architecture.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
ComPASS Summary, Budgets & Discussion Panagiotis Spentzouris, Fermilab ComPASS PI.
1 The ILC Control Work Packages. ILC Control System Work Packages GDE Oct Who We Are Collaboration loosely formed at Snowmass which included SLAC,
Peer review in the era of LHC experiments Experimental particle physics as a Big Science paradigm Rüdiger Voss Physics Department CERN, Geneva, Switzerland.
LHC Computing, CERN, & Federated Identities
U.S. ATLAS Computing Facilities Overview Bruce G. Gibbard Brookhaven National Laboratory U.S. LHC Software and Computing Review Brookhaven National Laboratory.
Computing Division FY03 Budget and budget outlook for FY04 + CDF International Finance Committee April 4, 2003 Vicky White Head, Computing Division.
INFSO-RI Enabling Grids for E-sciencE The EGEE Project Owen Appleton EGEE Dissemination Officer CERN, Switzerland Danish Grid Forum.
Run II Review Closeout 15 Sept., 2004 FNAL. Thanks! …all the hard work from the reviewees –And all the speakers …hospitality of our hosts Good progress.
1 (Brief) Introductory Remarks On Behalf of the U.S. Department of Energy ESnet Site Coordinating Committee (ESCC) W.Scott Bradley ESCC Chairman
Ian Bird WLCG Networking workshop CERN, 10 th February February 2014
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Marco Cattaneo, 3-June Event Reconstruction for LHCb  What is the scope of the project?  What are the goals (short+medium term)?  How do we organise.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Hall D Computing Facilities Ian Bird 16 March 2001.
Road Owners and PMS Christopher R. Bennett Senior Transport Specialist East Asia and Pacific Transport The World Bank Washington, D.C.
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Computing models, facilities, distributed computing
LHC experiments Requirements and Concepts ALICE
for the Offline and Computing groups
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
New strategies of the LHC experiments to meet
Nuclear Physics Data Management Needs Bruce G. Gibbard
LHCb thinking on Regional Centres and Related activities (GRIDs)
Development of LHCb Computing Model F Harris
Presentation transcript:

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL1 An incomplete view of the future of HEP Computing Matthias Kasemann Fermilab A view of the future of HEP Computing Some things I know for the future of HEP Computing Some things like to I know for the future of HEP Computing

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL2 Disclaimer Although we are here in the Root workshop I don’t present topics which are necessarily to be answered by Root in its current implementation or any derivative or future development in Root. I simply put down what worries me when I think about computing for future HEP experiments. (Speaking for myself and not for US, US DOE, FNAL nor URA.) (Product, trade, or service marks herein belong to their respective owners.)

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL3 Fermilab HEP Program Year: Neutrinos: MiniBooNE NuMI/Minos Run IIa Collider: Run IIb BTeV MI Fixed Target: Testbeam KaMI/CKM? Astrophysics: Sloan CDMS Auger LHC physics

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL4 The CERN Scientific Programme LEP ALEPH DELPHI L3 OPAL LHC ATLAS CMS ALICE LHCb Other LHC experiments (e.g. Totem) SPS & PS Heavy ions Compass NA48 Neutrino DIRAC HARP Other Facilities TOF Neutron AD ISOLDE Test beams North Areas West Areas East Hall Accelerators R&D Legend Approved Under consideration

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL5 HEP computing: The next 5 years…(1) u Data analysis for completed experiments continues u Challenges: u No major change to analysis model, code or infrastructure u Operation, continuity, maintaining expertise and effort u Data collection and analysis for ongoing experiments u Challenges: u Data volume, compute resources, software organization u Operation, continuity, maintaining expertise and effort

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL6 HEP computing: The next 5 years…(2) u Starting experiments: u Challenges: u Completion and verification of data and analysis model, u Data volume, compute resources, software organization, $$’s u Operation, continuity, maintaining expertise and effort u Experiments in preparation: u Challenges: u Definition and implementation of data and analysis model, u data volume, compute resources, software organization, $$’s u continuity, getting and maintaining expertise and effort u Build for change: applications, data models… u Build compute models which are adaptable to different local environments

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL7 Run 2 Data Volumes u First Run 2b costs estimates based on scaling arguments u Use predicted luminosity profile u Assume technology advance (Moore’s law) u CPU and data storage requirements both scale with data volume stored u Data volume depends on physics selection in trigger u Can vary between 1 – 8 PB (Run 2a: 1 PB) per experiment u Have to start preparation by 2002/2003

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL8 1 billion people surfing the Web How Much Data is Involved? Level 1 Rate (Hz) High Level-1 Trigger (1 MHz) High No. Channels High Bandwidth (500 Gbit/s) High Data Archive (PetaByte) LHCB KLOE HERA-B CDF IIa D0 IIa CDF H1 ZEUS UA1 LEP NA49 ALICE Event Size (bytes) ATLAS CMS KTeV

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL9 HEP computing: The next 5 years…(3) u Challenges in big collaborations u Long and difficult planning process u More formal procedure required to commit resources u Long lifetime, need flexible solutions which allow for change u Any state of experiment longer than typical PhD or postdoc time u Need for professional IT participation and support u Challenges in smaller collaborations u Limited in resources u Adapt and implement available solutions (“b-b-s”)

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL10 u Experiment in preparation at CERN/Switzerland u Strong US participation: ~20% u Startup: by 2005/2006, will run for 15+ years CMS Computing Challenges 1800 Physicists 150 Institutes 32 Countries Major challenges associated with: Communication and collaboration at a distance Distributed computing resources Remote software development and physics analysis R&D: New Forms of Distributed Systems

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL11 Role of computer networking (1) u State-of-the-art computer networking enables large international collaborations u needed for all aspects of collaborative work u to write the proposal, u produce and agree on the designs of the components and systems, u collaborate on overall planning and integration of the detector, confer on all aspects of the device, including the final physics results, and u provide information to collaborators and to the physics community and general public u Data from the experiment lives more-and-more on the network u All levels: raw, dst, aod, ntuple, draft-paper, paper

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL12 Role of computer networking (2) u HEP developed its own national network in the early 1980s u National research network backbones generally provide adequate support to HEP and other sciences. u Specific network connections are used where HEP has found it necessary to support special capabilities that could not be supplied efficiently or capably enough through more general networks. u US-CERN, several HEP links in Europe… u Dedicated HEP links are needed in special cases because u HEP requirements can be large and can overwhelm those of researchers in other fields u because regional networks do not give top priority to interregional connections

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL13 Data analysis in international collaborations: past u In the past analysis was centered at the experimental site u a few major external centers were used. u Up the mid 90s bulk data were transferred by shipping tapes, networks were used for programs and conditions data. u External analysis centers served the local/national users only. u Often staff (and equipment) from the external center being placed at the experimental site to ensure the flow of tapes. u The external analysis often was significantly disconnected from the collaboration mainstream.

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL14 u Why? u For one experiment looking ahead for a few years only centralized resources may be most cost effective, but: u national and local interests leads to massive national and local investments u For BaBar: u The total annual value of foreign centers to the US-based program is greatly in excess of the estimated cost to the US of creating the required high-speed paths from SLAC to the landing points of lines WAN funded by foreign collaborators u Future world-scale experimental programs must be planned with explicit support for a collaborative environment that allows many nations to be full participants in the challenges of data analysis. Data analysis in international collaborations: truly distributed

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL15 Distributed computing: u Networking is an expensive resource, should be minimized u Pre-emptive transfers can be used to improve responsiveness at the cost of some extra network traffic. u Multi-tiered architecture must become more general and flexible u to accommodate the very large uncertainties in the relative costs of CPU, storage and networking u To enable physicists to work effectively in the face of data having unprecedented volume and complexity u Aim for transparency and location independence of data access u the need for individual physicists to understand and manipulate all the underlying transport and task-management systems would be too complex

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL16 6/13/01: "It turns out that distributed computing is really hard," said Eric Schmidt, the chairman of Google, the Internet search engine company. "It's much harder than it looks. It has to work across different networks with different kinds of security, or otherwise it ends up being a single-vendor solution, which is not what the industry wants." Distributed Computing

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL17 LHC Data Grid Hierarchy (Schematic) Tier 1 FNAL/BNL T Tier 0 (CERN) Other Tier 1 centers

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL18 Many more technical questions to answer (1) u Operating system: u UNIX seems to be favored for data handling and analysis, u LINUX is most cost effective u Mainframe vs. commodity computing: u commodity computing can provide many solutions u Only affordable solution for future requirements u How to operate several thousand nodes? u How to write applications to benefit from several thousand nodes? u Data access and formats: u Metadata databases, event storage

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL19 Many more technical questions to answer (2) u Commercial vs. custom software, public domain u Programming languages: u Compiled languages for CPU intensive parts u Scripting languages provide excellent frameworks u How to handle and control big numbers in big detectors: u Number of channels, modules improves (several millions of channels, hundreds of modules u Need new automatic tools to calibrate, monitor and align channels

Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL20 Some more thoughts u Computing for HEP experiments is costly u In $$’s, people and time u Need R&D, prototyping and test-beds to develop solutions and validate choices u Improving the engineering aspect of computing for HEP experiments is essential u Treat computing and software as a project (see u Project lifecycles, milestones, resource estimates, reviews u Documenting conditions and work performed is essential for success u Track detector building for 20 years u Log data taking and processing conditions u Analysis steps, algorithms, cuts As transparent and automatic as possible