Brookhaven Analysis Facility Michael Ernst Brookhaven National Laboratory U.S. ATLAS Facility Meeting University of Chicago, Chicago 19 – 20 August, 2009.

Slides:



Advertisements
Similar presentations
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Advertisements

T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
1 Deployment of an LCG Infrastructure in Australia How-To Setup the LCG Grid Middleware – A beginner's perspective Marco La Rosa
Copyright © 2010 Platform Computing Corporation. All Rights Reserved.1 The CERN Cloud Computing Project William Lu, Ph.D. Platform Computing.
Scientific Data Infrastructure in CAS Dr. Jianhui Scientific Data Center Computer Network Information Center Chinese Academy of Sciences.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
Tier 1 Facility Status and Current Activities Rich Baker Brookhaven National Laboratory NSF/DOE Review of ATLAS Computing June 20, 2002.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Developing & Managing A Large Linux Farm – The Brookhaven Experience CHEP2004 – Interlaken September 27, 2004 Tomasz Wlodek - BNL.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Condor Usage at Brookhaven National Lab Alexander Withers (talk given by Tony Chan) RHIC Computing Facility Condor Week - March 15, 2005.
KISTI-GSDC SITE REPORT Sang-Un Ahn, Jin Kim On the behalf of KISTI GSDC 24 March 2015 HEPiX Spring 2015 Workshop Oxford University, Oxford, UK.
OSG Tier 3 support Marco Mambelli - OSG Tier 3 Dan Fraser - OSG Tier 3 liaison Tanya Levshina - OSG.
BNL Tier 1 Service Planning & Monitoring Bruce G. Gibbard GDB 5-6 August 2006.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
A PanDA Backend for the Ganga Analysis Interface J. Elmsheuser 1, D. Liko 2, T. Maeno 3, P. Nilsson 4, D.C. Vanderster 5, T. Wenaus 3, R. Walker 1 1: Ludwig-Maximilians-Universität.
11 November 2010 Natascha Hörmann Computing at HEPHY Evaluation 2010.
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
University user perspectives of the ideal computing environment and SLAC’s role Bill Lockman Outline: View of the ideal computing environment ATLAS Computing.
PERFORMANCE AND ANALYSIS WORKFLOW ISSUES US ATLAS Distributed Facility Workshop November 2012, Santa Cruz.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
ATLAS Distributed Computing perspectives for Run-2 Simone Campana CERN-IT/SDC on behalf of ADC.
U.S. ATLAS Facility Planning U.S. ATLAS Tier-2 & Tier-3 Meeting at SLAC 30 November 2007.
January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 1 SIMULATION OF DAILY ACTIVITITIES AT REGIONAL CENTERS MONARC Collaboration Alexander Nazarenko.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
A Service-Based SLA Model HEPIX -- CERN May 6, 2008 Tony Chan -- BNL.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
David Adams ATLAS ATLAS Distributed Analysis (ADA) David Adams BNL December 5, 2003 ATLAS software workshop CERN.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Response of the ATLAS Spanish Tier2 for.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
Main parameters of Russian Tier2 for ATLAS (RuTier-2 model) Russia-CERN JWGC meeting A.Minaenko IHEP (Protvino)
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
ATLAS TIER3 in Valencia Santiago González de la Hoz IFIC – Instituto de Física Corpuscular (Valencia)
ATLAS Physics Analysis Framework James R. Catmore Lancaster University.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
Belle II Physics Analysis Center at TIFR
Western Analysis Facility
A full demonstration based on a “real” analysis scenario
Luca dell’Agnello INFN-CNAF
Readiness of ATLAS Computing - A personal view
ATLAS Sites Jamboree, CERN January, 2017
TYPES OFF OPERATING SYSTEM
Presentation transcript:

Brookhaven Analysis Facility Michael Ernst Brookhaven National Laboratory U.S. ATLAS Facility Meeting University of Chicago, Chicago 19 – 20 August, – 20 August, 2009

20 August, 2009 M. Ernst Facility Workshop 2 Anticipated Analysis Workflow  Interactively develop and test code used in the athena frame work to access data via pAthena  Once the user has tested the code on a small number of events and she/he is satisfied w/ results the user will submit PanDA jobs to process the data located at the Tier-1 and/or Tier-2 centers. The output will be stored close to location where jobs ran  When user has created a DPD she/he wants to run over the data many times, she/he will use existing DQ2 tools to transfer the root files to the user’s site  Once the user data is at the analysis facility the user will analyze the DPD many times  User ntuples can be further reduced and shipped back to the home institution for further analysis

20 August, 2009 M. Ernst Facility Workshop 3 Aspects of Analysis Facility Implementation  Single machine access  Needed for “pre-PanDA” testing and Athena development work  Could be primary method for analyses with small data volume  Parallel processing w/ PROOF  Needed for interactive analyses with large data volume o One PROOF instance may last n hours o From interactive histo making to command line processing o Allows transparent use of multiple nodes (other than batch system) o Automatically distributes load and merges results  Distributed processing using Condor  Needed for long running analyses tasks (>1 day)  Needed for analyses that cannot use TSelector framework  Mature system that can be used for ntuple processing  Has queuing, prioritization and fair share mechanisms

20 August, 2009 M. Ernst Facility Workshop 4 Central Analysis Facility at BNL  Existing Computing Infrastructure of U.S. ATLAS meets requirements of the ATLAS CM  For Central Production and Distributed Analysis  Meet partially analysis needs of individual physicists and groups  Purpose: Serve the needs of data analysis at the “Post-AOD” stage  Allow members of the detector performance groups and individual users performing physics studies to analyze selected or reduced data with fast turnaround  Fraction of users will lack significant Tier-3 resources beyond a PC  Allow users to develop and debug code and run interactive jobs o Essential step prior to submitting pAthena and long analysis jobs o Need demonstrated by usage of interactive facilities at CERN and BNL  Selected information will be processed recursively many times at low latency before the data volume is further reduced o This may require several iterations  Primary intention is to help physicist to process large data volumes o Production of private datasets, generation of Toy MC will be discouraged  Not intended to replace the existing analysis model  Strengthen it by adding additional functionality (addressing needs of early analysis stages)  Users are strongly encouraged to make use of pAthena  AF useful as long as analysis code is in flux – once confidence in algor. reached produce DnPD

20 August, 2009 M. Ernst Facility Workshop 5 Central Analysis Facility at BNL  Will be kept separate from the (Tier-1) production facility  Functionality for parallel processing (interactive and batch)  Properly sized to satisfy the CPU and I/O intensive requirements  Will host selection of AOD & ESD, and primary DPD  Supports access of large data volumes multiple time w/o data movement, registration  Open to institutions opting to have BNL procure, install and maintain their equipment Analysis Facility will be an End-User facility Analysis Facility will be an End-User facility Combines functionalities currently in use but at larger scale A la ACAS with features available in PROOF (if applicable) and Condor Farms Analysis Facility can be thought of as a Tier-3 Facility Analysis Facility can be thought of as a Tier-3 Facility In addition to institutional Tier-3’s Providing support for users who do not have access to local Tier-3 resources Analysis Facility will allow users to work efficiently with “User Ntuples” Analysis Facility will allow users to work efficiently with “User Ntuples” Derived from a variety of data formats (produced using pAthena)

20 August, 2009 M. Ernst Facility Workshop 6 Data Management  AF will have finite amount of resources  Shared between ESD, AOD, DPD and User storage space  Data will be replicated from US ATLAS Tier-1 to AF  User tools to transfer data to home institutions  Data transfer into AF will be managed by group coordinator  Disk management of PROOF system handled by Xrootd  Tight binding of subjobs with associated data handled automatically

20 August, 2009 M. Ernst Facility Workshop 7 Central Analysis Facility GRID/PANDA AF Data transfer Tier1 Tier2s Users Multiple processing of data sets Interactive jobs Pathena Jobs Histos making Selected AOD ESD Tier 3 Selected D 1 PD D 2 PD D 3 PD Histos making DPD transfer DPD transfer User Ntuple Selected D 1 PD D 2 PD D 3 PD User Ntuple

20 August, 2009 M. Ernst Facility Workshop 8 Analysis Facility at BNL  Analysis Facility computer cluster with:  Excellent access to Atlas data  Easy access to Atlas databases (including conditions root files)  Available to any Atlas member (requires local account, currently ~750)  Augments Tier 3’s at Universities  Universities can contribute approved hardware o Fair share mechanism will allow Universities to use what they contributed  Open to U.S. ATLAS Institutions starting in September, 2009

20 August, 2009 M. Ernst Facility Workshop 9 Scientific Computing at BNL Joint BNL/Stony Brook Computational Science Center 100 TFlop BlueGene/L QCDOC RIKEN-BNL-Columbia: 10 TFlop DOE HEP: 10 TFlop RHIC/ATLAS Computing Facility (RACF) RHIC data and analysis center ATLAS Tier 1 Center Combined capacity: 20 MSI2k, 7 PB Disk, 9 PB Tape Platform for: ATLAS Tier 3 Computing LSST/astrophysics computing Computing for Daya Bay and LBNE experiments

20 August, 2009 M. Ernst Facility Workshop 10 RACF space and deployment planning Core structure: Brookhaven Computing Facility (13000 sq. ft.) Renovation (2000 sq. ft.) New addition (6400 sq. ft.) QCDOC NYBlue (7000 sq. ft. for RACF)

20 August, 2009 M. Ernst Facility Workshop 11 Analysis Facility at BNL  BNL has currently 1000 Batch slots available for user analysis  Usage of ~800 slots will be restricted to Group Analysis after start of LHC data taking (~mid November, 2009) according to ATLAS Computing Model  Acas nodes available for interactive analysis  Can be expanded if need – defined by user needs  Prototype Proof cluster – will expand as needed  150 TB (high quality, highly available) disk space for users  Non permanent – but not scratch space (managed)  University can contribute funds or “certified” hardware to place computing at BNL * Based on vendor price of $2,500 per compute server; incl. rack, cables, console, overhead, Acquisition ($) Operating ($ per yr) CPU (per 8-core node) 2,970*571 Disk (per TB) 80181

20 August, 2009 M. Ernst Facility Workshop 12 Hosting Services  Provide full integration of CPU and Disk into existing RACF infrastructure - Operating Cost covers  Space, Power and Cooling  “Racking and Stacking” including “out-of-band” console connection  Hardware service (typically next business day)  Line-rate 1 Gbps network connection for WNs and 10 Gbps for disk servers  Monitoring (system and services, nagios w/ notification as appr.)  OS installation, administration and maintenance (regular kernel patches (typically for cyber security), same as ATLAS farm)  Operating - typically 5days/week, 8 hours/day  Account Management  Shared file system access for ATLAS S/W releases  Batch queues (Condor)  Dedicated queues for institutional resources augmented by “General Queue” to share resources with the rest of the (US) ATLAS community in case cycles aren’t used  RACF provisions scalable gateway services for interactive access  Disk is procured in large quantities and distributed in TB quantities  Integration into Storage Management Infrastructure o