The PanDA Distributed Production and Analysis System Torre Wenaus Brookhaven National Laboratory, USA ISGC 2008 Taipei, Taiwan April 9, 2008 Torre Wenaus.

Slides:



Advertisements
Similar presentations
PanDA Integration with the SLAC Pipeline Torre Wenaus, BNL BigPanDA Workshop October 21, 2013.
Advertisements

Dec 14, 20061/10 VO Services Project – Status Report Gabriele Garzoglio VO Services Project WBS Dec 14, 2006 OSG Executive Board Meeting Gabriele Garzoglio.
 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
Implementing Finer Grained Authorization in the Open Science Grid Gabriele Carcassi, Ian Fisk, Gabriele, Garzoglio, Markus Lorch, Timur Perelmutov, Abhishek.
Validata Release Coordinator Accelerated application delivery through automated end-to-end release management.
Dynamics AX Technical Overview Application Architecture Dynamics AX Technical Overview.
Condor at Brookhaven Xin Zhao, Antonio Chan Brookhaven National Lab CondorWeek 2009 Tuesday, April 21.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
The Panda System Mark Sosebee (for K. De) University of Texas at Arlington dosar workshop March 30, 2006.
Assessment of Core Services provided to USLHC by OSG.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
Integrating HPC into the ATLAS Distributed Computing environment Doug Benjamin Duke University.
May 8, 20071/15 VO Services Project – Status Report Gabriele Garzoglio VO Services Project – Status Report Overview and Plans May 8, 2007 Computing Division,
Apr 30, 20081/11 VO Services Project – Stakeholders’ Meeting Gabriele Garzoglio VO Services Project Stakeholders’ Meeting Apr 30, 2008 Gabriele Garzoglio.
PanDA Multi-User Pilot Jobs Maxim Potekhin Brookhaven National Laboratory Open Science Grid WLCG GDB Meeting CERN March 11, 2009.
PanDA A New Paradigm for Computing in HEP Kaushik De Univ. of Texas at Arlington NRC KI, Moscow January 29, 2015.
Use of Condor on the Open Science Grid Chris Green, OSG User Group / FNAL Condor Week, April
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
OSG Area Coordinator’s Report: Workload Management April 20 th, 2011 Maxim Potekhin BNL
Mar 28, 20071/9 VO Services Project Gabriele Garzoglio The VO Services Project Don Petravick for Gabriele Garzoglio Computing Division, Fermilab ISGC 2007.
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Developing & Managing A Large Linux Farm – The Brookhaven Experience CHEP2004 – Interlaken September 27, 2004 Tomasz Wlodek - BNL.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Successful Common Projects: Structures and Processes WLCG Management.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
David Adams ATLAS ADA, ARDA and PPDG David Adams BNL June 28, 2004 PPDG Collaboration Meeting Williams Bay, Wisconsin.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
My Name: ATLAS Computing Meeting – NN Xxxxxx A Dynamic System for ATLAS Software Installation on OSG Sites Xin Zhao, Tadashi Maeno, Torre Wenaus.
Maarten Litmaath (CERN), GDB meeting, CERN, 2006/02/08 VOMS deployment Extent of VOMS usage in LCG-2 –Node types gLite 3.0 Issues Conclusions.
OSG Area Coordinator’s Report: Workload Management Maxim Potekhin BNL
Tarball server (for Condor installation) Site Headnode Worker Nodes Schedd glidein - special purpose Condor pool master DB Panda Server Pilot Factory -
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Evolution of a High Performance Computing and Monitoring system onto the GRID for High Energy Experiments T.L. Hsieh, S. Hou, P.K. Teng Academia Sinica,
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
A PanDA Backend for the Ganga Analysis Interface J. Elmsheuser 1, D. Liko 2, T. Maeno 3, P. Nilsson 4, D.C. Vanderster 5, T. Wenaus 3, R. Walker 1 1: Ludwig-Maximilians-Universität.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
OSG Area Coordinator’s Report: Workload Management May14 th, 2009 Maxim Potekhin BNL
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
PanDA & BigPanDA Kaushik De Univ. of Texas at Arlington BigPanDA Workshop, CERN October 21, 2013.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
GRID Security & DIRAC A. Casajus R. Graciani A. Tsaregorodtsev.
Eileen Berman. Condor in the Fermilab Grid FacilitiesApril 30, 2008  Fermi National Accelerator Laboratory is a high energy physics laboratory outside.
DIRAC Pilot Jobs A. Casajus, R. Graciani, A. Tsaregorodtsev for the LHCb DIRAC team Pilot Framework and the DIRAC WMS DIRAC Workload Management System.
ATLAS Database Access Library Local Area LCG3D Meeting Fermilab, Batavia, USA October 21, 2004 Alexandre Vaniachine (ANL)
OSG Area Coordinator’s Report: Workload Management Maxim Potekhin BNL May 8 th, 2008.
Parag Mhashilkar Computing Division, Fermi National Accelerator Laboratory.
OSG Area Coordinator’s Report: Workload Management October 6 th, 2010 Maxim Potekhin BNL
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
U.S. ATLAS Facility Planning U.S. ATLAS Tier-2 & Tier-3 Meeting at SLAC 30 November 2007.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Sep 17, 20081/16 VO Services Project – Stakeholders’ Meeting Gabriele Garzoglio VO Services Project Stakeholders’ Meeting Sep 17, 2008 Gabriele Garzoglio.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
CMS Experience with the Common Analysis Framework I. Fisk & M. Girone Experience in CMS with the Common Analysis Framework Ian Fisk & Maria Girone 1.
DIRAC for Grid and Cloud Dr. Víctor Méndez Muñoz (for DIRAC Project) LHCb Tier 1 Liaison at PIC EGI User Community Board, October 31st, 2013.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
Alessandro De Salvo Mayuko Kataoka, Arturo Sanchez Pineda,Yuri Smirnov CHEP 2015 The ATLAS Software Installation System v2 Alessandro De Salvo Mayuko Kataoka,
Job submission overview Marco Mambelli – August OSG Summer Workshop TTU - Lubbock, TX THE UNIVERSITY OF CHICAGO.
J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April ATLAS Grid Activities Preparing for Data Analysis Jim Shank.
Panda Monitoring, Job Information, Performance Collection Kaushik De (UT Arlington), Torre Wenaus (BNL) OSG All Hands Consortium Meeting March 3, 2008.
Multi User Pilot Jobs update
U.S. ATLAS Grid Production Experience
Readiness of ATLAS Computing - A personal view
Leigh Grundhoefer Indiana University
LHC Data Analysis using a worldwide computing grid
Presentation transcript:

The PanDA Distributed Production and Analysis System Torre Wenaus Brookhaven National Laboratory, USA ISGC 2008 Taipei, Taiwan April 9, 2008 Torre Wenaus Brookhaven National Laboratory, USA ISGC 2008 Taipei, Taiwan April 9, 2008

Torre Wenaus, BNL 2 Panda Overview Launched 8/05 by US ATLAS to achieve scalable data-driven WMS Designed for analysis as well as production Insulate users from distributed computing complexity Lower entry threshold US ATLAS production since end ‘05 Analysis since Spring ’06 ATLAS-wide production since early ‘08 OSG WMS program since 9/06 VO-neutrality Condor integration Security enhancements Launched 8/05 by US ATLAS to achieve scalable data-driven WMS Designed for analysis as well as production Insulate users from distributed computing complexity Lower entry threshold US ATLAS production since end ‘05 Analysis since Spring ’06 ATLAS-wide production since early ‘08 OSG WMS program since 9/06 VO-neutrality Condor integration Security enhancements Workload management system for Production ANd Distributed Analysis Core BNL, UT Arlington/CERN. 4 developers, 1 full time Workload management system for Production ANd Distributed Analysis Core BNL, UT Arlington/CERN. 4 developers, 1 full time

Torre Wenaus, BNL 3 PanDA Attributes Pilots for ‘just in time’ workload management Efficient ‘CPU harvesting’ prior to workload job release Insulation from grid submission latencies, failure modes and inhomogeneities Tight integration with data management and data flow Designed/developed in concert with the ATLAS DDM system Highly automated, extensive monitoring, low ops manpower Based on well proven, highly scalable, robust web technologies Can use any job submission service (CondorG, local batch, EGEE, Condor glide-ins,...) to deliver pilots Global central job queue and management Fast, fully controllable brokerage from the job queue Based on data locality, resource availability, priority, quotas,... Supports multiple system instances for regional partitioning Pilots for ‘just in time’ workload management Efficient ‘CPU harvesting’ prior to workload job release Insulation from grid submission latencies, failure modes and inhomogeneities Tight integration with data management and data flow Designed/developed in concert with the ATLAS DDM system Highly automated, extensive monitoring, low ops manpower Based on well proven, highly scalable, robust web technologies Can use any job submission service (CondorG, local batch, EGEE, Condor glide-ins,...) to deliver pilots Global central job queue and management Fast, fully controllable brokerage from the job queue Based on data locality, resource availability, priority, quotas,... Supports multiple system instances for regional partitioning

Torre Wenaus, BNL 4 Panda Operation Server Architecture LAMP stack Server Architecture LAMP stack T. Maeno

Torre Wenaus, BNL 5 Job Submission T. Maeno

Torre Wenaus, BNL 6 PanDA Server T. Maeno

Torre Wenaus, BNL 7 Pilots

Torre Wenaus, BNL 8 Pilot Scheduling Management Scheduling runs with a grid cert appropriate to and managed by the region and usage mode Scheduling runs with a grid cert appropriate to and managed by the region and usage mode

Torre Wenaus, BNL 9 PanDA Data Flow

Torre Wenaus, BNL 10 Analysis with PanDA: pathena Tadashi Maeno Running the ATLAS software: Locally: athena PanDA: pathena --inDS --outDS Running the ATLAS software: Locally: athena PanDA: pathena --inDS --outDS Outputs can be sent to xrootd/PROOF farm, directly accessible for PROOF analysis

Torre Wenaus, BNL 11 Production vs. Analysis Use the same PanDA and underlying grid infrastructure Same software, monitoring system, facilities No duplicated operations manpower Analysis benefits from problems found/resolved by shifters Separate computing resources Different queues mapping to different clusters Relative usages can be controlled without competition Different policies for data management Data delivery to sites is in production workflow, but not in analysis Analysis jobs go to where the data currently resides User can specify the site, or let the system decide Analysis outputs reside on the local SE until retrieved by user PanDA user-level accounting/quota system supports analysis; being extended to production (at physics working group and project levels) Priorities also supported; important basis for brokerage decisions Use the same PanDA and underlying grid infrastructure Same software, monitoring system, facilities No duplicated operations manpower Analysis benefits from problems found/resolved by shifters Separate computing resources Different queues mapping to different clusters Relative usages can be controlled without competition Different policies for data management Data delivery to sites is in production workflow, but not in analysis Analysis jobs go to where the data currently resides User can specify the site, or let the system decide Analysis outputs reside on the local SE until retrieved by user PanDA user-level accounting/quota system supports analysis; being extended to production (at physics working group and project levels) Priorities also supported; important basis for brokerage decisions

Torre Wenaus, BNL 12 PanDA in Operation : PanDA responsible for US ATLAS production and analysis, performed extremely well ~2500 CPUs, ~35% of total ATLAS production, well in excess of US share of ~20% through efficient resource use Event gen, simu, digitization, pile-up, mixing, reco Many hundreds of different physics processes simulated ~350 analysis users, 125 with >1k jobs, >2M total jobs 1 shifter, ~.1 FTE to manage PanDA ops itself 2008: PanDA responsible for ATLAS-wide production Roll-out smooth, and almost complete; no scaling issues seen But scaling studies ongoing, on web server and DB levels, to keep ahead of the growth curve Expansion includes a second full PanDA instance at CERN For ‘human’ more than technical reasons But a valuable further scaling knob for the future : PanDA responsible for US ATLAS production and analysis, performed extremely well ~2500 CPUs, ~35% of total ATLAS production, well in excess of US share of ~20% through efficient resource use Event gen, simu, digitization, pile-up, mixing, reco Many hundreds of different physics processes simulated ~350 analysis users, 125 with >1k jobs, >2M total jobs 1 shifter, ~.1 FTE to manage PanDA ops itself 2008: PanDA responsible for ATLAS-wide production Roll-out smooth, and almost complete; no scaling issues seen But scaling studies ongoing, on web server and DB levels, to keep ahead of the growth curve Expansion includes a second full PanDA instance at CERN For ‘human’ more than technical reasons But a valuable further scaling knob for the future

Torre Wenaus, BNL 13 US ATLAS PanDA Production ‘06-’07 Walltime CPU-months Walltime CPU-months January December 2007 Distribution among US Tier 1 (BNL) and 5 Tier 2s Distribution among US Tier 1 (BNL) and 5 Tier 2s K. De PanDA, EGEE, Nordic Jan-Nov ‘07 PanDA, EGEE, Nordic Jan-Nov ‘07

Torre Wenaus, BNL 14 Job Error Rates Apr-May 07 Errors from all sources. PanDA errors mainly application errors K. De

Torre Wenaus, BNL 15 ATLAS PanDA Production ‘08 Roll-out across ATLAS began fall ‘07 Today PanDA production spans all 10 ATLAS regions almost 200 sites/queues 10k-18k concurrent jobs Nordic grid (NDGF) just coming online Italy ramping up; served by distinct (CERN) PanDA system instance Roll-out across ATLAS began fall ‘07 Today PanDA production spans all 10 ATLAS regions almost 200 sites/queues 10k-18k concurrent jobs Nordic grid (NDGF) just coming online Italy ramping up; served by distinct (CERN) PanDA system instance Production snapshot yesterday

Torre Wenaus, BNL 16

Torre Wenaus, BNL 17 Production Snapshot Apr 8 ‘08

Torre Wenaus, BNL 18 Non-ATLAS OSG Usage Currently CHARMM protein folding application. Paper based on results User/VO does - job submission, using simple http-based Python client - pilot submission, such that pilots carry their DN identity - queue selection: tagging queues for inclusion in a logical ‘CHARMM’ queue BNL/ATLAS/OSG provides - PanDA service/DB infrastructure; same as used by ATLAS - Panda monitoring, VO customized - Machine(s) for VO pilot submission - Support from ~3 FTE pool at BNL - Future: Data mgmt and data-driven workflow User/VO does - job submission, using simple http-based Python client - pilot submission, such that pilots carry their DN identity - queue selection: tagging queues for inclusion in a logical ‘CHARMM’ queue BNL/ATLAS/OSG provides - PanDA service/DB infrastructure; same as used by ATLAS - Panda monitoring, VO customized - Machine(s) for VO pilot submission - Support from ~3 FTE pool at BNL - Future: Data mgmt and data-driven workflow

Torre Wenaus, BNL 19 Security in PanDA Uses grid certificate based security for the PanDA server and its client communications (https) OSG workload management effort at BNL in collaboration with BNL Tier 1 and OSG very active in security extensions Currently integrating glExec to switch pilot identity to that of the downloaded user job Prototype implemented and working; integration in production underway Fermilab acting as US/OSG test site; finding an EGEE site is an ongoing challenge! Implementing limited proxies (pilot role) to restrict pilot capabilities Prevent users from gaining access to extensive rights via the pilot proxy Pilot proxy also concealed from the application code (removed from disk) Client Server validation, payload validation to come Drawing on CMS/FNAL work Will prevent hijacking PanDA pilots to run unauthorized jobs Authorizing pilot submitters and pilots themselves (beyond simply holding valid certificate) with time-limited token also to come Uses grid certificate based security for the PanDA server and its client communications (https) OSG workload management effort at BNL in collaboration with BNL Tier 1 and OSG very active in security extensions Currently integrating glExec to switch pilot identity to that of the downloaded user job Prototype implemented and working; integration in production underway Fermilab acting as US/OSG test site; finding an EGEE site is an ongoing challenge! Implementing limited proxies (pilot role) to restrict pilot capabilities Prevent users from gaining access to extensive rights via the pilot proxy Pilot proxy also concealed from the application code (removed from disk) Client Server validation, payload validation to come Drawing on CMS/FNAL work Will prevent hijacking PanDA pilots to run unauthorized jobs Authorizing pilot submitters and pilots themselves (beyond simply holding valid certificate) with time-limited token also to come

Torre Wenaus, BNL 20 Near Term PanDA Activities, Plans Scaling. Have not hit limits yet and want to keep it that way. MySQL performance, load sharing, failover, partitioning Apache performance, tuning, redundancy, multi-host System partitioning (time, region, activity,...) Security -- as discussed ATLAS-wide deployment Largely complete but tuning and extensions based on operations experience continues -- principal developer activity Code organization, configuration management, packaging Housecleaning after 2.5 years! Maintainability, ease of component and service installation, easier pickup outside ATLAS OSG. PanDA is part of the OSG WMS program, managed from BNL Security, Condor pilot glideins (with CMS), Pilot Factory (Condor scheduler glideins), production OSG WMS service, OSG infrastructure services (site attribute discovery/publishing, resource usage management) Scaling. Have not hit limits yet and want to keep it that way. MySQL performance, load sharing, failover, partitioning Apache performance, tuning, redundancy, multi-host System partitioning (time, region, activity,...) Security -- as discussed ATLAS-wide deployment Largely complete but tuning and extensions based on operations experience continues -- principal developer activity Code organization, configuration management, packaging Housecleaning after 2.5 years! Maintainability, ease of component and service installation, easier pickup outside ATLAS OSG. PanDA is part of the OSG WMS program, managed from BNL Security, Condor pilot glideins (with CMS), Pilot Factory (Condor scheduler glideins), production OSG WMS service, OSG infrastructure services (site attribute discovery/publishing, resource usage management)

Torre Wenaus, BNL 21 Conclusion PanDA has delivered its objectives well Efficient resource utilization (pilots and data management) Ease of use and operation (automation, monitoring, global queue/broker) Makes both production and analysis users happy Current focus is on broadened deployment and supporting scale-up ATLAS-wide production deployment Expanding analysis deployment (glExec dependent in some regions) Also support for ever-lengthening list of production, analysis use cases Reconstruction reprocessing just implemented PROOF/xrootd integration implemented and expanding Data management/handling always the greatest challenge Extending automated data handling now that PanDA operates ATLAS-wide Leveraging OSG effort and expertise, especially in security and delivering services On track to provide stable and robust service for physics datataking Ready in terms of functionality, performance and scaling behavior Further development is incremental Starting to turn our scalability knobs, in advance of operational need PanDA has delivered its objectives well Efficient resource utilization (pilots and data management) Ease of use and operation (automation, monitoring, global queue/broker) Makes both production and analysis users happy Current focus is on broadened deployment and supporting scale-up ATLAS-wide production deployment Expanding analysis deployment (glExec dependent in some regions) Also support for ever-lengthening list of production, analysis use cases Reconstruction reprocessing just implemented PROOF/xrootd integration implemented and expanding Data management/handling always the greatest challenge Extending automated data handling now that PanDA operates ATLAS-wide Leveraging OSG effort and expertise, especially in security and delivering services On track to provide stable and robust service for physics datataking Ready in terms of functionality, performance and scaling behavior Further development is incremental Starting to turn our scalability knobs, in advance of operational need