PanDA@BBP R.Mashinistov (UTA) TIM@BNL July 20-21.

Slides:



Advertisements
Similar presentations
3rd Campus Grid SIG Meeting. Agenda Welcome OMII Requirements document Grid Data Group HTC Workshop Research Computing SIG? AOB Next meeting (AG)
Advertisements

P-GRADE and WS-PGRADE portals supporting desktop grids and clouds Peter Kacsuk MTA SZTAKI
1 Grid services based architectures Growing consensus that Grid services is the right concept for building the computing grids; Recent ARDA work has provoked.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
Testing PanDA at ORNL Danila Oleynik University of Texas at Arlington / JINR PanDA UTA 3-4 of September 2013.
System Center 2012 R2 Windows Azure Pack Service Management Automation 101.
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
A Brief Overview by Aditya Dutt March 18 th ’ Aditya Inc.
1 1 Hybrid Cloud Solutions (Private with Public Burst) Accelerate and Orchestrate Enterprise Applications.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Cloud Usage Overview The IBM SmartCloud Enterprise infrastructure provides an API and a GUI to the users. This is being used by the CloudBroker Platform.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
MSc. Miriel Martín Mesa, DIC, UCLV. The idea Installing a High Performance Cluster in the UCLV, using professional servers with open source operating.
Through the development of advanced middleware, Grid computing has evolved to a mature technology in which scientists and researchers can leverage to gain.
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
22 nd September 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Holding slide prior to starting show. A Portlet Interface for Computational Electromagnetics on the Grid Maria Lin and David Walker Cardiff University.
6 February 2009 ©2009 Cesare Pautasso | 1 JOpera and XtremWeb-CH in the Virtual EZ-Grid Cesare Pautasso Faculty of Informatics University.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Ricardo Rocha CERN (IT/GS) EGEE’08, September 2008, Istanbul, TURKEY Experiment.
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
The GridPP DIRAC project DIRAC for non-LHC communities.
LSF Universus By Robert Stober Systems Engineer Platform Computing, Inc.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
OpenStack Chances and Practice at IHEP Haibo, Li Computing Center, the Institute of High Energy Physics, CAS, China 2012/10/15.
The GridPP DIRAC project DIRAC for non-LHC communities.
Active-HDL Server Farm Course 11. All materials updated on: September 30, 2004 Outline 1.Introduction 2.Advantages 3.Requirements 4.Installation 5.Architecture.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
CMS Experience with the Common Analysis Framework I. Fisk & M. Girone Experience in CMS with the Common Analysis Framework Ian Fisk & Maria Girone 1.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
DIRAC for Grid and Cloud Dr. Víctor Méndez Muñoz (for DIRAC Project) LHCb Tier 1 Liaison at PIC EGI User Community Board, October 31st, 2013.
The LGI Pilot job portal EGI Technical Forum 20 September 2011 Jan Just Keijser Willem van Engen Mark Somers.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
Al Lilianstrom and Dr. Olga Terlyga NLIT 2016 May 4 th, 2016 Under the Hood of Fermilab’s Identity Management Service.
PANDA PILOT FOR HPC Danila Oleynik (UTA). Outline What is PanDA Pilot PanDA Pilot architecture (at nutshell) HPC specialty PanDA Pilot for HPC 2.
BY: SALMAN 1.
Consulting Services JobScheduler Architecture Decision Template
Use of HLT farm and Clouds in ALICE
Open OnDemand: Open Source General Purpose HPC Portal
Computing Clusters, Grids and Clouds Globus data service
BY: SALMAN.
StratusLab First Periodic Review
AWS Integration in Distributed Computing
Virtualisation for NA49/NA61
Blueprint of Persistent Infrastructure as a Service
Belle II Physics Analysis Center at TIFR
Working With Azure Batch AI
PanDA setup at ORNL Sergey Panitkin, Alexei Klimentov BNL
StratusLab Final Periodic Review
Consulting Services JobScheduler Architecture Decision Template
StratusLab Final Periodic Review
Provisioning 160,000 cores with HEPCloud at SC17
HEP Computing Tools for Brain Studies
Get the Most Out of GoAnywhere: Agents
Enable computational and experimental  scientists to do “more” computational chemistry by providing capability  computing resources and services at their.
David Cameron ATLAS Site Jamboree, 20 Jan 2017
Virtualisation for NA49/NA61
Ruslan Fomkin and Tore Risch Uppsala DataBase Laboratory
Recap: introduction to e-science
BigPanDA WMS for Brain Studies
WLCG Collaboration Workshop;
Advanced Computing Facility Introduction
Module 01 ETICS Overview ETICS Online Tutorials
H2020 EU PROJECT | Topic SC1-DTH | GA:
Presentation transcript:

PanDA@BBP R.Mashinistov (UTA) TIM@BNL July 20-21

Introduction In 2017 the pilot project between BigPanDA and Blue Brain Project (BBP) (Swiss Federal Institute of Technology in Lausanne) teams began. The proof of concept project aimed to demonstrate efficient application of the PanDA software for the supercomputer-based reconstructions and simulations offering a radically new approach for understanding the multilevel structure and function of the brain. In the first phase, the goal of this joint project is to support the execution of BBP software on a variety of distributed computing systems powered by PanDA. The targeted systems for demonstration include: Intel x86-NVIDIA GPU based BBP clusters located in Geneva (47 TFlops) and Lugano (81 TFlops), BBP IBM BlueGene/Q supercomputer ( 0.78 PFLops and 65 TB of DRAM memory) located in Lugano, the Titan Supercomputer with peak theoretical performance 27 PFlops operated by the Oak Ridge Leadership Computing Facility (OLCF), and Cloud based resources such as Amazon Cloud.

BBP computing infrastructure ssh slurm slurm slurm pbs api & sheduler BlueGene/Q viz cluster lxviz cluster Titan supercomputer On-demand cluster LUGANO GENEVE

Initial PanDA test @ BBP For initial test in March 2017 PanDA Server + Client were installed to the VM at the Campus Biotech (Geneva). Job submission has been done via the python based PanDA Client. Pilot were started manually on Geneve and Lugano clusters including BlueGene/Q. Titan & Amazon resources were integrated on later (on April). SQL https PanDA Server PanDA DB PanDA Client REST API REST API https Pilot Pilot Pilot Resource A Resource B Resource Z TIM@BNL July 20-21

PanDA Portal @ BBP (1/2) Authentication File Catalog Data Transfer System Data Management System Web Services Pre/post processing SQL SQL https PanDA Server PanDA DB PanDA Monitor PanDA Client REST API REST API https Pilot Scheduler Pilot Scheduler Pilot Pilot Pilot Resource A Resource B Resource Z TIM@BNL July 20-21

PanDA Portal @ BBP (2/2) PanDA Portal Authentication File Catalog Data Transfer System Data Management System Web Services Pre/post processing SQL SQL https PanDA Server PanDA DB PanDA Monitor PanDA Client REST API REST API https Pilot Scheduler Pilot Scheduler Pilot Pilot Pilot Resource A Resource B Resource Z TIM@BNL July 20-21

PanDA for BlueBrainProject ssh https https Various interfaces The test ‘Hello world’ jobs were successfully submited to the targeted resources via PanDA portal. CLI REST API WEB UI PanDA Portal@BBP Queues to execute jobs on different resources PanDA queues BlueGene/Q viz cluster lxviz cluster Titan OLCF GENEVE LUGANO

User Interfaces Web forms to define, submit and monitor jobs Authentication LDAP as a first solution Will be replaced by SSO Possible to integrate with Jupyter notebooks (widely used by BlueBrain scientists) Easy to change computing cluster (jobs are not atomic, typically scientist knows which cluster should be used with his computational job) TIM@BNL July 20-21

Impersonation Need to track job submitter on computing cluster Pilot should have access only to the private files of job owner It’s possible to propagate name and group of user inside job definition Root access approach Run pilot under root, use name/group as sbatch params Root writes transformation wrapper with SUID, use name/group as sbatch params inside wrapper No-root access approach Use separate queue for every working group (cannot add cluster user for every project participant) Run pilot under working group manager account Working groups can be mapped with existing groups in LDAP or SSO TIM@BNL July 20-21

Conclusion Phase 1 of “Proof of the concept” was successfully finished The project demonstrated that the software tools and methods for processing large volumes of experimental data, which have been developed initially for experiments at the LHC accelerator, can be successfully applied to BBP. Phase 2 of “Pre-production” is under investigation currently Next most significant tasks are: Data management system Jupyter integration