PNPI HEPD seminar 4 th November 2003 1 Andrey Shevel Distributed computing in High Energy Physics with Grid Technologies (Grid tools at PHENIX)

Slides:



Advertisements
Similar presentations
Globus Workshop at CoreGrid Summer School 2006 Dipl.-Inf. Hamza Mehammed Leibniz Computing Centre.
Advertisements

FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
1 OBJECTIVES To generate a web-based system enables to assemble model configurations. to submit these configurations on different.
DataGrid is a project funded by the European Union 22 September 2003 – n° 1 EDG WP4 Fabric Management: Fabric Monitoring and Fault Tolerance
EU-GRID Work Program Massimo Sgaravatto – INFN Padova Cristina Vistoli – INFN Cnaf as INFN members of the EU-GRID technical team.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
1 GRID Workshop – Toulouse IAS – October 20th Ground segment & products Department Globus TK4 experiment for image data processing : security architecture,
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Data Management for Physics Analysis in PHENIX (BNL, RHIC) Evaluation of Grid architecture components in PHENIX context Barbara Jacak, Roy Lacey, Saskia.
Sun Grid Engine Grid Computing Assignment – Fall 2005 James Ruff Senior Department of Mathematics and Computer Science Western Carolina University.
Basic Grid Job Submission Alessandra Forti 28 March 2006.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
Sergey Belov, Tatiana Goloskokova, Vladimir Korenkov, Nikolay Kutovskiy, Danila Oleynik, Artem Petrosyan, Roman Semenov, Alexander Uzhinskiy LIT JINR The.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Apache Airavata GSOC Knowledge and Expertise Computational Resources Scientific Instruments Algorithms and Models Archived Data and Metadata Advanced.
BaBar WEB job submission with Globus authentication and AFS access T. Adye, R. Barlow, A. Forti, A. McNab, S. Salih, D. H. Smith on behalf of the BaBar.
Grid Data Management A network of computers forming prototype grids currently operate across Britain and the rest of the world, working on the data challenges.
Use of R-GMA in BOSS Henry Nebrensky (Brunel University) VRVS 26 April 2004 Some slides stolen from various talks at EDG 2 nd Review (
Tech talk 20th June Andrey Grid architecture at PHENIX Job monitoring and related stuff in multi cluster environment.
Grid Computing - AAU 14/ Grid Computing Josva Kleist Danish Center for Grid Computing
CoG Kit Overview Gregor von Laszewski Keith Jackson.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
Computational grids and grids projects DSS,
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
LCG Middleware Testing in 2005 and Future Plans E.Slabospitskaya, IHEP, Russia CERN-Russia Joint Working Group on LHC Computing March, 6, 2006.
CHEP Sep Andrey PHENIX Job Submission/Monitoring in transition to the Grid Infrastructure Andrey Y. Shevel, Barbara Jacak,
Stuart Wakefield Imperial College London Evolution of BOSS, a tool for job submission and tracking W. Bacchi, G. Codispoti, C. Grandi, INFN Bologna D.
HEPD sem 14-Dec Andrey History photos: A. Shevel reports on CSD seminar about new Internet facilities at PNPI (Jan 1995)
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
Andrey Meeting 7 October 2003 General scheme: jobs are planned to go where data are and to less loaded clusters SUNY.
13 May 2004EB/TB Middleware meeting Use of R-GMA in BOSS for CMS Peter Hobson & Henry Nebrensky Brunel University, UK Some slides stolen from various talks.
Giuseppe Codispoti INFN - Bologna Egee User ForumMarch 2th BOSS: the CMS interface for job summission, monitoring and bookkeeping W. Bacchi, P.
Claudio Grandi INFN Bologna CHEP'03 Conference, San Diego March 27th 2003 BOSS: a tool for batch job monitoring and book-keeping Claudio Grandi (INFN Bologna)
PHENIX and the data grid >400 collaborators Active on 3 continents + Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Data Manipulation with Globus Toolkit Ivan Ivanovski TU München,
LSF Universus By Robert Stober Systems Engineer Platform Computing, Inc.
Status of Globus activities Massimo Sgaravatto INFN Padova for the INFN Globus group
MGRID Architecture Andy Adamson Center for Information Technology Integration University of Michigan, USA.
STAR Scheduler Gabriele Carcassi STAR Collaboration.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
10 March Andrey Grid Tools Working Prototype of Distributed Computing Infrastructure for Physics Analysis SUNY.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
CSF4 Meta-Scheduler Zhaohui Ding College of Computer Science & Technology Jilin University.
BOSS: the CMS interface for job summission, monitoring and bookkeeping
BOSS: the CMS interface for job summission, monitoring and bookkeeping
BOSS: the CMS interface for job summission, monitoring and bookkeeping
Initial job submission and monitoring efforts with JClarens
Alice Software Demonstration
Gridifying the LHCb Monte Carlo production system
Presentation transcript:

PNPI HEPD seminar 4 th November Andrey Shevel Distributed computing in High Energy Physics with Grid Technologies (Grid tools at PHENIX)

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Topics  Grid/Globus  HEP Grid Projects  PHENIX as the example  Conceptions and scenario for widely distributed multi cluster computing environment  Job submission and job monitoring  Live demonstration

Andrey ShevelPNPI HEPD seminar 4 th November 2003 What is the Grid  “Dependable, consistent, pervasive access to [high-end] resources”.  Dependable: Can provide performance and functionality guarantees.  Consistent: Uniform interfaces to a wide variety of resources.  Pervasive: Ability to “plug in” from anywhere.

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Another Grid description  Quote from Information Power Grid (IPG) at NASA  entations/IPG.AvSafety.VG.1.1up.pdf  Grids are tools, middleware, and services for:  providing a uniform look and feel to a wide variety of computing and data resources;  supporting construction, management, and use of widely distributed application systems;  facilitating human collaboration and remote access and operation of scientific and engineering instrumentation systems;  managing and securing the computing and data infrastructure.

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Basic HEP requirements in distributed computing  Authentication/Authorization/Security  Data Transfer  File/Replica Cataloging  Match Making/Job submission/Job monitoring  System monitoring

Andrey ShevelPNPI HEPD seminar 4 th November 2003 HEP Grid Projects  European Data Grid   Grid Physics Nuclear   Particle Physics Data Grid   Many others.

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Possible Task  Here we are trying to gather computing power around many clusters. The clusters are located in different sites with different authorities. We use all local rules as they are:  Local schedulers, policies, priorities;  Other local circumstances.  One of many possible scenarios is discussed in this presentation.

Andrey ShevelPNPI HEPD seminar 4 th November 2003 General scheme: jobs are planned to go where data are and to less loaded clusters Remote cluster RCF Main Data Repository Partial Data Replica File Catalog User

Andrey ShevelPNPI HEPD seminar 4 th November 2003 GridFTP (Globus-url-copy) Globus job-manager/fork Package GSUNY Cataloging engine BOSS BODE User Jobs GT latest Base subsystems for PHENIX Grid

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Input/Output Sandbox(es) Master Job (script) submitted by user Major Data Sets ( physics or simulated data) Minor Data Sets (Parameters, scripts, etc.) Satellite Job (script) Submitted by Master Job Conceptions

Andrey ShevelPNPI HEPD seminar 4 th November 2003 The job submission scenario at remote Grid cluster  To determine (to know) qualified computing cluster: available disk space, installed software, etc.  To copy/replicate the major data sets to remote cluster.  To copy the minor data sets (scripts, parameters, etc.) to remote cluster.  To start the master job (script) which will submit many jobs with default batch system.  To watch the jobs with monitoring system – BOSS/BODE.  To copy the result data from remote cluster to target destination (desktop or RCF).

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Master job-script  The master script is submitted from your desktop and performed on the Globus gateway (may be in group account) with using monitoring tool (it is assumed BOSS).  It is supposed that the master script will find the following information in the environment variables:  CLUSTER_NAME – name of the cluster;  BATCH_SYSTEM – name of the batch system;  BATCH_SUBMIT – command for job submission through BATCH_SYSTEM.

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Local desktop Globus gateway Submission of MASTER job Through globus-jobmanager/fork Remote Cluster Job submission with Command $BATCH_SUBMIT MASTER job is performing On Globus gateway Job submission scenario

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Transfer the major data sets  There are a number of methods to transfer major data sets:  The utility bbftp (whithout use of GSI) can be used to transfer the data between clusters;  The utility gcopy (with use of GSI) can be used to copy the data from one cluster to another one.  Any third party data transfer facilities (e.g. HRM/SRM).

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Copy the minor data sets  There are at least two alternative methods to copy the minor data sets (scripts, parameters, constants, etc.):  To copy the data to /afs/rhic.bnl.gov/phenix/users/user_account/…  To copy the data with the utility CopyMinorData (part of package gsuny).

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Package gsuny List of scripts  General commands ( ftp://ram3.chem.sunysb.edu/pub/suny-gt-2/gsuny.tar.gz)  GPARAM – configuration description for set of remote clusters  GlobusUserAccountCheck – to check the Globus configuration for local user account.  gping – to test availability of the Globus gateways.  gdemo – to see the load of remote clusters.  gsub – to submit the job on less loaded cluster;  gsub-data – to submit the job where data are;  gstat, gget, gjobs – to get status of the job, standard output, detailed info about jobs.

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Package gsuny Data Transfer  gcopy – to copy the data from one cluster (local hosts) to another one.  CopyMinorData – to copy minor data sets from cluster (local host) to cluster.

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Job monitoring  After the initial development of the description of required monitoring tool ( eeting4Aug2003/jobsub.pdf ) it was found the packages:  Batch Object Submission System (BOSS) by Claudio Grandi  Web interface BO SS D ATABASE E XPLORER (BODE) by Alexei Filine

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Basic BOSS components boss executable: the BOSS interface to the user MySQL database: where BOSS stores job information jobExecutor executable: the BOSS wrapper around the user job dbUpdator executable: the process that writes to the database while the job is running Interface to Local scheduler

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Basic job flow boss submit boss query boss kill BOSS DB BOSS Local Scheduler Exec node n Exec node m Globus gateway BODE (Web interface) Here is cluster N gsub master-script Globus Space To wrap the job

Andrey ShevelPNPI HEPD seminar 4 th November 2003 gsub TbossSuny # submit to less loaded cluster shevel]$ cat TbossSuny. /etc/profile. ~/.bashrc echo " This is master JOB" printenv boss submit -jobtype ram3master -executable ~/andrey.shevel/TestRemoteJobs.pl -stdout \ ~/andrey.shevel/master.out -stderr ~/andrey.shevel/master.err shevel]$ CopyMinorData local:andrey.shevel unm: YOU are copying THE minor DATA sets --FROM-- --TO-- Gateway = 'localhost' 'loslobos.alliance.unm.edu' Directory = '/home/shevel/andrey.shevel' '/users/shevel/.' Transfer of the file '/tmp/andrey.shevel.tgz5558' was succeeded

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Status of the PHENIX Grid  Live info is available on the page  The group account ‘phenix’ is available now at  SUNYSB (rserver1.i2net.sunysb.edu)  UNM (loslobos.alliance.unm.edu)  IN2P3 (in process now)

Andrey ShevelPNPI HEPD seminar 4 th November User: guest Pass: Guest101 Live Demo for BOSS Job monitoring

Andrey ShevelPNPI HEPD seminar 4 th November 2003 Computing Utility (instead conclusion)  It is clear that computing utility (computing cluster built up for concrete collaboration tasks).  The computing utility can be implemented anywhere in the World.  The computing utility can be used from anywhere (France, USA, Russia, etc.).  Most important part of the computing utility is man power.