HEPD sem 14-Dec-2004 1 Andrey History photos: A. Shevel reports on CSD seminar about new Internet facilities at PNPI (Jan 1995)

Slides:



Advertisements
Similar presentations
4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
Advertisements

1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Data Management for Physics Analysis in PHENIX (BNL, RHIC) Evaluation of Grid architecture components in PHENIX context Barbara Jacak, Roy Lacey, Saskia.
K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
The B A B AR G RID demonstrator Tim Adye, Roger Barlow, Alessandra Forti, Andrew McNab, David Smith What is BaBar? The BaBar detector is a High Energy.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Enabling Grids for E-sciencE Medical image processing web portal : Requirements analysis. An almost end user point of view … H. Benoit-Cattin,
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
Using the WDK for Windows Logo and Signature Testing Craig Rowland Program Manager Windows Driver Kits Microsoft Corporation.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Tech talk 20th June Andrey Grid architecture at PHENIX Job monitoring and related stuff in multi cluster environment.
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
Andrew McNab - Manchester HEP - 5 July 2001 WP6/Testbed Status Status by partner –CNRS, Czech R., INFN, NIKHEF, NorduGrid, LIP, Russia, UK Security Integration.
WP9 Resource Management Current status and plans for future Juliusz Pukacki Krzysztof Kurowski Poznan Supercomputing.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
Tier 1 Facility Status and Current Activities Rich Baker Brookhaven National Laboratory NSF/DOE Review of ATLAS Computing June 20, 2002.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
PNPI HEPD seminar 4 th November Andrey Shevel Distributed computing in High Energy Physics with Grid Technologies (Grid tools at PHENIX)
11 SUPPORTING APPLICATIONS IN WINDOWS XP PROFESSIONAL Chapter 9.
CHEP Sep Andrey PHENIX Job Submission/Monitoring in transition to the Grid Infrastructure Andrey Y. Shevel, Barbara Jacak,
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
4/20/02APS April Meeting1 Database Replication at Remote sites in PHENIX Indrani D. Ojha Vanderbilt University (for PHENIX Collaboration)
St.Petersburg state university computing centre and the 1st results in the DC for ALICE V.Bychkov, G.Feofilov, Yu.Galyuck, A.Zarochensev, V.I.Zolotarev.
17th Sep 2007 Andrey Computing cluster at NCG Introduction Past upgrades Current state of the cluster Problems with cluster Where to find.
Developing & Managing A Large Linux Farm – The Brookhaven Experience CHEP2004 – Interlaken September 27, 2004 Tomasz Wlodek - BNL.
22 nd September 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
David Adams ATLAS ADA, ARDA and PPDG David Adams BNL June 28, 2004 PPDG Collaboration Meeting Williams Bay, Wisconsin.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
Andrey Meeting 7 October 2003 General scheme: jobs are planned to go where data are and to less loaded clusters SUNY.
PHENIX and the data grid >400 collaborators Active on 3 continents + Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep , 2014 Draft.
Virtual Batch Queues A Service Oriented View of “The Fabric” Rich Baker Brookhaven National Laboratory April 4, 2002.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
PPDG update l We want to join PPDG l They want PHENIX to join NSF also wants this l Issue is to identify our goals/projects Ingredients: What we need/want.
February 28, 2003Eric Hjort PDSF Status and Overview Eric Hjort, LBNL STAR Collaboration Meeting February 28, 2003.
Databases for data management in PHENIX Irina Sourikova Brookhaven National Laboratory for the PHENIX collaboration.
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
January 26, 2003Eric Hjort HRMs in STAR Eric Hjort, LBNL (STAR/PPDG Collaborations)
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Participation of JINR in CERN- INTAS project ( ) Korenkov V., Mitcin V., Nikonov E., Oleynik D., Pose V., Tikhonenko E. 19 march 2004.
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
Status of Globus activities Massimo Sgaravatto INFN Padova for the INFN Globus group
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
10 March Andrey Grid Tools Working Prototype of Distributed Computing Infrastructure for Physics Analysis SUNY.
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
Dave Newbold, University of Bristol14/8/2001 Testbed 1 What is it? First deployment of DataGrid middleware tools The place where we find out if it all.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
(Prague, March 2009) Andrey Y Shevel
U.S. ATLAS Grid Production Experience
Simulation use cases for T2 in ALICE
Michael P. McCumber Task Force Meeting April 3, 2006
Presentation transcript:

HEPD sem 14-Dec Andrey History photos: A. Shevel reports on CSD seminar about new Internet facilities at PNPI (Jan 1995)

HEPD sem 14-Dec Andrey

HEPD sem 14-Dec Andrey Distributed computing in HEP [Grid prospects] Andrey Y. Shevel

HEPD sem 14-Dec Andrey PHENIX Job Submission/Monitoring in transition to the Grid Infrastructure Andrey Y. Shevel, Barbara Jacak, Roy Lacey, Dave Morrison, Michael Reuter, Irina Sourikova, Timothy Thomas, Alex Withers

HEPD sem 14-Dec Andrey Brief info on PHENIX + Large, widely-spread collaboration (same scale as CDF and D0), more than 450 collaborators, 12 nations, 57 Institutions, 11 U.S. Universities, currently in fourth year of data-taking. + ~250 TB/yr of raw data. + ~230 TB/yr of reconstructed output. + ~370 TB/yr microDST + nanoDST. + In total about ~850TB+ of new data per year. + Primary event reconstruction occurs at BNL RCF (RHIC Computing Facility). + Partial copy of raw data is at CC-J (Computing Center in Japan) and part of DST output is at CC-F (France).

HEPD sem 14-Dec Andrey PHENIX Grid We could expect in total about 10 clusters in nearest years. RIKEN CCJ (Japan) University of New Mexico Job submission Cluster RAM Brookhaven National Lab Vanderbilt University PNPI (Russia) Stony Brook Data moving IN2P3 (France) CCJ

HEPD sem 14-Dec Andrey PHENIX multi cluster conditions + Computing Clusters have different: - computing power; - batch job schedulers; - details of administrative rules. + Computing Clusters have common: - OS Linux (there are clusters with different Linux versions); - Most of clusters have gateways with Globus toolkit; - Grid status board ( )

HEPD sem 14-Dec Andrey Other PHENIX conditions + Max number of the computing clusters is about Max number of the submitted at the same time Grid jobs is about 10**4 or less. + The amount of the data to be transferred (between BNL and remote cluster) for physics analysis is varied from about 2 TB/quarter to 5 TB/week. + We use PHENIX file catalogs: - centralized file catalog ( ); - cluster file catalogs (for example at SUNYSB is used slightly re- designed version MAGDA ).

HEPD sem 14-Dec Andrey Exporting the application software to run on remote clusters + The porting of PHENIX software in binary form is presumably most common port method in PHENIX Grid : - copying over AFS to mirror PHENIX directory structure on remote cluster (by cron job); - preparing PACMAN packages for specific class of tasks (e.g. specific simulation).

HEPD sem 14-Dec Andrey The requirements for job monitoring in multi cluster environment + What is job monitoring ? + To keep track of the submitted jobs - whether the jobs have been accomplished; - in which cluster the jobs are performed; - where the jobs were performed in the past (one day, one week, one month ago). + Obviously the information about the jobs must be written in the database and kept there. The same database might be used for job control purpose (cancel jobs, resubmit jobs, other job control operations in multi cluster environment) + PHENIX job monitoring tool was developed on the base of BOSS (

HEPD sem 14-Dec Andrey “Challenges for PHENIX Grid” + Admin service (where the user can complain if something is going wrong with his Grid jobs on some cluster?). + More sophisticated job control in multi cluster environment; job accounting. + Complete implementing technology for run-time installation for remote clusters. + More checking tools to be sure that most things in multi cluster environment are running well – i.e. automate the answer for the question “is account A on cluster N being PHENIX qualified environment?”. To check it every hour or so. + Portal to integrate all PHENIX Grid tools in one user window. [388] A Lightweight Monitoring and Accounting System for LHCb DC04 Production [476] CHOS, a method for concurrently supporting multiple operating system. [455] Application of the SAMGrid Test Harness for Performance Evaluation and Tuning of a Distributed Cluster Implementation of Data Handling Services [443] The AliEn Web PortalThe AliEn Web Portal [182] Grid Enabled Analysis for CMS: prototype, status and resultsGrid Enabled Analysis for CMS: prototype, status and results

HEPD sem 14-Dec Andrey

HEPD sem 14-Dec Andrey My Summary on CHEP The multi cluster environment is PHENIX reality and we need more user friendly tools for typical user to reduce the cost of clusters power integration. + In our condition the best way to do that is to use already developed subsystems as bricks to build up the robust PHENIX Grid computing environment. Most effective way to do that is to be AMAP cooperative with other BNL collaborations (STAR as good example). + Serious attention must be paid to automatic installation of the existing physics software.

HEPD sem 14-Dec Andrey Many flavors of grid systems (no 100% compatibility) + Grid SAM + EGEE + NORDUGRID …. SAM looks most working but … SAM development was started at 1987…

HEPD sem 14-Dec Andrey What was mentioned often … + Data handling issues: D-cache; xrootd; SRM ( [ 334] Production mode Data-Replication framework in STAR using the HRM Grid)Production mode Data-Replication framework in STAR using the HRM Grid + Security issues. + Grid Administration/Operation/Support centers. + Deployment issues.

HEPD sem 14-Dec Andrey Development hit – xrootd (Example: SLAC Configuration) client machines kan01kan02kan03kan04 kanxx bbr-olb03bbr-olb04 kanolb-a

HEPD sem 14-Dec Andrey Grid prospects + Many small problems are transformed into one big problem (Grid :-). + Advantages (point of balance of interests) - for funding authorities; - for institutes; - for collaborations; - for end users (physicists).

HEPD sem 14-Dec Andrey Estimates

HEPD sem 14-Dec Andrey Grid computing advantage (simulation versus analysis) + The simulation on Grid structure implies high volume data transfer (i.e. overheads); + On other hand the data analysis assumes limited data transfer (once for relatively long period, may be once per ½ year).

HEPD sem 14-Dec Andrey Conclusion PNPI role in Grid + Anybody who plans to participate in accelerator physics simulation/analysis has to learn the basics of Grid computing organization and collaboration rules where you plan to participate (to get Grid certificate as the first step). + In order to do so HEPD has to keep up to date own computing cluster facility (about 10 TB of disk space and appropriate computing power) and external data transfer throughput 1-5 MBytes/sec.