Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.

Slides:



Advertisements
Similar presentations
Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Advertisements

4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
Status of BESIII Distributed Computing BESIII Workshop, Mar 2015 Xianghu Zhao On Behalf of the BESIII Distributed Computing Group.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL March 25, 2003 CHEP 2003 Data Analysis Environment and Visualization.
STAR Software Walk-Through. Doing analysis in a large collaboration: Overview The experiment: – Collider runs for many weeks every year. – A lot of data.
The B A B AR G RID demonstrator Tim Adye, Roger Barlow, Alessandra Forti, Andrew McNab, David Smith What is BaBar? The BaBar detector is a High Energy.
Enabling Grids for E-sciencE Medical image processing web portal : Requirements analysis. An almost end user point of view … H. Benoit-Cattin,
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
J OINT I NSTITUTE FOR N UCLEAR R ESEARCH OFF-LINE DATA PROCESSING GRID-SYSTEM MODELLING FOR NICA 1 Nechaevskiy A. Dubna, 2012.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
MySQL and GRID Gabriele Carcassi STAR Collaboration 6 May Proposal.
100 Million events, what does this mean ?? STAR Grid Program overview.
Overview Why are STAR members encouraged to use SUMS ? Improvements and additions to SUMS Research –Job scheduling with load monitoring tools –Request.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May STAR Computing Doug Olson for: J é r ô me Lauret.
Tier 1 Facility Status and Current Activities Rich Baker Brookhaven National Laboratory NSF/DOE Review of ATLAS Computing June 20, 2002.
CHEP Sep Andrey PHENIX Job Submission/Monitoring in transition to the Grid Infrastructure Andrey Y. Shevel, Barbara Jacak,
STAR Analysis Meeting, BNL, Dec 2004 Alexandre A. P. Suaide University of Sao Paulo Slide 1 BEMC software and calibration L3 display 200 GeV February.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
HEPD sem 14-Dec Andrey History photos: A. Shevel reports on CSD seminar about new Internet facilities at PNPI (Jan 1995)
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
PHENIX and the data grid >400 collaborators Active on 3 continents + Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
STAR Software Walk-Through. Doing analysis in a large collaboration: Overview The experiment: – Collider runs for many weeks every year. – A lot of data.
STAR Collaboration Meeting, BNL, Feb 2005 Alexandre A. P. Suaide University of Sao Paulo Slide 1 BEMC software update L3 display 200 GeV February.
PPDG update l We want to join PPDG l They want PHENIX to join NSF also wants this l Issue is to identify our goals/projects Ingredients: What we need/want.
February 28, 2003Eric Hjort PDSF Status and Overview Eric Hjort, LBNL STAR Collaboration Meeting February 28, 2003.
David Adams ATLAS DIAL: Distributed Interactive Analysis of Large datasets David Adams BNL August 5, 2002 BNL OMEGA talk.
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
BaBar and the GRID Tim Adye CLRC PP GRID Team Meeting 3rd May 2000.
STAR Collaboration Meeting, BNL – march 2003 Alexandre A. P. Suaide Wayne State University Slide 1 EMC Update Update on EMC –Hardware installed and current.
Computing for Alice at GSI (Proposal) (Marian Ivanov)
LHCb report to LHCC and C-RSG Philippe Charpentier CERN on behalf of LHCb.
AliRoot survey: Analysis P.Hristov 11/06/2013. Are you involved in analysis activities?(85.1% Yes, 14.9% No) 2 Involved since 4.5±2.4 years Dedicated.
Alignment in real-time in current detector and upgrade 6th LHCb Computing Workshop 18 November 2015 Beat Jost / Cern.
U.S. ATLAS Facility Planning U.S. ATLAS Tier-2 & Tier-3 Meeting at SLAC 30 November 2007.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Production Mode Data-Replication Framework in STAR using the HRM Grid CHEP ’04 Congress Centre Interlaken, Switzerland 27 th September – 1 st October Eric.
Jiri Chudoba for the Pierre Auger Collaboration Institute of Physics of the CAS and CESNET.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Alien and GSI Marian Ivanov. Outlook GSI experience Alien experience Proposals for further improvement.
CCJ introduction RIKEN Nishina Center Kohei Shoji.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
ATLAS Computing: Experience from first data processing and analysis Workshop TYL’10.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
Using the Grid for the ILC Mokka and Marlin on the Grid ILC Software Meeting, Cambridge 2006.
David Adams ATLAS DIAL Distributed Interactive Analysis of Large datasets David Adams BNL May 19, 2003 BNL Technology Meeting.
Compute and Storage For the Farm at Jlab
Accessing the VI-SEEM infrastructure
U.S. ATLAS Grid Production Experience
ALICE Physics Data Challenge 3
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Artem Trunov and EKP team EPK – Uni Karlsruhe
ALICE Computing Upgrade Predrag Buncic
ATLAS DC2 & Continuous production
Production Manager Tools (New Architecture)
Presentation transcript:

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 BNL (2 sites) –~ 1100 CPU –~ 400 TB –LSF batch NERSC-PDSF –~ 500 CPU –~ 150 TB –SGE batch São Paulo –Test cluster 10 CPU, 3 TB – SGE batch Upgrade project ~ 50 CPU and ~ 40 TB

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 The size of the raw data STAR Au+Au event statistics (raw) –~ 2-3 MB/event –~ events/s –Total 2004 Au+Au M events ~ 65 TB Cu+Cu run –~ 70 M 200 GeV –~ 40 M 62 GeV –~ 4 M 22 GeV Plus all the p+p, d+Au and previous runs

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 The reconstruction, simulation, etc. Reconstruction –Basically done in BNL –Au+Au is estimated to take 18 months (only 60% is complete) Compare with 1 new run every year –A physics ready production needs ~ 2 production rounds (calibrations, improvements, etc) Simulation and embedding –Done at PDSF –Simulation is transferred to BNL STAR takes more data that it currently can make available for analysis

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 Analysis Real data analysis is done in RCF Simulation and embedding analysis is done in PDSF Small fractions of datasets are scattered over many institutions mainly for analysis PDSF

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 Why do we need grid? If STAR wants to keep the production and analysis running in a speed compatible with data taking, other institutions need to share computer power –Next run STAR will take at least one order of magnitude more events than last year –The RCF/PSDF farm does not grow in the same rate The user point of view –More time available for physics Data will be available earlier –More computing power for analysis Analysis will run faster –Submit the jobs from your home institution and get the output in there No need to know where the data is No need to log on RCF or PDSF You manage your disk space

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid Three level structure Tier0 sites (BNL) –Dedicated to reconstruction, simulation and analysis Tier1 sites (PDSF) –Runs reconstruction on demand –Receives all the reconstructed files for analysis –Simulations and embedding Tier2 sites (all other facilities, including São Paulo) –Receives a fraction of files for analysis –Eventually runs reconstruction depending on demand

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 Needs Reconstruction and file distribution – Tier0 production ALL EVENT files get copied on HPSS at the end of a job Strategy implies dataset IMMEDIATE replication –As soon as a file is registered, it becomes available for “distribution” –2 Levels of data distributions – Local and Global – Local All analysis files are on disks –Notions of distributed disk – Cost effective solution – Global Tier1 (all) and tier2 (partial) sites Cataloging is fundamental –Must know where the files are –The only central connection between users and files –Central and local catalogs Database should be updated right after file transfer Customized scheduler –Find out where data is upon user request –Redirect jobs to cluster where data is saved –Job submission should not be random but highly coordinated with other users requests

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 What is STAR doing on grid? For STAR, grid computing is EVERY DAY Production used – Data transfer using SRM, RRS,.. – We run simulation production on the Grid (easy) – Resource reserved for DATA production (still done traditionally) No real technical difficulties Mostly fears related to un-coordinated access and massive transfers – User analysis Chaotic in nature, requires accounting, quota, privilege, etc … Increase interest from some institutions – Already success under controlled conditions

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR jobs in the grid

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 Accomplishments in the last few months Full database mirrors over many institutions –Hold detector conditions, calibrations, status, etc… –Highly used during user analisys File catalog and scheduler available outside BNL –User can query files and submit jobs using grid Still some pitfalls for general user analysis Integration between sites –Tools to keep grid certificates, batch systems and local catalogs updated –Library distribution automatically done using AFS or local copy (updated in a daily basis) Full integration of the 3 sites (BNL, PDSF and SP) with OSG

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 User analysis in the grid STAR analysis schema –99% based on ROOT applications –User develops personal analysis code that process the data Steps to properly submit analysis jobs in the grid –Select the proper cluster in the grid –Transfer and compile the analysis code to that cluster –Use the file catalog to select the files –Run the jobs (as many as necessary) The node the job runs and the number of jobs is defined by the scheduler and depends on the cluster size, number of events and time to process each event. All this information is managed by the file catalog –Transfer the output to the local site Many of these steps are not yet fully functional but progressing fast

Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 Current status and to do list The GRID between PSDF and RCF works quite well –Mainly used for simulation jobs São Paulo, BNL and LBL are fully integrated –Libraries, file catalog, scheduler, OSG, etc. –Being used to test user analysis under the grid Activities for the next few months –Integrate the SGE batch system in the grid framework Still some problems with respect to report right numbers to gridCat Problems keeping jobs alive after few hours –Developments of authentication tools RCF (BNL) and PDSF (LBL) are part of DOE labs –User analysis