Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May 2005 1 STAR Computing Doug Olson for: J é r ô me Lauret.

Slides:



Advertisements
Similar presentations
RHIC, STAR computing towards distributed computing on the Open Science Grid Jérôme LAURET RHIC/STAR.
Advertisements

Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Grid Collector: Enabling File-Transparent Object Access For Analysis Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
Grid Data Management A network of computers forming prototype grids currently operate across Britain and the rest of the world, working on the data challenges.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Central Reconstruction System on the RHIC Linux Farm in Brookhaven Laboratory HEPIX - BNL October 19, 2004 Tomasz Wlodek - BNL.
The Data Grid: Towards an Architecture for the Distributed Management and Analysis of Large Scientific Dataset Caitlin Minteer & Kelly Clynes.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
100 Million events, what does this mean ?? STAR Grid Program overview.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Using Bitmap Index to Speed up Analyses of High-Energy Physics Data John Wu, Arie Shoshani, Alex Sim, Junmin Gu, Art Poskanzer Lawrence Berkeley National.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
PHENIX and the data grid >400 collaborators Active on 3 continents + Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
1 GCA Application in STAR GCA Collaboration Grand Challenge Architecture and its Interface to STAR Sasha Vaniachine presenting for the Grand Challenge.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
PPDG update l We want to join PPDG l They want PHENIX to join NSF also wants this l Issue is to identify our goals/projects Ingredients: What we need/want.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
February 28, 2003Eric Hjort PDSF Status and Overview Eric Hjort, LBNL STAR Collaboration Meeting February 28, 2003.
1D. Olson, SDM-ISIC Mtg, 26 Mar 2002 Scientific Data Management: An Incomplete Experimental HENP Perspective D. Olson, LBNL 26 March 2002 SDM-ISIC Meeting.
STAR Collaboration, July 2004 Grid Collector Wei-Ming Zhang Kent State University John Wu, Alex Sim, Junmin Gu and Arie Shoshani Lawrence Berkeley National.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
International Symposium on Grid Computing (ISGC-07), Taipei - March 26-29, 2007 Of 16 1 A Novel Grid Resource Broker Cum Meta Scheduler - Asvija B System.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
Jerome Lauret – ROOT Workshop 2005, CERN, CH ROOT4STAR: a ROOT based framework for user analysis and data mining J é r ô me Lauret -
Workflows and Data Management. Workflow and DM Run3 and after: conditions m LHCb major upgrade is for Run3 (2020 horizon)! o Luminosity x 5 ( )
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Dynamic staging to a CAF cluster Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop,
Production Mode Data-Replication Framework in STAR using the HRM Grid CHEP ’04 Congress Centre Interlaken, Switzerland 27 th September – 1 st October Eric.
CHEP ‘06Eric Hjort, Jérôme Lauret Data and Computation Grid Decoupling in STAR – An Analysis Scenario using SRM Technology Eric Hjort (LBNL) Jérôme Lauret,
STAR Scheduler Gabriele Carcassi STAR Collaboration.
Dynamic Data Placement: the ATLAS model Simone Campana (IT-SDC)
1 Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Efficient Data Access for Distributed Computing at RHIC A. Vaniachine Lawrence.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
ATLAS Distributed Computing Tutorial Tags: What, Why, When, Where and How? Mike Kenyon University of Glasgow.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
LHC experiments Requirements and Concepts ALICE
INFN-GRID Workshop Bari, October, 26, 2004
for the Offline and Computing groups
ALICE Physics Data Challenge 3
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Readiness of ATLAS Computing - A personal view
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Ákos Frohner EGEE'08 September 2008
LCG middleware and LHC experiments ARDA project
INSPIRE Geoportal Thematic Views Application
LHC Data Analysis using a worldwide computing grid
Nuclear Physics Data Management Needs Bruce G. Gibbard
An innovative campus grid prototype
Presentation transcript:

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May STAR Computing Doug Olson for: J é r ô me Lauret

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May STAR experiment... ● The Solenoidal Tracker At RHIC ● is an experiment located at the Brookhaven National Laboratory (BNL), USA ● A collaboration of 586 people wide, spanning over 12 countries for a total of 52 institutions ● A Pbytes scale experiment overall (raw+reconstructed) with several Million of files

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May The Physics … ● A multi-purpose detector system ● For Heavy Ion (Au+Au, Cu+Cu, d+Au, …) ● For Spin program p+p

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Zhangbu Xu, DNP2004

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Carl Gagliardi, Hard Probes 2004 ?

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Data Acquisition Prediction ● 150 MB/sec ● 60% Live, 3-4 months running => 1+ PB of data / run ● Possible rates x10 by ● x2 net output requested, the rest will be trigger ● Is needed to satisfy the Physics program … ● But pose some challenges ahead (RHIC-II era)

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Data Sets sizes - Year4 ● Raw Data Size ● <> ~ 2-3 MB/event - All on Mass Storage (HPSS as MSS) ● Needed only for calibration, production – Not centrally or otherwise stored ● Real Data size ● Data Summary Tape+QA histos+Tags+run information and summary: <> ~ 2-3 MB/event ● Micro-DST: KB/event ● Total Year4 DAQ DST/event MuDST Data production Analysis

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Data analysis ● Offline ● A single framework (root4star) for ● Simulation ● Data mining ● User analysis ● Real-Data Production ● Follows a Tier0 model ● Redistribution of MuDST to Tier1 sites ● Simulation production ● On the Grid …

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May How much data – What does this mean ??

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Data Sets sizes Tier0 Projections 2003/2004 data

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May ● An evolution and projections for the next 10 years (tier0) ● All hardware becomes obsolete ● Includes a 1/4 replacement every year CPU need projections

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May How long ? Year scale production cycles This is “new” since Year4 for RHIC experiments accustom to fast production turn around … NON-STOP data production and data acquisition

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Consequences on overall strategy

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Needs & Principles ● Cataloguing important ● Must be integrated with framework and tools ● Catalogue MUST be ● The central connection to datasets for users ● Moving model from PFN to LFN to DataSets, cultural issue at first ● STAR has a (federated) Catalog of its own brew… ● Production cycles are long ● Does not leave room for mistakes ● Planning, phase, convergence ● Data MUST be available ASAP to Tier1/Tier2 sites ● Access to data cannot be random but optimized at ALL levels ● Access to MSS is a nightmare when un-coordinated ● Is access to “named” PFN still an option ? ● Need for a data-access coordinator, SRM (??)

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Data distribution As immediately accessible as possible ● Tier0 production ● ALL EVENT files get copied on MSS (HPSS) at the end of a production job ● Strategy implies dataset IMMEDIATE replication ● As soon as a file is registered, it becomes available for “distribution” ● 2 Levels of data distributions – Local and Global ● Local ● All analysis files (MuDST) are on disks ● Ideally: One copy on centralized storage (NFS), one in MSS (HPSS) ● Practically: Storage do not allow to have all files “live” on NFS ● Notions of distributed disk – Cost effective solution ● Global ● Tier1 (LBNL) -- Tier2 sites (“private” resources for now) local/global relation through SE/MSS strategy needs to be consistent Grid STARTS from your backyard on …

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Distributed disks SE attached to specific CE at a site D D DD Client Script adds records Pftp on local disk FileCatalog Management Mark {un-}available Spider and update * Control Nodes VERY HOMEMADE VERY “STATIC” CE Data Mining Farm

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Distributed disks, possible model? D D DD Pftp on local disk Seeking to replace this with XROOTD/SRM XROOTD ● load balancing + scalability ● a way to avoid LFN/PFN translation (Xrootd dynamically discovers PFN based on LFN to PFN mapping)... Coordinated access to SE/MSS STILL needed - “A” coordinator would cement access consistency by providing policies, control, … Could it be DataMover/SRM ???

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Data transfer Off-site in STAR - SDM Data-Mover ● STAR started with ● A Tier-0 site - all “raw” files are transformed into pass1 (DST), pass2 (MuDST) files ● Tier-1 site - Receives all pass2 files, some “raw” and some pass1 files ● STAR is working on replicating this to other sites

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Data transfer flow Call RRS at each file transferred

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Experience with - SRM/HRM/RRS ● Extremely reliable ● Ronko's rotisserie feature “Set it, and forget it !” ● Several 10k files transferred, multiple TB for days, no losses ● Project was (IS) extremely useful, production usage in STAR ● Data availability at remote site as it is produced ● We need this NOW (resource constrained => distributed analysis and best use of both sites) ● Faster analysis yield to better science sooner ● Data safety ● Since RRS (prototype in use ~ 1 year) ● 250k files, 25 TB transferred AND Cataloged ● 100% reliability ● Project deliverables on-time

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Note on Grid ● For STAR, Grid computing is EVERY DAY Production used ● Data transfer using SRM, RRS,.. ● We run simulation production on the Grid (easy) ● Resource reserved for DATA production (still done traditionally) ● No real technical difficulties ● Mostly fears related to un-coordinated access and massive transfers ● Did not “dare” to touch user analysis ● Chaotic in nature, requires more solid SE, accounting, quota, privilege, etc …

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May More on Grid SUMS The STAR Unified Meta-Scheduler, A front end around evolving technologies for user analysis and data production GridCollector a framework addition for transparent access of event collection

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May SUMS (basics) ● STAR Unified Meta-Scheduler ● Gateway to user batch-mode analysis ● User writes an abstract job description ● Scheduler submits where files are, where CPU is,... ● Collects usage statistics ● User DO NOT need to know about the RMS layer ● Dispatcher and Policy engines ● DataSet driven - Full catalog implementation & Grid-aware ● Used to run simulation on grid (RRS on the way) ● Seamless transition of users to Grid when stability satisfactory ● Throttles IO resources, avoid contentions, optimizes on CPU ● Most advanced features include: self-adapt to site condition changes using ML modules Makes heavy use of ML

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May SUMS input From U-JDL to RDL ● SUMS: a way to unify diverse RMS ● An abstract way to describe jobs as input ● Datasets, file lists or event catalogues lead to job splitting ● A request is defined as a set or series of “operations” = A dataset could be subdivided in N operations Query/Wildcard /star/data09/reco/productionCentral/FullFie sched _1.list /.csh /star/data09/reco/productionCentral/FullFie sched _2.list /.csh rootMacros/numberOfEventsList.C\ <stdout /> <input etype=daq_reco_mudst" preferStorage="local" nFiles="all"/> toURL="file:/star/u/xxx/scheduler/out/" /> /star/data09/reco/productionCentral/FullFie sched _0.list /.csh /star/data09/reco/productionCentral/FullFie... /star/data09/reco/productionCentral/FullFie... / star/data09/reco/productionCentral/FullFie... /star/data09/reco/productionCentral/FullFie resolution root4star -q -b (\"$FILELIST\"\) URL="file:/star/u/xxx/scheduler/out/$JOBID.out" URL="catalog:star.bnl.gov?production=P02gd,fil <output fromScratch="*.root" Job description test.xml User Input … () … Policy …. dispatcher Extending proof of principle U-JDL to a feature reach Request Description Language (RDL) SBIR Phase I submitted to Phase II Supports workflow, multi-job, … Allows multiple datasets …

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May SUMS future ● Multiple scheduler ● Will replace with submission WS ● Could replace with other Meta-Scheduler (MOAB, …) Job control and GUI Mature enough (3 years) for spending time on GUI interface “appealing” application for any environment, easy(ier) to use

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May GridCollector “Using an Event Catalog to Speed up User Analysis in Distributed Environment” “tags” (bitmap index) based ● need to be define a-priori [production] ● Current version mix production tags AND FileCatalog information (derived from event tags) The compressed bitmap index is at least 10X faster than B-tree and 3X faster than the projection index

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May GridCollector Usage in STAR –Rest on now well tested and robust SRM (DRM+HRM) deployed in STAR anyhow Immediate Access and managed SE Files moved transparaentely by delegation to SRM service –Easier to maintain, prospects are enormous “Smart” IO-related improvements and home-made formats no faster than using GridCollector (a priori) –Physicists could get back to physics –And STAR technical personnel better off supporting GC It is a WORKING prototype of Grid interactive analysis framework

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Network needs in future ● Grid is a production reality ● To support it, the projections are as follow ● How does this picture looks like for user jobs support ?? ● Philosophy versus practical ● If network allows, send jobs to ANY CE and move data … ● Minor issue of finding the “closest” available data, advanced reservation, etc … ● If bandwidth do not allow, continue with placement ASAP …as we do now … and move jobs where files are (long lifetime data placement, re-use)

Jérôme Lauret – STAR – ICFA Workshop – Daegu/Korea May Moving from “dedicated” resources to “On Demand”  OpenScienceGrid ● Have been using grid tools in production at sites with STAR software pre-installed. ● Success rate was 100% when Grid infrastructure was “up” ● Only recommend to be careful with coordination local/global SE ● Moving forward … ● The two features to be achieved in the transition to OSG are ● Install necessary environment with jobs ● Enables Computing On Demand ● Integrate SRM/RRS into compute job workflow ● Makes cataloging generated data seamless with compute work (not yet achieved for all STAR compute modes)