Download presentation
Presentation is loading. Please wait.
Published byEthan Woolridge Modified over 9 years ago
1
RHIC, STAR computing towards distributed computing on the Open Science Grid Jérôme LAURET RHIC/STAR
2
2 Jérôme LAURET IWLSC, Kolkata India 2006 Outline The RHIC program, complex and experiments An overview of the RHIC Computing facility Expansion model Local Resources, remote usage Disk storage, a “distributed” paradigm Phenix and STAR STAR Grid program & tools SRM / DataMover SUMS GridCollector Brief overview of the Open Science Grid STAR on OSG
3
3 Jérôme LAURET IWLSC, Kolkata India 2006 The RHIC program, complex and experiments An overview of the RHIC Computing facility Expansion model Local Resources, remote usage Disk storage, a “distributed” paradigm Phenix and STAR STAR Grid program & tools SRM / DataMover SUMS GridCollector Brief overview of the Open Science Grid STAR on OSG
4
4 Jérôme LAURET IWLSC, Kolkata India 2006 The Relativistic Heavy Ion Collider (RHIC) complex & experiments A world-leading scientific program in Heavy-Ion and Spin program The Largest running NP experiment Located in Long Island, New York, USA Flexibility is key to understanding complicated systems Polarize protons sqrt(s) = 50-500 GeV Nuclei from d to Au, sqrt(s NN ) = 20-200 GeV Physics runs to date Au+Au @ 20, 62, 130, 200 GeV Polarized p+p @ 62, 200 GeV D+Au @ 200 GeV RHIC It is becoming the world leader in the scientific quest toward understanding how mass and spin combine into a coherent picture of the fundamental building blocks nature uses for atomic nuclei. It is also providing unique insight into how quark and gluons behaved collectively at the very first moment our universe was born.
5
5 Jérôme LAURET IWLSC, Kolkata India 2006 The experiments PHENIX BRAHMS &PP2PPPHOBOS STAR 1.2 km RHIC
6
6 Jérôme LAURET IWLSC, Kolkata India 2006 The RHIC program, complex and experiments An overview of the RHIC Computing facility Expansion model Local Resources, remote usage Disk storage, a “distributed” paradigm Phenix and STAR STAR Grid program & tools SRM / DataMover SUMS GridCollector Brief overview of the Open Science Grid STAR on OSG
7
7 Jérôme LAURET IWLSC, Kolkata India 2006 The RHIC Computing Facility (RCF) RHIC Computing Facility (RCF) at BNL Tier0 for the RHIC program Online recording of Raw data Production reconstruction of all (most) Raw data Facility for data selection (mining) and analysis Long term archiving and serving of all data … but not sized for Monte Carlo generation Equipment refresh funding (~25% annual replacement) Addressing obsolescence Results in important collateral capacity growth
8
8 Jérôme LAURET IWLSC, Kolkata India 2006 Tier1, Tier2, … remote facilities Remote Facilities Primary source of Monte Carlo data Significant analysis activity (equal in the case of STAR) Such sites are operational – the top 3 STAR NERSC/PDSF, LBNL Wayne State University Sao Paolo PHENIX RIKEN, Japan Center for High Performance Computing, University of New Mexico VAMPIRE cluster, Vanderbilt University Grid Computing Promising new direction in remote (distributed) computing STAR and, to a lesser extent, PHENIX are now active in Grid computing
9
9 Jérôme LAURET IWLSC, Kolkata India 2006 Key sub-systems Mass Storage System Hierarchical Storage Management by HPSS 4 StorageTek robotic tape silos ~4.5 PBytes 40 StorageTek 9940b tape drives ~1.2 GB/sec Change to technology to LTO drive this year CPU Intel/Linux dual racked processor systems ~ 2300 CPU’s for ~1800 kSPECint2000 Mixed of Condor & LSF based LRMS Central Disk 170 TBytes of RAID 5 storage Other storage solution: PANASAS, … 32 Sun/Solaris SMP NFS servers ~1.3 GByte/sec Distribute disk ~ 400 TBytes x2.3 more than centralized storage !!!
10
10 Jérôme LAURET IWLSC, Kolkata India 2006 How does it look like … Not like these … although …
11
11 Jérôme LAURET IWLSC, Kolkata India 2006 MSS, CPUs, Central Store … but like these or similar (the chairs do not seem more comfortable)
12
12 Jérôme LAURET IWLSC, Kolkata India 2006 Data recording rates Run4 set a first record STAR PHENIX 120MBytes/sec
13
13 Jérôme LAURET IWLSC, Kolkata India 2006 DAQ rates comparative A very good talk from Martin Putchke - CHEP04 Concepts and technologies used in contemporary DAQ systems ~ 25 ~ 100 ~ 40 150 MB/sec ~ 300 ~1250 Heavy Ion Experiments are in the > 100 MB/sec range All in MB/sec, approximate... STAR moving to x 10 capabilities in outer years (2008+)
14
14 Jérôme LAURET IWLSC, Kolkata India 2006 Mid to long term computing needs Computing projection model Goal is to estimate CPU, disk, mass storage and network capacities Model based on raw data scaling Moore’s law used for cost recession Feedback from the experimental groups Annual meetings, model refined if necessary (has been stable for a while) Estimate based on beam use plans May be offset by experiment, by year Integral consistent Maturity factor for codes Number of reconstruction passes “richness” factor for the data (density of interesting events)
15
15 Jérôme LAURET IWLSC, Kolkata India 2006 Projected needs STAR Phenix
16
16 Jérôme LAURET IWLSC, Kolkata India 2006 Discussion of model Data amount is accurate at 20% close i.e. model was adjusted to 20% lower-end Upper-end has larger impact in the outer years DAQ1000 for STAR enabling Billions of event capabilities a major (cost) factor driven by Physics demand Cost will be beyond current provision Tough years start as soon as 2008 Gets better in the outer years (Moore’s law catches up) Uncertainties grows with time however … Cost versus Moore’s law Implies “aggressive” technology upgrades (HPSS for example) Strategy heavily based on low cost distributed disk (cheap, CE attached)
17
17 Jérôme LAURET IWLSC, Kolkata India 2006 The RHIC program, complex and experiments An overview of the RHIC Computing facility Expansion model Local Resources, remote usage Disk storage, a “distributed” paradigm Phenix and STAR STAR Grid program & tools SRM / DataMover SUMS GridCollector Brief overview of the Open Science Grid STAR on OSG
18
18 Jérôme LAURET IWLSC, Kolkata India 2006 Disk storage – distributed paradigm The ratio is striking x2.3 ratio now, moves to x6 in outer years Requires SE strategy CPU shortfall Tier1 use (Phenix, STAR) Tier2 user analysis and data on demand (STAR)
19
19 Jérôme LAURET IWLSC, Kolkata India 2006 Phenix – dCache model Tier1 / CC-J / RIKEN MSS = HPSS Central stores Tier 0 – Tier 1 model Provides scalability for centralized storage Smooth(er) distributed disk model
20
20 Jérôme LAURET IWLSC, Kolkata India 2006 Phenix – Data transfer to RIKEN Network transfer rates of 700-750 Mbits/s could be achieved (i.e. ~90 MB/sec)
21
21 Jérôme LAURET IWLSC, Kolkata India 2006 STAR – SRM, GridCollector, Xrootd Different approach Large (early) pool of distributed disks, early adoption of dd model dd model too home-grown Did not scale well when mixing dd and central disks Tier 0 – Tier X (X=1 or 2) model Need something easy to deploy, easy to maintain Leveraging on SRM experience Data on demand Embryonic event level (GridCollector) Xrootd could benefit from an SRM back-end olbdxrootdolbdxrootdolbd xrootd olbd xrootd olbd xrootd olbd xrootd Manager (Head Node) Supervisor (Intermediate Node) Data Server (Leaf Node)
22
22 Jérôme LAURET IWLSC, Kolkata India 2006 STAR dd Evolution – From this Where does this data go ?? D D D D DataCarousel Client Script adds records Pftp on local disk FileCatalog Management Update FileLocations Mark {un-}available Spider and update * Control Nodes VERY HOMEMADE VERY “STATIC”
23
23 Jérôme LAURET IWLSC, Kolkata India 2006 STAR dd Evolution – … to that … Entire layer for Cataloguing is gone Layer for restore from MSS to dd gone D D DD Pftp on local disk DATA ON DEMAND XROOTD provides load balancing, possibly scalability, a way to avoid LFN/PFN translation... But does NOT fit within our SRM invested directions... BUT IS IT REALLY SUFFICIENT !!??
24
24 Jérôme LAURET IWLSC, Kolkata India 2006 Coordination of requests needed Un-coordinated requests to MSS is a disaster This applies to ANY SE- related tools Gets worst if the environment combine technologies (shared infrastructure) Effect of performance is drastic
25
25 Jérôme LAURET IWLSC, Kolkata India 2006 The RHIC program, complex and experiments An overview of the RHIC Computing facility Expansion model Local Resources, remote usage Disk storage, a “distributed” paradigm Phenix and STAR STAR Grid program & tools SRM / DataMover SUMS GridCollector Brief overview of the Open Science Grid STAR on OSG
26
26 Jérôme LAURET IWLSC, Kolkata India 2006 STAR Grid program - Motivation Tier0 production ALL EVENT files get copied on HPSS at the end of a production job Data reduction DAQ to Event to Micro-DST All MuDST are on “disks” One copy temporarily on centralized storage (NFS), one permanently in HPSS Script checks consistency (job status, presence of files in one and the other) If “sanity” checks (integrity / checksum), register files in Catalog Re-distribution If registered, MuDST may be “distributed” Distributed disk on Tier0 sites Tier1 (LBNL) -- Tier2 sites (“private” resources for now) Use of SRM since 2003... Strategy implies dataset IMMEDIATE replication Allows balancing of analysis Tie0 to Tier1 Data on demand enable Tier2 with capabilities
27
27 Jérôme LAURET IWLSC, Kolkata India 2006 Needed for immediate exploitation of resources Short / medium term strategy To distribute data Take advantage of the static data (schedulers, workflow, …) Advanced strategies Data-on-demand (planner, dataset balancing, data placement …) Selection of sub-sets of data (datasets of datasets, …) Consistent strategy (interoperability?, publishing?, ) Less naïve considerations Job Tracking Packaging Automatic Error recovery, Help desk Networking Advanced workflow, … SRM / DataMover STAR Unified Meta-Scheduler Xrootd, … GridCollector SRM back-ends Would enable Xrootd with Object on demand Will leverage existing to come to existence middleware or address one by one …
28
28 Jérôme LAURET IWLSC, Kolkata India 2006 SRM / DataMover SRMs are middleware components whose function is to provide dynamic space allocation file management of shared storage components on the Grid SRM Enstore JASMine Client USER/APPLICATIONS Grid Middleware SRM dCache SRM Castor SRM Unix-based disks SRM SE CCLRC RAL http://osg-docdb.opensciencegrid.org/0002/000299/001/GSM-WG-GGF15-SRM.pphttp://osg-docdb.opensciencegrid.org/0002/000299/001/GSM-WG-GGF15-SRM.pp t
29
29 Jérôme LAURET IWLSC, Kolkata India 2006 SRM / DataMover Layer on top of SRM In use for BNL/LBNL data transfer for years All MuDST moved to Tier1 this way Extremely reliable “Set it, and forget it !” Several 10k files transferred, multiple TB for days, no losses Project was (IS) extremely useful, production usage in STAR Data availability at remote site as it is produced We need this NOW Faster analysis is better science and sooner Data safety Caveat/addition in STAR: RRS (Replica Registration Service) 250k files, 25 TB transferred AND Catalogued 100% reliability Project deliverables on-time
30
30 Jérôme LAURET IWLSC, Kolkata India 2006 SRM / DataMover – Flow diagram Being deployed at Wayne State University Sao Paolo DRM used in data analysis scenario as light weight SE service (deployable on the fly) All the benefits from SRM (advanced reservation, …) If we know there IS a storage space, we can take it No heavy duty SE deployment NEW
31
31 Jérôme LAURET IWLSC, Kolkata India 2006 CE/SE decoupling Srm-copy from execution site DRM back to submission site Submission site DRM is called from execution site WN Requires outgoing, but not incoming, connection on the WN Srm-copy callback disabled (asynchronous transfer) Batch slot released immediately after srm-copy call Final destination of files is HPSS or disk, owned by user DRM Client /scratch...... DRM cache Submission Site Job execution Site
32
32 Jérôme LAURET IWLSC, Kolkata India 2006 SUMS – The STAR Unified Meta- Scheduler STAR Unified Meta-Scheduler Gateway to user batch-mode analysis User writes an abstract job description Scheduler submits where files are, where CPU is,... Collects usage statistics User DO NOT need to know about the RMS layer Dispatcher and Policy engines DataSet drive - Full catalog implementation & Grid-aware Throttles IO resources, avoid contentions, optimizes on CPU Avoid specifying data location …
33
33 Jérôme LAURET IWLSC, Kolkata India 2006 SUMS – The STAR Unified Meta- Scheduler STAR Unified Meta-Scheduler Gateway to user batch-mode analysis User writes an abstract job description Scheduler submits where files are, where CPU is,... Collects usage statistics User DO NOT need to know about the RMS layer Dispatcher and Policy engines DataSet drive - Full catalog implementation & Grid-aware Throttles IO resources, avoid contentions, optimizes on CPU BEFORE – VERY choppy As NFS would impact computational performances AFTER modulo remaining farm instability, smoother
34
34 Jérôme LAURET IWLSC, Kolkata India 2006 SUMS – The STAR Unified Meta- Scheduler, the next generation … NEW FEATURES RDL in addition of U-JDL Testing grid submission is OVER. SUMS is production and user analysis ready Light SRM helping tremendously Need scalability test Made aware of multiple packaging methods (from ZIP archive to PACMAN) Tested for simple analysis, need finalizing mixed archiving technology (detail) Versatile configuration Site can “plug-and-play” Possibility of Multi-VO support within ONE install An issue since we have multi 10k jobs/day NOW with spikes at 100k (valid) jobs from nervous users …
35
35 Jérôme LAURET IWLSC, Kolkata India 2006 GridCollector Using an Event Catalog to Speed up User Analysis in Distributed Environment STAR – event catalog … Based on TAGS produced at reco time Rest on now well tested and robust SRM (DRM+HRM) deployed in STAR anyhow Immediate Access and managed SE Files moved transparently by delegation to SRM service BEHING THE SCENE Easier to maintain, prospects are enormous “Smart” IO-related improvements and home-made formats no faster than using GridCollector (a priori) Physicists could get back to physics And STAR technical personnel better off supporting GC It is a WORKING prototype of Grid interactive analysis framework VERY POWERFULL Event “server” based (no longer files) GAIN ALWAYS > 1, regardless of selectivity root4star -b -q doEvents.C'(25,"select MuDst where Production=P04ie \ and trgSetupName=production62GeV and magScale=ReversedFullField \ and chargedMultiplicity>3300 and NV0>200", "gc,dbon")'
36
36 Jérôme LAURET IWLSC, Kolkata India 2006 GridCollector – The next step Can push functionalities “down” Index BitMap technology in ROOT framework Make “a” coordinator “aware” of events (i.e. objects) Xrootd a good candidate ROOT framework preferred Both would serve as a demonstrator (immediate benefit to a few experiments …) Object-On Demand: from files to Object Management - Science Application Partnership (SAP) – SciDAC-II In the OSG program of work as leveraging technologies to achieve goals
37
37 Jérôme LAURET IWLSC, Kolkata India 2006 The RHIC program, complex and experiments An overview of the RHIC Computing facility Expansion model Local Resources, remote usage Disk storage, a “distributed” paradigm Phenix and STAR STAR Grid program & tools SRM / DataMover SUMS GridCollector Brief overview of the Open Science Grid STAR on OSG
38
38 Jérôme LAURET IWLSC, Kolkata India 2006 The Open Science Grid In the US, Grid is moving to the Open Science-Grid An interesting adventure comparable similar European efforts EGEE interoperability at its heart Character of OSG Distributed ownership of resources. Local Facility policies, priorities, and capabilities need to be supported. Mix of agreed upon performance expectations and opportunistic resource use. Infrastructure deployment based on the Virtual Data Toolkit. Will incrementally scale the infrastructure with milestones to support stable running of mix of increasingly complex jobs and data management. Peer collaboration of computer and application scientists, facility, technology and resource providers “end to end approach”. Support for many VOs from the large (thousands) to the very small and dynamic (to the single researcher & high school class) Loosely coupled consistent infrastructure - “Grid of Grids”.
39
39 Jérôme LAURET IWLSC, Kolkata India 2006 STAR and the OSG STAR could not run on Grid3 Was running at PDSF, a Grid3 site setup in collaboration with our resources STAR on OSG = a big improvement OSG for an Open Science, not as strongly LHC sole focus Expanding to other science: revisit of needs and requirements More resources Greater stability Currentely Run MC on regular basis (nightly tests, standard MC) Recently focused on user analysis (light weight SRM) Helped other site deploy OSG stack And it shows … FIRST Functional site in Brazil, Universidade de Sao Paolo, a STAR institution … http://www.interactions.org/sgtw/2005/0727/star_saopaulo_more.html
40
40 Jérôme LAURET IWLSC, Kolkata India 2006 Summary RHIC computing facility provides adequate resources in the short-term Model imperfect for long term projections Problematic years starting 2008, driven by high data throughput & physics demands Mid-term issue This will impact Tier1 as well, assuming a refresh and planning along the same model Out-sourcing? Under data “stress” and increasing complexity RHIC experiments have integrated at one level or another distributed computing principles Data distribution and management Job scheduling, selectivity, … STAR intends to Take full advantage of OSG & help bring more institutions into the OSG Address the issue of batch oriented user analysis (opportunistic, …)
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.