Proposal for a DØ Remote Analysis Model (DØRAM)

Slides:



Advertisements
Similar presentations
4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
Advertisements

Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Belle computing upgrade Ichiro Adachi 22 April 2005 Super B workshop in Hawaii.
Reliability Week 11 - Lecture 2. What do we mean by reliability? Correctness – system/application does what it has to do correctly. Availability – Be.
The Google File System. Why? Google has lots of data –Cannot fit in traditional file system –Spans hundreds (thousands) of servers connected to (tens.
Rates and Billing for New ITS Services Financial Unit Liaison Meeting February 16, 2011 Barry D. MacDougall Information Technology Service.
Gordon: Using Flash Memory to Build Fast, Power-efficient Clusters for Data-intensive Applications A. Caulfield, L. Grupp, S. Swanson, UCSD, ASPLOS’09.
Report : Zhen Ming Wu 2008 IEEE 9th Grid Computing Conference.
Data Center Infrastructure
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
EarthLink Cloud Server Backup. Typical Business Challenges Does my tape back system provide the instantaneous access and rapid recovery that I need? How.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
+ discussion in Software WG: Monte Carlo production on the Grid + discussion in TDAQ WG: Dedicated server for online services + experts meeting (Thusday.
VO Sandpit, November 2009 e-Infrastructure to enable EO and Climate Science Dr Victoria Bennett Centre for Environmental Data Archival (CEDA)
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Hosted by Designing a Backup Architecture That Actually Works W. Curtis Preston President/CEO The Storage Group.
Snapshot of the D0 Computing and Operations Planning Process Amber Boehnlein For the D0 Computing Planning Board.
DØSAR, State of the Organization Jae Yu DOSAR, Its State of Organization 7th DØSAR (3 rd DOSAR) Workshop University of Oklahoma Sept. 21 – 22, 2006 Jae.
Status of DØ Computing at UTA Introduction The UTA – DØ Grid team DØ Monte Carlo Production The DØ Grid Computing –DØRAC –DØSAR –DØGrid Software Development.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
A Design for KCAF for CDF Experiment Kihyeon Cho (CHEP, Kyungpook National University) and Jysoo Lee (KISTI, Supercomputing Center) The International Workshop.
Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) IntroductionIntroduction Remote Analysis Station ArchitectureRemote.
DØ RAC Working Group Report Progress Definition of an RAC Services provided by an RAC Requirements of RAC Pilot RAC program Open Issues DØRACE Meeting.
DØ RACE Introduction Current Status DØRAM Architecture Regional Analysis Centers Conclusions DØ Internal Computing Review May 9 – 10, 2002 Jae Yu.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
Status of UTA IAC + RAC Jae Yu 3 rd DØSAR Workshop Apr. 7 – 9, 2004 Louisiana Tech. University.
Spending Plans and Schedule Jae Yu July 26, 2002.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
DØSAR a Regional Grid within DØ Jae Yu Univ. of Texas, Arlington THEGrid Workshop July 8 – 9, 2004 Univ. of Texas at Arlington.
From DØ To ATLAS Jae Yu ATLAS Grid Test-Bed Workshop Apr. 4-6, 2002, UTA Introduction DØ-Grid & DØRACE DØ Progress UTA DØGrid Activities Conclusions.
CDMS Computing Project Don Holmgren Other FNAL project members (all PPD): Project Manager: Dan Bauer Electronics: Mike Crisler Analysis: Erik Ramberg Engineering:
DØRACE Workshop Jae Yu Feb.11, 2002 Fermilab Introduction DØRACE Progress Workshop Goals Arrangements.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Feb. 14, 2002DØRAM Proposal DØ IB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) Introduction Partial Workshop Results DØRAM Architecture.
Feb. 13, 2002DØRAM Proposal DØCPB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) IntroductionIntroduction Partial Workshop ResultsPartial.
A UK Computing Facility John Gordon RAL October ‘99HEPiX Fall ‘99 Data Size Event Rate 10 9 events/year Storage Requirements (real & simulated data)
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
July 10, 2002DØRACE Status Report, Jae Yu DØ IB Meeting, OU Workshop 1 DØ RACE Status Report Introduction Software Distribution (DØRACE Setup) Hardware.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Hall D Computing Facilities Ian Bird 16 March 2001.
Video Security Design Workshop:
Experience of Lustre at QMUL
Data Center Infrastructure
VirtualGL.
Experience of Lustre at a Tier-2 site
5th DOSAR Workshop Louisiana Tech University Sept. 27 – 28, 2007
UK GridPP Tier-1/A Centre at CLRC
Welcome! Thank you for joining us. We’ll get started in a few minutes.
Grid Computing.
NGS Oracle Service.
TYPES OF SERVER. TYPES OF SERVER What is a server.
Grid Canada Testbed using HEP applications
Nuclear Physics Data Management Needs Bruce G. Gibbard
DØ Internal Computing Review
DØ RAC Working Group Report
Cluster Computers.
Presentation transcript:

Proposal for a DØ Remote Analysis Model (DØRAM) DØRACE Meeting Jan. 17, 2002 Jae Yu Introduction  Remote Analysis Station Architecture Requirement for Regional Analysis Centers Suggested Storage Equipment Design What Do I Think We Should Do? Conclusions Jan. 17, 2002 DØRAM Proposal DØRACE Meeting, Jae Yu

Why do we need a DØRAM? Total Run IIa data sizes are 350TB for RAW 200-400 TB for Reco + root 1.4x109 Events total At the fully optimized 10sec/event reco.1.4x1010 Seconds for one time reprocessing Takes one full year w/ 500 machines Takes ~8mos to transfer raw data for dedicated gigabit network Centralized system will do a lot of good but not sufficient (DØ analysis model proposal should be complemented with DØRAM) Need to allow remote locations to work on analysis efficiently Sociological benefits within the institutes Regional Analysis Centers should be established Jan. 17, 2002 DØRAM Proposal DØRACE Meeting, Jae Yu

…. Proposed DØRAM Architecture Central Analysis Center (CAC) Regional RAC …. Regional Analysis Centers IAC Institutional Desktop Stations DAS Store & Process 10~20% of All Data Normal Interaction Communication Path Occasional Interaction Jan. 17, 2002 DØRAM Proposal DØRACE Meeting, Jae Yu

Regional Analysis Centers A few geographically selected sites that satisfy requirements Provide almost the same level of service as FNAL to a few institutional analysis centers Analyses carried out within the regional center Store 10~20% of statistically random data permanently Most the analyses performed on these samples with the regional network Refine the analyses using the smaller but unbiased data set When the entire data set is needed  Underlying Grid architecture provide access to remaining data set Jan. 17, 2002 DØRAM Proposal DØRACE Meeting, Jae Yu

Regional Analysis Center Requirements Become a Mini-CAC Sufficient computing infrastructure Large bandwidth (gagibit or better) Sufficient Storage Space to hold 10~20% of data permanently and expandable to accommodate data increase >30TB just for Run IIa RAW data Sufficient CPU resources to provide regional or Institutional analysis requests and reprocessing Geographically located to avoid unnecessary network traffic overlap Software Distribution and Support Mirror copy of CVS database for synchronized update between RAC’s and CAC Keep the relevant copies of data bases Act as SAM service station Jan. 17, 2002 DØRAM Proposal DØRACE Meeting, Jae Yu

Regional Storage Cache IDE Hard drives are $1.5~$2./Gb Each IDE RAID array gives up to ~1TByte – hot swappable Can be configured to have up to 10TB in a rack Modest server can manage the entire system Gbit network switch provide high throughput transfer to outside world Flexible and scalable system Need an efficient monitoring and error recovery system Communication to resource management Gbit Switch IDE-RAID IDE-RAID . IDE-RAID IDE-RAID Disk Server Jan. 17, 2002 DØRAM Proposal DØRACE Meeting, Jae Yu

What Do I Think We Should Do? Most the students and postDocs are at FNAL, thus it is important to provide them sufficient computing and cache resources for their analysis.  The Current suggestion for backend analysis clusters should be built!! In the mean time, we should select a few sites as RACs and prepare sufficient hardware and infrastructure My rough map scan gives FNAL+3RACs in the US, and a few in Europe Software effort for Grid should proceed as fast as we can to supplement the hardware We cannot afford to spend time for Test beds Our set ups should be the Test Bed and the actual Grid A committee to determine number of RAC sites, their requirements, and select RACs within the next couple of months. Jan. 17, 2002 DØRAM Proposal DØRACE Meeting, Jae Yu

Conclusions DØ must prepare for large data set era Need to expedite analyses in timely fashion Need to distribute data set throughout the collaboration Establishing regional analysis centers will be the first step toward DØ Grid Will write up a proposal for your perusal Will ask a few of you to serve as RAC committee members Jan. 17, 2002 DØRAM Proposal DØRACE Meeting, Jae Yu