SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Computing Model for SuperBelle Outline Scale and Motivation Definitions.

Slides:



Advertisements
Similar presentations
Jens G Jensen Atlas Petabyte store Supporting Multiple Interfaces to Mass Storage Providing Tape and Mass Storage to Diverse Scientific Communities.
Advertisements

Large Scale Computing Systems
Chapter 4 Infrastructure as a Service (IaaS)
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
6/4/20151 Introduction LHCb experiment. LHCb experiment. Common schema of the LHCb computing organisation. Common schema of the LHCb computing organisation.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
EventStore Managing Event Versioning and Data Partitioning using Legacy Data Formats Chris Jones Valentin Kuznetsov Dan Riley Greg Sharp CLEO Collaboration.
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 4.
DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Introduction. Readings r Van Steen and Tanenbaum: 5.1 r Coulouris: 10.3.
December 17th 2008RAL PPD Computing Christmas Lectures 11 ATLAS Distributed Computing Stephen Burke RAL.
HEP GRID CHEP, KNU 11/9/2002 Youngjoon Kwon (Yonsei Univ.) 1 Belle Computing / Data Handling  What is Belle and why we need large-scale computing?
W.Smith, U. Wisconsin, ZEUS Computing Board Zeus Executive, March 1, ZEUS Computing Board Report Zeus Executive Meeting Wesley H. Smith, University.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Graduate Student Department Of CSE 1.
Nick Brook Current status Future Collaboration Plans Future UK plans.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
Belle MC Production on Grid 2 nd Open Meeting of the SuperKEKB Collaboration Soft/Comp session 17 March, 2009 Hideyuki Nakazawa National Central University.
Belle II Data Management System Junghyun Kim, Sunil Ahn and Kihyeon Cho * (on behalf of the Belle II Computing Group) *Presenter High Energy Physics Team.
The Computing System for the Belle Experiment Ichiro Adachi KEK representing the Belle DST/MC production group CHEP03, La Jolla, California, USA March.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
GridPP18 Glasgow Mar 07 DØ – SAMGrid Where’ve we come from, and where are we going? Evolution of a ‘long’ established plan Gavin Davies Imperial College.
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
Event Data History David Adams BNL Atlas Software Week December 2001.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne A Computing Model for SuperBelle This is an idea for discussion only!
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
The Advanced Data Searching System The Advanced Data Searching System with 24 February APCTP 2010 J.H Kim & S. I Ahn & K. Cho on behalf of the Belle-II.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
KISTI & Belle experiment Eunil Won Korea University On behalf of the Belle Collaboration.
High Energy FermiLab Two physics detectors (5 stories tall each) to understand smallest scale of matter Each experiment has ~500 people doing.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Computing for LHC Physics 7th March 2014 International Women's Day - CERN- GOOGLE Networking Event Maria Alandes Pradillo CERN IT Department.
NA62 computing resources update 1 Paolo Valente – INFN Roma Liverpool, Aug. 2013NA62 collaboration meeting.
ATLAS Distributed Computing perspectives for Run-2 Simone Campana CERN-IT/SDC on behalf of ADC.
14/03/2007A.Minaenko1 ATLAS computing in Russia A.Minaenko Institute for High Energy Physics, Protvino JWGC meeting 14/03/07.
Ian Bird WLCG Networking workshop CERN, 10 th February February 2014
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
David Adams ATLAS ATLAS Distributed Analysis (ADA) David Adams BNL December 5, 2003 ATLAS software workshop CERN.
Main parameters of Russian Tier2 for ATLAS (RuTier-2 model) Russia-CERN JWGC meeting A.Minaenko IHEP (Protvino)
16 September 2014 Ian Bird; SPC1. General ALICE and LHCb detector upgrades during LS2  Plans for changing computing strategies more advanced CMS and.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
A Data Handling System for Modern and Future Fermilab Experiments Robert Illingworth Fermilab Scientific Computing Division.
Belle II Computing Fabrizio Bianchi INFN and University of Torino Meeting Belle2 Italia 17/12/2014.
Recent Evolution of the Offline Computing Model of the NOA Experiment Talk #200 Craig Group & Alec Habig CHEP 2015 Okinawa, Japan.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
ATLAS Computing: Experience from first data processing and analysis Workshop TYL’10.
DØ Grid Computing Gavin Davies, Frédéric Villeneuve-Séguier Imperial College London On behalf of the DØ Collaboration and the SAMGrid team The 2007 Europhysics.
M. Sevior, DPHEP July Belle and the Amazon EC2 cloud Martin Sevior, University of Melbourne University of Melbourne ( Tom Fifield University.
Website Deployment Week 12. Software Engineering Practices Consider the generic process framework – Communication – Planning – Modeling – Construction.
Status: ATLAS Grid Computing
Computing Operations Roadmap
Overview of the Belle II computing
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Readiness of ATLAS Computing - A personal view
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
ALICE Computing Upgrade Predrag Buncic
New strategies of the LHC experiments to meet
 YongPyong-High Jan We appreciate that you give an opportunity to have this talk. Our Belle II computing group would like to report on.
DØ MC and Data Processing on the Grid
Computing at the HL-LHC
Development of LHCb Computing Model F Harris
Presentation transcript:

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Computing Model for SuperBelle Outline Scale and Motivation Definitions of Computing Model Interplay between Analysis Model and Computing Model Options for the Computing Model Strategy to choose the Model

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne SuperBelle will operate at 8x10 35 cm 2 sec -1 Physics BB rate ~ 800 events/sec Physics Continuum rate ~ 800 events/sec Physics tau rate ~ 800 events/sec Calibration ~ 500 events/sec Total ~ 3 Khz Event size ~ 30 KB Approximately 2 PetaBytes of Physics Data/Year + MC Data x 3 = 8 PetaBytes of Physics Data/Year This is well into the range of an LHC experiment Cost of computing for the LHC ~ the same as the costs of the experiments ~ $500 Million

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Computing Models Current Belle Computing model is for almost all data to be located at KEK. Reconstruction and large scale analysis conducted at KEK. Creation of variety of skims of reduced size Final Analysis conducted in Root/PAW on users workstations Some MC Generated offsite LHC computing model has data and analysis conducted at a distributed collection of clusters worldwide – the LHC GRID Cloud Computing – use commercial facilities for data processing and MC generation

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Analysis Models Belle Analysis Model - BASF Panther Banks are C++ data objects Skims are text files pointing to events in a Belle database Output of skims are ntuples. The skim file system works extremely well provided the data remain at KEK

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Analysis Models – LHC (ATLAS)‏ The ATLAS analysis model does not require data to be stored centrally. Furthermore the ATLAS Athena analysis framework makes it possible to recover original data from derived data. Johannes Elmsheuser (LMU M ̈ unchen) (ATLAS Users Analysis)‏ The Full set of AOD and ESD exist at multiple sites over the GRID

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne The GRID now basically works Our Graduate students routinely use it to do Analysis.

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne ATLAS monitoring site as of 9/12/2008

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne The LHC GRID over the past year

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne “Cloud Computing” Cloud Computing is the latest Buzzword Cloud computing makes large scale computing resources available on a commercial basis Internet Companies have massive facilities and scale LHC produces 330 TB in a week. Google processes 1 PB of data every 72 minutes! A simple SOAP request creates a “virtual computer” instance with which one can compute as they wish

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Questions for the Computing/Analysis Model Can we afford to place all the computing resources we need for SuperBelle at KEK? If so should we use the current skim file system for SuperBelle? Should we employ GRID technology for SuperBelle? Should we employ "Cloud Computing" for SuperBelle? How do we formulate a plan to decide among these options? When do we need to decide?/Do we need to decide? How does the computing model ties in with the replacement for BASF?

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Can we afford to place almost all the computing resources we need for SuperBelle at KEK? The earliest turn-on time for SuperBelle is It could be that by 2013, placing all the computing resources we need for SuperBelle at KEK will be a feasible solution. Back of the envelope estimate follows scaling from “New” KEK Computing system

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne New KEK Computer System has 4000 CPU cores Storage ~ 2 PetaBytes Data size ~ 1 ab -1 Initial rate of 2x10 35 cm 2 sec -1 => 4 ab -1 /year Current KEKB Computer System Design rate of 8x10 35 cm 2 sec -1 => 16 ab -1 /year SuperBelle Requirements CPU Estimate 10 – 80 times current depending on reprocessing rate So 4x10 4 – 3.4x10 5 CPU cores Storage 10 PB in 2013, rising to 40 PB/year after 2016

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Spreadsheet CPU (8 -32)x10 4 cpus over 5 years ~ 500$ per core (2008) Storage costs over 5 years ( ) PB (Disk, no tape) $800/TB (2008) Electricity ~ 100 W/CPU (2008), Price $0.2/KWhr (2008)

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Price in 2008 of SuperBelle Cluster (At best 100% uncertainty!)‏ CPU (8 -32)x10 4 cpus over 5 years ~ 500$ per core => $40 Million/Year Storage costs over 5 years ( ) PB (Disk, no tape) $800/TB => $(8 - 32) Million/Year Electricity ~ 100 W/CPU (64 – 256) TWHr=> $( ) Million/year Moores Law – Double Performance every 18 months Rough Estimate over 5 years $(61, 82,102,123,83) Million/Year Rough Estimate over 5 years $(11,12,10,8,7) Million/Year This is a defensible solution but Total Cost over 5 years ~ $50 Million needs more study...

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Should we use the current skim file system for SuperBelle? Current skim file system works over a total database size of around 1 PB, at 50 ab -1 the dataset will rise to 140 PB. Can we maintain performance with this 2 orders of magnitude increase in size? Needs study... Skim file system does not allow data replication. Primitive metadata system associated with the data. Do we need a file catalogue or metadata catalogue of some kind? Derived data does not know it’s parent data in Panther. We need to either keep this data with the derived data or make guesses as to which data will be needed later. (cf. ECL timing information) Newer Analysis models allow this.

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Should we employ GRID technology for SuperBelle? There is a good chance we can keep the majority of CPU power at Belle However elements of GRID technology could still be useful. (eg SRB, SRM) At this point ordinary Graduate Students are using GRID tools to actually do distributed data analysis over a globally distributed data set. ATLAS employs the GAUDI analysis framework along with, LHCb, HARP (Hardron Production rates) and GLAST (Gamma Ray Astronomy). This provides a persistent and distributed data storage model which has demonstrated scalability to the level required by SuperBelle The LHC Computing grid exists. New CPU power and storage can be added by just expanding an existing cluster or by creating a new one. Belle GRID for MC production is just about ready to start.

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Distributed Data The GRID plus analysis model allows users to replicate data and use local resources. Allows relatively easy use of distributed MC production. GRID will be maintained and developed over the lifetime of SuperBelle (by other people!) The GAUDI analysis Model allows derived data to locate it’s parent data.

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Should we employ "Cloud Computing" for SuperBelle? Commercial internet companies like Google and Amazon have computing facilities orders of magnitude larger than HEP. They have established a Business based on CPU power on demand, one could imagine that they could provide the compute and storage we need at a lower cost than dedicated facilities. User Request CPU appears Data stored Returned Cloud Resources are deployed as needed. Pay as you go. Standards? Propriety lock in? Do we want our data stored on Commercial Company? What cost? What is the evolution?

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne How do we formulate a plan to decide among these options? Form a Computing/Network/Framework working group! First meeting Thursday at 9:00 AM in room KEK 3-go-kan, room# 425 (TV-conf : 30425)‏ When do we need to decide?/Do we need to decide? How does the computing model tie in with the replacement for BASF? Come and join us!

SuperBelle Collaboration Meeting December 2008 Martin Sevior University of Melbourne Agenda of the first meeting of Computing Group Date : Dec. 11th (Thrs) 9:00-11:00 Place: KEK 3-go-kan, room# 425 (TV-conf : 30425)‏ 9:00 - 9:10 Introduction (T.Hara : Osaka)‏ 9:10 - 9:30 Current GRID system in Belle (H.Nakazawa :NCU)‏ 9:30 - 9:50 General intro./Data Farm Activities in KISTI (S.B.Park : KISTI)‏ 9: :10 Computing ideas (M.Sevior : Melbourne)‏ 10: :30 Software Framework (R.Itoh : KEK)‏ 10: min. discussion (everybody)‏