Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep. 12-13, 2014 Draft.

Slides:



Advertisements
Similar presentations
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Advertisements

Status of BESIII Distributed Computing BESIII Workshop, Mar 2015 Xianghu Zhao On Behalf of the BESIII Distributed Computing Group.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
Bookkeeping data Monitoring info Get jobs Site A Site B Site C Site D Agent Production service Monitoring service Bookkeeping service Agent © Andrei Tsaregorodtsev.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep ,
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
BESIII distributed computing and VMDIRAC
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
YAN, Tian On behalf of distributed computing group Institute of High Energy Physics (IHEP), CAS, China CHEP-2015, Apr th, OIST, Okinawa.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
Belle MC Production on Grid 2 nd Open Meeting of the SuperKEKB Collaboration Soft/Comp session 17 March, 2009 Hideyuki Nakazawa National Central University.
CHEP Sep Andrey PHENIX Job Submission/Monitoring in transition to the Grid Infrastructure Andrey Y. Shevel, Barbara Jacak,
Status of StoRM+Lustre and Multi-VO Support YAN Tian Distributed Computing Group Meeting Oct. 14, 2014.
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
HEPD sem 14-Dec Andrey History photos: A. Shevel reports on CSD seminar about new Internet facilities at PNPI (Jan 1995)
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
9 th Weekly Operation Report on DIRAC Distributed Computing YAN Tian From to
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
The GridPP DIRAC project DIRAC for non-LHC communities.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
Status of BESIII Distributed Computing BESIII Workshop, Sep 2014 Xianghu Zhao On Behalf of the BESIII Distributed Computing Group.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI VO auger experience with large scale simulations on the grid Jiří Chudoba.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
Alien and GSI Marian Ivanov. Outlook GSI experience Alien experience Proposals for further improvement.
The GridPP DIRAC project DIRAC for non-LHC communities.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
StoRM + Lustre Proposal YAN Tian On behalf of Distributed Computing Group
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
Gang Chen, Institute of High Energy Physics Feb. 27, 2012, CHAIN workshop,Taipei Co-ordination & Harmonisation of Advanced e-Infrastructures Research Infrastructures.
DIRAC for Grid and Cloud Dr. Víctor Méndez Muñoz (for DIRAC Project) LHCb Tier 1 Liaison at PIC EGI User Community Board, October 31st, 2013.
Workload Management Status DIRAC User Meeting Marseille, Oct 2012.
The status of IHEP Beijing Site WLCG Asia-Pacific Workshop Yaodong CHENG IHEP, China 01 December 2006.
Progress of Work on SE and DMS YAN Tian April. 16, 2014.
Status of BESIII Distributed Computing BESIII Collaboration Meeting, Nov 2014 Xiaomei Zhang On Behalf of the BESIII Distributed Computing Group.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
Distributed computing and Cloud Shandong University (JiNan) BESIII CGEM Cloud computing Summer School July 18~ July 23, 2016 Xiaomei Zhang 1.
CEPC software & computing study group report
WLCG IPv6 deployment strategy
Status of BESIII Distributed Computing
BESIII data processing
Distributed Computing in IHEP
The Beijing Tier 2: status and plans
Xiaomei Zhang CMS IHEP Group Meeting December
Overview of the Belle II computing
Belle II Physics Analysis Center at TIFR
Report of Dubna discussion
Work report Xianghu Zhao Nov 11, 2014.
Introduction to CVMFS A way to distribute HEP software on cloud
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Status of Storm+Lustre and Multi-VO Support
Computing at CEPC Xiaomei Zhang Xianghu Zhao
Simulation use cases for T2 in ALICE
Discussions on group meeting
Status of MC production on the grid
Xiaomei Zhang On behalf of CEPC software & computing group Nov 6, 2017
R. Graciani for LHCb Mumbay, Feb 2006
 YongPyong-High Jan We appreciate that you give an opportunity to have this talk. Our Belle II computing group would like to report on.
The CMS Beijing Site: Status and Application
Production Manager Tools (New Architecture)
The LHCb Computing Data Challenge DC06
Presentation transcript:

Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep , 2014 Draft Version

Outline  Introduction  Experience of BES-DIRAC Distributed Computing Computing model Computing resources list Official MC production Data transfer system  Distributed Computing for CEPC – A test bed established – Sharing resources with BES – User job workflow – Physical validation – To do

INTRODUCTION Part I

About Distributed Computing Distributed computing plays an import role in discovery of Higgs – “Without the LHC Computing Grid, the discovery could not have occurred” ---- Foster Many HEP experiments have employed distributed computing to integrate resources contribued by collaboration members. – such as LHCb, Belle 2, CTA, ILC, BES 3, etc.. Large HEP experiments need plenty of computing resources, which may not be afforded by only one institution or university.

DIRAC: a Interware DIRAC (Distributed Infrastructure with Remote Agent Control) is a interware for grid computing. It’s powfull, flexible and widely used as central component of grid solution. More info: DIRAC Homepage: DIRAC Github:

DIRAC Users: LHCb, Belle, CTA, ILC, etc… ILC: ~ 3,000 CPU Cores LHCb: ~ 40,000 CPU Cores Belle2: ~ 12,000 CPU Cores CTA: ~ 5,000 CPU Cores

EXPERIENCE OF BES-DIRAC DISTRIBUTED COMPUTING Part II

BES-DIRAC: Computing Model Detector IHEP Data Center DIRAC Central SE Remote Site Raw dst & ramdomtrg Raw data Remote Site MC dst Remote Users IHEP Users All dst

BES-DIRAC: Computing Resources List #ContributorsCE TypeCPU CoresSE TypeSE CapacityStatus 1IHEPCluster + Cloud144dCache214 TBActive 2Univ. of CASCluster152Active 3USTCCluster200 ~ 1280dCache24 TBActive 4Peking Univ.Cluster100Active 5Wuhan Univ.Cluster100 ~ 300StoRM39 TBActive 6Univ. of MinnesotaCluster768BeStMan50 TBActive 7JINRgLite + Cloud100 ~ 200dCache8 TBActive 8INFN & Torino Univ.gLite + Cloud26420 TBActive 9CERNCloud20Active 10Soochow Univ.Cloud20Active Total1868 ~ TB 11Shandong Univ.Cluster100Preparing 12BUAACluster256Preparing 13SJTUCluster TBPreparing Total TB

BES-DIRAC: Official MC Production #TimeTaskBOSS Ver.Total EventsJobsData Output J/psi inclusive (round 05) M32, TB ~ Psi3770 (round 03,04)6.6.4.p M69, TB Total M102, TB job 2 nd batch of 2 nd productionPhysical Validation Check of 1 st production keep run ~1350 jobs for one week 2 nd batch: Dec.7~15

BES-DIRAC: Simulation+Reconstruction Simulation + reconstruction jobs are supported. Randomtrg data has been distributed to remoted sites with SE. Job download randomtrg data from local SE, or directly read from SE mounted to nodes.

BES-DIRAC: Data Trasfer System Data transfered from March to July 2014, total 85.9 TB DataSource SEDestination SEPeak SpeedAverage Speed randomtrg r04USTC, WHUUMN96 MB/S76.6 MB/s (6.6 TB/day) randomtrg r07IHEPUSTC, WHU191 MB/s115.9 MB/s (10.0 TB/day) Data TypeDataData SizeSource SEDestination SE DST xyz24.5 TBIHEPUSTC psippscan2.5 TBIHEPUMN Random trigger data round TBIHEPUSTC, WHU, UMN, JINR round TBIHEPUSTC, WHU, UMN round TBIHEPUSTC, WHU, UMN round TBIHEPUSTC, WHU, UMN round TBIHEPUSTC, WHU, UMN, JINR round TBIHEPUSTC, WHU high quality ( > 99% one-time success rate) high transfer speed ( ~ 1 Gbps to USTC, WHU, UMN; 300Mbps to JINR):

IHEP  USTC, 10.0 TB/day USTC, WHU  6.6 TB/day one-time success > 99%

DISTRIBUTED COMPUTING FOR CEPC part III

A Test Bed Established BES-DIRAC Servers Job flow *.stdhep input data *.slcio output data BUAA Site OS: SL 5.8 Remote WHU Site OS: SL 6.4 Remote IHEP PBS Site OS: SL 5.5 IHEP-OpenStack Site IHEP Lustre WHU SE IHEP Local Resources IHEP DB DB mirror CVMFS Server

Sharing Resources with BES Which resources can be shared? – Central DIRAC Servers & Mantainers. (hope CEPC coll. can contribute manpower) – Computing & Storage resources contributed by sites who wish to support both BES and CEPC, such as IHEP, WHU, BUAA, Soochow Univ., etc … Multi-VO (Virutal Organization) support technology is under development – It’s a grid framework for managing resources for multi collaborations. – VOMS server has been configured, tested, now is ready to use. – multi-VO workload management system is under testing. – StoRM SE with multi-VO support is under developing.

User Job Workflow Submit a User Job Step by Step: (1) upload input data to SE (2) prepare a JDL file: job.jdl (3) prepare job.sh (4) submit job to DIRAC (5) monitoring job status in web portal (6) Download output data to Lustre

Physical Validation Check Under going… will be finished before Sep.10

To Do List Add and test new sites; Deploy remote mirror MySQL database; Development a frontend module for massive job splitting, submission, monitoring & data management; Refine multi-VO suport to manage BES&CEPC sharing resources;

Thanks Thank you for your attention! Q & A Further questions and cooperations, please contact ZHANG Xiaomei and YAN Tian