Xiaomei Zhang CMS IHEP Group Meeting December

Slides:



Advertisements
Similar presentations
Status of BESIII Distributed Computing BESIII Workshop, Mar 2015 Xianghu Zhao On Behalf of the BESIII Distributed Computing Group.
Advertisements

1 User Analysis Workgroup Update  All four experiments gave input by mid December  ALICE by document and links  Very independent.
Enabling Grids for E-sciencE Medical image processing web portal : Requirements analysis. An almost end user point of view … H. Benoit-Cattin,
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES News on monitoring for CMS distributed computing operations Andrea.
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep ,
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
Claudio Grandi INFN Bologna CMS Operations Update Ian Fisk, Claudio Grandi 1.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
YAN, Tian On behalf of distributed computing group Institute of High Energy Physics (IHEP), CAS, China CHEP-2015, Apr th, OIST, Okinawa.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Monitoring in EGEE EGEE/SEEGRID Summer School 2006, Budapest Judit Novak, CERN Piotr Nyczyk, CERN Valentin Vidic, CERN/RBI.
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
UMD TIER-3 EXPERIENCES Malina Kirn October 23, 2008 UMD T3 experiences 1.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep , 2014 Draft.
CERN Using the SAM framework for the CMS specific tests Andrea Sciabà System Analysis WG Meeting 15 November, 2007.
Outline: Tasks and Goals The analysis (physics) Resources Needed (Tier1) A. Sidoti INFN Pisa.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
1 Andrea Sciabà CERN The commissioning of CMS computing centres in the WLCG Grid ACAT November 2008 Erice, Italy Andrea Sciabà S. Belforte, A.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
Stephen Burke – Sysman meeting - 22/4/2002 Partner Logo The Testbed – A User View Stephen Burke, PPARC/RAL.
Alien and GSI Marian Ivanov. Outlook GSI experience Alien experience Proposals for further improvement.
The GridPP DIRAC project DIRAC for non-LHC communities.
Monitoring the Readiness and Utilization of the Distributed CMS Computing Facilities XVIII International Conference on Computing in High Energy and Nuclear.
StoRM + Lustre Proposal YAN Tian On behalf of Distributed Computing Group
Gestion des jobs grille CMS and Alice Artem Trunov CMS and Alice support.
ATLAS Computing Wenjing Wu outline Local accounts Tier3 resources Tier2 resources.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
Vendredi 27 avril 2007 Management of ATLAS CC-IN2P3 Specificities, issues and advice.
HEPiX IPv6 Working Group David Kelsey (STFC-RAL) GridPP33 Ambleside 22 Aug 2014.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
KEK Computing Resources after Earthquake Akiya Miyamoto 30-March-2011 ILD Software WG meeting Status as of today.
The ALICE Analysis -- News from the battlefield Federico Carminati for the ALICE Computing Project CHEP 2010 – Taiwan.
Claudio Grandi INFN Bologna Workshop congiunto CCR e INFNGrid 13 maggio 2009 Le strategie per l’analisi nell’esperimento CMS Claudio Grandi (INFN Bologna)
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
Daniele Bonacorsi Andrea Sciabà
Grid-Ireland Gateway Architecture
Kevin Thaddeus Flood University of Wisconsin
Status of BESIII Distributed Computing
Real Time Fake Analysis at PIC
The Beijing Tier 2: status and plans
LCG Service Challenge: Planning and Milestones
Farida Naz Andrea Sciabà
Bulk production of Monte Carlo
Data Challenge with the Grid in ATLAS
The Status of Beijing site, and CMS local DBS
INFN-GRID Workshop Bari, October, 26, 2004
GSIAF & Anar Manafov, Victor Penso, Carsten Preuss, and Kilian Schwarz, GSI Darmstadt, ALICE Offline week, v. 0.8.
David Cameron ATLAS Site Jamboree, 20 Jan 2017
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Readiness of ATLAS Computing - A personal view
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
1 VO User Team Alarm Total ALICE ATLAS CMS
Analysis Operations Monitoring Requirements Stefano Belforte
Simulation use cases for T2 in ALICE
Survey on User’s Computing Experience
3.2 Virtualisation.
CernVM Status Report Predrag Buncic (CERN/PH-SFT).
Discussions on group meeting
Bernd Panzer-Steindel CERN/IT
WLCG Collaboration Workshop;
TYPES OFF OPERATING SYSTEM
Grid Canada Testbed using HEP applications
N. De Filippis - LLR-Ecole Polytechnique
UM D0RACE STATION Status Report Chunhui Han June 20, 2002
The CMS Beijing Site: Status and Application
The LHCb Computing Data Challenge DC06
Presentation transcript:

Xiaomei Zhang CMS IHEP Group Meeting December 14 2008 CMS Computing Xiaomei Zhang CMS IHEP Group Meeting   December 14 2008

Outline T3 resource and discussion T2 resource and discussion CRAB and Local Batch Submission functions CRAB Server Site monitoring

T3 resource and discussion CPU resource (96 cores) Old: 6 machines, 48 cores; New: 6 machines, 48 cores Storage resource cmsd01 and cmsd02 (old:/cms/cms1, /cms/cms2, /cms/cmsmc1,/cms/cmsmc2) /home/cms/xxx(old:/home/lhc/cms/xxx) Backup: /home/cms/xxx, /cms/cmsd01/public/cms-software(<100G) Login cmsui01, lxslc(old lxplus, redhat) Plan Balance between local disk and SE Directly read and write SE after SE increase to 100TB Shut down VNC to use resource more reasonably and efficiently

T2 resource and discussion CPU resource (468 cores) Old: 104 cores; New: 364 cores Shared with Atlas, biomed, cms CMS 50% production, 50% analysis In theory, 100 or so for analysis(including physics group) In practice, 200 or more , you can get “running and queue” information from http://ganglia.ihep.ac.cn/ganglia/addons/job_monarch/?c=BEIJING-LCG2%20WN Storage resource (20TB ->100TB) Plans Give local users higher priority Establish special role for local users in CERN voms (/cms/Role=NULL->/cms/Role=Beijing) 100 or less SE will be too high in the future Phedex transfer(debug+prod), production, global analysis, local analysis

T2 resource and physics group A nominal T2 resource: ~500 batch slots, 200TB ~ 468 batch slots(sharing) , 100TB 200 batch slots in T2_CN_Beijing CPU Half for simulation and half for analysis 100 for analysis in T2_CN_Beijing Storage Central Space 30TB(?) MC space 20TB Physics Group Space 60~90TB Local Storage Space 30~60TB User Space 1TB (20 members) 100TB in T2_CN_Beijing soon( MC space 20TB, Local Storage 30~60TB, User Space 20TB, tmp 1TB)

T2 association to physics group

Should we join physics group? T2 physics group meeting this workshop Advantage: Share cpu and space resource Higher priority More support Crab server Stage out to an official area More experience Space management Batch slots management From physics point of view,… Disadvantage: Requirements on resources

CRAB and Local Batch Submission functions CRAB used for local batch submission Plan to release next year Still a plan, no real actions until now The functions included in the future Support the input dataset reading from and the output dataset writing to the local disk Support dataset registering in local DBS Read or Write to SE with certificate or without certificate? Still in discussion

CRAB Server standalone and client-server mode client-server mode a server is placed between the user and the Grid Encourage analysis users to use it Standalone mode Directly submit to grid as now Maintain or not support in the future? Still in discussion Used for local batch submission? guide and questions: https://twiki.cern.ch/twiki/bin/view/CMS/SWGuideCrab Get support from hn-cms-crabFeedback mailing-list Grid Crab server

CRAB Server Goals: Status: make users to submit a large number of jobs easier don't wait for submission, strong resubmission functionality email you when jobs are finished know the job status easily in the web page No limitations on the size of inputbox Easy to maintain, extend, support crab client would be kept as stable as possible, as simple as possible crab server responsible for most of maintaining and extending tasks(physics group) Status: The first release in May this year 14 May 2008: CRABSERVER_1_0_0 has been released Two crab servers available now Bari and Pisa Servers will be managed by physics groups A common server in CERN set up soon

The progress in batch submission advantage and disadvantage Easy for functions control and debug Difficult to maintain in the future, time consuming plans and problems I plan two support ways Support the dataset reading from and writing to local disk but can not support disk data to be published in DBS Support the dataset reading from and writing to SE require dcap open to nodes without certificate or require srmcp with certificate

Site Monitoring New transferring link tests At the rate of 2.5Mb/s Too many parallel links keep SE busy Information counted in site monitoring CMS SAM availability CMS JobRobot CMS Commission links Methods used in site monitoring Site flag: UNCOMMISSIONED: site unusable or very unstable WARNING: unusable today but ok before COMMISSIONED: usable and stable The flag for the day depends on the metrics of the previous 7 days

Site Monitoring