YAN, Tian On behalf of distributed computing group Institute of High Energy Physics (IHEP), CAS, China CHEP-2015, Apr. 13-17th, OIST, Okinawa.

Slides:



Advertisements
Similar presentations
Status of BESIII Distributed Computing BESIII Workshop, Mar 2015 Xianghu Zhao On Behalf of the BESIII Distributed Computing Group.
Advertisements

CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
Jiri Chudoba for the Pierre Auger Collaboration Institute of Physics of the CAS and CESNET.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
Minerva Infrastructure Meeting – October 04, 2011.
Bookkeeping data Monitoring info Get jobs Site A Site B Site C Site D Agent Production service Monitoring service Bookkeeping service Agent © Andrei Tsaregorodtsev.
Enabling Grids for E-sciencE Medical image processing web portal : Requirements analysis. An almost end user point of view … H. Benoit-Cattin,
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Pilots 2.0: DIRAC pilots for all the skies Federico Stagni, A.McNab, C.Luzzi, A.Tsaregorodtsev On behalf of the DIRAC consortium and the LHCb collaboration.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep ,
BESIII distributed computing and VMDIRAC
Grid Initiatives for e-Science virtual communities in Europe and Latin America DIRAC TEAM CPPM – CNRS DIRAC Grid Middleware.
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
Status of StoRM+Lustre and Multi-VO Support YAN Tian Distributed Computing Group Meeting Oct. 14, 2014.
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
Developing & Managing A Large Linux Farm – The Brookhaven Experience CHEP2004 – Interlaken September 27, 2004 Tomasz Wlodek - BNL.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Successful Common Projects: Structures and Processes WLCG Management.
Cloud Status Laurence Field IT/SDC 09/09/2014. Cloud Date Title 2 SaaS PaaS IaaS VMs on demand.
9 th Weekly Operation Report on DIRAC Distributed Computing YAN Tian From to
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep , 2014 Draft.
Getting started DIRAC Project. Outline  DIRAC information system  Documentation sources  DIRAC users and groups  Registration with DIRAC  Getting.
Conference name Company name INFSOM-RI Speaker name The ETICS Job management architecture EGEE ‘08 Istanbul, September 25 th 2008 Valerio Venturi.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
CHEP 2006, February 2006, Mumbai 1 LHCb use of batch systems A.Tsaregorodtsev, CPPM, Marseille HEPiX 2006, 4 April 2006, Rome.
INFSO-RI Enabling Grids for E-sciencE Ganga 4 – The Ganga Evolution Andrew Maier.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
The GridPP DIRAC project DIRAC for non-LHC communities.
Status of BESIII Distributed Computing BESIII Workshop, Sep 2014 Xianghu Zhao On Behalf of the BESIII Distributed Computing Group.
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
1 LHCb view on Baseline Services A.Tsaregorodtsev, CPPM, Marseille Ph.Charpentier CERN Baseline Services WG, 4 March 2005, CERN.
1 Cloud Services Requirements and Challenges of Large International User Groups Laurence Field IT/SDC 2/12/2014.
1 DIRAC agents A.Tsaregorodtsev, CPPM, Marseille ARDA Workshop, 7 March 2005, CERN.
OpenNebula: Experience at SZTAKI Peter Kacsuk, Sandor Acs, Mark Gergely, Jozsef Kovacs MTA SZTAKI EGI CF Helsinki.
Jiri Chudoba for the Pierre Auger Collaboration Institute of Physics of the CAS and CESNET.
The GridPP DIRAC project DIRAC for non-LHC communities.
The VOMS and the SE in Tier2 Presenter: Sergey Dolgobrodov HEP Meeting Manchester, January 2009.
Building Virtual Scientific Computing Environment with Openstack Yaodong Cheng, CC-IHEP, CAS ISGC 2015.
Virtual Cluster Computing in IHEPCloud Haibo Li, Yaodong Cheng, Jingyan Shi, Tao Cui Computer Center, IHEP HEPIX Spring 2016.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
StoRM + Lustre Proposal YAN Tian On behalf of Distributed Computing Group
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
VMDIRAC DEVELOPMENT PROPOSAL Zhao Xianghu Oct 14, 2014.
DIRAC for Grid and Cloud Dr. Víctor Méndez Muñoz (for DIRAC Project) LHCb Tier 1 Liaison at PIC EGI User Community Board, October 31st, 2013.
DIRAC Distributed Computing Services A. Tsaregorodtsev, CPPM-IN2P3-CNRS FCPPL Meeting, 29 March 2013, Nanjing.
Workload Management Status DIRAC User Meeting Marseille, Oct 2012.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Parag Mhashilkar (Fermi National Accelerator Laboratory)
Multi-community e-Science service connecting grids & clouds R. Graciani 1, V. Méndez 2, T. Fifield 3, A. Tsaregordtsev 4 1 University of Barcelona 2 University.
Enabling Grids for E-sciencE Claudio Cherubino INFN DGAS (Distributed Grid Accounting System)
Status of BESIII Distributed Computing BESIII Collaboration Meeting, Nov 2014 Xiaomei Zhang On Behalf of the BESIII Distributed Computing Group.
Progress on Design and Implement of Job Management System Suo Bing, Yan Tian, Zhao Xianghu
CEPC software & computing study group report
Status of BESIII Distributed Computing
Status of BESIII Computing
The advances in IHEP Cloud facility
Distributed Computing in IHEP
Xiaomei Zhang CMS IHEP Group Meeting December
Elastic Computing Resource Management Based on HTCondor
Yaodong CHENG Computing Center, IHEP, CAS 2016 Fall HEPiX Workshop
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Computing at CEPC Xiaomei Zhang Xianghu Zhao
Discussions on group meeting
Xiaomei Zhang On behalf of CEPC software & computing group Nov 6, 2017
Status and prospects of computing at IHEP
The LHCb Computing Data Challenge DC06
Presentation transcript:

YAN, Tian On behalf of distributed computing group Institute of High Energy Physics (IHEP), CAS, China CHEP-2015, Apr th, OIST, Okinawa

Distributed computing for BESIII Other experiments wish to use DIRAC Multi-VO support architecture refinement configurations in DIRAC configurations in StoRM SE frontend system monitoring and accounting system 2

Start running at 2012 Based on DIRAC Figures: 10 sites ~ 3000 CPU cores ~ 400 TB storage ~ 50k jobs/month 3

DIRAC as middleware ganga as frontend, with extension for BESIII dCache with 126TB disk array as central SE A high level data transfer system for data transferring between SEs VMDIRAC plugin for cloud sites Use CERN’s cvmfs server for software deployment 4

Large High Altitude Air Shower Observatory (LHAASO) 5 Jiangmen Underground Neutrino Observatory (JUNO) Circular Electron Positron Collider (CEPC) Hard X-ray Modulation Telescope (HXMT)

One DIRAC setup for one VO is expensive: need dedicated hardware need expert manpower to maintain small VOs are not willing to afford that Universities in China joined several experiments above one site belongs to several VOs they may have quota and priority policies a single DIRAC setup will be easy to manage these resources BES-DIRAC is already running and can be extended without too much effort Easily to support new experiments in future 6

multiple DIRAC servers for load balances dedicated DB server for DFC, accounting and monitoring lightweight frontend JSUB local CVMFS server for software deployment StoRM+Lustre for integration of grid and local data 7

VO and VOMS settings disable the global VO add VO items in the configuration set VOMS server and URL for each VO User Group settings add new user group for each VO add a generic pilot group for each VO specify VO in user group and generic pilot group 8

Cloud site cloud site has different setting scheme we plan to add direct VO control for cloud sites Temporary solutions for cloud site specify the groups belong to the VO in requirements only cloud from that group can launch VMs and accept jobs 9 non-Cloud site VO can be set for each site in /CE/queue multi-VO in one site can be specified

SiteDirector agents one agent for one VO, with it’s own configuration file /opt/dirac/pro/etc/WorkloadManagemen t_SiteDirector.cfg can be distributed in slaver DIRAC server for load balance Multi-VO’s job control SiteDirector will send generic pilot to job queue under the same VO as the job The generic pilot of the VO is set in the configuration Generic pilot can pull all jobs in the same VO 10

Credential management in StoRM rely on user credential for what concern user authentication and authorization use VOMS extension to define access policy Multi-VO configurations in StoRM list of supported VOs: VOS variable in storm.def storage areas for each VO: STORM_STORAGEAREA_LIST STORM_{SA}_VONAME STORM_{SA}_ONLINE_SIZE STORM_{SA}_DEFAULT_ACL_LIST VO specific users and groups in user.conf, group.conf files VOMS information for each VO in siteinfo/vo.d/ 11

A lightweight frontend system (JSUB) is designed ganga is too complicated for us it takes a long time for adding plugin for new VOs Features of JSUB easy for adding a plugin when new VO is joined support massive job splitting and submission support workflow control support task-based job management support Condor backend A Prototype is completed support bes/cepc/juno VO has been used by juno and cepc users 12

Monitoring system system admin will benefit from site and server monitoring help find problems quickly and locating exactly save manpower during the management of several VO’s resources VO-related Accounting can give statistics of resource usage for different users, and different VOs more functionality can be added beyond DIRAC’s accounting Progress a dedicated DB server is prepared prototype is under design 13

Based on BES-DIRAC platform, we are making DIRAC as a service for multi-VOs (bes, cepc, juno, etc.). DIRAC and StoRM SE are configured to support multi-VO workload management and data management. Frontend and monitoring system are under developing with multi- VO considerations. 14