ALICE data access WLCG data WG revival 4 October 2013.

Slides:



Advertisements
Similar presentations
Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Advertisements

ALICE G RID SERVICES IP V 6 READINESS
Ningning HuCarnegie Mellon University1 Optimizing Network Performance In Replicated Hosting Peter Steenkiste (CMU) with Ningning Hu (CMU), Oliver Spatscheck.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
Torrent-based Software Distribution in ALICE.
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
ALICE Operations short summary and directions in 2012 Grid Deployment Board March 21, 2011.
ALICE Operations short summary and directions in 2012 WLCG workshop May 19-20, 2012.
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
ALICE DATA ACCESS MODEL Outline ALICE data access model - PtP Network Workshop 2  ALICE data model  Some figures.
Advanced Network Architecture Research Group 2001/11/149 th International Conference on Network Protocols Scalable Socket Buffer Tuning for High-Performance.
Grid Data Management A network of computers forming prototype grids currently operate across Britain and the rest of the world, working on the data challenges.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES P. Saiz (IT-ES) AliEn job agents.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
File and Object Replication in Data Grids Chin-Yi Tsai.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
Costin Grigoras ALICE Offline. In the period of steady LHC operation, The Grid usage is constant and high and, as foreseen, is used for massive RAW and.
Sejong STATUS Chang Yeong CHOI CERN, ALICE LHC Computing Grid Tier-2 Workshop in Asia, 1 th December 2006.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
Advanced Network Architecture Research Group 2001/11/74 th Asia-Pacific Symposium on Information and Telecommunication Technologies Design and Implementation.
N EWS OF M ON ALISA SITE MONITORING
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team 12 th CERN-Korea.
Site operations Outline Central services VoBox services Monitoring Storage and networking 4/8/20142ALICE-USA Review - Site Operations.
ALICE – networking LHCONE workshop 10/02/ Quick plans: Run 2 data taking Both for Pb+Pb and p+p – Reach 1 nb -1 integrated luminosity for rare.
Update on replica management
Caitriana Nicholson, CHEP 2006, Mumbai Caitriana Nicholson University of Glasgow Grid Data Management: Simulations of LCG 2008.
Status of PDC’07 and user analysis issues (from admin point of view) L. Betev August 28, 2007.
CGW 04, Stripped replication for the grid environment as a web service1 Stripped replication for the Grid environment as a web service Marek Ciglan, Ondrej.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES A. Abramyan, S. Bagansco, S. Banerjee, L. Betev, F. Carminati,
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team ATHIC2012, Busan,
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
Xrootd Monitoring and Control Harsh Arora CERN. Setting Up Service  Monalisa Service  Monalisa Repository  Test Xrootd Server  ApMon Module.
Point-to-point Architecture topics for discussion Remote I/O as a data access scenario Remote I/O is a scenario that, for the first time, puts the WAN.
JAliEn Java AliEn middleware A. Grigoras, C. Grigoras, M. Pedreira P Saiz, S. Schreiner ALICE Offline Week – June 2013.
ALICE DATA ACCESS MODEL Outline 05/13/2014 ALICE Data Access Model 2  ALICE data access model  Infrastructure and SE monitoring.
+ AliEn site services and monitoring Miguel Martinez Pedreira.
Efi.uchicago.edu ci.uchicago.edu Ramping up FAX and WAN direct access Rob Gardner on behalf of the atlas-adc-federated-xrootd working group Computation.
August 28, 2003APAN, Logistical Networking WS DiDaS Distributed Data Storage Ludek Matyska Masaryk University, Institute of Comp. Sci. and CESNET, z.s.p.o.
NeST: Network Storage John Bent, Venkateshwaran V Miron Livny, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau.
Efi.uchicago.edu ci.uchicago.edu Storage federations, caches & WMS Rob Gardner Computation and Enrico Fermi Institutes University of Chicago BigPanDA Workshop.
THE GLUE DOMAIN DEPLOYMENT The middleware layer supporting the domain-based INFN Grid network monitoring activity is powered by GlueDomains [2]. The GlueDomains.
ALICE Grid operations +some specific for T2s US-ALICE Grid operations review 7 March 2014 Latchezar Betev 1.
Analysis efficiency Andrei Gheata ALICE offline week 03 October 2012.
1 R. Voicu 1, I. Legrand 1, H. Newman 1 2 C.Grigoras 1 California Institute of Technology 2 CERN CHEP 2010 Taipei, October 21 st, 2010 End to End Storage.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
GRID interoperability and operation challenges under real load for the ALICE experiment F. Carminati, L. Betev, P. Saiz, F. Furano, P. Méndez Lorenzo,
Status of GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
ALICE computing Focus on STEP09 and analysis activities ALICE computing Focus on STEP09 and analysis activities Latchezar Betev Réunion LCG-France, LAPP.
Activities and Perspectives at Armenian Grid site The 6th International Conference "Distributed Computing and Grid- technologies in Science and Education"
Efi.uchicago.edu ci.uchicago.edu Sharing Network Resources Ilija Vukotic Computation and Enrico Fermi Institutes University of Chicago Federated Storage.
Storage discovery in AliEn
1 LCG-France 22 November 2010 Tier2s connectivity requirements 22 Novembre 2010 S. Jézéquel (LAPP-ATLAS)
The ALICE Analysis -- News from the battlefield Federico Carminati for the ALICE Computing Project CHEP 2010 – Taiwan.
Federating Data in the ALICE Experiment
Dynamic Extension of the INFN Tier-1 on external resources
Extending the farm to external sites: the INFN Tier-1 experience
Data Formats and Impact on Federated Access
ALICE internal and external network
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
ALICE Monitoring
INFN-GRID Workshop Bari, October, 26, 2004
Bernd Panzer-Steindel, CERN/IT
Update on Plan for KISTI-GSDC
A Messaging Infrastructure for WLCG
Torrent-based software distribution
Storage elements discovery
Simulation use cases for T2 in ALICE
Ákos Frohner EGEE'08 September 2008
ALICE Computing Upgrade Predrag Buncic
Publishing ALICE data & CVMFS infrastructure monitoring
Presentation transcript:

ALICE data access WLCG data WG revival 4 October 2013

Outline 2  ALICE data model  Some figures & policies  Infrastructure monitoring  Replica discovery mechanism

The AliEn catalogue 3  Central catalogue of logical file names (LFN)  With owner:group and unix-style permissions  Size, MD5 of files, metadata on sub-trees  Each LFN has a GUID  Any number of PFNs can be associated to an LFN  Like root:// // / / HH and hhhhh are hashes of the GUID

ALICE data model (2) 4  Data files are accessed directly  Jobs go to where a copy of the data is – job brokering by AliEn  Reading from the closest working replica to the job  All WAN/LAN i/o through xrootd  while also supporting http, ftp, torrent for downloading other input files  At the end of the job N replicas are uploaded from the job itself (2x ESDs, 3xAODs, etc...)  Scheduled data transfers for raw data with xrd3cp  T0 -> T1

Storage elements and rates 5  60 disk storage elements + 8 tape-backed (T0 and T1s)  28PB in 307M files (replicas included)  2012 averages:  31PB written (1.2GB/s) 2.4PB RAW, ~70MB/s average raw data replication  216PB read back (8.6GB/s) - 7x the amount written  Sustained periods of 3-4x the above

Data Consumers 6  Last month analysis tasks (mix of all types of analysis)  14.2M input files  87.5% accessed from the site local SE at 3.1MB/s  12.5% read from remote at 0.97MB/s  Average processing speed ~2.8MB/s  Analysis job efficiency ~70% for the Grid average CPU power of HepSpec06  => 0.4MB/s/HepSpec06 per job

Data access from analysis jobs 7  Transparent fallback to remote SEs works well  Penalty for remote i/o, buffering essesntial  The external connection is a minor issue … IO-intensive analysis train instance

Aggregated SE traffic 8 Period of the IO-intensive train

Monitoring and decision making 9  On all VoBox-es a MonALISA service collects  Job resource consumption, WN host monitoring …  Local SEs host monitoring data (network traffic, load, sockets etc)  VoBox to VoBox network measurements  traceroute / tracepath / bandwidth measurement  Results are archived and used to create network topology of all-to-all

Network topology view in MonALISA 10

Available bandwidth per stream 11 Funny ICMP throttling Discreet effect of the congestion control algorithm on links with packet loss (x 8.3Mbps) Suggested larger-than-default buffers (8MB) Default buffers

Bandwidth test matrix 12  4 years of archived results for 80x80 sites matrix 

Replica discovery mechanism 13  Closest working replicas are used for both reading and writing  Sorting the SEs by the network distance to the client making the request  Combining network topology data with the geographical one  Weighted by reliability test results  Writing is slightly randomized for more ‘democratic’ data distribution

Plans 14  Work with sites to improve local infrastructure  Eg. tuning of xrootd gateways for large GPFS clusters, insufficient backbone capacity  Provide only relevant information (too much is not good) to resolve uplink problems  Deploy a similar (throughput) test suite on the data servers  (Re)enable icmp where it is missing  (Re)apply TCP buffer settings …  We only see the end-to-end results  Complete WAN infrastructure not yet revealed

Conclusions 15  ALICE tasks use all resources in democratic way  No dedicated SEs or sites for particular tasks With the small exception of RAW  The model is adaptive to the network capacity and performance  Uniform use of xrootd  Tuning needed to accommodate better i/o hungry analysis tasks – this is the largest consumer of disk and network  Coupled with site storage and network tuning of every individual site  The LHCONE initiative has already shown positive effect