Www.egi.eu EGI-InSPIRE RI-261323 EGI-InSPIRE www.egi.eu EGI-InSPIRE RI-261323 Data Management Highlights in TSA3.3 Services for HEP Fernando Barreiro Megino,

Slides:



Advertisements
Similar presentations
LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
Advertisements

Summary of issues and questions raised. FTS workshop for experiment integrators Summary of use  Generally positive response on current state!  Now the.
Ian Fisk and Maria Girone Improvements in the CMS Computing System from Run2 CHEP 2015 Ian Fisk and Maria Girone For CMS Collaboration.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
Status of WLCG Tier-0 Maite Barroso, CERN-IT With input from T0 service managers Grid Deployment Board 9 April Apr-2014 Maite Barroso Lopez (at)
ALICE data access WLCG data WG revival 4 October 2013.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Grid Data Management A network of computers forming prototype grids currently operate across Britain and the rest of the world, working on the data challenges.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
CERN IT Department CH-1211 Geneva 23 Switzerland t The Experiment Dashboard ISGC th April 2008 Pablo Saiz, Julia Andreeva, Benjamin.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
Take on messages from Lecture 1 LHC Computing has been well sized to handle the production and analysis needs of LHC (very high data rates and throughputs)
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES PhEDEx Monitoring Nicolò Magini CERN IT-ES-VOS For the PhEDEx.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Successful Common Projects: Structures and Processes WLCG Management.
CERN IT Department CH-1211 Genève 23 Switzerland t Monitoring: Tracking your tasks with Task Monitoring PAT eLearning – Module 11 Edward.
Caitriana Nicholson, CHEP 2006, Mumbai Caitriana Nicholson University of Glasgow Grid Data Management: Simulations of LCG 2008.
Author: Andrew C. Smith Abstract: LHCb's participation in LCG's Service Challenge 3 involves testing the bulk data transfer infrastructure developed to.
Julia Andreeva, CERN IT-ES GDB Every experiment does evaluation of the site status and experiment activities at the site As a rule the state.
WebFTS File Transfer Web Interface for FTS3 Andrea Manzi On behalf of the FTS team Workshop on Cloud Services for File Synchronisation and Sharing.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Ricardo Rocha CERN (IT/GS) EGEE’08, September 2008, Istanbul, TURKEY Experiment.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Recent improvements in HLRmon, an accounting portal suitable for national Grids Enrico Fattibene (speaker), Andrea Cristofori, Luciano Gaido, Paolo Veronesi.
EGI-InSPIRE EGI-InSPIRE RI DDM solutions for disk space resource optimization Fernando H. Barreiro Megino (CERN-IT Experiment Support)
The CMS Top 5 Issues/Concerns wrt. WLCG services WLCG-MB April 3, 2007 Matthias Kasemann CERN/DESY.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
ATP Future Directions Availability of historical information for grid resources: It is necessary to store the history of grid resources as these resources.
JAliEn Java AliEn middleware A. Grigoras, C. Grigoras, M. Pedreira P Saiz, S. Schreiner ALICE Offline Week – June 2013.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
Data Placement Intro Dirk Duellmann WLCG TEG Workshop Amsterdam 24. Jan 2012.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI How to integrate portals with the EGI monitoring system Dusan Vudragovic.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI User-centric monitoring of the analysis and production activities within.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Monitoring of the LHC Computing Activities Key Results from the Services.
Stephen Gowdy FNAL 9th Feb 2015CMS Computing Model Simulation 1.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Disk Space publication Simone Campana Fernando Barreiro Wahid.
Gennaro Tortone, Sergio Fantinel – Bologna, LCG-EDT Monitoring Service DataTAG WP4 Monitoring Group DataTAG WP4 meeting Bologna –
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Dynamic Data Placement: the ATLAS model Simone Campana (IT-SDC)
WLCG Transfers Dashboard A unified monitoring tool for heterogeneous data transfers. Alexandre Beche.
The Grid Storage System Deployment Working Group 6 th February 2007 Flavia Donno IT/GD, CERN.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
XRootD Monitoring Report A.Beche D.Giordano. Outlines  Talk 1: XRootD Monitoring Dashboard  Context  Dataflow and deployment model  Database: storage.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Storage Accounting John Gordon, STFC OMB August 2013.
Using HLRmon for advanced visualization of resource usage Enrico Fattibene INFN - CNAF ISCG 2010 – Taipei March 11 th, 2010.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
WLCG Transfers monitoring EGI Technical Forum Madrid, 17 September 2013 Pablo Saiz on behalf of the Dashboard Team CERN IT/SDC.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI John Gordon EGI Virtualisation and Cloud Workshop Amsterdam 12 th May 2011.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
1 LCG-France 22 November 2010 Tier2s connectivity requirements 22 Novembre 2010 S. Jézéquel (LAPP-ATLAS)
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
Federating Data in the ALICE Experiment
Daniele Bonacorsi Andrea Sciabà
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Introduction to Data Management in EGI
Proposal for obtaining installed capacity
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Evolution of the distributed computing model The case of CMS
Ákos Frohner EGEE'08 September 2008
LHC Data Analysis using a worldwide computing grid
The LHCb Computing Data Challenge DC06
Presentation transcript:

EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Data Management Highlights in TSA3.3 Services for HEP Fernando Barreiro Megino, Domenico Giordano, Maria Girone, Elisa Lanciotti, Daniele Spiga on behalf of CERN-IT-ES-VOS and SA3 EGI Technical Forum – Data management highlights

EGI-InSPIRE RI Outline Introduction: WLCG today LHCb Accounting Storage Element and File Catalogue consistency ATLAS Distributed Data Management: Breaking cloud boundaries CMS Popularity and Automatic Site Cleaning Conclusions EGI Technical Forum – Data management highlights

EGI-InSPIRE RI Outline Introduction: WLCG today LHCb Accounting Storage Element and File Catalogue consistency ATLAS Distributed Data Management: Breaking cloud boundaries CMS Popularity and Automatic Site Cleaning Conclusions EGI Technical Forum – Data management highlights

EGI-InSPIRE RI WLCG today EGI Technical Forum – Data management highlights experiments ALICE ATLAS CMS LHCb Over 140 sites ˜150k CPU cores >50 PB disk Few thousand users O(1M) file transfers/day O(1M) jobs/day

EGI-InSPIRE RI Outline Introduction: WLCG today LHCb Accounting Storage Element and File Catalogue consistency ATLAS Distributed Data Management: Breaking cloud boundaries CMS Popularity and Automatic Site Cleaning Conclusions EGI Technical Forum – Data management highlights

EGI-InSPIRE RI LHCb Accounting EGI Technical Forum – Data management highlights Reports used are currently the main input for clean-up campaigns Agent that on a daily basis generates an accounting report based on the information available on the book-keeping system Metadata breakdown Location Data type Event type File type Display information in dynamic web-page

EGI-InSPIRE RI Outline Introduction: WLCG today LHCb Accounting Storage Element and File Catalogue consistency ATLAS Distributed Data Management: Breaking cloud boundaries CMS Popularity and Automatic Site Cleaning Conclusions EGI Technical Forum – Data management highlights

EGI-InSPIRE RI Storage element and file catalogue consistency Grid Storage Elements (SEs) are decoupled from the File Catalogue (FC). Inconsistencies can arise: 1.Dark data: Waste of disk space 1.Dark data: Data in the SEs, but not in the FC. Waste of disk space 2.Lost/corrupted files: Operational problems, e.g. failing jobs 2.Lost/corrupted files: Data in the FC, but not in the SEs. Operational problems, e.g. failing jobs Dark datafull storage dumpsDark data is identified through consistency checks using full storage dumps one common format and procedureNeed one common format and procedure that covers various SEs: DPM, dCache, StoRM and CASTOR three experiments: ATLAS, CMS and LHCb Decision Text format and XML format Required information is: Spacetoken, LFN (or PFN), file size, creation time and checksum The storage dump should be provided on a weekly/monthly basis or on demand EGI Technical Forum – Data management highlights

EGI-InSPIRE RI Example of good synchronization: LHCb storage usage at CNAF CNAF provides storage dumps daily Checks are done centrally with LHCb Data Management tools Good SE-LFC agreementGood SE-LFC agreement Preliminary results: EGI Technical Forum – Data management highlights Small discrepancies (O(1TB)) are not a real problem. They can be due to a delay between uploading to the SE and registration to LFC and delay to refresh the information in the LHCb database

EGI-InSPIRE RI Outline Introduction: WLCG today LHCb Accounting Storage Element and File Catalogue consistency ATLAS Distributed Data Management: Breaking cloud boundaries CMS Popularity and Automatic Site Cleaning Conclusions EGI Technical Forum – Data management highlights

EGI-InSPIRE RI Original data distribution model Hierarchical tier organization based on Monarc network topology Developed over a decade ago Sites are grouped into clouds for organizational reasons Possible communications: Optical Private Network T0-T1 T1-T1 National networks Intra-cloud T1-T2 Restricted communications: General public network Inter-cloud T1-T2 Inter-cloud T2-T2 But the network capabilities are not the same anymore! Many use-cases require breaking these boundaries! EGI Technical Forum – Data management highlights

EGI-InSPIRE RI Machinery in place 12 Purpose: Generate full mesh transfer statistics for monitoring, site commissioning and to feed back the system EGI Technical Forum – Data management highlights

EGI-InSPIRE RI Consequences Link commissioning –Sites optimizing network connections E.g. UK experience –Revealed different network issues E.g. asymmetric network throughput for various sites (affecting also other experiments) Definition of T2Ds: “Directly connected T2s” Commissioned sites with good network connectivity These sites benefit from closer transfer policies Gradual flattening of the ATLAS Computing Model in order to reduce limitations on –Dynamic data placement –Output collection of multi-cloud analysis Current development of generic, detailed FTS monitor –FTS servers publishing file level information (CERN-IT-GT) –Expose info through generic web interface and API (CERN-IT-ES) EGI Technical Forum – Data management highlights

EGI-InSPIRE RI Outline Introduction: WLCG today LHCb Accounting Storage Element and File Catalogue consistency ATLAS Distributed Data Management: Breaking cloud boundaries CMS Popularity and Automatic Site Cleaning Conclusions EGI Technical Forum – Data management highlights

EGI-InSPIRE RI CMS Popularity EGI Technical Forum – Data management highlights In order to understand how to manage storage more efficiently, it is important to know what data (i.e. which files) is being accessed most and what are the access patterns 30PB of files 50 sites The CMS Popularity service now tracks the utilization of 30PB of files over more than 50 sites CRAB CMS distributed analysis framework CRAB CMS distributed analysis framework Input files Input Blocks LumiRanges Dashboard DB Pull and translate jobs to file level entities Popularity DB Popularity information Popularity web frontend 15 External systems (e.g. cleaning agent) External systems (e.g. cleaning agent)

EGI-InSPIRE RI CMS Popularity Monitoring EGI Technical Forum – Data management highlights

EGI-InSPIRE RI Automatic site cleaning Victor Group pledges & PheDEX Popularity service & PheDEX PheDEX 1. Selection of groups filling their pledge on T2s 2. Selection of unpopular replicas 3. Publication of decisions Used&pledged space Replica popularity Space information Replicas to delete Deleted replicas, Group-site association information Popularity Web Agent running daily on a dedicated machine Project initially developed for ATLAS, now extended for CMS Plug-in architecture Common core Experiment specific plug-ins wrapping their Data Management API calls Project initially developed for ATLAS, now extended for CMS Plug-in architecture Common core Experiment specific plug-ins wrapping their Data Management API calls EGI Technical Forum – Data management highlights Equally important to know what data is not accessed! Automatic procedures for site clean up

EGI-InSPIRE RI Outline Introduction: WLCG today LHCb Accounting Storage Element and File Catalogue consistency ATLAS Distributed Data Management: Breaking cloud boundaries CMS Popularity and Automatic Site Cleaning Conclusions EGI Technical Forum – Data management highlights

EGI-InSPIRE RI Conclusions First 2 years of data taking experiences on the LHC were successful Data volumes and user activity keep increasing We are learning how to operate the infrastructure efficiently Common challenges for all experiments Automate daily operations Optimize the usage of the storage and network resources Evolving computing models Improving data placement strategies 19 EGI Technical Forum – Data management highlights