CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/i t Plans and Architectural Options for Physics Data Analysis at CERN D. Duellmann, A. Pace.

Slides:



Advertisements
Similar presentations
Storage Review David Britton,21/Nov/ /03/2014 One Year Ago Time Line Apr-09 Jan-09 Oct-08 Jul-08 Apr-08 Jan-08 Oct-07 OC Data? Oversight.
Advertisements

Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS CASTOR Status Alberto Pace.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
EU-GRID Work Program Massimo Sgaravatto – INFN Padova Cristina Vistoli – INFN Cnaf as INFN members of the EU-GRID technical team.
CASTOR Upgrade, Testing and Issues Shaun de Witt GRIDPP August 2010.
CERN - IT Department CH-1211 Genève 23 Switzerland t Oracle and Streams Diagnostics and Monitoring Eva Dafonte Pérez Florbela Tique Aires.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
CERN IT Department CH-1211 Genève 23 Switzerland t Tape-dev update Castor F2F meeting, 14/10/09 Nicola Bessone, German Cancio, Steven Murray,
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS CERN and Computing … … and Storage Alberto Pace Head, Data.
CERN IT Department CH-1211 Geneva 23 Switzerland t The Experiment Dashboard ISGC th April 2008 Pablo Saiz, Julia Andreeva, Benjamin.
CERN IT Department CH-1211 Genève 23 Switzerland t EIS section review of recent activities Harry Renshall Andrea Sciabà IT-GS group meeting.
CERN IT Department CH-1211 Genève 23 Switzerland t Data Management Evolution and Strategy at CERN G. Cancio, D. Duellmann, A. Pace With input.
Mass Storage System Forum HEPiX Vancouver, 24/10/2003 Don Petravick (FNAL) Olof Bärring (CERN)
Your university or experiment logo here NextGen Storage Shaun de Witt (STFC) With Contributions from: James Adams, Rob Appleyard, Ian Collier, Brian Davies,
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
WLCG Service Report ~~~ WLCG Management Board, 24 th November
CERN IT Department CH-1211 Geneva 23 Switzerland t Storageware Flavia Donno CERN WLCG Collaboration Workshop CERN, November 2008.
CERN IT Department CH-1211 Genève 23 Switzerland t Tier0 Status - 1 Tier0 Status Tony Cass LCG-LHCC Referees Meeting 18 th November 2008.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
CERN IT Department CH-1211 Genève 23 Switzerland t Castor development status Alberto Pace LCG-LHCC Referees Meeting, May 5 th, 2008 DRAFT.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
Report from CASTOR external operations F2F meeting held at RAL in February Barbara Martelli INFN - CNAF.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
CASTOR evolution Presentation to HEPiX 2003, Vancouver 20/10/2003 Jean-Damien Durand, CERN-IT.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
Future home directories at CERN
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Priorities update Andrea Sciabà IT/GS Ulrich Schwickerath IT/FIO.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS Castor incident (and follow up) Alberto Pace.
Evolution of storage and data management Ian Bird GDB: 12 th May 2010.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
The CMS Top 5 Issues/Concerns wrt. WLCG services WLCG-MB April 3, 2007 Matthias Kasemann CERN/DESY.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
CERN IT Department CH-1211 Genève 23 Switzerland t HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 The CASTOR SRM2 Interface Status and plans.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
CERN IT Department CH-1211 Genève 23 Switzerland t CERN IT Monitoring and Data Analytics Pedro Andrade (IT-GT) Openlab Workshop on Data Analytics.
Author - Title- Date - n° 1 Partner Logo WP5 Status John Gordon Budapest September 2002.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
SRM v2.2 Production Deployment SRM v2.2 production deployment at CERN now underway. – One ‘endpoint’ per LHC experiment, plus a public one (as for CASTOR2).
8 August 2006MB Report on Status and Progress of SC4 activities 1 MB (Snapshot) Report on Status and Progress of SC4 activities A weekly report is gathered.
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
Grid Deployment Board 5 December 2007 GSSD Status Report Flavia Donno CERN/IT-GD.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
CERN IT Department CH-1211 Genève 23 Switzerland t CCRC’08 Review from a DM perspective Alberto Pace (With slides from T.Bell, F.Donno, D.Duelmann,
CERN IT Department CH-1211 Genève 23 Switzerland t Increasing Tape Efficiency Original slides from HEPiX Fall 2008 Taipei RAL f2f meeting,
Summary of SC4 Disk-Disk Transfers LCG MB, April Jamie Shiers, CERN.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Evolution of WLCG Data & Storage Management Outcome of Amsterdam.
Top 5 Experiment Issues ExperimentALICEATLASCMSLHCb Issue #1xrootd- CASTOR2 functionality & performance Data Access from T1 MSS Issue.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Standard Protocols in DPM Ricardo Rocha.
CERN IT Department CH-1211 Genève 23 Switzerland t DPM status and plans David Smith CERN, IT-DM-SGT Pre-GDB, Grid Storage Services 11 November.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS CASTOR and EOS status and plans Giuseppe Lo Presti on behalf.
Riccardo Zappi INFN-CNAF SRM Breakout session. February 28, 2012 Ingredients 1. Basic ingredients (Fabric & Conn. level) 2. (Grid) Middleware ingredients.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS Storage plans for CERN and for the Tier 0 Alberto Pace (and.
CASTOR: possible evolution into the LHC era
StoRM: a SRM solution for disk based storage systems
Database Readiness Workshop Intro & Goals
StoRM Architecture and Daemons
Introduction to Data Management in EGI
CERN Lustre Evaluation and Storage Outlook
Ákos Frohner EGEE'08 September 2008
CTA: CERN Tape Archive Overview and architecture
WLCG Storage planning Andrea Sciabà
Data Management cluster summary
LHC Data Analysis using a worldwide computing grid
Presentation transcript:

CERN IT Department CH-1211 Genève 23 Switzerland t Plans and Architectural Options for Physics Data Analysis at CERN D. Duellmann, A. Pace

CERN IT Department CH-1211 Genève 23 Switzerland t 2 Agenda Short summary of –Data Management Software Status for the Tier-0 –Ongoing developments for Tier-0 activities Requirements for Analysis received so far Current areas of developments and plans D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008

CERN IT Department CH-1211 Genève 23 Switzerland t 3 Software Status for the Tier-0 For all Tier-0 data operations, CERN is using CASTOR for –Migration of LHC data to tape –Replication of the LHC data to Tier-1 sites CCRC’08 Data Challenges have validated this architecture which is now in production (version 2.1.7)... but few improvements are necessary as Tier-0 data must also be accessible for analysis at CERN ALICE ATLAS CMS LHCb ASGC - Taiwan BNL - USA CNAF - Italy FNAL - USA FZK-GridKA – Germany IN2P3 - France NDGF - Nordic Countries NIKHEF - The Netherlands PIC - Spain RAL - United Kingdom TRIUMF - Canada CERN Tier-0 D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008

CERN IT Department CH-1211 Genève 23 Switzerland t 4 Ongoing developments for Tier-0 During 2008, several simplifications and improvements were made to CASTOR 1.Improvements on Security, SRM and Monitoring 2.Improvements in the tape handling area 3.Reduced latency when accessing files already on disk and improved concurrent access See D.Duellmann at Hepix Fall D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008

CERN IT Department CH-1211 Genève 23 Switzerland t 5 Security, SRM and Monitoring Strong authentication now implemented –End user access to CASTOR has been secured (nameserver, disk cache, protocols) –Support both PKI (Grid certificates) and Kerberos Consolidation ongoing, Performance optimizations planned –See Frohner, Garcia, Ponce at Hepix Fall Castor Monitoring and SRM interface being improved –See De Witt, Lo Presti, Pokorski, Rekatsinas at Hepix Fall D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008 Today (on v2.1.8) Spring 2009 Mid 2009

CERN IT Department CH-1211 Genève 23 Switzerland t 6 Increasing Tape Efficiency New tape queue management supporting recall / migration policies, access control lists, and user / group based priorities New tape format with no performance drop when handling small files Data aggregation to allow managed tape migration / recall to try to match the drive speed (increased efficiency) See Bessone, Cancio, Murray, Taurelli at Hepix Fall See plan and timelines at D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008 Mid / End 2009 End 2009 Today (on v2.1.8)

CERN IT Department CH-1211 Genève 23 Switzerland t 7 The access latency problem... Not a problem for the Tier-0 operation –Scheduled I/O guarantees Tier-0 data streaming to tape and to Tier-1 which are not affected by the high LSF latency... but a concern for interactive data access and analysis –Unacceptable to wait seconds to open a file for read when it is already on disk D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008

CERN IT Department CH-1211 Genève 23 Switzerland t 8 Reduced access latency Removed the LSF job scheduling for reading disk files –Latency reduces from seconds to milliseconds (like a file system) –Plan to remove job scheduling also for write operation Focussed to offer direct access to storage on Castor disk pools –Direct access offered with the XROOT interface which is already part of the current service definition –Excellent integration with root –Mountable file system possible –Additional access protocols can be added (eg. NFS 4.1) Performances can be increased by improving the name server See A.Peters at ACAT D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008 Mid / End 2009 Today (on v2.1.8) End 2010 (if agreed) Today using FUSE (on v2.1.8) Today (on v2.1.8)

CERN IT Department CH-1211 Genève 23 Switzerland t 9 New requirements for analysis Working groups on analysis requirements led by B. Panzer (Jul – Oct ‘08) for Tier-0, by M. Schulz (Nov’08 - ) for Tiers-1/2 (simplified) requirements collected so far –Analysis made on disk pools only Tape access limited to managed and aggregated recall / migration requests from the experiments No end-user recalls or direct access to tapes –Strong demand for direct file access No unnecessary latencies Mountable file system accessible from both interactive desktops and batch systems –Integration with the existing physics analysis tools (grid middleware and root) –Secure / manageable (Authentication, ACLs, Quota, Accounting, etc...) –File catalogue and name server consistency D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008

CERN IT Department CH-1211 Genève 23 Switzerland t 10 Areas being CERN Deploy a new file system –High performance file system access –Mainstream software and services Offer direct file access to Castor files –Limited duplication of data –Consistency guaranteed and data management possible –Performance impacted by name server lookup D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008

CERN IT Department CH-1211 Genève 23 Switzerland t 11 Area 1: A new file system High performance, highly scalable –Gpfs, Lustre, NFS... No end-client software necessary (posix) for basic file operations Long-term dependencies if advanced (specific) file operations features are exposed to end users Areas of research –Integration with existing infrastructure –Wide Area Transfer / Global namespace –Data replication (replicated pools / hot spots) –Tape integration –integration with the file catalogue –SRM interface (based on StoRM ?, BeStMan ?) Realistic timeline: End 2010 for the file system D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008

CERN IT Department CH-1211 Genève 23 Switzerland t 12 Area 2: Castor direct access Direct file system access to Castor files with similar performance in term of requests/sec and latency –A new file system could also be considered as a storage technology for Castor file servers –But... name server lookup generates performance overhead Access protocol –XROOT, RFIO, GridFTP (available now) –Mountable file system possible using Fuse (available now) –NFS 4.1 (as a future option) Name server / file catalogue consistency ensured –Global namespace –Same logical filename across pools and tapes Integration with tapes (in place) data management possible No long-term dependencies: separation between access protocols and storage technologies. D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008

CERN IT Department CH-1211 Genève 23 Switzerland t 13D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008 The options under investigation Technology independent data protocol Centrally managed data solution Storage technology change transparent (Lustre / GPFS /...) Internal architecture change transparent (Tape / Disk / Flash...) End-to-end consistency (eg checksums) Higher performance Data managed by the experiments Namespace on tape differs from namespace on disk Two files with same name can have different content Two files with different name can have same content IT-organized data migrations more heavy Castor RFIO GridFTP XROOT MFS (using FUSE) Posix MFS (ex: NFS 4.1) File System Posix Mountable File System Access control, Managed migrations DB Name server DB Name server

CERN IT Department CH-1211 Genève 23 Switzerland t 14 End of 2010 If planned D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008 Possible timescale Base functionality available today Spring 2009 – Security, Access control and Quotas End of 2009 – improved name servers End of 2010 – Native file system access as a new service Spring 2009 – Prototype End of 2009 – Visible from lxplus End of 2010 – Running as a new service Castor RFIO GridFTP XROOT MFS (using FUSE) Posix MFS (NFS 4.1) File System Posix Mountable File System Access control, Managed migrations DB Name server TODAY Requires for unscheduled I/O Spring 2009 End 2009 Spring 2009 No service approved yet These dates are only effort estimations

CERN IT Department CH-1211 Genève 23 Switzerland t 15 Conclusion Mountable file system is considered a strategic direction –Either with Castor or with a new file system Feedback welcome to understand to what extent experiments require a unique name space between files on disks and files on tape The two approaches implies different levels of consistency and data management automation experiments expect from IT D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008

Questions / Discussion D.Duellmann, A.Pace – Grid Deployment Board, 12 November 2008