LHCC Referees Meeting – 28 June 2004 - 1 LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.

Slides:



Advertisements
Similar presentations
HEPiX Edinburgh 28 May 2004 LCG les robertson - cern-it-1 Data Management Service Challenge Scope Networking, file transfer, data management Storage management.
Advertisements

HEPiX GFAL and LCG data management Jean-Philippe Baud CERN/IT/GD.
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
Plateforme de Calcul pour les Sciences du Vivant SRB & gLite V. Breton.
RLS Production Services Maria Girone PPARC-LCG, CERN LCG-POOL and IT-DB Physics Services 10 th GridPP Meeting, CERN, 3 rd June What is the RLS -
1 Grid services based architectures Growing consensus that Grid services is the right concept for building the computing grids; Recent ARDA work has provoked.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Summary of issues and questions raised. FTS workshop for experiment integrators Summary of use  Generally positive response on current state!  Now the.
The LCG File Catalog (LFC) Jean-Philippe Baud – Sophie Lemaitre IT-GD, CERN May 2005.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
Data Management Kelly Clynes Caitlin Minteer. Agenda Globus Toolkit Basic Data Management Systems Overview of Data Management Data Movement Grid FTP Reliable.
RLS Tier-1 Deployment James Casey, PPARC-LCG Fellow, CERN 10 th GridPP Meeting, CERN, 3 rd June 2004.
Grid Deployment Data challenge follow-up & lessons learned Ian Bird LCG Deployment Area Manager LHCC Comprehensive Review 22 nd November 2004.
LCG and HEPiX Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002.
The LCG File Catalog (LFC) Jean-Philippe Baud – Sophie Lemaitre IT-GD, CERN May 2005.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
Replica Management Services in the European DataGrid Project Work Package 2 European DataGrid.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
INFSO-RI Enabling Grids for E-sciencE Experience of using gLite for analysis of ATLAS combined test beam data A. Zalite / PNPI.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
Managing Data DIRAC Project. Outline  Data management components  Storage Elements  File Catalogs  DIRAC conventions for user data  Data operation.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
Distributed Data for Science Workflows Data Architecture Progress Report December 2008.
WP3 Information and Monitoring Rob Byrom / WP3
Database authentication in CORAL and COOL Database authentication in CORAL and COOL Giacomo Govi Giacomo Govi CERN IT/PSS CERN IT/PSS On behalf of the.
David Adams ATLAS ATLAS distributed data management David Adams BNL February 22, 2005 Database working group ATLAS software workshop.
The CMS Top 5 Issues/Concerns wrt. WLCG services WLCG-MB April 3, 2007 Matthias Kasemann CERN/DESY.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
1 Active Directory Service in Windows 2000 Li Yang SID: November 2000.
Baseline Services Group Report LHCC Referees Meeting 27 th June 2005 Ian Bird IT/GD, CERN.
10 May 2001WP6 Testbed Meeting1 WP5 - Mass Storage Management Jean-Philippe Baud PDP/IT/CERN.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
ATLAS Database Access Library Local Area LCG3D Meeting Fermilab, Batavia, USA October 21, 2004 Alexandre Vaniachine (ANL)
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
Site Services and Policies Summary Dirk Düllmann, CERN IT More details at
VOX Project Status T. Levshina. 5/7/2003LCG SEC meetings2 Goals, team and collaborators Purpose: To facilitate the remote participation of US based physicists.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
CMS Experience with the Common Analysis Framework I. Fisk & M. Girone Experience in CMS with the Common Analysis Framework Ian Fisk & Maria Girone 1.
CERN - IT Department CH-1211 Genève 23 Switzerland t Service Level & Responsibilities Dirk Düllmann LCG 3D Database Workshop September,
LHCbComputing Update of LHC experiments Computing & Software Models Selection of slides from last week’s GDB
J Jensen / WP5 /RAL UCL 4/5 March 2004 GridPP / DataGrid wrap-up Mass Storage Management J Jensen
EGEE Data Management Services
Federating Data in the ALICE Experiment
Evolution of storage and data management
Jean-Philippe Baud, IT-GD, CERN November 2007
Bob Jones EGEE Technical Director
EGEE Middleware Activities Overview
ATLAS Use and Experience of FTS
StoRM: a SRM solution for disk based storage systems
Vincenzo Spinoso EGI.eu/INFN
Status of the SRM 2.2 MoU extension
LCG Security Status and Issues
Introduction to Data Management in EGI
Ákos Frohner EGEE'08 September 2008
LCG middleware and LHC experiments ARDA project
Data Management cluster summary
POOL/RLS Experience Current CMS Data Challenges shows clear problems wrt to the use of RLS Partially due to the normal “learning curve” on all sides in.
Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002
INFNGRID Workshop – Bari, Italy, October 2004
Presentation transcript:

LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004

LHCC Referees Meeting – 28 June Background  Continue to develop LCG-2 service to deploy and validate basic underlying infrastructure services essential to have in place  Cannot wait for new EGEE developments – but ensure we are aligned  What we do now may/will be replaced but there is still much to learn and understand  What we propose is consistent with EGEE developments  Underlying system-level issues (firewalls, security, network behaviour, error handling, …) need to be addressed now  Much is learned in the DC’s – need to validate solutions to those problems  Intend to deploy/validate EGEE solutions in parallel (on pre-production service)

LHCC Referees Meeting – 28 June Data Management – 3 areas  Reliable Data Transfer Service  Essential to have in place and verify by end 2004 that we are able to reliably distribute data at significant fraction of data rate expected at LHC start-up Series of service challenges associated with this  File Catalogue  Based on lessons learned in DC’s in last few months – address some fundamental issues of performance and scalability  Valuable input for EGEE developments – with appropriate interfaces could also be alternative implementation of EGEE model (?)  Lightweight disk pool manager  Recognised as necessary – (LCG, Grid3, EGEE) – will be a collaborative effort

LHCC Referees Meeting – 28 June Reliable Data Transfer service  Two areas of work:  Basic underlying infrastructure for service challenges  Management software  Underlying infrastructure:  Load-balanced gridftp service at each end point (500 MB/s would require several gridftp servers ~ >5?)  Disk pools in place Disk management policies – garbage collection, etc.  Routing for data transfers around firewalls vs control channel  This is being set up now by ADC/CS together with FNAL and Nikhef

LHCC Referees Meeting – 28 June Reliable Data Transfer – management  Implementation:  Currently investigating/testing 3 possibilities: TMDB (from CMS) – together with EGEE and CMS –We could use “as-is”, EGEE want to adapt to new architecture Stork (from VDT) pyRFT (python implementation of Globus RFT)  Decide within a week – TMDB looks a good candidate solution  All of these could be used with little adaptation, allowing us to focus on system-level issues  Optimising performance, security issues, etc  Effort:  1-2 people in GD team, together with CMS and EGEE – assuming TMDB  Work in testing has started, set up test framework to FNAL and Nikhef Already being done in context of basic network infrastructure testing

LHCC Referees Meeting – 28 June File catalogues  Proposal: - Key is to simplify, concentrate on functionality and performance  Single central file catalogue: GUID  PFN mappings – no attributes on PFNs LFN  GUID mappings – no user-definable attributes (they are in metadata catalogue) System attributes on GUID – file size, checksum, etc Hierarchical LFN namespace Multiple LFNs for a GUID – compatible implementation with EGEE & Alien Bulk inserts of LFN  GUID  PFN Bulk queries, and cursors for large queries Transactions, Control of transaction exposed to user  Metadata catalogue: Assume most metadata is in experiment catalogues For VO that need it – simple catalogue of “name-value” pair on GUID – separate from file catalogue

LHCC Referees Meeting – 28 June File catalogues – 2  Other issues to be addressed:  Fix naming scheme (has been source of problems)  Transactions  Cursors for efficient and consistent large queries  Collections – in file catalogue – seen as directories/symlinks (or as GUID)  GSI authentication …  … simple C clients (extend existing C clients)  Management tools – logging, accounting, browsing (web based)  Availability  Short term: assume fail-over between instances on several sites Use Oracle tools, db clients look in IS for current primary Oracle DataGuard makes them consistent  Longer term: multi-master database would fit this logic also – using IS

LHCC Referees Meeting – 28 June File catalogues – 3  Has been discussed with POOL team  Will be discussed in PEB tomorrow  Effort is identified in Deployment team  Estimate 1 month for basic efficient implementation using existing catalogue  Begins now if PEB agrees – some up-front work has been done to investigate potential solutions  Prototype in mid-August  Not addressed directly:  Replication – Consider conflict resolution in implementation Expect replication to use DB tools (Oracle) – subject of separate project  WAN interaction – Several ideas (RRS, DB proxy from SAM) Needed to provide connection re-use, timeouts, retries

LHCC Referees Meeting – 28 June Lightweight disk pool manager  Recent experience and current thinking gives following strategy for storage access:  LCG-2, EGEE, Grid3 all see a need for a lightweight dpm  SRM is common interface to storage; 3 cases: 1)Integration of large (tape) MSS (at Tier 1 etc) – Responsibility of site to make the integration – this is the case 2)Large Tier 2’s – sites with large disk pools (10’s Terabytes, many fileservers), need a flexible system dCache provides a good solution, but needs effort to integrate and manage 3)Sites with smaller disk pools, less available management effort Need a lightweight (install, manage) solution  We propose to develop 3)

LHCC Referees Meeting – 28 June Lightweight DPM  Implementation  Re-use same catalogue infrastructure/name server as file catalogues  SRM interface leveraging what exists now  Re-use Globus gridftp server if possible  Local I/O using rfio  Effort  in GDA/EGEE at CERN and Orsay  Interest from Grid3/OSG also  Timescale:  Catalogue infrastructure in August  SRM implementation can start in parallel

LHCC Referees Meeting – 28 June Summary  Propose to address 3 data management issues:  Reliable data delivery  File catalogues  Lightweight disk pool manager  Focus on basic essential services – leave higher levels to experiments  All important to have in place to understand basic system  Data transfer and file catalogues have priority …  … but a simple dpm is missing  Work can start now