Storage Workshop Summary Wahid Bhimji University Of Edinburgh On behalf all of the participants…

Slides:



Advertisements
Similar presentations
GridPP9 – 5 February 2004 – Data Management DataGrid is a project funded by the European Union GridPP is funded by PPARC GridPP2: Data and Storage Management.
Advertisements

GridPP Meeting, 28/6/06 UB Overview à Role of the UB n Represent user community within GridPP management n Request and allocate.
Storage Review David Britton,21/Nov/ /03/2014 One Year Ago Time Line Apr-09 Jan-09 Oct-08 Jul-08 Apr-08 Jan-08 Oct-07 OC Data? Oversight.
Partner Logo Tier1/A and Tier2 in GridPP2 John Gordon GridPP6 31 January 2003.
User Board - Supporting Other Experiments Stephen Burke, RAL pp Glenn Patrick.
HEPiX Edinburgh 28 May 2004 LCG les robertson - cern-it-1 Data Management Service Challenge Scope Networking, file transfer, data management Storage management.
Wahid Bhimji SRM; FTS3; xrootd; DPM collaborations; cluster filesystems.
Nick Brook University of Bristol The LHC Experiments & Lattice EB News Brief overview of the expts  ATLAS  CMS  LHCb  Lattice.
LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
Data Management TEG Status Dirk Duellmann & Brian Bockelman WLCG GDB, 9. Nov 2011.
Storage: Futures Flavia Donno CERN/IT WLCG Grid Deployment Board, CERN 8 October 2008.
Project Overview GridPP Storage J Jensen GridPP storage workshop RHUL, April 2010.
e-ScienceTalk: Supporting Grid and High Performance Computing Reporting across Europe GA No September 2010 – 31 May 2013.
Birmingham Particle Physics Masterclass 23 th April 2008 Birmingham Particle Physics Masterclass 23 th April 2008 The Grid What & Why? Presentation by:
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
Filesytems and file access Wahid Bhimji University of Edinburgh, Sam Skipsey, Chris Walker …. Apr-101Wahid Bhimji – Files access.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
INFSO-RI Enabling Grids for E-sciencE SRMv2.2 experience Sophie Lemaitre WLCG Workshop.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Grid Data Management A network of computers forming prototype grids currently operate across Britain and the rest of the world, working on the data challenges.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
StoRM Some basics and a comparison with DPM Wahid Bhimji University of Edinburgh GridPP Storage Workshop 31-Mar-101Wahid Bhimji – StoRM.
SRM 2.2: status of the implementations and GSSD 6 th March 2007 Flavia Donno, Maarten Litmaath INFN and IT/GD, CERN.
Your university or experiment logo here NextGen Storage Shaun de Witt (STFC) With Contributions from: James Adams, Rob Appleyard, Ian Collier, Brian Davies,
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Storage Wahid Bhimji DPM Collaboration : Tasks. Xrootd: Status; Using for Tier2 reading from “Tier3”; Server data mining.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
GridPP18 Glasgow Mar 07 DØ – SAMGrid Where’ve we come from, and where are we going? Evolution of a ‘long’ established plan Gavin Davies Imperial College.
GridPP3 Project Management GridPP20 Sarah Pearce 11 March 2008.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
Δ Storage Middleware GridPP10 What’s new since GridPP9? CERN, June 2004.
Your university or experiment logo here Storage and Data Management - Background Jens Jensen, STFC.
Your university or experiment logo here GridPP Storage Future Jens Jensen GridPP workshop RHUL, April 2010.
Testing the UK Tier 2 Data Storage and Transfer Infrastructure C. Brew (RAL) Y. Coppens (Birmingham), G. Cowen (Edinburgh) & J. Ferguson (Glasgow) 9-13.
LCG Service Challenges: Planning for Tier2 Sites Update for HEPiX meeting Jamie Shiers IT-GD, CERN.
Light weight Disk Pool Manager experience and future plans Jean-Philippe Baud, IT-GD, CERN September 2005.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Caitriana Nicholson, CHEP 2006, Mumbai Caitriana Nicholson University of Glasgow Grid Data Management: Simulations of LCG 2008.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
Grid Deployment Board – 10 February GD LCG Workshop Goals Give overview where we are Stimulate cooperation between the centres Improve the communication.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Pre-GDB Storage Classes summary of discussions Flavia Donno Pre-GDB.
WLCG Grid Deployment Board, CERN 11 June 2008 Storage Update Flavia Donno CERN/IT.
Documentation (& User Support) Issues Stephen Burke RAL DB, Imperial, 12 th July 2007.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
RUP Deployment RUP Deployment Workflow
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
Data Placement Intro Dirk Duellmann WLCG TEG Workshop Amsterdam 24. Jan 2012.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
Report from GSSD Storage Workshop Flavia Donno CERN WLCG GDB 4 July 2007.
GridPP storage status update Joint GridPP Board Deployment User Experiment Update Support Team, Imperial 12 July 2007,
Storage Interfaces and Access pre-GDB Wahid Bhimji University of Edinburgh On behalf of all those who participated.
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
SRM v2.2 Production Deployment SRM v2.2 production deployment at CERN now underway. – One ‘endpoint’ per LHC experiment, plus a public one (as for CASTOR2).
Testing Infrastructure Wahid Bhimji Sam Skipsey Intro: what to test Existing testing frameworks A proposal.
Grid Deployment Board 5 December 2007 GSSD Status Report Flavia Donno CERN/IT-GD.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
The Grid Storage System Deployment Working Group 6 th February 2007 Flavia Donno IT/GD, CERN.
J Jensen/J Gordon RAL Storage Storage at RAL Service Challenge Meeting 27 Jan 2005.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
Gene Oleynik, Head of Data Storage and Caching,
Status of the SRM 2.2 MoU extension
WLCG Service Interventions
Harvard Web Publishing Web Publishing for the Harvard Community
Presentation transcript:

Storage Workshop Summary Wahid Bhimji University Of Edinburgh On behalf all of the participants…

What was it? Monday / Tuesday this week. Focused on Tier 2 Grid Storage in the UK. Good attendance from UK storage community: – All our SRM technologies dCache/DPM/ Storm – Expertise in the demanding customers - CMS/ ATLAS – but not forgetting the others

Stated Aims Data at LHC – Tier 2 storage facing significant challenges. Workshop - take a step back and: 1.Identify significant issues from site experiences. 2. Provide training and spread good practice for current SRM/Storage technologies. 3.Consider other/emerging technologies: – Do these actually address the challenges – If so – what do we need to do to test/develop them.

Overview Sessions on: Site Experiences – bottlenecks SRMs – experiences/tutorials/plans/ comparisons Hardware Filesystems and File Access The Future Some highlights here – see upcoming write-up for full info…

Site Experiences Sites (big and small) running well Dealing with all kinds of operational issues Issues: Upgrade schedule – can gridpp manage process, to getting features and stability Thin(ner) SRM layers -­ do all use cases need the full SRM spec -­ longer term discussion

SRM – Storage Resource Manager Variety of SRM implementations (3 in UK Tier2s) All increasingly mature – no losers dropping out Do have different strengths

SRMs dCache: Getting easier to use Well liked by current sites Certain key valuable features such as replication DPM: Most used in UK ( and world) Support from CERN improved Development of some features desirable (replication, draining, checking) we can push /help them StoRM Thin SRM - exploits filesystem Now mature and more widely used Questions over support and feature implementation Needs increasing UK expertise

SRMs Converging (in some ways)– eg NFS4, file:// But is there a point in making DPM more like dCache (in features) or Storm (in Lustre support) …

Hardware SSDs: Ongoing Studies at Glasgow Performance for sure but is it needed if LHC data structures and access improve Networking: 10 Gig for those who can afford it… channel bonding for others

Procurement Discussions Advantages of teaming together with others (inc university or T2 or T1) for hardware purchases to get bulk rates. Even sharing purchase information (where allowed) is useful.

Filesystems – File Access File access: Changes in LHC file access have big impacts on site storage efficiencies, tuning etc. Need to continue studies and keep in close contact with developments

Filesystems Other advanced filesystems - GPFS, HDFS can offer features that can be exploited for HEP storage As can models of analysis such as Hadoop Map-Reduce

The Future Brainstorming session covered both short term needs and longer term strategy Compile to proper output -> workplan but, to give some examples…

Discussions Local file access User space expectations Upgrade schedule getting the features and stability Super lightweight SRMs Current solutions made more easily deployable and configurable Support: Recipes gridpp certified? Coordinated testing of implementations Regular Workshops! Emerging Issues / Questions Dream storage solutions Actions for GRIPP STORAGE

Conclusions Valuable workshop – In and outside sessions Indications of challenges ahead – And work to be done

Conclusions People say storage is hard – So we are lucky to have the GridPP Storage Community – We need to maintain fora for it.

Some Issues Local file access – random file patterns cause load Priorities for Evaluation – Hardware / SRM-Filesystems Which SRM is best for a Tier 2- not just a big=dCache small= DPM issue. Do I buy SSDs? 10 Gig E? More reliable disks? User space expectations- I have 2 TB in my pocket why do you give me 1 on the grid Should GridPP package/certify its own releases – ensure they are stable + get early access to features Do we need (to use) a lighter way to access some things on a grid – at Tier 3s etc. Can we get more value through shared procurements