Future Plans at RAL Tier 1 Shaun de Witt. Introduction Current Set-Up Short term plans Final Configuration How we get there… How we plan/hope/pray to.

Slides:



Advertisements
Similar presentations
Hardware Reliability at the RAL Tier1 Gareth Smith 16 th September 2011.
Advertisements

Applications Area Issues RWL Jones GridPP13 – 5 th June 2005.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS CASTOR Status Alberto Pace.
Exporting Raw/ESD data from Tier-0 Tier-1s Wrap-up.
Steve Traylen Particle Physics Department Experiences of DCache at RAL UK HEP Sysman, 11/11/04 Steve Traylen
LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
CASTOR Upgrade, Testing and Issues Shaun de Witt GRIDPP August 2010.
1 RAL Status and Plans Carmine Cioffi Database Administrator and Developer 3D Workshop, CERN, November 2009.
CERN IT Department CH-1211 Genève 23 Switzerland t Tier0 Status Tony Cass (With thanks to Miguel Coelho dos Santos & Alex Iribarren) LCG-LHCC.
CERN IT Department CH-1211 Genève 23 Switzerland t Plans and Architectural Options for Physics Data Analysis at CERN D. Duellmann, A. Pace.
Data management in grid. Comparative analysis of storage systems in WLCG.
CERN IT Department CH-1211 Genève 23 Switzerland t EIS section review of recent activities Harry Renshall Andrea Sciabà IT-GS group meeting.
RAL Site Report Castor F2F, CERN Matthew Viljoen.
Your university or experiment logo here NextGen Storage Shaun de Witt (STFC) With Contributions from: James Adams, Rob Appleyard, Ian Collier, Brian Davies,
Operation of CASTOR at RAL Tier1 Review November 2007 Bonny Strong.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
CERN IT Department CH-1211 Genève 23 Switzerland t Tier0 Status - 1 Tier0 Status Tony Cass LCG-LHCC Referees Meeting 18 th November 2008.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
CASTOR Databases at RAL Carmine Cioffi Database Administrator and Developer Castor Face to Face, RAL February 2009.
Your university or experiment logo here CEPH at the Tier 1 Brain Davies On behalf of James Adams, Shaun de Witt & Rob Appleyard.
Steve Traylen PPD Rutherford Lab Grid Operations PPD Christmas Lectures Steve Traylen RAL Tier1 Grid Deployment
RAL Site Report Castor Face-to-Face meeting September 2014 Rob Appleyard, Shaun de Witt, Juan Sierra.
CASTOR: CERN’s data management system CHEP03 25/3/2003 Ben Couturier, Jean-Damien Durand, Olof Bärring CERN.
An Agile Service Deployment Framework and its Application Quattor System Management Tool and HyperV Virtualisation applied to CASTOR Hierarchical Storage.
Report from CASTOR external operations F2F meeting held at RAL in February Barbara Martelli INFN - CNAF.
User Board Input Tier Storage Review 21 November 2008 Glenn Patrick Rutherford Appleton Laboratory.
Review of Recent CASTOR Database Problems at RAL Gordon D. Brown Rutherford Appleton Laboratory 3D/WLCG Workshop CERN, Geneva 11 th -14 th November 2008.
CASTOR status Presentation to LCG PEB 09/11/2004 Olof Bärring, CERN-IT.
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Pre-GDB Storage Classes summary of discussions Flavia Donno Pre-GDB.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
CERN SRM Development Benjamin Coutourier Shaun de Witt CHEP06 - Mumbai.
CERN IT Department CH-1211 Genève 23 Switzerland t Load Testing Dennis Waldron, CERN IT/DM/DA CASTOR Face-to-Face Meeting, Feb 19 th 2009.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
CMS Issues. Background – RAL Infrastructure TM Nsd Xrd- mgr TM Nsd Xrd- mgr TM Rhd stagerd TGW Rhd stagerd TGW Cupv Vmgr Vdqm nsd Cupv Vmgr Vdqm nsd Cupv.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
Your university or experiment logo here Future Disk-Only Storage Project Shaun de Witt GridPP Review 20-June-2012.
The CMS Top 5 Issues/Concerns wrt. WLCG services WLCG-MB April 3, 2007 Matthias Kasemann CERN/DESY.
Storage Interfaces Introduction Wahid Bhimji University of Edinburgh Based on previous discussions with Working Group: (Brian Bockelman, Simone Campana,
CERN IT Department CH-1211 Genève 23 Switzerland t HEPiX Conference, ASGC, Taiwan, Oct 20-24, 2008 The CASTOR SRM2 Interface Status and plans.
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
Storage Interfaces and Access pre-GDB Wahid Bhimji University of Edinburgh On behalf of all those who participated.
CASTOR project status CASTOR project status CERNIT-PDP/DM October 1999.
SRM-2 Road Map and CASTOR Certification Shaun de Witt 3/3/08.
ASGC Site Report Jason Shih ASGC Grid Ops CASTOR External Operation Face to Face Meeting.
CASTOR Status at RAL CASTOR External Operations Face To Face Meeting Bonny Strong 10 June 2008.
Patricia Méndez Lorenzo Status of the T0 services.
ASCC Site Report Eric Yen & Simon C. Lin Academia Sinica 20 July 2005.
CASTOR in SC Operational aspects Vladimír Bahyl CERN IT-FIO 3 2.
Bonny Strong RAL RAL CASTOR Update External Institutes Meeting Nov 2006 Bonny Strong, Tim Folkes, and Chris Kruk.
Improving Performance using the LINUX IO Scheduler Shaun de Witt STFC ISGC2016.
First Experiences with Ceph on the WLCG Grid Rob Appleyard Shaun de Witt, James Adams, Brian Davies.
An Analysis of Data Access Methods within WLCG Shaun de Witt, Andrew Lahiff (STFC)
Security recommendations DPM Jean-Philippe Baud CERN/IT.
Dissemination and User Feedback Castor deployment team Castor Readiness Review – June 2006.
Data & Storage Services CERN IT Department CH-1211 Genève 23 Switzerland t DSS CASTOR and EOS status and plans Giuseppe Lo Presti on behalf.
CASTOR at RAL in 2016 Rob Appleyard. Contents Current Status Staffing Upgrade plans Questions Conclusion.
Jean-Philippe Baud, IT-GD, CERN November 2007
Tom Byrne, Bruno Canning
CASTOR Giuseppe Lo Presti on behalf of the CASTOR dev team
BNL Tier1 Report Worker nodes Tier 1: added 88 Dell R430 nodes
IT-DB Physics Services Planning for LHC start-up
Service Challenge 3 CERN
Castor services at the Tier-0
Ákos Frohner EGEE'08 September 2008
RAL Tom Byrne George Vasilakakos, Bruno Canning, Alastair Dewhurst, Ian Johnson, Alison Packer.
The INFN Tier-1 Storage Implementation
CTA: CERN Tape Archive Overview and architecture
Data Management cluster summary
CASTOR: CERN’s data management system
Presentation transcript:

Future Plans at RAL Tier 1 Shaun de Witt

Introduction Current Set-Up Short term plans Final Configuration How we get there… How we plan/hope/pray to use CEPH

Current Infrastructure Common CMS ATLAS Gen LHcB Common Services (nsd, Cupv, vmgr, vdqm) Instances Disk Layer (Tape and Disk ‘Pools’) TapeLayer (at least 1 dedicated drive)

ATLAS Instance Exploded nsd01 nsd02 SRM01 SRM02 SRM03 SRM04 HeadNode01 RH Stager TapeGateway HeadNode01 RH Stager TapeGateway HeadNode02 TransferMgr HeadNode02 TransferMgr HeadNode03 TransferMgr NSD Xroot Mgr HeadNode03 TransferMgr NSD Xroot Mgr atlasTape atlasDataDisk atlasScratchDisk Xroot proxy x12 x1

Database Configuration Repack Nameserver Cupv … Repack Nameserver Cupv … CMS SRM CMS STGR CMS SRM CMS STGR LHCb SRM LHCb STGR LHCb SRM LHCb STGR ATLAS SRM ATLAS STGR Gen SRM Gen STGR Gen SRM Gen STGR Repack Nameserver Cupv … Repack Nameserver Cupv … CMS SRM CMS STGR CMS SRM CMS STGR LHCb SRM LHCb STGR LHCb SRM LHCb STGR ATLAS SRM ATLAS STGR Gen SRM Gen STGR Gen SRM Gen STGR Primaries Standbys DataGuard

Short Term Plans… Improve tape cache performance for ATLAS –Tape rates limited by disk –Currently heavy IO (read/write from grid/tape) –Currently configured with 10(7) ‘small’ server in RAID6 –Would RAID-1(0) help?

The Future Swift/S3 XROOT/ gridFTP CASTOR

What is Erasure-coded CEPH High-throughput Objectstore EC uses 16+3 ALL user data planned to use erasure coding (no replication) S3/SWIFT recommended interfaces –Xroot and gridFTP for legacy support … more later

The Plan… Move –The data Modify Merge

The ‘Plan’ – Phase 1 Current disk purchases are usable for CEPH and classic CASTOR Start moving atlasScratchDisk over to echo –Lifetime of files should be ~2 weeks –Allows us to monitor production use of echo Work with VOs to migrate relevant data using FTS Maintain classic CASTOR

The Plan – Phase 2 Once all disk-only (on echo?) –Consolidate to single castor instance With single shared diskpool Tape drive dedication…(Tim knows) Clear all diskcopies Single 2/3 node RAC + stdby –Common headnodes supporting all services –Maintain 3-4 SRMs Will probably be phased in

Accessing ECHO VOs use gridFTP and xroot ATM –Write using 1 protocol, read using either But not S3/SWIFT –Proposed gridFTP url (writing?) gsiftp://gateway.domain.name/ / Steers transfer to pool Certificate (VO and Role) for AA –Xroot URL As suggested by Seb… –But what about access?

The Known Unknowns S3/SWIFT interoperability? Will CASTOR/CEPH support EC pools? Partial writes verboten? Do users need to supply Support for CEPH plugins