KIT – University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association STEINBUCH CENTRE FOR COMPUTING - SCC www.kit.edu.

Slides:



Advertisements
Similar presentations
STEINBUCH CENTRE FOR COMPUTING - SCC KIT – University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association.
Advertisements

Proposed Storage Area Network Facilities For Discussion.
Netbus: A Transparent Mechanism for Remote Device Access in Virtualized Systems Sanjay Kumar PhD Student Advisor: Prof. Karsten Schwan.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Jos van Wezel Doris Ressmann GridKa, Karlsruhe TSM as tape storage backend for disk pool managers.
Introduction to Storage Area Network (SAN) Jie Feng Winter 2001.
System Area Network Abhiram Shandilya 12/06/01. Overview Introduction to System Area Networks SAN Design and Examples SAN Applications.
Brocade VDX 6746 switch module for Hitachi Cb500
“It’s going to take a month to get a proof of concept going.” “I know VMM, but don’t know how it works with SPF and the Portal” “I know Azure, but.
Silicon Graphics, Inc. Cracow ‘03 Grid Workshop SAN over WAN - a new way of solving the GRID data access bottleneck Dr. Wolfgang Mertz Business Development.
Storage area Network(SANs) Topics of presentation
SQL Server, Storage And You Part 2: SAN, NAS and IP Storage.
Storage Task Force Intermediate pre report. History GridKa Technical advisory board needs storage numbers: Assemble a team of experts. 04/05 At HEPiX.
COEN 180 NAS / SAN. NAS Network Attached Storage (NAS) Each storage device has its own network interface. Filers: storage device that interfaces at the.
SAN Design Considerations Hylton Leigh Senior Consultant Novell Consulting, UK Stuart Thompson Senior Consultant Novell Consulting, UK.
Electronics Confidential Information © 3M 2005, All Rights Reserved Enterprise Computing Applications Charlie Staley January 2007.
Emerging Storage Options for Server Blade Architectures Server Blade Summit 2005.
COEN 180 NAS / SAN. Storage Trends Storage Trends: Money is spend on administration Morris, Truskowski: The evolution of storage systems, IBM Systems.
Product Manager Networking Infrastructure Choices for Storage.
National Energy Research Scientific Computing Center (NERSC) The GUPFS Project at NERSC GUPFS Team NERSC Center Division, LBNL November 2003.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
5.3 HS23 Blade Server. The HS23 blade server is a dual CPU socket blade running Intel´s new Xeon® processor, the E5-2600, and is the first IBM BladeCenter.
Storage Area Networks The Basics. Storage Area Networks SANS are designed to give you: More disk space Multiple server access to a single disk pool Better.
Storage Survey and Recent Acquisition at LAL Michel Jouvin LAL / IN2P3
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Windows Server 2012 VSP Windows Kernel Applications Non-Hypervisor Aware OS Windows Server 2008, 2012 Windows Kernel VSC VMBus Emulation “Designed for.
HEPiX Storage Task Force Roger Jones Lancaster CHEP06, Mumbai, February 2006.
CASPUR SAN News Andrei Maslennikov Orsay, April 2001.
School of EECS, Peking University Microsoft Research Asia UStore: A Low Cost Cold and Archival Data Storage System for Data Centers Quanlu Zhang †, Yafei.
Storage Systems Market Analysis Dec 04. Storage Market & Technologies.
InfiniSwitch Company Confidential. 2 InfiniSwitch Agenda InfiniBand Overview Company Overview Product Strategy Q&A.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Storage Networking Evolution Jim Morin VP Strategic Planning June 2001.
Virtualization for the LHCb Online system CHEP Taipei Dedicato a Zio Renato Enrico Bonaccorsi, (CERN)
Using NAS as a Gateway to SAN Dave Rosenberg Hewlett-Packard Company th Street SW Loveland, CO 80537
Storage, Networks, Data Management Report on Parallel Session OSG Meet 8/2006 Frank Würthwein (UCSD)
KIT – The cooperation of Forschungszentrum Karlsruhe GmbH und Universität Karlsruhe (TH) Hadoop on HEPiX storage test bed at FZK Artem Trunov.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Mr. P. K. GuptaSandeep Gupta Roopak Agarwal
Tier-2 storage A hardware view. HEP Storage dCache –needs feed and care although setup is now easier. DPM –easier to deploy xrootd (as system) is also.
Architecting Private Clouds Using Windows Server 2012
Deployment options for Fluid Cache for SAN with VMware
Internet Protocol Storage Area Networks (IP SAN)
SMUCSE 8344 Storage Area Networks. SMUCSE 8344 What’s SANs.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
© 2007 EMC Corporation. All rights reserved. Internet Protocol Storage Area Networks (IP SAN) Module 3.4.
Status: Central Storage Services CD/LSC/CSI/CSG June 26, 2007.
By Harshal Ghule Guided by Mrs. Anita Mahajan G.H.Raisoni Institute Of Engineering And Technology.
New directions in storage | ISGC 2015, Taipei | Patrick Fuhrmann | 19 March 2015 | 1 Presenter: Patrick Fuhrmann dCache.org Patrick Fuhrmann, Paul Millar,
Centre de Calcul de l’Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief.
GPFS Parallel File System
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Validation tests of CNAF storage infrastructure Luca dell’Agnello INFN-CNAF.
KIT - University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association Xrootd SE deployment at GridKa WLCG.
Ryan Leonard Storage and Solutions Architect
Storage Area Networks The Basics.
Video Security Design Workshop:
Experience of Lustre at QMUL
Enterprise Computing Applications
NL Service Challenge Plans
Kako sa novim tehnologijama graditi datacentar budućnosti?
Experience of Lustre at a Tier-2 site
Luca dell’Agnello INFN-CNAF
Introduction to Networks
Oracle Storage Performance Studies
Co-designed Virtual Machines for Reliable Computer Systems
Cost Effective Network Storage Solutions
Presentation transcript:

KIT – University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association STEINBUCH CENTRE FOR COMPUTING - SCC Preview of a Novel Architecture for Large Scale Storage Andreas Petzold, Christoph-Erdmann Pfeiler, Jos van Wezel Steinbuch Centre for Computing

2 Storage Management Systems at GridKa dCache for ATLAS, CMS, LHCb 6 PB disk-only 3 PB tape-buffers 287 pools on 58 servers Agnostic to underlying storage technology Scalla xrootd for ALICE 2.7 PB disk-only/tape buffer 15 servers Agnostic to underlying storage technology CHEP 2013, Amsterdam SRM Storage Management System File System Linux Storage Controller Disks

3 Current GridKa Disk Storage Technologies 9 x DDN S2AA enclosures 9000 disks 796 LUNs SAN Brocade DCX 1 x DDN SFA10K 10 enclosures 600 disks 1 x DDN SFA12K 5 enclosures 360 disks CHEP 2013, Amsterdam

4 Current GridKa Tape Storage Technologies 2 Oracle/Sun/STK SL x slots 22 LTO5, 16 LTO4 1 IBM TS slots 24 LTO4 1 GRAU XL 5376 slots 16 LTO3, 8 LTO4 CHEP 2013, Amsterdam

5 GridKa Storage Overview CHEP 2013, Amsterdam gateway nodes Compute nodes storage nodes

6 GridKa Storage Units CHEP 2013, Amsterdam Servers connect directly to storage or via SAN but not explicitly needed Cluster Filesystem (GPFS) connects several servers filesystem visible on all nodes predictable IO throughput nice storage virtualisation layer Currently evaluating alternative filesystems XFS, ZFS, BTRFS, EXT4 10G Ethernet Fibre Channel

7 Novel Storage Solutions for GridKa Expect large resource increase in 2015 Chance to look at new solutions during LHC LS1 Simplification in operations and deployment required Solution 1: DataDirectNetworks SFA12K-E Server VMs run embedded in storage controller Solution 2: Rausch Netzwerktechnik BigFoot More conventional setup; server directly connected to local disks CHEP 2013, Amsterdam

8 Shortening Long IO Paths CHEP 2013, Amsterdam Worker node NIC Switch Router Server HBA SAN Switch Storage Controller disk SAS/FC fabric application begin { do_io } end application begin { do_io } end Worker node NIC Switch Router Server disk SAS/FC fabric application begin { do_io } end application begin { do_io } end Traditional storage at GridKa Previewed storage at GridKa disk Worker node NIC Switch Router Storage Controller disk SAS/FC fabric application begin { do_io } end application begin { do_io } end disk BigFoot DDN SFA12K-E Worker node NIC Switch Router Server IB HCA IB HCA Storage Controller disk SAS/FC fabric application begin { do_io } end application begin { do_io } end disk IB HCA IB HCA potentially redundant components network DDN SFA12K

9 DDN SFA E Architecture CHEP 2013, Amsterdam 10 GE NIC dedicated to VM 8 VMs per SFA pair DDN VM driver dma access to storage Worker node SFA OS Cache sfablkdriver SAS Switch VM1 xrootd VM1 xrootd VM2 VM3 VM4 IO bridge IO bridge IO bridge IO bridge IO bridge IO bridge IO bridge IO bridge NIC SSD SFA OS Cache sfablkdriver SAS Switch VM1 VM2 VM3 VM4 IO bridge IO bridge IO bridge IO bridge IO bridge IO bridge IO bridge IO bridge SSD

10 BigFoot Architecture CHEP 2013, Amsterdam 10 GE NIC Worker node SAS HBA/RAID Controller OS Cache SSD NIC

11 Thoughts on Effect of SAN and Long Distance Latency CHEP 2013, Amsterdam

12 Thoughts on Effect of SAN and Long Distance Latency CHEP 2013, Amsterdam

13 Expected Benefits Reduced Latency HBA and SAN: 2.1 μs (e.g. Brocade DCX, blade to blade) storage controller Improved IO rates Reduced power Server and controller HBA, SAN Switch ~ W = ~600Euro/server/year Reduced investment €/HBA, €/switch port = € Improved MTBF Less components CHEP 2013, Amsterdam

14 Possible Drawbacks Loss in flexibility w/o SAN storage building blocks are larger Limited server access to storage blocks Storage systems are only connected via LAN VMs inside storage controller (DDN SFA12K-E) Competition for resources Limited number of VMs limits “server/TB” ratio Loss of redundancy Simple server attached storage (BigFoot) Limited by simple hardware controller HW admin doesn’t scale to 100s of boxes No redundancy CHEP 2013, Amsterdam

15 Glimpse at Performance Preliminary performance evaluation IOZONE testing parallel threads on XFS filesystem Xrootd data server in VM, performance similar to IOZONE Out-of-the-box settings, no tuning Performance below expectations, reasons still to be understood ZFS tested on BigFoot CHEP 2013, Amsterdam IOZONE

16 Conclusions Storage extension at GridKa requires expensive upgrade of GridKa disk SAN – or novel storage solution Tight integration of server and storage looks promising Many possible benefits – further evaluation required Less components Less power consumption Less complexity Performance needs to be understood together with vendors More tests with other vendors in near future CHEP 2013, Amsterdam

17 CHEP 2013, Amsterdam