POW : System optimisations for data management 11 November 2004.

Slides:



Advertisements
Similar presentations
IBM Software Group ® Integrated Server and Virtual Storage Management an IT Optimization Infrastructure Solution from IBM Small and Medium Business Software.
Advertisements

XenData SX-10 LTO Archive Appliance An Archive Appliance based on IT standards, designed for the demanding requirements of the media and entertainment.
Virtualisation From the Bottom Up From storage to application.
Exadata Distinctives Brown Bag New features for tuning Oracle database applications.
MUNIS Platform Migration Project WELCOME. Agenda Introductions Tyler Cloud Overview Munis New Features Questions.
Introduction to DBA.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Vorlesung Speichernetzwerke Teil 2 Dipl. – Ing. (BA) Ingo Fuchs 2003.
Novell Server Linux vs. windows server 2008 By: Gabe Miller.
Storage area Network(SANs) Topics of presentation
SQL Server, Storage And You Part 2: SAN, NAS and IP Storage.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
© 2008 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice HP StorageWorks LeftHand update Marcus.
PetaByte Storage Facility at RHIC Razvan Popescu - Brookhaven National Laboratory.
© 2006 EMC Corporation. All rights reserved. Managing the Data Center Section 5.2.
Data Storage Willis Kim 14 May Types of storages Direct Attached Storage – storage hardware that connects to a single server Direct Attached Storage.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Backup Rationalisation Reorganisation of the CERN Computer Centre Backups David Asbury IT/DS Friday 6 December 2002.
Frangipani: A Scalable Distributed File System C. A. Thekkath, T. Mann, and E. K. Lee Systems Research Center Digital Equipment Corporation.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Storage Area Networks The Basics. Storage Area Networks SANS are designed to give you: More disk space Multiple server access to a single disk pool Better.
HEPIX 3 November 2000 Current Mass Storage Status/Plans at CERN 1 HEPIX 3 November 2000 H.Renshall PDP/IT.
Parallel Computing The Bad News –Hardware is not getting faster fast enough –Too many architectures –Existing architectures are too specific –Programs.
The Mass Storage System at JLAB - Today and Tomorrow Andy Kowalski.
DMF Configuration for JCU HPC Dr. Wayne Mallett Systems Manager James Cook University.
Online Systems Status Review of requirements System configuration Current acquisitions Next steps... Upgrade Meeting 4-Sep-1997 Stu Fuess.
Best Practices for Backup in SAN/NAS Environments Jeff Wells.
STEALTH Content Store for SharePoint using Caringo CAStor  Boosting your SharePoint to the MAX! "Optimizing your Business behind the scenes"
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
Planning and Designing Server Virtualisation.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
1 Computer and Network Bottlenecks Author: Rodger Burgess 27th October 2008 © Copyright reserved.
IT253: Computer Organization
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
20-22 September 1999 HPSS User Forum, Santa Fe CERN IT/PDP 1 History  Test system HPSS 3.2 installation in Oct 1997 IBM AIX machines with IBM 3590 drives.
MOUNT10: Company, Products and Solutions ABAKUS Distribution, a.s. Jaroslav Techl
Integrating JASMine and Auger Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
Large Scale Parallel File System and Cluster Management ICT, CAS.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
Clustering In A SAN For High Availability Steve Dalton, President and CEO Gadzoox Networks September 2002.
CASTOR: CERN’s data management system CHEP03 25/3/2003 Ben Couturier, Jean-Damien Durand, Olof Bärring CERN.
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
CASPUR Site Report Andrei Maslennikov Lead - Systems Amsterdam, May 2003.
 The End to the Means › (According to IBM ) › 03.ibm.com/innovation/us/thesmartercity/in dex_flash.html?cmp=blank&cm=v&csr=chap ter_edu&cr=youtube&ct=usbrv111&cn=agus.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
Building and managing production bioclusters Chris Dagdigian BIOSILICO Vol2, No. 5 September 2004 Ankur Dhanik.
Mass Storage at SARA Peter Michielse (NCF) Mark van de Sanden, Ron Trompert (SARA) GDB – CERN – January 12, 2005.
01. December 2004Bernd Panzer-Steindel, CERN/IT1 Tape Storage Issues Bernd Panzer-Steindel LCG Fabric Area Manager CERN/IT.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
26. Juni 2003Bernd Panzer-Steindel, CERN/IT1 LHC Computing re-costing for for the CERN T0/T1 center.
CDP Technology Comparison CONFIDENTIAL DO NOT REDISTRIBUTE.
GDB Meeting 12. January Bernd Panzer-Steindel, CERN/IT 1 Mass Storage at CERN GDB meeting, 12. January 2005.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
XenData SX-10 LTO Archive Appliance
Open-E Data Storage Software (DSS V6)
Managing Explosive Data Growth
Chapter 10: Mass-Storage Systems
Video Security Design Workshop:
Robotics and Tape Drives
NL Service Challenge Plans
Chapter 12: Mass-Storage Structure
Chapter III, Desktop Imaging Systems and Issues: Lesson III Moving Image Data
TYPES OFF OPERATING SYSTEM
Chapter 1: Introduction
XenData SX-550 LTO Archive Servers
CASTOR: CERN’s data management system
Chapter 2: Operating-System Structures
Chapter 2: Operating-System Structures
Presentation transcript:

POW : System optimisations for data management 11 November 2004

It says ‘probable user error..’

Before optimising, you need a workable foundation Hardware: optimisation of use feasible –May mean virtualisation is soon needed Software: optimisation is possible, –Timescales long and rather unpredictable –Undermanning a principal cause User applications: who knows?

Plenty of previous experience…mostly bad choices IBM 3495, (no SCSI) TL820, (too small) IBM 3494, (too small) GRAU ‘J’ multimedia test, 98 ‘don’t need it’ Sony Petasite test, 98 (wrong drive)

Current usage of devices is inefficient You have seen JvE’s evidence Write: ~75% effective duty cycle feasible, because –? s drive assigned but waiting for robotics –30 s pick/place in drive –45 s to first byte –10 GB migration, 300 s writing –45 s rewind/unload –? s wait for robotics to dismount and free drive –Drive reassigned to next task Read: ~20% or less effective duty cycle seen –As above, but typically read 1 file of < 1 GB –120 s overhead, ~30 s reading Savage marshalling of reads? Faster robotics?

Easy things Current devices could do (part/most of ) the job –4 GBytes/s write is ~ B, ~6 MF once –15 PB/yr is ~15 Powderhorns/yr, ~4 MF/yr ~90 K vols/yr, ~9 MF/yr –Reading, as done today, 4 GBytes/s ~500 drives, ~25 MF once –Need more space (2 new buildings?) and ~30 MF –Need ~15 MF/yr

Hard things All current drives will be old, by 2007 –IBM 3592, 300GB 40 Mbytes/s, out already 1 yr –STK 9940B, 200GB 30 Mbytes/s, successor 2005? All current libraries will be old, too –IBM 3584, ~5 K vols, maybe 100 EPH –STK Powderhorn, 5.5 K vols 190 EPH IBM 3584 obsolete (slow, no passthrough..)

Major uncertainties Will it soon all be on disk? ~15 PB/yr? ~ 15 MF? –If so, use drives designed for backup/recall: LTO x –LTO use means must replace Powderhorns Or, will tape reads be still very important, if not the predominant load? –If so, need very many drives: ~ What high end commercial devices will be available then? This is NOT commodity, no mass market… –Will HP, IBM, STK or Quantum still be making drives? –What sort? –All directed to corporate backup/recall –All seem designed to be used via FC fabrics, virtualisation….

Is the current model tenable? Three layers of ‘separated functions’…. CPU servers, ~10K Disk servers, ~1 K at ~n TB each –Tape servers, ~n x 100 at ~ MB/s –Tape devices, ~n x 100 at ~ MB/s –Tape libraries, ~2, ~ 10 year lifetime –Tapes, ~30 K vols of ~500 GB/yr, ~3-5 year lifetime – Current disk servers cannot really drive tape units capable of >100 Mbyte/s Current GBit networking and FC HBAs can’t handle this either Maybe obliged to use a virtualised tape layer?

Sensible to have a flexible start point Install at least two suppliers on the floor This means in practice IBM and STK –Both offer ‘their own brand’ and at least LTO as well This means at least 2 types of library –ADIC (multimedia, ‘Scalar 1000’, bought out GRAU) too expensive?? –IBM library offering no use, but can’t be avoided presently –STK SL8500 does not offer IBM 3592, but is SDLT/LTO compatible, very fast and very extensible And then there is virtualisation… this should be tried

Why virtualise? Backup notoriously does not share resources and uses a large amount of these: virtualising the resources used solves this for good. Today backup claims –2 silos of 3590, E, B –1 STK Timberwolf 9710 library, 7 DLT7000s, –CASTOR-based AFS backup –Soon the L700 (in Prevessin)….. User applications and ‘user’ disk servers are not very efficient, and would struggle to use high performance devices There are products that can virtualise all these libraries and drives, and the real resources are shared (STK VMS Open, ‘Indigo’…) The real drives would be exploited at maximum capability, and.. –No problematic tape driver maintenance or performance mods needed –No high speed disk server to drive attach is needed –An efficient hidden disk buffer resides between user and real drives –Media and drive migration not a user (ie CASTOR) concern –Multicopy, remote vaulting, ‘policies’ are offered STK presentation exceprt……

A New Foundation for our Data Storage “Services” Common Platform –As much “off the shelf” technology as practical I/F via generally available HBAs Intel processors, GigE interconnect (mesh) –StorageTek specific value add Cluster optimized for high data throughput and failover High performance write biased disk buffer Software –Linux based –StorageTek IP Journaling, virtual tape/library, data lifecycle and policy management STK

New Storage “Services” Buffer 1 – 12 Node Pairs Common Platform Virtual Tape “Indigo Project” Virtual Tape “Indigo Project” Continuous Protect Continuous Protect Library Management StreamLine Library Manager Library Management StreamLine Library Manager Future Data Storage Services Appliance Buffer STK

Common Platform Node Architecture 4x Gigabit Ethernet Disk Cont- roller ~600 MB/s Data Mover Data Mover SATA disk buffer Up to ~45 TB 2x 2Gb FC Customer Tape applications STK

Scalability via the Common Platform Tape drives Disk Client System Client System Disk Gigabit Switches (Mesh) Node pair STK

Hardware Configurations 1-node2-node (node pair) 4-node (dual node pair) 8-node (quad-node pair) 24-node (12 node pairs) Throughput max~300 MB/s (~1.2 TB/hr) 600 MB/s (~2 TB/hr) 1200 MB/s (~4 TB/hr) 2400 MB/s (>8 TB/hr) >7200 MB/s (>25 TB/hr) Front end FC ports Max virtual devices ,576 B.E. disk FC ports B.E. tape FC ports Minimum usable disk buffer ~6 TB ~12 TB~24 TB~72 TB Maximum usable disk buffer * ~45 TB ~90 TB~180 TB~540 TB RAS99.9%99.999% STK

So next… Will we get an SL8500? Will we get enough equipment to use the IBM 3592s, after initial test? –8 drives (if bought after ‘test’) vs B –Must provide ~10% of workload support –This means ~2,500 cartridges (~0.5 MF) –This means a much larger 3584, or an IBM maintained and separate Powderhorn…..

Another project, which turned out unexpectedly..