NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.

Slides:



Advertisements
Similar presentations
Computing Infrastructure
Advertisements

What’s New: Windows Server 2012 R2 Tim Vander Kooi Systems Architect
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
RAID- Redundant Array of Inexpensive Drives. Purpose Provide faster data access and larger storage Provide data redundancy.
1 CSC 486/586 Network Storage. 2 Objectives Familiarization with network data storage technologies Understanding of RAID concepts and RAID levels Discuss.
CSE 486/586 CSE 486/586 Distributed Systems Case Study: Facebook f4 Steve Ko Computer Sciences and Engineering University at Buffalo.
RAID Technology. Use Arrays of Small Disks? 14” 10”5.25”3.5” Disk Array: 1 disk design Conventional: 4 disk designs Low End High End Katz and Patterson.
© 2007 DataDirect Networks, Inc. All Rights Reserved. Simple Petabyte-Level Scaling Industry Leading Performance SATAssure Data Integrity Power & Space.
Symantec De-Duplication Solutions Complete Protection for your Information Driven Enterprise Richard Hobkirk Sr. Pre-Sales Consultant.
1 Recap (RAID and Storage Architectures). 2 RAID To increase the availability and the performance (bandwidth) of a storage system, instead of a single.
High Performance Computing Course Notes High Performance Storage.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
How to Cluster both Servers and Storage W. Curtis Preston President The Storage Group.
CSE 451: Operating Systems Winter 2010 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura.
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
Data Storage Willis Kim 14 May Types of storages Direct Attached Storage – storage hardware that connects to a single server Direct Attached Storage.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Peter Stefan, NIIF 29 June, 2007, Amsterdam, The Netherlands NIIF Storage Services Collaboration on Storage Services.
Lustre at Dell Overview Jeffrey B. Layton, Ph.D. Dell HPC Solutions |
© 2013 Mellanox Technologies 1 NoSQL DB Benchmarking with high performance Networking solutions WBDB, Xian, July 2013.
Lecture 4 1 Reliability vs Availability Reliability: Is anything broken? Availability: Is the system still available to the user?
Two or more disks Capacity is the same as the total capacity of the drives in the array No fault tolerance-risk of data loss is proportional to the number.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
© Copyright 2010 Hewlett-Packard Development Company, L.P. 1 HP + DDN = A WINNING PARTNERSHIP Systems architected by HP and DDN Full storage hardware and.
Interposed Request Routing for Scalable Network Storage Darrell Anderson, Jeff Chase, and Amin Vahdat Department of Computer Science Duke University.
High Performance Computing G Burton – ICG – Oct12 – v1.1 1.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
Data Storage CPTE 433 John Beckett. The Paradox “If I can go to a computer store and buy 1000 gigabytes for $50, why does it cost more in your server.
1 - Q Copyright © 2006, Cluster File Systems, Inc. Lustre Networking with OFED Andreas Dilger Principal System Software Engineer
Hosted by Case Study - Storage Consolidation Steve Curry Yahoo Inc.
Small File File Systems USC Jim Pepin. Level Setting  Small files are ‘normal’ for lots of people Metadata substitute (lots of image data are done this.
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
Virtualization in the NCAR Mass Storage System Gene Harano National Center for Atmospheric Research Scientific Computing Division High Performance Systems.
Virtualization for Storage Efficiency and Centralized Management Genevieve Sullivan Hewlett-Packard
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
IMDGs An essential part of your architecture. About me
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
Switched Storage Architecture Benefits Computer Measurements Group November 14 th, 2002 Yves Coderre.
Large Scale Parallel File System and Cluster Management ICT, CAS.
SAN DIEGO SUPERCOMPUTER CENTER SDSC's Data Oasis Balanced performance and cost-effective Lustre file systems. Lustre User Group 2013 (LUG13) Rick Wagner.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
R. Scott Studham, Associate Director Advanced Computing April 13, 2004 HPC At PNNL March 2004.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
Infrastructure for Data Warehouses. Basics Of Data Access Data Store Machine Memory Buffer Memory Cache Data Store Buffer Bus Structure.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
BlueWaters Storage Solution Michelle Butler NCSA January 19, 2016.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
SATA In Enterprise Storage Ron Engelbrecht Vice President and General Manager Engineering and Manufacturing Operations September 21, 2004.
Data Evolution: 101. Parallel Filesystem vs Object Stores Amazon S3 CIFS NFS.
1 CEG 2400 Fall 2012 Network Servers. 2 Network Servers Critical Network servers – Contain redundant components Power supplies Fans Memory CPU Hard Drives.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
Presented by Robust Storage Management On Desktop, in Machine Room, and Beyond Xiaosong Ma Computer Science and Mathematics Oak Ridge National Laboratory.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
E VALUATION OF GLUSTER AT IHEP CHENG Yaodong CC/IHEP
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
Storage HDD, SSD and RAID.
Experience of Lustre at QMUL
RAID Redundant Arrays of Independent Disks
Steve Ko Computer Sciences and Engineering University at Buffalo
Steve Ko Computer Sciences and Engineering University at Buffalo
Experience of Lustre at a Tier-2 site
CS 554: Advanced Database System Notes 02: Hardware
JDAT Production Hardware
CSE 451: Operating Systems Winter 2009 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura 1.
Mark Zbikowski and Gary Kimura
CSE 451: Operating Systems Winter 2012 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura 1.
CS 295: Modern Systems Organizing Storage Devices
Presentation transcript:

NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham

2 Science Drivers Three different domains with different requirements High Performance Computing – Chemistry Low storage volumes (10 TB) High performance storage (>500MB/s per client, GB/s aggregate) POSIX access High Throughput Proteomics – Biology Large storage volumes (PB’s) and exploding Write once, read rarely if used as an archive Modes latency okay (<10s to data) If analysis could be done in place it would require faster storage Atmospheric Radiation Measurement - Climate Modest side storage requirements (100’s TB) Shared with community and replicated to ORNL

3 Overview The proteomics driven storage explosion is casing us to: Developing filesystems that enable lower cost hardware Continued write on fileserver failure (route around) Mirrored fileservers so we can use direct attached disk Increasing filesystem technology to meet scalability and performance metrics needed by the science 10,000+ clients accessing a POSIX 10+PB filesystem >500MB/s single client rate Add advanced technologies into the filesystem to increase performance and make it “smarter” Scalable content management Move the computation into the storage It must work in production (not a research project)

4 EMSL’s Current Storage Strategy EMSL’s Storage Strategy has focused on capacity We want to be here Our storage sales rep want us here We use tools like Lustre to help us bridge this gap. Estimated $/TB as a function of time and technology

5 EMSL’s Current Storage Strategy Developing filesystems that enable lower cost hardware Our experience has shown that expensive disks fail about as often as cheap disks. We have a large sampling of disks: 1,000 FC-SAN drives to make a 53TB filesystem 20% duty cycle – The drives don’t fail much (1-3 disks per month) Entire filesystem (all 1,000 drives) down once every two months. Mostly due to vendor required firmware updates to SAN switches or hardware failures. 7,500 SCSI drives to provide ½ PB of scratch space 100% duty cycle. Average ~3 disk failures per day (should be 0.5 per day). Experiencing bugs in the Seagate disks 1,000 ATA/SAN to provide 200TB archive 10% duty cycle. Average 1-3 disk failures per month

6 NWfs Hardware Low Cost, High Performance Storage We have replaced all our tapes with low-cost ATA storage. NWfs Project: Includes; Lustre, Cluster mgt tools, minor Metadata capturing tools and a custom client side GUI to support gridFTP, striped and parallel data transfers. Linux-based OSTs Containing: 2 CPU’s & RAM Multiple 3Ware ATA RAID Adapters 16 SATA Disk Drives “Hot-Swap” RAID5 with multiple hot spares per node. $3.5K/TB after RAID5 Infiniband 4X backbone New SATA drives include rotational vibration safeguard EMSL’s Current Storage Strategy 400TB ≈ $1.5M

7 Increasing filesystem technology to meet scalability and performance metrics needed by the science Lustre has been in full production since last Aug and used for aggressive IO from our supercomputer. Highly stable Still hard to manage We are expanding our use of Lustre to act as the filesystem for our archival storage. Deploying a ~400TB filesystem 660MB/s from a single client with a simple “dd” is faster than any local or global filesystem we have tested. We are finally in the era where global filesystems provide faster access

8 EMSL’s Current Storage Strategy Scalable Content Management Storage Pool1 Remote Storage Pool2 Client Remote Index Index3Index3 Index2Index2 MetaData Server Cluster MetaData

9 EMSL’s Current Storage Strategy Looks a lot like Lustre OSTOST Client Index3Index3 Index2Index2 MDSMDS

10 EMSL’s Current Storage Strategy Add replication to support DAS & collaboration OST Client Remote Index Index3 Index2 MDS

11 Active Storage Moving the computation into the storage rather than moving the data to the compute power. Data StreamParallel file systemReassemble & post process Data StreamPost process in object based parallel file system Classic parallel file systems stripe at the block level. This requires the distributed data to be reassembled in order to post process PNNL is developing code that will allow post processing to be performed on objects inside the file system and make use of the computational power on the file servers. Classical Storage Active Storage Demonstrated 1.3GB/s FT stream EMSL’s Current Storage Strategy

12 EMSL’s Current Storage Strategy NWFS V3.0 Lustre with replication, Content Mgt, Active StorageOSTOST OST Client Remote Index Index3Index3 Index2Index2 MDSMDS API