Small File File Systems USC Jim Pepin. Level Setting  Small files are ‘normal’ for lots of people Metadata substitute (lots of image data are done this.

Slides:



Advertisements
Similar presentations
Computing Infrastructure
Advertisements

Data Storage & Security Dr Alastair F. Brown Head of Computing MRC Human Genetics Unit MRC Institute of Genetics and Molecular Medicine The University.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
Orin
2013 Summer Institute: Discover Big Data, August 5-9, San Diego, California SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Dealing.
HPC USER FORUM I/O PANEL April 2009 Roanoke, VA Panel questions: 1 response per question Limit length to 1 slide.
Chapter 5: Server Hardware and Availability. Hardware Reliability and LAN The more reliable a component, the more expensive it is. Server hardware is.
CSE 486/586 CSE 486/586 Distributed Systems Case Study: Facebook f4 Steve Ko Computer Sciences and Engineering University at Buffalo.
RAID CS5493/7493. RAID : What is it? Redundant Array of Independent Disks configured into a single logical storage unit.
The Stanford Directory Architecture for Shared Memory (DASH)* Presented by: Michael Bauer ECE 259/CPS 221 Spring Semester 2008 Dr. Lebeck * Based on “The.
A match made in heaven?. Who am I? Richard Barlow Systems Architect and Engineering Manager for the Virginia Credit Union Worked in IT for almost 20 years.
Scale-out Central Store. Conventional Storage Verses Scale Out Clustered Storage Conventional Storage Scale Out Clustered Storage Faster……………………………………………….
HELICS Petteri Johansson & Ilkka Uuhiniemi. HELICS COW –AMD Athlon MP 1.4Ghz –512 (2 in same computing node) –35 at top500.org –Linpack Benchmark 825.
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
Sun FIRE Jani Raitavuo Niko Ronkainen. Sun FIRE 15K Most powerful and scalable Up to 106 processors, 576 GB memory and 250 TB online disk storage Fireplane.
CSE 451: Operating Systems Winter 2010 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura.
Servers Redundant Array of Inexpensive Disks (RAID) –A group of hard disks is called a disk array FIGURE Server with redundant NICs.
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
Data Storage Willis Kim 14 May Types of storages Direct Attached Storage – storage hardware that connects to a single server Direct Attached Storage.
Peter Stefan, NIIF 29 June, 2007, Amsterdam, The Netherlands NIIF Storage Services Collaboration on Storage Services.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
BACKUP/MASTER: Immediate Relief with Disk Backup Presented by W. Curtis Preston VP, Service Development GlassHouse Technologies, Inc.
Lustre at Dell Overview Jeffrey B. Layton, Ph.D. Dell HPC Solutions |
Storage Area Networks The Basics. Storage Area Networks SANS are designed to give you: More disk space Multiple server access to a single disk pool Better.
© 2013 Mellanox Technologies 1 NoSQL DB Benchmarking with high performance Networking solutions WBDB, Xian, July 2013.
SANPoint Foundation Suite HA Robert Soderbery Sr. Director, Product Management VERITAS Software Corporation.
Interposed Request Routing for Scalable Network Storage Darrell Anderson, Jeff Chase, and Amin Vahdat Department of Computer Science Duke University.
pNFS extension for NFSv4 IETF 61 November, 2004
DMF Configuration for JCU HPC Dr. Wayne Mallett Systems Manager James Cook University.
Computers in the real world Objectives Explain the need for secondary storage devices Understand the three main storage types – Optical – Magnetic – Solid.
Ceph Storage in OpenStack Part 2 openstack-ch,
HPC USER FORUM I/O PANEL April 2009 Roanoke, VA Panel questions: 1 response per question Limit length to 1 slide.
Latest Relevant Techniques and Applications for Distributed File Systems Ela Sharda
School of EECS, Peking University Microsoft Research Asia UStore: A Low Cost Cold and Archival Data Storage System for Data Centers Quanlu Zhang †, Yafei.
Virtualization for Storage Efficiency and Centralized Management Genevieve Sullivan Hewlett-Packard
Confidential1 Introducing the Next Generation of Enterprise Protection Storage Enterprise Scalability Enhancements.
1 Moshe Shadmon ScaleDB Scaling MySQL in the Cloud.
Amy Apon, Pawel Wolinski, Dennis Reed Greg Amerson, Prathima Gorjala University of Arkansas Commercial Applications of High Performance Computing Massive.
HPC system for Meteorological research at HUS Meeting the challenges Nguyen Trung Kien Hanoi University of Science Melbourne, December 11 th, 2012 High.
Storage Networking Evolution Jim Morin VP Strategic Planning June 2001.
Public Relations Interim Image Archive Goal: Provide and INTERIM image archive solution for Public Relations 2 to 4 TB of images currently spread across.
March, 2003 SOS 7 Jim Harrell Unlimited Scale Inc.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
CASPUR Site Report Andrei Maslennikov Lead - Systems Rome, April 2006.
HADOOP DISTRIBUTED FILE SYSTEM HDFS Reliability Based on “The Hadoop Distributed File System” K. Shvachko et al., MSST 2010 Michael Tsitrin 26/05/13.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Scott Koranda, UWM & NCSA 14 January 2016www.griphyn.org Lightweight Data Replicator Scott Koranda University of Wisconsin-Milwaukee & National Center.
 Introduction  Architecture NameNode, DataNodes, HDFS Client, CheckpointNode, BackupNode, Snapshots  File I/O Operations and Replica Management File.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
Parallel IO for Cluster Computing Tran, Van Hoai.
Tackling I/O Issues 1 David Race 16 March 2010.
Architecture of a platform for innovation and research Erik Deumens – University of Florida SC15 – Austin – Nov 17, 2015.
Truly Distributed File Systems Paul Timmins CS 535.
Predrag Buncic CERN Data management in Run3. Roles of Tiers in Run 3 Predrag Buncic 2 ALICEALICE ALICE Offline Week, 01/04/2016 Reconstruction Calibration.
GPFS Parallel File System
29/04/2008ALICE-FAIR Computing Meeting1 Resulting Figures of Performance Tests on I/O Intensive ALICE Analysis Jobs.
Canadian Bioinformatics Workshops
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
Scott Koranda, UWM & NCSA 20 November 2016www.griphyn.org Lightweight Replication of Heavyweight Data Scott Koranda University of Wisconsin-Milwaukee &
Storage Area Networks The Basics.
Integrating Disk into Backup for Faster Restores
Steve Ko Computer Sciences and Engineering University at Buffalo
CERN Lustre Evaluation and Storage Outlook
Lesson Objectives Aims You should be able to:
Clouds & Containers: Case Studies for Big Data
CSE 451: Operating Systems Winter 2009 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura 1.
CSE 451: Operating Systems Winter 2012 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura 1.
Hard disk basics Prof:R.CHARLES SILVESTER JOE Departmet of Electronics St.Joseph’s College,Trichy.
Presentation transcript:

Small File File Systems USC Jim Pepin

Level Setting  Small files are ‘normal’ for lots of people Metadata substitute (lots of image data are done this way)  Comes from ‘pc’/desktop world  These users have discovered ‘hpc’ but don’t want to change programs (not even MPI)  Find ways to help (best is ‘rewrite’ but that is not reasonable to expect)  Small files are deadly to most file systems Some more than others Impact of ‘custer’ systems

Level Setting  Disks Sata  Not fast.  Reliability issues  Cheap Fast disk (15k etc)  Not cheap  Fast People are looking at ‘cheap’  Drives better backup/maintainability solutions Distributed doesn’t mean ‘faster’ Virtualization can be your enemy (in some ways)

Basics  1800 node cluster Presents special problems  Myrinet ‘interconnect’  Ethernet (gb) data plane  Fiber channel disk/tape data plane (2Gb/s) 256+ disk/tape devices  15+ file servers  250+ TB disk  Tape Backup DR site

Basics  QFS base FS Archiving and distributed access Sun thing  Local parallel FS on nodes  NFS Issues around it  “Condo” disk versus Condo nodes

Basics  Three types of File systems  Parallel FS on compute nodes (temp) Exception on ‘condo’ nodes  Small files More directory transactions Small frames win No stripes  Large files More data transactions Jumbo frames win Stripes win  Tuning is stripe factors and blocksizes

Small Files  Examples  Genomics Group 10ks of files in single directory  Natural Language Group k files in directory Many nodes accessing same stuff  Dictionaries  Backups are ‘slower’ / ‘harder’ Reasons Updating directory data Blocking of data on tape

Small Files  Ways to help “Faster” disk (helps metadata/directory space) Distributed file access (qfs)  Metadata still a ‘block’. Read/write locks Updating for distributed access Next version scales better (lock improvements) No free lunch Special Purpose File Systems and/or local space on cluster nodes (replication)

Next generation  Why change needed NFS doesn’t cut it  Why  GPFS Helps some  10Gb hosts on ‘data plane’ Next month  Ram disk for ‘metadata’?

Next generation  Storage management solutions SRB and friends Database based solutions Lustre possible Object storage Performance for small files/objects is question in my mind  All these have potential but… Back to don’t change code “Virtualization” conundrum  How to build massively parallel data spaces HPCS/other projects