16 th May 2006Alessandra Forti Storage Alessandra Forti Group seminar 16th May 2006.

Slides:



Advertisements
Similar presentations
By Rashid Khan Lesson 6-A Place for Everything: Storage Management.
Advertisements

RAID (Redundant Arrays of Independent Disks). Disk organization technique that manages a large number of disks, providing a view of a single disk of High.
RAID Oh yes Whats RAID? Redundant Array (of) Independent Disks. A scheme involving multiple disks which replicates data across multiple drives. Methods.
NAS vs. SAN 10/2010 Palestinian Land Authority IT Department By Nahreen Ameen 1.
High Performance Computing Course Notes Grid Computing.
“Redundant Array of Inexpensive Disks”. CONTENTS Storage devices. Optical drives. Floppy disk. Hard disk. Components of Hard disks. RAID technology. Levels.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
RAID- Redundant Array of Inexpensive Drives. Purpose Provide faster data access and larger storage Provide data redundancy.
1 CSC 486/586 Network Storage. 2 Objectives Familiarization with network data storage technologies Understanding of RAID concepts and RAID levels Discuss.
R.A.I.D. Copyright © 2005 by James Hug Redundant Array of Independent (or Inexpensive) Disks.
Fault tolerance and disaster recovery
REDUNDANT ARRAY OF INEXPENSIVE DISCS RAID. What is RAID ? RAID is an acronym for Redundant Array of Independent Drives (or Disks), also known as Redundant.
Distributed components
High Performance Computing Course Notes High Performance Storage.
Oxford Jan 2005 RAL Computing 1 RAL Computing Implementing the computing model: SAM and the Grid Nick West.
I/O Systems and Storage Systems May 22, 2000 Instructor: Gary Kimura.
CSE 451: Operating Systems Winter 2010 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura.
Servers Redundant Array of Inexpensive Disks (RAID) –A group of hard disks is called a disk array FIGURE Server with redundant NICs.
Session 3 Windows Platform Dina Alkhoudari. Learning Objectives Understanding Server Storage Technologies Direct Attached Storage DAS Network-Attached.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
ICOM 6005 – Database Management Systems Design Dr. Manuel Rodríguez-Martínez Electrical and Computer Engineering Department Lecture 6 – RAID ©Manuel Rodriguez.
Configuring File Services Lesson 6. Skills Matrix Technology SkillObjective DomainObjective # Configuring a File ServerConfigure a file server4.1 Using.
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
Day 10 Hardware Fault Tolerance RAID. High availability All servers should be on UPSs –2 Types Smart UPS –Serial cable connects from UPS to computer.
CSC 456 Operating Systems Seminar Presentation (11/13/2012) Leon Weingard, Liang Xin The Google File System.
Copyright © 2006 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Technology Education Copyright © 2006 by The McGraw-Hill Companies,
Guide to Linux Installation and Administration, 2e 1 Chapter 9 Preparing for Emergencies.
4.1 © 2004 Pearson Education, Inc. Exam Managing and Maintaining a Microsoft® Windows® Server 2003 Environment Lesson 4: Organizing a Disk for Data.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
"1"1 Introduction to Managing Data " Describe problems associated with managing large numbers of disks " List requirements for easily managing large amounts.
Andrew C. Smith – Storage Resource Managers – 10/05/05 Functionality and Integration Storage Resource Managers.
Architecture and ATLAS Western Tier 2 Wei Yang ATLAS Western Tier 2 User Forum meeting SLAC April
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
RAID Disk Arrays Hank Levy. 212/5/2015 Basic Problems Disks are improving, but much less fast than CPUs We can use multiple disks for improving performance.
VMware vSphere Configuration and Management v6
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
WINDOWS SERVER 2003 Genetic Computer School Lesson 12 Fault Tolerance.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Distributed File Systems Questions answered in this lecture: Why are distributed file systems useful? What is difficult about distributed file systems?
T3g software services Outline of the T3g Components R. Yoshida (ANL)
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
1 CEG 2400 Fall 2012 Network Servers. 2 Network Servers Critical Network servers – Contain redundant components Power supplies Fans Memory CPU Hard Drives.
Hands-On Microsoft Windows Server 2008 Chapter 7 Configuring and Managing Data Storage.
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
Open Science Grid Consortium Storage on Open Science Grid Placing, Using and Retrieving Data on OSG Resources Abhishek Singh Rana OSG Users Meeting July.
RAID TECHNOLOGY RASHMI ACHARYA CSE(A) RG NO
Network-Attached Storage. Network-attached storage devices Attached to a local area network, generally an Ethernet-based network environment.
Answer to Summary Questions
Configuring File Services
Introduction to Data Management in EGI
CSE451 I/O Systems and the Full I/O Path Autumn 2002
Artem Trunov and EKP team EPK – Uni Karlsruhe
Storage Virtualization
CSE 451: Operating Systems Spring 2006 Module 18 Redundant Arrays of Inexpensive Disks (RAID) John Zahorjan Allen Center.
RAID RAID Mukesh N Tekwani
CSE 451: Operating Systems Winter 2009 Module 13 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura 1.
TECHNICAL SEMINAR PRESENTATION
UNIT IV RAID.
Mark Zbikowski and Gary Kimura
CSE 451: Operating Systems Winter 2012 Redundant Arrays of Inexpensive Disks (RAID) and OS structure Mark Zbikowski Gary Kimura 1.
CSE 451: Operating Systems Winter 2007 Module 18 Redundant Arrays of Inexpensive Disks (RAID) Ed Lazowska Allen Center 570.
INFNGRID Workshop – Bari, Italy, October 2004
RAID RAID Mukesh N Tekwani April 23, 2019
IT 344: Operating Systems Winter 2007 Module 18 Redundant Arrays of Inexpensive Disks (RAID) Chia-Chi Teng CTB
Improving performance
Presentation transcript:

16 th May 2006Alessandra Forti Storage Alessandra Forti Group seminar 16th May 2006

Alessandra Forti Introduction Different applications, different data, different environments, different solutions. –Root, paw ≠ Athena, BetaMiniApp or Dzero applications –AOD ≠ user ntuples –Tier2 ≠ department –Home directories ≠ user space on data servers We don’t need to stick to one solution for everything So I guess the basic fundamental questions to answer are: –What data do will we have on our storage? –How much space do we need? –What do we want from the our storage i.e redundancy, performance? Can we obtain this in different ways? –How can we access this space?

16 th May 2006Alessandra Forti Solutions Accessibility –SRM –GFAL, lcg-utils –AFS/NFS Classic storage –RAID (Redundant Array of Inexpensive disks) on local machines Grid Storage –Dcache –Xrootd –/grid (HTTP based file system)

16 th May 2006Alessandra Forti Accessibility SRM: the grid middleware component whose function is to provide dynamic space allocation and file management on shared distributed storage systems –Manage space Negotiate and assign space to users and manage lifetime of spaces –Manage files on behalf of user Pin files in storage till they are released Manage lifetime of files –Manage file sharing Policies on what should reside on a storage or what to evict –Bring the files from remote locations Manage multi-file requests queue file requests, pre-stage

16 th May 2006Alessandra Forti Accessibility GFAL is a library that can be linked to the applications to access data on a grid system. –It supports SRM APIs and the majority of grid protocols. Lcg-utils also can talk to SRM to copy, replicate and list data. –Lcg utils are the way to copy data on a grid system and register the copy in the file catalogs.

16 th May 2006Alessandra Forti Accessibility AFS/NFS (briefly) are shared file systems that can help sharing small amount of data. –AFS on a WAN would be really good if used for software distribution and I think ATLAs is supporting it. –NFS cannot be used outside the local site and it doesn’t scale very well with a number large (few hundred) of clients writing at the same time. Reading is fine.

16 th May 2006Alessandra Forti Classic storage Classic storage consist in 1 or more data servers normally with RAIDED disks accessible by local machines normally via NFS. Accessible some times (mostly at bigger labs) by remote machines via transfer protocols like scp, ftp or else, but not by applications for direct data reading. There are no file catalogs attached. Files are not replicated somewhere else –Need of local redundancy The file name space is local and normally offered by NFS

16 th May 2006Alessandra Forti RAID There are different RAID levels depending on the purpose. –Most used: RAID0, RAID1, RAID5 RAID 0: clusters 2 or more disks, data are written in blocks (striped) across the disks, there is no redundancy. –enhanced read/write performance but no reliability if one disk dies all data are lost –Good for access of temporary data a WEB cache for example. RAID 1: mirrors two or more disks –Exponentially enhanced reliability –Linearly enhanced read performance (data striping for reading but not for writing) –Partitions can be mirrored instead of disks –Good for servers: home dirs, WEB servers, Computing element, dcache head node, sw servers

16 th May 2006Alessandra Forti RAID RAID 2,3,4 data are striped across the disks at respectively bit, bytes, block level –They have parity disks for reliability Parity is a way of tracking changes using single bits or blocks of bits. Parity alone is not enough to do error recovery and reconstruction. – They are not very popular: if the parity disk dies the whole raid is unrecoverable. –They require minimum 3 disks RAID 5 like RAID4 (block-level striping) but the parity is distributed across disks –Enhanced reliability parity and data blocks are distributed. If one disk dies it can be rebuilt, if two die the whole array is lost. –In theory unlimited number of disks in practice it is better to limit them. –Poorer write performance due to the way parity must be maintained consistent with each write. Raid 5 is what is normally used on data servers where reads are more frequent than writes.

16 th May 2006Alessandra Forti Grid Storage Grid Storage consist in any device that has a space. data servers, Worker Nodes, tapes…. It is accessible to local CPUs via a number of different protocols depending on what storage management software the site administrator has installed. It is accessible from anywhere in the world to copy data in and out using grid utilities. It has all the supported VO file catalogs attached Files can be easily replicated at other sites –No real need for local redundancy File name space has to span across multiple machines. In Manchester we have 400 TB of distributed disks on the worker nodes. –Dcache xrootd and other solutions are a way to exploit it.

16 th May 2006Alessandra Forti dcache Dcache has been developed by Fermi lab and DESY to deal with their tape storage system and the staging of data on disk but it has evolved in a more general storage system manager tool. Advantages –It is SRM integrated so it has most of the space management features. –Combines several hundred nodes disks under a single file name space. –Load balance. –Data only removed if space is running short (no threshold) –Takes care that at least 'n' but not more than 'm' copies of a single dataset exists within one dCache instance. –Takes care that this rule is still true if nodes go down (schedules or even unexpected)

16 th May 2006Alessandra Forti dcache(3) Disadvantages –It is not POSIX compliant files cannot be accessed as on a normal unix file system –Supported protocols are rewritten in dcache language –It is written in java –Sources are not available –The file name space is implemented using a database in the middle –Support is, for various reasons, inadequate Unfortunately up to now it was the only solution availbable for a system like Manchester one Other viable solutions could be xrootd and StoRM

16 th May 2006Alessandra Forti Xrootd(1) XROOTD: file server which provides high performance file-based access. It was developed by BaBar/SLAC as an extension of rootd. It is now distributed as part of the standard ROOT. It is now being adopted by two LHC experiments (Alice and CMS) Advantages: –data are located within xrootd process there is no need of a database to catalog files on the system –It supports load balancing xrootd determines which server is the best for client’s request to open a file –It is fault tollerant Fault tolerance feature missing data can be again restored from other disks –Authorization plugin resolve "trusted/untrusted" user for write access Disadvantages –It is not integrated with SRM so all the space management isn’t there –lcg-utils and GFAL cannot talk to xrootd (yet)

16 th May 2006Alessandra Forti Discussion