Presentation on theme: "Managing the Unimaginable: A Practical Approach to Petabyte Data Storage Randy Cochran, Infrastructure Architect, IBM Corporation,"— Presentation transcript:
Managing the Unimaginable: A Practical Approach to Petabyte Data Storage Randy Cochran, Infrastructure Architect, IBM Corporation, TOD Information on Demand Infrastructure
1 Data Storage is Getting Out-of-Hand Are storage demands starting to overpowering you?
2 Most Research Firms Agree It is projected that just four years from now, the worlds information base will be doubling in size every 11 hours. (The toxic terabyte; How data-dumping threatens business efficiency, Paul Coles, Tony Cox, Chris Mackey, and Simon Richardson, IBM Global Technology Services white paper, July 2006) Our two-year terabyte CAGR of 52% is 3ppt (percentage points) below rolling four quarter results of 55%. ("Enterprise Hardware: storage forecast & views from CIOs", Richard Farmer and Neal Austria, Merrill Lynch Industry Overview, 03 January 2007) With a 2006–2011 CAGR nearing 60%, there is no lack in demand for storage… ("Worldwide Disk Storage Systems 2007–2011 Forecast: Mature, But Still Growing and Changing", Research Report # IDC206662, Natalya Yezhkova, Electronics.ca Publications, May 2007) According to TheInfoPro…..the average installed capacity in Fortune 1000 organizations has jumped from 198 TB in early 2005 to 680 TB in October …..TIP found that capacity is doubling every 10 months. (InfoStor Magazine, Kevin Komiega, October 19, 2006)
3 Whats Driving Petabyte Level Storage? The Perfect Storm General Increase in demand New digital data technologies More regulatory requirements Better protection from litigation Disaster Recovery plans Proliferation of Sophisticated applications Declining storage media costs A desire for greater storage efficiency Storage technical skills scarcity A growing understanding of retained datas business value According to IDC, between 2006 and 2010 information added annually to the digital universe will increase more than six fold from 161 to 988 exabytes.
4 Just How Big is a Petabyte? Petabyte storage had been around for years – online Petabyte storage has not. Ninety-two percent of new information is stored on magnetic media, primarily hard disks.How Much Information 2003, UC Berkeley's School of Information Management and SystemsSchool of Information Management and Systems
5 How Big is That in Human Terms? According to Britannica.com the U.S. Library of Congress contains approximately 18 million books, 2.5 million recordings, 12 million photographs, 4.5 million maps, and more than 54 million manuscripts.
6 Why is Petabyte Storage a Challenge? Areas Impacted by Petabyte Storage: Content and File Management Application & Database Characteristics Storage Management Architectural Design Strategy Performance and Capacity SAN Fabric Design Backup and Recovery Methods Security System Complexity Compliance with Regulatory Requirements Operational Policies and Processes Maintenance Requirements
7 Content and File Management
8 Management Starts With Data Classification Data Classification Assumptions Not all data is created equal The business value of data changes over time Performance can be improved by re-allocating data to an optimized storage configuration The value of most business data is not fixed; it is expected to change over time Understanding the business value of data is a crucial in designing an effective data management strategy Which data has a greater value to the business - a clients purchase record, or a memo about last years phone system upgrade?
9 Data Classification Example There are no universally accepted standard definitions for Tier Levels.
10 Control Your File Content Implement file aging Set data retention periods Eliminate low value data Clean out old backup files Eliminate outdated information Deploy de-duplication technology Reduce storage of low value data Locate and purge corrupt files Crack down on unauthorized storage usage Periodically review log files and archive or delete obsolete information
11 Application and Database Characteristics
12 Know your applications needs User expectations Workload complexity Read or write intensity Sequential files usage IOPS dependence Stripe size optimization Throughput requirements Service prioritization Growth expectations Dont allow databases to lock up vast amounts of storage Know Your Application and Database Needs
13 Applications characteristics will drive storage decisions Value to the business Number of users Usage patterns Steady Bursty Cyclical Variable 7x24 or 9x5 access Domestic or global access Distributed or self-contained High or low security data Architectural constraints Significant performance gains (or losses) can be achieved by matching requirements to storage characteristics Applications Will Drive Storage Requirements
14 Storage Management
15 Large Storage Systems Must Be Managed Information Lifecycle Management (ILM) Hierarchical Storage Management (HSM) Storage Resource Management (SRM) Storage Virtualization "Enterprises can achieve better and more targeted utilization of resources by first establishing the value of their information assets and then using storage management software to execute the policies that define how resources are utilized." Noemi Greyzdorf, research manager, Storage Software, IDC
16 Information Lifecycle Management (ILM is) the process of managing business data throughout its lifecycle from conception until disposition across different storage media, within the constraints of the business process. (courtesy of Veritas Corporation, Nov. 2004) ILM is not a commercial product, but a complete set of products and processes for managing data from its initial inception to its final disposition.
17 Information Lifecycle Management Information has business value Its value changes over time It ages at different rates It has a finite life-cycle As data ages its performance needs change Some Information is subject to different security requirements, due to government regulatory or legal enforcements Outdated information has different disposal criteria A combination of processes and technologies that determine how information flows through a corporate environment Encompasses management of information from its creation until it becomes obsolete and is destroyed
18 Best Practices for ILM Implementations Know exactly where information is stored Be able to retrieve information quickly and efficiently Limit access to only those who need to view data Create policies for managing and maintaining data Do not destroy important documents Avoid keeping multiple copies of the same data Retain information only until it is no longer useful Destroy outdated files on a regular basis Document all processes and keep them up-to-date
19 Hierarchical Storage Management HSM is a policy-based data storage management system that automatically moves data between high- cost and low-cost storage media, without requiring the knowledge or involvement of the user. (courtesy of IBM has been involved in providing HSM solutions for over 30-years and offer a wide variety of products with automated data movement capabilities.
20 File Access Activity Over Time
21 Hierarchical Storage Management HSM Concepts Only 10%-15% of most data is actively accessed The business value of data changes over time Between 80% and 90% of all stored data is inactive High performance storage (FC disks) are expensive Lower performance media (tape, optical platters, and SATA disk) are comparatively inexpensive 10% 20% 70% Archive
22 Hierarchical Storage Management HSM Concepts (cont.) Enterprise class storage is not required for all data Policies can be set to establish the proper frequency for transitioning aging data to less expensive media HSM allows optimal utilization of expensive disk storage Low cost, high density disks consume fewer resources Overall storage system performance may improve $$$$ $$$ $$ $
23 IBM Products with HSM Capabilities General Parallel File System (GPFS) IBM Content Manager for Multiplatforms Tivoli Storage Manager HSM for Windows Tivoli Storage Manager for Space Management (AIX) SAN File System (SFS) DFSMShsm (Mainframe) High Performance Storage System (HPSS)
24 Storage Resource Management Storage Resource Management (SRM) is the process of optimizing the efficiency and speed with which the available drive space is utilized in a storage area network (SAN). Functions of an SRM program include data storage, data collection, data backup, data recovery, SAN performance analysis, storage virtualization, storage provisioning, forecasting of future needs, maintenance of activity logs, user authentication, protection from hackers and worms, and management of network expansion. An SRM solution may be offered as a stand-alone product, or as part of an integrated program suite. (Definition Courtesy of IBMs primary tool for Storage Resource Management is their TotalStorage Productivity Center suite of tools for disk, data, fabric, and replication.
25 Storage Resource Management Functions
26 Storage Virtualization Virtualization The act of integrating one or more (back end) services or functions with additional (front end) functionality for the purpose of providing useful abstractions. Typically virtualization hides some of the back end complexity, or adds or integrates new functionality with existing back end services. Virtualization can be nested or applied to multiple layers of a system. (Definition Courtesy of Virtualization allows most of the complexity of a storage infrastructure to be hidden from the user.
27 Virtualization Makes Storage One Large Pool Virtualization Characteristics Makes storage configuration details invisible to the user Improves overall manageability of the system Aggregates isolated storage islands into a unified view Facilitates greater flexibility and scalability Optimizes utilization of storage capacity Provides the ability to move data on-the-fly Improves storage subsystems flexibility Allows rapid re-allocation of storage resources Improves performance by providing another layer of caching May provide additional functionality for the SAN
28 Architectural Design Strategy
29 Key Architectural Design Considerations Resource Consumption Storage Economics RAID Allocation Performance Objectives Other Design Issues The integrity of the architectural design will determine the overall performance, stability, economic efficiency, manageability and future scalability of the system.
30 Power Consumption vs. Storage Capacity ** National retail price of electricity per KwH from Power, Cooling, Space Efficient Storage, page 2, ESG white paper, Enterprise Strategy Group, July These disks all have very similar power consumption requirements, even though the largest one features 28 times the capacity of the smaller one. In addition, each disk will require approximately watts of electrical power to cool each BTU of heat produced.
31 Comparing Storage Subsystem Power Costs Significant power savings may be realized by redistributing data to the appropriate type and size of disk drive.
32 Comparing Storage Subsystem Cooling Costs Additional power savings may be realized from the reduced cooling requirements provided by high capacity, lower wattage disk drives.
33 Comparing Storage Floor-Space Cost The DS4800 and DS4200 storage subsystems include the required number of disk expansion trays mounted in standard equipment racks.
34 How Do the Costs Add Up? DS8300DS4800 DS4200s with SATA Disk Traditional Approach Everything on DS8300s Tiered Storage Approach Savings : $614,935 / yr.
35 A Look at Older Disk Subsystem Efficiency Storing 100 TB of data on more modern storage subsystems results in 50% less power consumption, a 53% reduction in BTUs per hr., and a reduction in required floor space of 38%. In addition, a DS8300 system has over 7x the throughput of the ESS800.
36 Why is Tiered Storage Important? Maps datas business value to disk characteristics Places data on storage appropriate to its usage Incorporates lower cost disks Reduces resource usage (power, cooling, etc.) Matches user access needs to storage characteristics Capitalizes on higher capacity disk drive technology Increases overall performance of the system
37 A Typical Tiered Storage Architecture Business Critical High Performance / Very High Availability Business Important Good Performance / High Availability Business Standard Average Performance / Standard Availability Reference / Historical Near-line or Off-line Normally a tiered storage strategy is based on datas business value. DS8300 DS4800 DS4200s with SATA Disk TS3500 Tape Library
38 Choosing the Right Controller Frame
39 Choosing the Right Disk Characteristics
40 Comparing Disk Drive Attributes
41 The Cost Impact of Adding Disk Trays Note: Calculations based on 146 GB, 10K RPM Drives
42 Tiered Storage Design Pros and Cons Advantages Lower initial purchase price Higher capacity per square foot Reduced power consumption Decreased requirement for cooling Increased equipment flexibility Potentially a higher performance solution Disadvantages Inherently a more complex architecture Greater up-front effort to design and implement Requires advanced storage design skills and knowledge
43 RAID Selection Decision Drivers Application or Database characteristics Read/write mix Dependency on IOPS RAID Performance characteristics Appropriate RAID level Number of disks per array Stripe size Available bandwidth Configuration rules and recommendations Loss from data parity and hot sparing Disk failure probability RAID parity rebuild times
44 Loss from Mirroring, Striping, and Sparing RAID10 = Mirror plus Stripe RAID1 = Mirror Only
45 Loss from RAID5 Parity and Sparing Note: The second tray has one 2+P array to allow for one spare drive per two trays. Note: The second tray has one 6+P array to allow for one spare drive per two trays. Note: Each tray has one spare drive per tray.
46 Other Architectural Considerations Compatibility High availability Architectural robustness Flexibility and scalability Stability of the technology Vendors financial standing Well defined product line roadmap Support for industry standards
47 Performance and Throughput
48 Storage Subsystem Performance Drivers Business objectives and user expectations Applications and database characteristics Server characteristics SAN fabric characteristics Storage controller characteristics Caching characteristics Configuration characteristics Disk latency characteristics "We can't solve problems by using the same kind of thinking we used when we created them." Albert Einstein
49 Storage Performance Enhancers Data Mover – Reassigning data transfer tasks to a specialized engine reduces the workload on the host processing system. Search Engines – Systems dedicated to executing searches in vast amounts of stored data to satisfy specific requests. Directory Services – Stores and organizes information about resources and data objects. High Speed Interconnections – Dedicated behind the scenes networks dedicated to the transfer of large amounts of data. Autonomic Computing – must have an ability to reconfigure itself under varying and possibly unpredictable conditions.
50 Other Storage Performance Tips
51 SAN Fabric Network
52 SAN Fabric Overview SAN Fabric is the interconnecting structure between associated servers and storage devices Proper fabric design will directly impact on: Performance Availability Equipment cost Manageability Maintainability Communications protocol can be either Fibre Channel, Ethernet, or a combination of both Ability of the fabric to scale is critical Monitoring of SAN fabric traffic is a necessity
53 Designing a High Performance Fabric Select an optimal SAN fabric topology Mesh Waterfall Core / Edge Fat tree Butterfly Torus Hypercube Hybrid Ensure the design is driven by business application requirements Keep it as simple as possible for manageability
54 Common SAN Fabric Examples
55 SAN Fabric Design Considerations SAN Fabric design issues: Throughput requirements Potential bottlenecks Port speed / port count Port subscription rate Maximum hop count Redundancy for High Availability Modularity (flexibility/scalability) Future upgradeability Complexity vs. overall manageability Isolation vs. unification Wide Area Network interconnections Power consumption and footprint Component cost
56 Backup and Recovery
57 Backup and Recovery Challenges Size of the backup window Ability to recover files The time to recover files Integrity of the data backups Required frequency of backups Stored data retention period Functional Legal Available bandwidth for backup resources Media deterioration over time Technological obsolescence
58 The Traditional Storage Backup Approach 1.0 PB of Storage TS3500 Tape Library (192) of the newest LTO 4 tape drives running at a maximum native transfer rate of 120 MB/sec. would need at least 13-hours to back up 1.0 PB of data.
59 Fast Tape Drives can Saturate the Network One LTO-4 tape drive run in native mode is 20% faster than the effective usable bandwidth of Gigabit Ethernet! Four LTO-4 tape drives run at a 2:1 compressed mode will take most of the usable bandwidth of 10Gbps Ethernet!
60 Large Systems Backup Approaches Point-in-Time Copies and Replication Snapshots Most popular method of large storage backup Snapshots create an exact copy of the source data Once a bitmap is created, storage access can resume While copying, new data is written to both source and target Requires minimal downtime for production systems Replication Replication creates a mirror image of data over distance May be synchronous (consistent) or asynchronous (lagging) Synchronous is distance-limited, asynchronous is not
61 Point-in-Time Copy
62 Data Replication Structures
63 Other Large Systems Backup Approaches Object-based Backups Backs up only new blocks that have changed Copies only files it has never seen before Inserts pointers if file exists somewhere else Provides instant recoveries by presenting a mountable volume Delta-Block Incremental Backups Evaluates changed data by breaking a file down into discrete blocks Block-for-block comparison of a modified file with an existing file When a difference is detected it extracts a copy of that block only Usually copies a number of blocks, but not the entire file Continuous Data Protection (CDP) Copies blocks to the backup system as soon as they change Stores data in a log that allows recovery from any point in time Performs fast recoveries by restoring only blocks that have changed Provides instant recoveries by presenting a mountable volume
64 Security Requirements
65 Impact of Petabyte Storage on Security Traditional distributed access control techniques are designed for smaller systems with general or random workloads Petabyte storage may service tens of thousands of clients and hundreds of storage devices Storage design must be capable of supporting I/O patterns that are highly parallel and very bursty by nature Security solutions must be kept highly scalable to keep up with storage growth patterns
66 Impact of Petabyte Storage on Security (Cont.) Authentication and authorization requirements can dramatically impact server performance Performance could be further reduced if data is encrypted Traditional security protocols perform poorly because they do not scale well The number of security operations is closely tied to the number of devices and requests
67 Regulatory Compliance
68 The Challenge Of Regulatory Compliance Whats driving storage regulatory legislation? Corporate fraud and illegal practices Increased emphasis on the security of personal data The threat of terrorist activities The global impact of the Internet Increased reliance on stored data for defense against litigation Increased business dependence on electronic communications ( , digital voice messaging, instant messaging, VoIP, etc.)
69 Regulatory Requirements Continue to Grow According to the Storage Networking Industry Association (SNIA) there are over 20,000 regulations worldwide addressing the storage of data If you do business overseas, you must also be aware of the applicable foreign regulatory requirements The number of government regulatory requirements increases every year There is little chance this upward trend will reverse itself in the future Regulatory guidelines do not dictate how you should maintain your data, only what the expected outcome should be.
70 Common Regulatory Compliance Goals Most regulatory requirements are based upon: Security: Maintain data in a secure environment Efficiency: Rapid location and retrieval of data Legibility: Recovered documents must be in a readable format that is clear and concise Authenticity: The recovered document must be verifiable as the original Validation: Documentation process must be available for review by a neutral third party Regulatory compliance becomes more challenging as storage subsystems grow in size and complexity.
71 Regulatory Legislation Examples Sarbanes-Oxley HIPAA USA Patriot Act Gramm-Leach-Bliley Act FRCP CFR a(f) NASD 3010 and CFR Part 11 (FDA) DOD California Senate Bill 1386 Florida Sunshine Law PCI ISO CFR Title 18 (FERC) E-SIGN EU Directive 95/46/EC Basel II NARA GRS20 CFR Title 47, Part 42 NASD 2711/NYSE Rule 472 JCAHO FPC 65 COOP compliance
72 Maintenance Requirements
73 Disk Drive Reliability Misconceptions Actual disk failure rate is usually higher than published –Vendors indicate a.58% -.88% failure rate –Actual field usage suggests a 1.5% - 3.5% (or greater) failure rate Field studies show no appreciable difference in reliability between SCSI, FC, and SATA drives Heat and high duty cycles do not appear to have as detrimental of an effect on disk life as once thought
74 Disk Drive Reliability Misconceptions (Cont.) Infant Mortality doesnt appear to be a significant issue for newly installed disks Disks exhibit a fairly linear rate of failure over time, which is contrary to the standard bathtub model Self-Monitoring, Analysis and Reporting Technology (S.M.A.R.T.) diagnostics appear to predict only about 50% of all disk failures
75 Disk Failures Are an Expected Occurrence Using the disk count from our previous Traditional vs. Tiered models, its easy to see that disk failures will occur on a regular basis.
76 Other Considerations
77 Other Issues to Consider Design for minimal human intervention Maintain extensive monitoring of the environment Exercise control over information propagation Architect for maximum performance and throughput Ensure robustness and high availability Configure for scalability and flexibility Develop well defined SLA objectives Implement a structured support operation
78 Emerging Technologies
79 Emerging Technologies to Watch Thin Provisioning Data de-duplication SAS Interface InfiniBand NPIV (N_Port ID Virtualization) Large-platter storage technologies 2.5 disk technologies Solid state disk drives Virtualized file systems (i.e.- ZFS, SOFS) Grid Storage
81 Some Parting Thoughts Online multi-petabyte storage is a reality Data will double every two to three years Storage media cost-per-GB will continue to decline Storage Operational management is a growing issue Governmental regulations will increase over time New technologies will demand additional storage Experienced storage designers and administrators will grow increasingly harder to find Scarce data center resources (bandwidth, floor space, power, cooling, etc.) will become more expensive A carefully designed architecture is the key to efficient storage operations
82 Putting It All Together Captain Kirk: Scotty - We need more power!!! Mr. Scott: Capn, I'm gi'in ya all I got, she can na take much more!
83 Questions? Randy Cochran, Infrastructure Architect IBM Global Technical Services - Cell: (630)