November 2, 2000HEPiX/HEPNT FERMI SAN Effort Lisa Giacchetti Ray Pasetes GFS information contributed by Jim Annis.

Slides:



Advertisements
Similar presentations
Die Kooperation von Forschungszentrum Karlsruhe GmbH und Universität Karlsruhe (TH) StorNextFS, a fast global Filesysteme in a heterogeneous Cluster Environment.
Advertisements

Archive Task Team (ATT) Disk Storage Stuart Doescher, USGS (Ken Gacke) WGISS-18 September 2004 Beijing, China.
Copyright © 2014 EMC Corporation. All Rights Reserved. Linux Host Installation and Integration for Block Upon completion of this module, you should be.
1 CSC 486/586 Network Storage. 2 Objectives Familiarization with network data storage technologies Understanding of RAID concepts and RAID levels Discuss.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Silicon Graphics, Inc. Cracow ‘03 Grid Workshop SAN over WAN - a new way of solving the GRID data access bottleneck Dr. Wolfgang Mertz Business Development.
Silicon Graphics, Inc. Poster Presented by: SGI Proprietary Technologies for Breakthrough Research Rosario Caltabiano North East Higher Education & Research.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility Newport News, Virginia USA
Comparison and Performance Evaluation of SAN File System Yubing Wang & Qun Cai.
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
VMware Infrastructure Alex Dementsov Tao Yang Clarkson University Feb 28, 2007.
Storage Networking. Storage Trends Storage growth Need for storage flexibility Simplify and automate management Continuous availability is required.
National Energy Research Scientific Computing Center (NERSC) The GUPFS Project at NERSC GUPFS Team NERSC Center Division, LBNL November 2003.
Peter Stefan, NIIF 29 June, 2007, Amsterdam, The Netherlands NIIF Storage Services Collaboration on Storage Services.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
Frangipani: A Scalable Distributed File System C. A. Thekkath, T. Mann, and E. K. Lee Systems Research Center Digital Equipment Corporation.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Storage Area Networks The Basics. Storage Area Networks SANS are designed to give you: More disk space Multiple server access to a single disk pool Better.
Storage Survey and Recent Acquisition at LAL Michel Jouvin LAL / IN2P3
Terabyte IDE RAID-5 Disk Arrays David A. Sanders, Lucien M. Cremaldi, Vance Eschenburg, Romulus Godang, Christopher N. Lawrence, Chris Riley, and Donald.
Interposed Request Routing for Scalable Network Storage Darrell Anderson, Jeff Chase, and Amin Vahdat Department of Computer Science Duke University.
Baydel Founded in 1972 Headquarters: Surrey, England North American Headquarters: San Jose, CA Engineering Driven Organization Specialize in Computer Storage.
A Linux PC Farm for Physics Analysis at the ZEUS Experiment Marek Kowal, Krzysztof Wrona, Tobias Haas, Ingo Martens, Rainer Mankel DESY, Notkestrasse 85,
Storage for Opteron Server Sun Microsystems Inc.
Online Systems Status Review of requirements System configuration Current acquisitions Next steps... Upgrade Meeting 4-Sep-1997 Stu Fuess.
CASPUR SAN News Andrei Maslennikov Orsay, April 2001.
Slide 1 DESIGN, IMPLEMENTATION, AND PERFORMANCE ANALYSIS OF THE ISCSI PROTOCOL FOR SCSI OVER TCP/IP By Anshul Chadda (Trebia Networks)-Speaker Ashish Palekar.
DAC-FF The Ultimate Fibre-to-Fibre Channel External RAID Controller Solution for High Performance Servers, Clusters, and Storage Area Networks (SAN)
CASPUR Site Report Andrei Maslennikov Sector Leader - Systems Orsay, April 2001.
CASPUR Site Report Andrei Maslennikov Lead - Systems Karlsruhe, May 2005.
Sensitivity of Cluster File System Access to I/O Server Selection A. Apon, P. Wolinski, and G. Amerson University of Arkansas.
Amy Apon, Pawel Wolinski, Dennis Reed Greg Amerson, Prathima Gorjala University of Arkansas Commercial Applications of High Performance Computing Massive.
Linux Servers with JASMine K. Edwards, A. Kowalski, S. Philpott HEPiX May 21, 2003.
MA8000/EMA12000 Fibre Channel Product Set Storage controllers, cache, batteries: B21 HSG80 RAID Array Controller w/ 256MB Cache B21 HSx80.
1 U.S. Department of the Interior U.S. Geological Survey Contractor for the USGS at the EROS Data Center EDC CR1 Storage Architecture August 2003 Ken Gacke.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Computer Systems Lab The University of Wisconsin - Madison Department of Computer Sciences Linux Clusters David Thompson
Copyright © 2014 EMC Corporation. All Rights Reserved. Windows Host Installation and Integration for Block Upon completion of this module, you should be.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
CASPUR Storage Lab Andrei Maslennikov CASPUR Consortium Catania, April 2002.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
The II SAS Testbed Site Jan Astalos - Institute of Informatics Slovak Academy of Sciences.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Probe Plans and Status SciDAC Kickoff July, 2001 Dan Million Randy Burris ORNL, Center for.
Storage and Storage Access 1 Rainer Többicke CERN/IT.
Jefferson Lab Site Report Sandy Philpott Thomas Jefferson National Accelerator Facility (formerly CEBAF - The Continuous Electron Beam Accelerator Facility)
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
James Annis Gabriele Garzoglio Peretz Partensky Chris Stoughton The Experimental Astrophysics Group, Fermilab The Terabyte Analysis Machine Project data.
CASPUR Site Report Andrei Maslennikov Lead - Systems Amsterdam, May 2003.
Disk Farms at Jefferson Lab Bryan Hess
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
CASPUR Site Report Andrei Maslennikov Lead - Systems Rome, April 2006.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
CASPUR Site Report Andrei Maslennikov Group Leader - Systems RAL, April 1999.
November 1, 2000HEPiX/HEPNT FERMI SITE REPORT Lisa Giacchetti.
Randy MelenApril 14, Stanford Linear Accelerator Center Site Report April 1999 Randy Melen SLAC Computing Services/Systems HPC Team Leader.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
COMPASS Computerized Analysis and Storage Server Iain Last.
Mass Storage at SARA Peter Michielse (NCF) Mark van de Sanden, Ron Trompert (SARA) GDB – CERN – January 12, 2005.
Storage Networking. Storage Trends Storage grows %/year, gets more complicated It’s necessary to pool storage for flexibility Intelligent storage.
GPFS Parallel File System
12/19/01MODIS Science Team Meeting1 MODAPS Status and Plans Edward Masuoka, Code 922 MODIS Science Data Support Team NASA’s Goddard Space Flight Center.
NL Service Challenge Plans
Storage Networking.
Storage Area Network
Storage Networking.
Presentation transcript:

November 2, 2000HEPiX/HEPNT FERMI SAN Effort Lisa Giacchetti Ray Pasetes GFS information contributed by Jim Annis

November 2, 2000HEPiX/HEPNT Overview Motivation Current Problems Future goals Evaluation CXFS SANergy GFS Current Status

November 2, 2000HEPiX/HEPNT Motivation Current Problems Unbalanced use of central UNIX cluster Large dataset(s) need to be shared in a large distributed compute environment Current solutions lack performance throughput Future goals Linux analysis cluster with SMP feel

November 2, 2000HEPiX/HEPNT Evaluation: CXFS Currently SGI-only Currently requires RAID True(er) SAN FS Commitment to Linux port Equipment 1 Origin Origin 200s 1 Brocade F-C switch 1 SGI (Clarion) RAID ~1TB raw

November 2, 2000HEPiX/HEPNT Evaluation: SANergy Heterogeneous solution -- Solaris, WinNT, WIN2K, IRIX, Tru-64, MacOS, AIX Works with RAID or JBOD Pseudo SAN FS with NFS look Linux port in future (11/00, both MDC and Client) Equipment 1 Sun Sparc20: RAID management box 1 Ultra 60, 3 Linux, 1 NT4: MDC and client 1 O2200 (client only) 1 16 port Brocade switch 1 Metastor E4400 RAID ~720GB raw

November 2, 2000HEPiX/HEPNT Evaluation: GFS GFS Open source ( GPL’d ) Sistina Software (ex-University of Minnesota) High performance 64-bit files and file system Distributed, server-less metadata Data synchronization via global, disk based locks Journaling and node cast-out Three major pieces: The network storage pool driver The file system The locking modules

November 2, 2000HEPiX/HEPNT Evaluation: GFS (equipment) System integrator Linux NetworX Cluster control box Compute Nodes Linux NetworX Dual 600 MHz Pentium III ASUS motherboard 1 Gig RAM 2x36 Gig EIDE disks Qlogic 2100 HBA Ethernet Cisco Catalyst 2948G Fibre Channel Gadzoox Capellix 3000 Global Disk DotHill SanNet 4200 Dual Fibre Channel controllers 10x73 Gig Seagate Cheetah SCSI disk Software Linux Qlogic drivers GFS V3.0 Condor

November 2, 2000HEPiX/HEPNT Current Status: CXFS Config: 1 file system, GB disk RAID 5 LUNs striped together; Each system w/ 1 HBA 3 writes and 3 reads simultaneous of 1GB files at 64K blocks (6 different files) READ 11.5/11.6/11.9 MB/s WRITE 36.5/28.4/28.4 MB/s SGI Clarion RAID biases towards writes Aggregate: 128 MB/s / 200 MB/s 64% utilization Peak single write for 2GB file 64K blk = 45MB/s Peak single read for 2GB file 64K blk = 28MB/s Simultaneous writes to same file = MB/sec

November 2, 2000HEPiX/HEPNT Current Status: CXFS Stability Issues Cluster can hang when unmounting file systems Problem on one machine can affect all nodes resulting in need to reboot entire cluster Simple reboot often does not work and will need to execute a hard reset. Java GUI Occasionally hangs Occasionally reports erroneous cluster status

November 2, 2000HEPiX/HEPNT Current Status: SANergy Equipment almost in place MetaStor hardware raid tested w/out SANergy Pleased with performance Worked as AFS file server central disk store Used this hardware with CXFS test Config: disk RAID 5 Results: 95+MB/s read ; 90+MB/s write Limited by HBA Software yet to be received

November 2, 2000HEPiX/HEPNT Current Status: GFS Config: 5 machines, 1 5-disk RAID-5 2 reads and 1 write, simultaneous of 1 GB files at 64k blocks Write: 5.1 MB/s Read: 30.0, 30.0 MB/s Aggregate 65 MB/s / 90 MB/s 72% utilization