Presentation is loading. Please wait.

Presentation is loading. Please wait.

San Diego Supercomputer Center & NPACIAHM March 2002 Storage Resource Broker Case Studies George Kremenek

Similar presentations


Presentation on theme: "San Diego Supercomputer Center & NPACIAHM March 2002 Storage Resource Broker Case Studies George Kremenek"— Presentation transcript:

1 San Diego Supercomputer Center & NPACIAHM March 2002 Storage Resource Broker Case Studies George Kremenek

2 San Diego Supercomputer Center & NPACIAHM March 2002 Projects Digital Sky Project (NPACI) {NVO (NSF)} Hayden Planetarium Simulation & Visualization ASCI - Data Visualization Corridor (DOE) Visual Embryo Project (NLM) Long Term Archiving Project (NARA) Information Power Grid (NASA) Particle Physics Data Grid (DOE) {GrPhyN (NSF)} Biomedical Information Research Network (NIH) RoadNet (NSF) Grid Portal (NPACI) NSDL – National Science Digital Library (NSF) Knowledge Network for BioComplexity (NSF) Tera Scale Computing (NSF) Hyper LTER Earth System Sciences – CEED, Bionome, SIO Explorer Education – Transana (NPACI) Mol Science – JCSG, AfCS Digital Libraries – ADL, Stanford, UMichigan, UBerkeley, CDL

3 San Diego Supercomputer Center & NPACIAHM March 2002 Hayden Planetarium

4 San Diego Supercomputer Center & NPACIAHM March 2002 Problem Data Transfer and Share TearaBytes of Data Across the Internet –Simulation Data produced at NSCA –Visualized at SDSC –Validated at AMNH, NCSA, UVa & other places –Consumed at AMNH Data sizes ranged from 3 TB to 10 TB Other sites (CalTech, BIRN) used as cache resources

5 San Diego Supercomputer Center & NPACIAHM March 2002 Hayden Planetarium Project “A Search for Life: Are We Alone?” The animations was done for the new planetarium show “A Search for Life: Are We Alone?” narrated by Harrison Ford. The show opened Saturday, March 2nd. Sites involved in the project : AMNH = American Museum of Natural History NCSA = National Center for Supercomputing Applications SDSC = San Diego Supercomputer Center University of Virginia CalTech, NASA, UCSD

6 San Diego Supercomputer Center & NPACIAHM March 2002

7 San Diego Supercomputer Center & NPACIAHM March 2002 Hayden Credits People involved AMNH : Producer, Anthony Braun, Director, Carter Emmart, Erik Wesselak, Clay Budin, Ryan Wyatt, Asst. Curator, Dept of Astrophysics, Mordecai Mac Low NCSA : Stuart Levy, Bob Patterson SDSC : David R. Nadeau, Erik Enquist, George Kremenek, Larry Diegel, Eva Hocks U. Virginia: Professor, John F. Hawley

8 San Diego Supercomputer Center & NPACIAHM March 2002 Hayden Data in SRB. Disk accretion: –Simulation run at SDSC by John Hawley. Data stored in SRB. Jet imagery : –Images from Hubble Space Telescope. Data stored in SRB. Flight path: –Planned at NCSA and AMNH. Data stored in SRB.

9 San Diego Supercomputer Center & NPACIAHM March 2002 Hayden Data Flow NCSA SDSC AMNH NYC GPFS 7.5 TB IBM SP2 SGI production,parameters, movies, images data simulation visualization HPSS 7.5 TB 2.5 TB UniTree UVa NY CalTech BIRN

10 San Diego Supercomputer Center & NPACIAHM March 2002 Hayden, Data involved ISM = Interstellar Medium Simulation – run by Mordecai Mac Low of AMNH at NCSA : 2.5 Terabytes sent from NCSA to SDSC. Data stored in SRB (HPSS, GPFS). Ionization : –Simulation run at AMNH, 117 Gigabytes sent from AMNH to SDSC. Data stored in SRB. Star motion: –Simulation run at AMNH by Ryan Wyatt.38 Megabytes sent from AMNH to SDSC.

11 San Diego Supercomputer Center & NPACIAHM March 2002 Hayden totals Data –total 3 * 2.5 TB = 7.5 TB Files –3 * 9827 files + miscellaneous files Duration –December 2001, January, February 2002

12 San Diego Supercomputer Center & NPACIAHM March 2002 Hayden Conclusions The SRB was used as a central repository for all original, processed or rendered data. Location transparency crucial for data storage, data sharing and easy collaborations. SRB successfully used for a commercial project in “impossible” production deadline situation dictated by marketing department. Collaboration across sites made feasible with SRB

13 San Diego Supercomputer Center & NPACIAHM March 2002 ASCI - DOE

14 San Diego Supercomputer Center & NPACIAHM March 2002 Advanced Simulation and Computing (ASCI) Area –Advanced computations, three-dimensional modeling, simulation and visualization. Problem –evaluating SRB as an advanced data handling platform for the DOE data visualization corridor. Requirements –SRB working well with HPSS for handling large files as well and large number of small files. –Data movement in “bulk” by researchers

15 San Diego Supercomputer Center & NPACIAHM March 2002 ASCII and DataCutter ASCI is currently evaluating the DataCutter technology in SRB DataCutter –handles multidimensional data subset-ing and filtering developed by U of Maryland Ohio State. –ASCI is interested in the integration of DataCutter with SRB for the advanced visualization corridor.

16 San Diego Supercomputer Center & NPACIAHM March 2002 ASCI Data Flow Data movement across 3 hosts HPSS SRB server data cache SRB server MCAT Oracle applications SRB clients local FS

17 San Diego Supercomputer Center & NPACIAHM March 2002 ASCI People ASCI project - LLNL –Celeste Matarazzo, Punita Sinha The Storage Resource Broker (SRB): SDSC – Michael Wan, Arcot Rajasekar, Reagan Moore Datacutter : Univ. Maryland, OSU –Joel Saltz, Tahsin Kurc, Alan Sussman

18 San Diego Supercomputer Center & NPACIAHM March 2002 ASCI Time-line – Dec 2002 Data Sizes –Very large files (multi GB) –Large number of small files (over a million files) –Total size exceeding 2 TB for each run SRB Solution –SRB/HPSS interoperation – highly integrated –SRB data mover protocol adapted to HPSS parallel mover protocol

19 San Diego Supercomputer Center & NPACIAHM March 2002 ASCI Parallel Protocol HPSS server –directs the parallel data transfer scheme –uses the class of service HPSS feature SRB server –is utilizing the HPSS's parallel mover protocol. –transfer rates of up to 40 MB/sec –speedup of 2 to 5 times using multiple threads can be achieved.

20 San Diego Supercomputer Center & NPACIAHM March 2002 ASCI Small Files Ingesting a very large number of small files into SRB –is time consuming if the files are ingested one at a time –greatly improved with the use of bulk ingestion. –ingestion was broken down into two parts the registration of files with MCAT the I/O operations (file I/O and network data transfer) –multi-threading was used for both the registration and I/O operations. –new utility - Sbload was created for this purpose. –reduced the ASCI benchmark time of ingesting ~2,100 files from ~2.5 hours to ~7 seconds.

21 San Diego Supercomputer Center & NPACIAHM March 2002 ASCI Conclusions Very large number (2 million) of small/average files can be ingested into SRB (HPSS) in short time Sbload (with bulk SRB registration) can load and register up to 300 files a second Sbload will be included in next SRB release Sbload can be used for other resources also

22 San Diego Supercomputer Center & NPACIAHM March 2002 Digital Sky Project

23 San Diego Supercomputer Center & NPACIAHM March 2002 Digital Sky 2MASS (2 Microns All Sky Survey): –Bruce Berriman, IPAC, Caltech; John Good, IPAC, Caltech, Wen-Piao Lee, IPAC, Caltech NVO (National Virtual Observatory): –Tom Prince, Caltech, Roy Williams CACR, Caltech, John Good, IPAC, Caltech SDSC – SRB : –Arcot Rajasekar, Mike Wan, George Kremenek, Reagan Moore

24 San Diego Supercomputer Center & NPACIAHM March 2002

25 San Diego Supercomputer Center & NPACIAHM March 2002

26 San Diego Supercomputer Center & NPACIAHM March 2002 Digital Sky - 2MASS The input data was on tapes in a random order. Ingestion nearly 1.5 year - almost continuous SRB performed a spatial sort on data insertion. The disc cache (800 GB) for the HPSS containers was utilized.

27 San Diego Supercomputer Center & NPACIAHM March 2002 Digital Sky Data Ingestion Informix SUN SRB SUN E10K HPSS …. 800 GB 10 TB SDSC IPAC CALTECH input tapes from telescopes star catalog Data Cache

28 San Diego Supercomputer Center & NPACIAHM March 2002 Digital Sky Data Ingestion 4 parallel streams (4 MB/sec per stream), 24*7*365 Total 10+TB, 5 million, 2 MB images in 147,000 containers. Ingestion speed limited by input tape reads –Only two tapes per day can be read work flow incorporated persistent features to deal with network outages and other failures. C API was utilized for fine grain control and to be able to manipulate and insert metadata into Informix catalog at IPAC Caltech.

29 San Diego Supercomputer Center & NPACIAHM March 2002 Data Sorting Sorting of 5 million files on the fly Input tape files: temporal order Stored SRB Containers: spatial order –Scientists view/analyze data by neighborhood Data Flow: –Files from tape streamed to SRB –SRB puts them in proper ‘bins’ (containers) –Container cache-management a big problem –Files from a tape may go into more than 1000 bins –Cache space limitations ( GB) made for a lot of trashing –SRB Daemon managed cache - watermarks

30 San Diego Supercomputer Center & NPACIAHM March 2002 Digital Sky Data Retrieval average 3000 images a day Informix SUNs SRB SUN E10K HPSS …. 800 GB 10 TB SDSC IPAC CALTECH WEB JPL SUNsSGIs WEB

31 San Diego Supercomputer Center & NPACIAHM March 2002 Digital Sky Apps (Eg.: sky mosaic, classification, …) Processing 10 TB on thousands of nodes SRB SUN E15K HPSS …. 10 TB SDSC IBM SP2 (DTF) SAN disks, shared 10+TB

32 San Diego Supercomputer Center & NPACIAHM March 2002 DigSky Conclusion SRB can handle large number of files Metadata access is still less than ½ sec delay Replication of large collections Single command for geographical replication On-the-fly sorting (out-of-tape sorting) Availability of data otherwise not possible Near-line access to 5 million files (10 TB) Successfully used in web-access & large scale analysis (daily)

33 San Diego Supercomputer Center & NPACIAHM March 2002 Thank you for your attention. Any questions?


Download ppt "San Diego Supercomputer Center & NPACIAHM March 2002 Storage Resource Broker Case Studies George Kremenek"

Similar presentations


Ads by Google