Why Build Image Mosaics for Wide Area Surveys? An All-Sky 2MASS Mosaic Constructed on the TeraGrid A. C. Laity, G. B. Berriman, J. C. Good (IPAC, Caltech);

Slides:



Advertisements
Similar presentations
3 September 2004NVO Coordination Meeting1 Grid-Technologies NVO and the Grid Reagan W. Moore George Kremenek Leesa Brieger Ewa Deelman Roy Williams John.
Advertisements

Aug 3-4, 2003 Victoria1 Montage, Pegasus and ROME G. B. Berriman, J.C. Good, M. Kong, A. Laity IPAC/Caltech J. Jacob, A. Bergou, D. S. Katz JPL R. Williams.
National Partnership for Advanced Computational Infrastructure San Diego Supercomputer Center Data Grids for Collection Federation Reagan W. Moore University.
Grid Astronomy with Image Federation Roy Williams Michael Feldmann California Institute of Technology.
The ADAMANT Project: Linking Scientific Workflows and Networks “Adaptive Data-Aware Multi-Domain Application Network Topologies” Ilia Baldine, Charles.
Optimal Photometry of Faint Galaxies Kenneth M. Lanzetta Stony Brook University.
GGF-17 Astro Workshop Preservation Environment Working Group Officers: Bruce Barkstrom (NASA Langley) Reagan Moore (SDSC) Goals  Demonstrate.
Grist: Grid Data Mining for Astronomy Joseph Jacob, Daniel Katz, Craig Miller, and Harshpreet Walia (JPL) Roy Williams, Matthew Graham, and Michael Aivazis.
Applying Data Grids to Support Distributed Data Management Storage Resource Broker Reagan W. Moore Ian Fisk Bing Zhu University of California, San Diego.
A Web service for Distributed Covariance Computation on Astronomy Catalogs Presented by Haimonti Dutta CMSC 691D.
Science Applications of Montage: An Astronomical Image Mosaic Engine ESTO G. Bruce Berriman, John Good and Anastasia Laity.
The Montage Image Mosaic Service: Custom Mosaics on Demand ESTO John Good, Bruce Berriman Mihseh Kong and Anastasia Laity IPAC, Caltech
Montage: Experiences in Astronomical Image Mosaicking on the TeraGrid ESTO.
Pegasus: Mapping complex applications onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
An Astronomical Image Mosaic Service for the National Virtual Observatory
A Grid-Enabled Engine for Delivering Custom Science- Grade Images on Demand
An Astronomical Image Mosaic Service for the National Virtual Observatory / ESTO.
Introduction to Spitzer and some applications Data products Pipelines Preliminary work K. Nilsson, J.M. Castro Cerón, J.P.U. Fynbo, D.J. Watson, J. Hjorth.
Ewa Deelman Using Grid Technologies to Support Large-Scale Astronomy Applications Ewa Deelman Center for Grid Technologies USC Information.
The Use of Infrared Color-Color Plots to Identify Rare Objects in the Galactic Mid-Plane Jessica Fuselier Dr. Robert Benjamin, advisor.
Astronomical Image Mosaicking ESTO.
CREATING A MULTI-WAVELENGTH GALACTIC PLANE ATLAS WITH AMAZON WEB SERVICES G. Bruce Berriman, John Good IPAC, California Institute of Technolog y Ewa Deelman,
UNIVERSITY of MARYLAND GLOBAL LAND COVER FACILITY High Performance Computing in Support of Geospatial Information Discovery and Mining Joseph JaJa Institute.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Introduction The réseau grid of 27 horizontal and 27 vertical lines was originally superimposed over Carte du Ciel (CdC) plates to assist the process of.
Digitized Sky Survey Update Brian McLean : Archive Sciences Branch / Operations and Engineering Division.
Pegasus A Framework for Workflow Planning on the Grid Ewa Deelman USC Information Sciences Institute Pegasus Acknowledgments: Carl Kesselman, Gaurang Mehta,
Supported by the National Science Foundation’s Information Technology Research Program under Cooperative Agreement AST with The Johns Hopkins University.
San Diego Supercomputer CenterUniversity of California, San Diego Preservation Research Roadmap Reagan W. Moore San Diego Supercomputer Center
MASSACHUSETTS INSTITUTE OF TECHNOLOGY NASA GODDARD SPACE FLIGHT CENTER ORBITAL SCIENCES CORPORATION NASA AMES RESEARCH CENTER SPACE TELESCOPE SCIENCE INSTITUTE.
Appraisal and Data Mining of Large Size Complex Documents Rob Kooper, William McFadden and Peter Bajcsy National Center for Supercomputing Applications.
Innovations in the Multimission Archive at STScI (MAST) M. Corbin, M. Donahue, C. Imhoff, T. Kimball, K. Levay, P. Padovani, M. Postman, M. Smith, R. Thompson.
Large-Scale Science Through Workflow Management Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
Semantic Grid + Data Federation US National Virtual Observatory Roy Williams California Institute of Technology NVO co-director.
Pegasus: Planning for Execution in Grids Ewa Deelman Information Sciences Institute University of Southern California.
Pegasus: Mapping Scientific Workflows onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
Ten Years of Software Sustainability at The Infrared Processing and Analysis Center G. Bruce Berriman and John Good NASA Exoplanet Science Institute, Infrared.
Wiss. Beirat AIP, ClusterFinder & VO-Methods H. Enke German Astrophysical Virtual Observatory ClusterFinder VO Methods for Astronomical Applications.
Kurt Mueller San Diego Supercomputer Center NPACI HotPage Updates.
Rule-Based Preservation Systems Reagan W. Moore Wayne Schroeder Mike Wan Arcot Rajasekar Richard Marciano {moore, schroede, mwan, sekar,
National Partnership for Advanced Computational Infrastructure San Diego Supercomputer Center Persistent Management of Distributed Data Reagan W. Moore.
Pegasus: Running Large-Scale Scientific Workflows on the TeraGrid Ewa Deelman USC Information Sciences Institute
Common Archive Observation Model (CAOM) What is it and why does JWST care?
Pegasus: Mapping complex applications onto the Grid Ewa Deelman Center for Grid Technologies USC Information Sciences Institute.
Hyperatlas Coregistered Federated Imagery Roy Williams Bruce Berriman George Djorgovski John Good Reagan Moore Caltech CACR Caltech IPAC Caltech Astronomy.
1 The DES Calibrations Effort Douglas L. Tucker (DES Calibrations Scientist) DES NSF/DOE Review, 8-9 June 2009 The DES Calibrations Effort has connections.
The Global Land Cover Facility is sponsored by NASA and the University of Maryland.The GLCF is a founding member of the Federation of Earth Science Information.
The COMPASS (Catalogs of Objects and Measure Parameters for All Sky Surveys) Database Overview Gretchen Greene, Brian McLean, David Wolfe, and Charles.
Data Archives: Migration and Maintenance Douglas J. Mink Telescope Data Center Smithsonian Astrophysical Observatory NSF
Large Area Surveys - I Large area surveys can answer fundamental questions about the distribution of gas in galaxy clusters, how gas cycles in and out.
Virtual Observatories, Astronomical Telescopes and Instrumentation (August 2002)1 An Architecture for Access To A Compute Intensive Image Mosaic Service.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Kevin Cooke.  Galaxy Characteristics and Importance  Sloan Digital Sky Survey: What is it?  IRAF: Uses and advantages/disadvantages ◦ Fits files? 
Pegasus: Planning for Execution in Grids Ewa Deelman, Carl Kesselman, Gaurang Mehta, Gurmeet Singh, Karan Vahi Information Sciences Institute University.
Super Computing 2000 DOE SCIENCE ON THE GRID Storage Resource Management For the Earth Science Grid Scientific Data Management Research Group NERSC, LBNL.
Tackling I/O Issues 1 David Race 16 March 2010.
GRITS II, IPAC 11 Jun 2010VO Inventory Service1 Virtual Observatory Inventory Service VO I NVENTORY S ERVICE John Good VAO System Engineer IPAC / Caltech.
SAN DIEGO SUPERCOMPUTER CENTER Replication Policies for Federated Digital Repositories Robert H. McDonald Chronopolis Project Manager
Wide-field Infrared Survey Explorer (WISE) is a NASA infrared- wavelength astronomical space telescope launched on December 14, 2009 It’s an Earth-orbiting.
1 Performance Impact of Resource Provisioning on Workflows Gurmeet Singh, Carl Kesselman and Ewa Deelman Information Science Institute University of Southern.
Data Grids, Digital Libraries and Persistent Archives: An Integrated Approach to Publishing, Sharing and Archiving Data. Written By: R. Moore, A. Rajasekar,
Pegasus WMS Extends DAGMan to the grid world
Cloudy Skies: Astronomy and Utility Computing
Montage: An On-Demand Image Mosaic Service for the NVO
Pegasus and Condor Gaurang Mehta, Ewa Deelman, Carl Kesselman, Karan Vahi Center For Grid Technologies USC/ISI.
COMPASS Database SPACE TELESCOPE SCIENCE INSTITUTE Gretchen Greene
TeraScale Supernova Initiative
Pegasus Workflows on XSEDE
High Throughput Computing for Astronomers
Frieda meets Pegasus-WMS
Presentation transcript:

Why Build Image Mosaics for Wide Area Surveys? An All-Sky 2MASS Mosaic Constructed on the TeraGrid A. C. Laity, G. B. Berriman, J. C. Good (IPAC, Caltech); D. S. Katz, J. C. Jacob (JPL); L. Brieger, R. W. Moore (SDSC); R. Williams (Caltech); E. Deelman, G. Singh, M.-H. Su (ISI) Acknowledgements: This work made use of Montage, funded by the National Aeronautics and Space Administration's Earth Science Technology Office, Computational Technologies Project, under Cooperative Agreement Number NCC5-626 between NASA and the California Institute of Technology.“ The 2MASS project is a collaboration between The University of Massachusetts and the Infrared Processing and Analysis Center (JPL/ Caltech). Funding is provided primarily by NASA and the NSF. 2MASS image data products generated by Montage are not endorsed by the 2MASS project. Part of the work was done under funding support from the National Science Foundation “National Virtual Observatory”. Visualize the content and observational coverage of a wide area Science question: I am planning an observing proposal of a cluster of galaxies. What observations have been made in this area? Superimpose footprints of observations on a mosaic of the cluster Science Question: I am studying a large molecular cloud. Are there candidate sources not recorded in individual source catalogs? Superimpose locations of sources from existing catalogs on mosaics from surveys at different wavelengths Examine multi-wavelength mosaics for new candidate sources seen in at least two wavelengths Science Question: Have I found a high proper motion object or an artifact in an image? Is it a T-dwarf candidate or is it in a region of higher than average extinction? Examine mosaics from 2MASS, SDSS and DPOSS, all at different epochs Examine mosaics at different wavelengths to study distribution of color across images. Quality assurance of wide area imaging surveys Artifacts present over spatial scales of several degrees Diffraction spikes, ghost images, and filter and dichroic glints Wide area mosaics expedite and simplify identification of artifacts Computation of an All-Sky 2MASS Mosaic To provide a “value-added” 2MASS product that combines overlapping images to improve sensitivity; To demonstrate applicability of computing at-scale to astronomical missions and surveys, especially projects such as LSST; and To demonstrate the utility of the NVO Hyperatlas format. The Montage image mosaic toolkit: Highly parallelizable Processes FITS images in any WCS projection Reprojects images while preserving flux and astrometry Rectifies variable sky emission to a common level Coadds reprojected and background-rectified images Visit Processing Summary: Input files: - The 2MASS All Sky Survey Atlas Image Data Set. 4,121,440 files in the J, H and Ks bands, SIN projection, each 2 MB in size (32-bit). All-sky mosaic performed on the Distributed Terascale Facility node at San Diego Supercomputer Center (SDSC) ( Parallelization of Montage: used the Message Passing Interface. Images distributed to multiple nodes of a cluster for parallel steps. - The input files that fell on mosaic boundaries were opened, read, and used multiple times in the processing of adjacent mosaics, so that a total of 14 TB in 6,275,494 files were actually opened and read. - processing time to complete the mosaics was more than 100,000 CPU-hours on a Teragrid Linux cluster (1.5 Ghz Intel Itanium 2 processors). Organization and Storage Of All-Sky Mosaic: The total size over all three bands is 10 TB (64-bit). Output consists of degree plates per band, 5205 plates total, 2 GB in size, each tiled into a 12x12 array of 13-MB files for ease of handling. There are 750,000 files total. Output organized in National Virtual Observatory (NVO) Hyperatlas structure ( implemented using the SDSC Storage Resource Broker data grid ( The Hyperatlas organizes images in an Atlas consisting of multiple pages that are standard WCS sphere-to-plane projection mappings defined by a pointing center, projection type, rotation, and scale. The All-Sky Mosaic uses the TM-5-TAN-20 Atlas of 1734 pages covering the celestial sphere, with every point of the sphere at most 3.5 degrees (= 5/√2) from a pointing center, a scale of 1 arcsecond per pixel, and the WCS projection TAN. Processing and Data Management: The entire 2MASS image archive was replicated onto Teragrid disk. An archive of the images was created on the IBM HPSS High Performance Storage System after aggregating the images into 147,000 containers to minimize the impact on the HPSS name space. Checksums were used to validate the integrity of the images. Continued validation was essential to recognize and correct problems with data transport, RAID controllers, and media failure. The computation was decomposed into hour runs on 128 processors. The mosaic process generated 64-bit image files, resulting in 20 TB of output data. The output images were truncated back to 32-bit image files, registered into the Storage Resource Broker, and replicated into the IBM HPSS archive. Current Status: Will process second pass with addition of 1.8-TB unpublished 2MASS survey data Validation will compare source-extraction on mosaic with source extraction on original images Product will be made accessible to astronomers Goals of the 2MASS All-Sky Mosaic Project The Montage Mosaic Engine 3-color mosaic centered on Wolf 359, showing artifacts 3-color 2MASS mosaic of a section of the Galactic Plane covering l = º, b= ± 4º