Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop SKA Data Flow and Processing – a key SKA design driver Paul Alexander and Jaap.

Slides:



Advertisements
Similar presentations
What does LOFAR have to do with the Virtual Observatory (VO)? LOFAR Science Day 16 December 2003 Melbourne David Barnes The University of Melbourne.
Advertisements

The Australian Virtual Observatory e-Science Meeting School of Physics, March 2003 David Barnes.
SKADSMTAvA A. van Ardenne SKADS Coördinator ASTRON, P.O. Box 2, 7990 AA Dwingeloo The Netherlands SKADS; The.
SADC HPC Workshop, 2 Dec 2013, Cape Town
Paul Alexander DS3 & DS3-T3 SKADS Review 2006 DS3 The Network and its Output Data Paul Alexander.
Cape Town 2013 Dr Paul Calleja Director Cambridge HPC Service SKA The worlds largest Radio Telescope streaming data processor.
ASKAP Status Update CSIRO ASTRONOMY AND SPACE SCIENCE David Brodrick EPICS [Spring] Collaboration Meeting 22nd October 2014.
SA Bursary Conference December 2009 SKA Design. SA Bursary Conference December 2009 SKA Design Designing an optimal SKA around AAs at low frequencies.
HARP-B/ACSIS on the JCMT: spatially resolved chemistry of warm gas John Richer (Cavendish, Cambridge) On behalf of HARP and ACSIS teams.
ASKAP Central Processor: Design and Implementation Calibration and Imaging Workshop 2014 ASTRONOMY AND SPACE SCIENCE Ben Humphreys | ASKAP Software and.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Paul Alexander, Peter Hall Design Issues and Implementation ChallengesAAVP 2010 Design issues and implementation challenges Paul Alexander and Peter Hall.
Netherlands Institute for Radio Astronomy 1 ASTRON is part of the Netherlands Organisation for Scientific Research (NWO) From LOFAR design to SKA1 System.
NWO Meeting 24 May 04 a Virtual Survey System AstroWise- OmegaCAM – Lofar – Euro-VO a Virtual Survey System AstroWise- OmegaCAM – Lofar – Euro-VO Edwin.
Sascha D-PAD Sparse Aperture Array.
Signal Processing for Aperture Arrays. AAVS1 256 antenna elements distributed over –4 stations –64 elements each.
Performance of station array configurations Sparse vs. Dense, Regular vs Random Jaap D. Bregman AAVP Workshop,Cambridge,
1 Large Synoptic Survey Telescope Review Kirk Gilmore - SLAC DOE Review June 15, 2005.
Keith Grainge Calibration requirementsAAVP 2010 Calibration and Imaging Requirements Keith Grainge (plus teams at ASTRON, Oxford, Cambridge)
Paul Alexander Dynamic RangeAAVP 2010 Overview of Calibration and Dynamic Range Challenges Paul Alexander.
Big Data Imperial June 2013 Dr Paul Calleja Director HPCS The SKA The worlds largest big-data project.
Name1 SKA(DS) System Design Aspects 4 th SKADS Workshop, Lisbon, 2-3 October 2008 SKA(DS) System Design Aspects: building a system Laurens Bakker.
Steve Torchinsky SKADS Science Overview Lisbon 2007 Jan 12 SKADS Science Overview.
National Center for Supercomputing Applications Observational Astronomy NCSA projects radio astronomy: CARMA & SKA optical astronomy: DES & LSST access:
Phased Array Feeds John O’Sullivan SKANZ 2012 CSIRO Astronomy and Space Science,
Copyright, 1996 © Dale Carnegie & Associates, Inc. Tim Cornwell Kumar Golap Sanjay Bhatnagar Wide field imaging - computational challenges.
Controlling Field-of-View of Radio Arrays using Weighting Functions MIT Haystack FOV Group: Lynn D. Matthews,Colin Lonsdale, Roger Cappallo, Sheperd Doeleman,
1 The OSKAR Simulator (Version 2!) AAVP Workshop, ASTRON, 15 th December 2011 Fred Dulwich, Ben Mort, Stef Salvini.
Rosie Bolton1 SKADS Costing work 4 th SKADS Workshop, Lisbon, 2-3 October 2008 SKADS Costing work: Spreadsheets to scalable designs Rosie Bolton Dominic.
Andrew Faulkner1 SKADS and SKA 4 th SKADS Workshop, Lisbon SKADS and SKA Andrew Faulkner.
Design of a Software Correlator for the Phase I SKA Jongsoo Kim Cavendish Lab., Univ. of Cambridge & Korea Astronomy and Space Science Institute Collaborators:
3rd September 2006SKADS Workshop 2006Mike Garrett 1 DS2 - Science & Data Simulations Mike Garrett (JIVE)
Casper Signal Processing Workshop 2009 SKA Signal Processing (Preliminary) Wallace Turner Domain Specialist for Signal Processing.
CRISP & SKA WP19 Status. Overview Staffing SKA Preconstruction phase Tiered Data Delivery Infrastructure Prototype deployment.
Rosie Bolton SKADS Workshop April 10th 2008 SKADS Costing work: beyond the benchmark scenario Rosie Bolton Paul Alexander, Andy Faulkner and SKADS Costing.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Integration of Astro-WISE with Grid storage.
SKA System Design Kobus Cloete 9 December 2010 AAVP Workshop "Exploring the Universe with the world's largest radio telescope"
© ASTRON On the Fly LOFAR Station Correlator André W. Gunst.
Integrated receivers for mid-band SKA Suzy Jackson Engineer, Australia Telescope National Facility SKADS FP6 Meeting – Chateau de Limelette – 4-6 November,
Research Networks and Astronomy Richard Schilizzi Joint Institute for VLBI in Europe
Andrew Faulkner 2 nd SKADS Workshop October 2007 SKADS Benchmark Scenario Andrew Faulkner.
ASKAP Capabilities John Reynolds on behalf of the SEIC and ASKAP team.
● Radio telescope arrays – km diameter – Resolution arcmin to micro-arcsec at radio wavelengths ● Similar (baseline/ wavelength) for other regimes.
Large Aperture Experiment to Detect the Dark Ages (LEDA) Jonathon Kocz CfA LWA Users Meeting 26th-27th July 2012.
Large Area Surveys - I Large area surveys can answer fundamental questions about the distribution of gas in galaxy clusters, how gas cycles in and out.
ASKAP: Setting the scene Max Voronkov ASKAP Computing 23 rd August 2010.
Summary of the Paris SKA Meeting 4-8 September 2006 R. T. Schilizzi.
The Allen Telescope Array Douglas Bock Radio Astronomy Laboratory University of California, Berkeley Socorro, August 23, 2001.
Paul AlexanderSKA Computational Challenges Square Kilometre Array Computational Challenges Paul Alexander.
EURO-VO: GRID and VO Lofar Information System Design OmegaCEN Kapteyn Institute TARGET- Computing Center University Groningen Garching, 10 April 2008 Lofar.
Philippe Picard 2 nd SKADS Workshop October 2007 Station Processing Philippe Picard Observatoire de Paris Meudon, 11th October 2007.
Paul Alexander 2 nd SKADS Workshop October 2007 SKA and SKADS Costing The Future Paul Alexander Andrew Faulkner, Rosie Bolton.
Rosie Bolton 2 nd SKADS Workshop October 2007 SKADS System Design and Costing: Update and next steps Rosie Bolton University of Cambridge.
Gijs Verdoes Kleijn Edwin Valentijn Marjolein Cuppen for the Astro-WISE consortium.
Blurring boundaries data stewardship and observatory processing for the next generation of radio-telescopes Marco de Vos - ASTRON.
Square Kilometre Array eInfrastructure: Requirements, Planning, Future Directions Duncan Hall SPDO Software and Computing EGEE 2009.
Paul Alexander1 DS3 Deliverable Status 4 th SKADS Workshop, Lisbon, 2-3 October 2008 DS3 Deliverables Review.
E-MERLIN & Radio issues Simon Garrington Wide-fields (ie ~10x10 arcmin) uJy sensitivity; mas resolution key science driver for e-MERLIN ~1500 sources.
Netherlands Institute for Radio Astronomy 1 APERTIF beamformer and correlator requirements Laurens Bakker.
The Science Data Processor and Regional Centre Overview Paul Alexander UK Science Director the SKA Organisation Leader the Science Data Processor Consortium.
Korea Astronomy and Space Science Institute
The CSIRO ASKAP Science Data Archive Progress and Plans
Mid Frequency Aperture Arrays
Signal Processing for Aperture Arrays
Computing Architecture
EGEE NA4 Lofar Lofar Information System Design OmegaCEN
Imaging and Calibration Challenges
Aperture Array Station Processing
Technical Foundations & Enabling Technologies – DS4
SKADS White Paper - Technologies -
Presentation transcript:

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop SKA Data Flow and Processing – a key SKA design driver Paul Alexander and Jaap Bregman

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Overview We “know” that the SKA will produce vast amounts of data Following the flow of data and information is crucial to the overall system design of the SKA Can we process the data the SKA will produce How does the data processing tension the SKA design? This analysis based only on imaging requirements

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop.. Sparse AA Dense AA.. Mass Storage TimeStandard Central Processing Facility - CPF User interface via Internet... To 250 AA Stations... DSP... DSP To 2400 Dishes m Dishes Correlator – AA & Dish 16 Tb/s 80 Gb/s Control Processors & User interface Post Processor Data Time Control MHz Wide FoV GHz Wide FoV GHz Single Pixel or Focal plane array Tile & Station Processing SKA Overall System

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Fastest HPC today – Roadrunner ~ 1.2 PFlop Some numbers to remember Total internet data rate ~ 20 Tb/s BUT on your desktop you can have ~1 TFlop

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop CorrelatorVisibility processors switch AA: 250 x 16 Tb/s Dish: 3000 x 64 Gb/s Image formation Science analysis & archive 12-15m Dishes Sparse AA Data rates to the correlator Data rate from each collector G 1 = 2 N p  f N bit N b = 4  f N bit N b Dense AA Dishes G 1 ~ 64 Gbs PAFs FoV is constant across the band N b ~ 7 to give 20 sq-deg across the band G 1 ~ 60 Gb/s (Memo 100) AA, Number of elements N e ~ 65000; N b << N e limited by data rate 250 sq-deg across band N b ~ 1200 (Memo 100) G 1 ~ 16 Tb/s

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Data rates from the correlator Standard results for integration/dump time and channel width Naive data rate then given by Can reduce this using baseline-dependent integration times and channel widths

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Reduction in data rate using baseline-dependence (> 0.06) 1/3B = B 1 B >> B 1 Data rate out of correlator exceeds input data rate for 15-m dishes for baselines exceeding ~ 130km (36km if single integration time) At best for dishes output data rate ~ input; AA’s reduction by ~10 4 Data rates from the correlator

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Data rates from the correlator

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Processing model

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Post correlator Processing on the raw UV data is highly parallel UV data needs to be buffered based on current algorithms UV processing includes: –Gridding and de-gridding from current sky model –Peeling and subtraction of current sky model Image size: a 2 N ch (B/D) 2 N b Ratio UV to “image” data Major reduction in data rate occurs between UV data and image data

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop The SKA Processing Challenge CorrelatorVisibility processors switch AA: 250 x 16 Tb/s Dish: 3000 x 60 Gb/s Image formation Science analysis, user interface & archive ~10 PFlop~ ? PFlop Software complexity ~1 – 500 Tb/s ~ 200 Pflop to 2.5 Eflop

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Model for UV processor Highly parallel – consider something achievable – NVIDIA promises 20 TFlop in 2 years – assume 50 TFlop Approximate analysis of ops/sample: 20000/calibration loop Processor: €1000, 5 calibration loops, 50% efficiency, each processor processes ~ 1 GB/s of data Requirement: 100 PFlop (AA)2 EFlop (dishes) Buffer 1 hr of data therefore we need to buffer 7.2 TB in a fast store Assume €200 per TB. UV-processing cost:Cost = €2.5m (TB/s) AA< €10m Dishes ~ €125m

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop The SKA Processing Challenge CorrelatorVisibility processors switch AA: 250 x 16 Tb/s Dish: 3000 x 60 Gb/s Image formation Science analysis, user interface & archive ~10 PFlop~ ? PFlop Software complexity ~1 – 500 Tb/s ~ 200 Pflop to 2.5 Eflop

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Data Products ~0.5 – 10 PB/day of image data Source count ~10 6 sources per square degree ~10 10 sources in the accessible SKA sky, 10 4 numbers/record ~1 PB for the catalogued data 100 Pbytes – 3 EBytes / year of fully processed data

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Processing challenge depends critically on overall system and experiment For dishes, the data rate drops significantly only after the UV processor  For many experiments the data rate can increase through the correlator. Design of processing system needs a complete analysis of the data flow  Separate UV-processor and imaging processors  Processing UV data is a highly parallel excellent scaleability  Local buffer for UV processor Maximum baseline and number of correlated elements drive costs Processing is expensive, but achievable BUT must consider whole system – the processor MUST match the requirements of the system – PrepSKA Summary

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Non PDRA resources Data flow and overview Alexander, Diamond, Garrington, ASTRON (de Vos, de Bruyn?) Pipeline processing: ASTRON LOFAR team CASU – Cambridge Astronomical Survey Unit – Irwin TDP – to be determined Data products and VO UK VO team – Walton, Richards; Astrowise Groningen Hardware architecture and implementation Centre for scientific computing (Cambridge), Oxford e-Science centre TDP - Kembal

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop Data from the correlator Data rate depends on the experiment f = 1000 MHz. Minimum data rate for dishes  0.8 square degrees, m dishes out to 200 km baseline  f = 500MHz High data rate for Aperture Array  75 square degrees, m stations out to 200 km baseline  f = 500MHz >24 PB temporary buffer required

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop For each sample Processing Subtract N cs confusing bright sources from global sky model DB Grid sample allowing for sky curvature Form image FFT + deconvolution Effective ~20000 ops/sample using current algorithms Calibrate: solve for telescope errors Repeat N loop Update global sky model Store ~50TB/day of Observations

Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop.. Sparse AA Dense AA.. Mass Storage TimeStandard Central Processing Facility - CPF User interface via Internet... To 250 AA Stations... DSP... DSP To 2400 Dishes m Dishes Correlator – AA & Dish 16 Tb/s 80 Gb/s Control Processors & User interface Post Processor Data Time Control MHz Wide FoV GHz Wide FoV GHz Single Pixel or Focal plane array Beam Data Tile & Station Processing SKA Overall Structure