Low-Frequency Pulsar Surveys and Supercomputing Matthew Bailes.

Slides:



Advertisements
Similar presentations
E-VLBI down-under Tasso Tzioumis ATNF, CSIRO July 2005 Towards e-VLBI.
Advertisements

Ben Barsdell Matthew Bailes Christopher Fluke David Barnes.
E-VLBI progress in the South Tasso Tzioumis ATNF, CSIRO September 2006 “Towards e-VLBI”  “e-VLBI”
April 8/9, 2003 Green Bank GBT PTCS Conceptual Design Review John Ford August 5, 2008 CICADA Project The NRAO is operated for the National Science Foundation.
Baseband Processing and SKA Simulations using Supercomputers: Enhancing Australia‘s radio astronomy facilities and its bid for the SKA Steven Tingay Swinburne.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
ASKAP Central Processor: Design and Implementation Calibration and Imaging Workshop 2014 ASTRONOMY AND SPACE SCIENCE Ben Humphreys | ASKAP Software and.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
A bright millisecond radio burst of Extragalactic origin Duncan Lorimer, Matthew Bailes, Maura McLaughlin, David Narkevic and Froney Crawford Science (in.
Real Parallel Computers. Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra, Meuer, Simon Parallel.
Atacama Large Millimeter/submillimeter Array Expanded Very Large Array Robert C. Byrd Green Bank Telescope Very Long Baseline Array Digital Signal Processing.
Atacama Large Millimeter/submillimeter Array Expanded Very Large Array Robert C. Byrd Green Bank Telescope Very Long Baseline Array Spectrometer PDR John.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
HPCC Mid-Morning Break Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery Introduction to the new GPU (GFX) cluster.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Current LBA Developments Chris Phillips CSIRO ATNF 13/7/2005.
DiFX: Software Correlation at Swinburne for the LBA Adam Deller Swinburne University/CSIRO Australia Telescope National Facility Supervisors: A/Prof Steven.
Backend electronics for radioastronomy G. Comoretto.
DELL PowerEdge 6800 performance for MR study Alexander Molodozhentsev KEK for RCS-MR group meeting November 29, 2005.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
Recent results from the Swinburne supercomputer VLBI software correlator(s) Steven Tingay, Craig West, Adam Deller, Shinji Horiuchi Centre for Astrophysics.
Atacama Large Millimeter/submillimeter Array Expanded Very Large Array Robert C. Byrd Green Bank Telescope Very Long Baseline Array New VLBA capabilities.
Hardware Trends. Contents Memory Hard Disks Processors Network Accessories Future.
Design of a Software Correlator for the Phase I SKA Jongsoo Kim Cavendish Lab., Univ. of Cambridge & Korea Astronomy and Space Science Institute Collaborators:
Paul Alexander & Jaap BregmanProcessing challenge SKADS Wide-field workshop SKA Data Flow and Processing – a key SKA design driver Paul Alexander and Jaap.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Next Generation Digital Back-ends at the GMRT Yashwant Gupta Yashwant Gupta National Centre for Radio Astrophysics Pune India CASPER meeting Cambridge.
Pulsar Searching + Receivers Matthew Bailes (CAASTRO), FPGAs: Peter Mc Mahon, Terry Fibila, Jonathon Kocz. CPUs: Willem van Straten, Andrew Jameson, Mike.
APSR: digital signal processing at Parkes Willem van Straten, Andrew Jameson and Matthew Bailes Centre for Astrophysics & Supercomputing Third ATNF Gravitational.
GPU DAS CSIRO ASTRONOMY AND SPACE SCIENCE Chris Phillips 23 th October 2012.
Pulsar search and timing Pulsar search and timing 22/10/2011 INDIGO Bhal Chandra Joshi Bhal Chandra Joshi.
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
2009/4/21 Third French-Japanese PAAP Workshop 1 A Volumetric 3-D FFT on Clusters of Multi-Core Processors Daisuke Takahashi University of Tsukuba, Japan.
Astrophysical Applications on Superclusters Matthew Bailes Swinburne Centre for Astrophysics and Supercomputing.
DiFX Performance Testing Chris Phillips eVLBI Project Scientist 25 June 2009.
ASKAP Capabilities John Reynolds on behalf of the SEIC and ASKAP team.
Pulsar surveys at Arecibo and Green Bank David Champion Gravity Wave Meeting, Marsfield, Dec 2007.
An FX software correlator for VLBI Adam Deller Swinburne University Australia Telescope National Facility (ATNF)
GSB : A real-time Software back-end for the GMRT Jayanta Roy National Centre for Radio Astrophysics Pune, India 12 th December 2008 Collaborators.
Ramesh Bhat Centre for Astrophysics & Supercomputing Swinburne University of Technology Time Domain Astronomy Meeting, Marsfield, 24 October 2011 Searching.
News from Alberto et al. Fibers document separated from the rest of the computing resources
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
Short introduction Pulsar Parkes. Outline PDFB – Single beam pulsar timing system CASPER – Single beam pulsar coherent dedispersion system.
Distributed FX software correlation Adam Deller Swinburne University/CSIRO Australia Telescope National Facility Supervisors: A/Prof Steven Tingay, Prof.
Pulsar Data analysis Desh, Anish RRI MWA Meeting Canberra, January
Review of developments in Australasia and mainland Asia Steven Tingay Swinburne University of Technology Next Generation correlator meeting, JIVE 27 -
A real-time software backend for the GMRT : towards hybrid backends CASPER meeting Capetown 30th September 2009 Collaborators : Jayanta Roy (NCRA) Yashwant.
APSR Matthew Bailes Swinburne University Of Technology.
Coherent Dedispersion Pulsar Timing Machines Matthew Bailes + Swinburne, Caltech, ATNF, CASPER.
The Green Bank Telescope Frank Ghigo, National Radio Astronomy Observatory 7 th US VLBI Technical Meeting, Haystack, Nov 2009.
ATCA GPU Correlator Strawman Design ASTRONOMY AND SPACE SCIENCE Chris Phillips | LBA Lead Scientist 17 November 2015.
April 8/9, 2003 Green Bank GBT PTCS Conceptual Design Review John Ford September 28, 2009 Casper Instrumentation at Green Bank The NRAO is operated for.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
1 Next Generation Correlators, June 26 th −29 th, 2006 The LOFAR Blue Gene/L Correlator Stichting ASTRON (Netherlands Foundation for Research in Astronomy)
Square Kilometre Array eInfrastructure: Requirements, Planning, Future Directions Duncan Hall SPDO Software and Computing EGEE 2009.
VLBI Developments in Australia ASTRONOMY AND SPACE SCIENCE Chris Phillips| LBA Lead Scientist 11 November 2014.
Hardware Architecture
Sobolev(+Node 6, 7) Showcase +K20m GPU Accelerator.
Brief introduction about “Grid at LNS”
NFV Compute Acceleration APIs and Evaluation
Korea Astronomy and Space Science Institute
Virtualization OVERVIEW
ASKAP High Time Res Capabilities
Y.Chen1, B.J.Xiao1, 2, F.Wang1, S.Li1, F.Yang1, 3
Pulsar Timing with ASKAP Simon Johnston ATNF, CSIRO
The High Time Resolution Universe Survey Backend
The Uniboard  FPGA Processing for Astronomy
Presentation transcript:

Low-Frequency Pulsar Surveys and Supercomputing Matthew Bailes

Outline: Baseband Instrumentation MultiBOB MWA survey vs PKSMB survey Data rates CPU times Low-Frequency Pulsar Monitoring The Future Supercomputers

Pulsar “Dedispersion” Incoherent

Coherent Dedispersion Unresolved on us timescales From young or millisecond pulsars Power-law distribution of energies PSR J

Pulsar Timing (Kramer et al.)

CPSR2 Timing (Hotan, Bailes & Ord)

Swinburne Baseband Recorders etc 1998: Canadian S2 to computer (16 MHz x 2)  100K system + video tapes 2000: CPSR  20 MHz x 2 + DLT7000 drives x : CPSR2  128 MHz x 2 + real-time supercomputer (60 cores) 2006: DiFX (Deller, Tingay, Bailes & West)  Software Correlator (ATNF adopted) 2007: APSR  1024 MHz x 2 + real-time supercomputer (160 cores) 2008: MultiBOB  13 x 1024 ch x 64us + fibre core supercomputer

dspsr software Mature Delivers < 100 ns timing on selected pulsars Total power estimation every 8us with RFI excision Write a “loader” Can do:  Giant pulse work  Pulsar searching (coherent filterbanks)  Pulsar timing/polarimetry  Interferometry with pulsar gating

PSRDADA (van Straten) psrdada.sourceforge.net Generic UDP data capture system (APSR/MultiBOB) Ring Buffer(s)  Can attach threads to fold/dedisperse etc  Hierachical buffers  Shares available CPU resources/disk  Web-based control/monitoring Free! + hooks to dspsr & psrchive.

APSR Takes 8 Gb/s voltages Forms:  16 x 128 channels (with coherent dedispersion)  4 Stokes, umpteen pulsars  Real-time fold to DM=250 pc/cc. O(100) Ops/sample  Sustaining >>100 Gflops ~100K computers. June MHz 4bits 768 MHz 2bits

Coherent Dedispersion BW/time x x x x (100K) (300K) BW year

Coherent Dedispersion Now “trivial” FFT ease ~ B -2 / 3

MultiBOB High Resolution Universe Survey (PALFA of the South) Werthimer’s iBOB boards  1024 channels, down to 10us sampling  Two pols FPGA coding hard…  Use software gain equalizer/summer ~5 MB/s beam 1 Gb/s Fibre to Swinburne (>1000 km fibre) Real time searching!

New PKS MB Survey: Bailes 13 beams 9 minutes/pointing 1024 channels 300 MHz BW 64 us sampling +/- 15 deg Kramer 13 beams 70 minutes/pointing 1024 channels 300 MHz BW 64 us sampling +/- 3.5 deg Johnston 13 beams 4.5 minutes/pointing 1024 channels 300 MHz BW 32 us sampling The rest

MWA Samples  Takes (24x1.3MHz=32 MHz) x 2 x 512  “Just” 32 GB/s (64 Gsamples/s) FFTs it  (5 N log2 ops/pt = 2.2 Tflops) XMultiplies & adds  (512)*256*B*4 = 16 TMACs

Sensitivity: ~3-5x PKS 32 vs 288 MHz 350 vs 25 K 700 vs 0.6 deg 2 (folded factor)

PKS vs MWA G ~ 3-5 x better T sys ~ 14 x worse ? B 1/2 ~ 3 x worse Flux ~ 25 x better (1400 vs 200 MHz) t 1/2 ~ 32 x better ~ Parity Single Pulse work ~ Comparable Coherent search ~ 32x improvement! But: There is a limit to the time you can observe a pulsar! 4m vs 144m -> 5x deeper.

Scattering b=0 1,10,100,1000ms

Scattering b=5d 1,10,50,100ms

b=30 0.5,1ms

36 GB/s Search instrumentation? 32 MHz... FX GB/s 5 bits x 512 Grid... 2D FFT Volts SpectraVisibilities uv FBanks Dedisp... Spectra FFT Fold Pulsars <1 bit/s 200 GB/s 32 bits x GB/s 32 bits x 512 x GB/s x GB/s Correlator Us ? ?

Search Timings 36,000 “coherent beams” (768m/4m=192) 2 36 gigapixels/s Dedisperse/CPU core  Gigapixel/120s  36 x 120 = 4320 cores = 500 machines = 250 kW N FFT = 36,000 * 1024 (DMs)/8192 = 4608 FFTs/sec Seek (3s / 8192 x 1024 pt FFT)  14,000 cores ~ 1800 machines = MW. (M$/yr)

Swinburne The Green Machine  installed May/June 2007  185 Dell PowerEdge1950 nodes  2 quad-core processors (Clovertown: Intel Xeon 64-bit 2.33 GHz)  16GB RAM  1TB disk -> 300 TB total  1640 cores/14 Tflops  dual channel gigabit ethernet  CentOS Linux OS  job queue submission  20 Gb infiniband (Q1 2008)  83 kW.vs. 130 kW cooling Machines: ~1.2M Fuel: ~100K/yr

Search Times: Depend only upon:  Npixels x Nchans x Tsamp -1 Requires:  No acceleration trials PSR J  In 8192s, small width from acceleration

Search Timings (32x32 tiles) >1024 “coherent beams” 36->1 gigapixels/s Dedisperse/core  Gigapixel/120s  120 = 120 cores = 15 machines = 7 kW N FFT = 1024 * 1024 (DMs)/8192(s/FFT) = 128 FFTs/sec Seek (3s / (8192 x 1024) pt FFT)  378 cores ~ 50 machines = 25 kW.

RRATs Log N - Log S (helps with long pointings…) 1000 x integration time. Maybe good RRAT finder.

Monitoring: Monitoring?

Monitoring:

Build Your Own Telescope? May be cheaper to build dedicated PSR telescope than attempt to process everything from existing telescopes! 32x32 tile: (2D FFT - 1D FFT - dedisperse - FFT)  ~2M telescopes  ~2M “beamformer/receivers”  ~1M correlator  ~1M Supercomputer  ~1M construction  ~7-8M

Next-Gen Supercomputers (IO or Tflops?) Infiniband 20 Gb (40Gb)  288 port switch  ~10 Tb/s IO Capacity (1-2K/node) Teraflop CPU capacities/node (140 Gflops now) Teraflop Server or Tflop GPU?  10 GB/s vs 76 GB/s Power (0.1W/$)  2M = 200 kW

Architecture (2011??): 288 Ports 40 Gb/s 288 Ports 40 Gb/s 144 Tflops 300K ~1M FX

Summary: Strong motivation for multiple (~100) tied array beams  PSRs/deg^2 Surveys only possible with compact configurations  At present Future Supercomputers may allow search even with MWA-like telescopes