Presentation is loading. Please wait.

Presentation is loading. Please wait.

Cape Town 2013 Dr Paul Calleja Director Cambridge HPC Service SKA The worlds largest Radio Telescope streaming data processor.

Similar presentations


Presentation on theme: "Cape Town 2013 Dr Paul Calleja Director Cambridge HPC Service SKA The worlds largest Radio Telescope streaming data processor."— Presentation transcript:

1 Cape Town 2013 Dr Paul Calleja Director Cambridge HPC Service SKA The worlds largest Radio Telescope streaming data processor

2 Cape Town 2013 Introduction to Cambridge HPCS Overview of the SKA project SKA streaming data processing challenge The SKA SDP consortium Overview

3 Cape Town 2013 Cambridge University The University of Cambridge is a world leading teaching & research institution, consistently ranked within the top 3 Universities world wide Annual income of £1200M - 40% is research related - one of the largest R&D budgets within the UK HE sector students, 9,000 staff Cambridge is a major technology centre –1535 technology companies in surrounding science parks –£12B annual revenue –53000 staff The HPCS has a mandate to provide HPC services to both the University and wider technology company community

4 Cape Town 2013 Four domains of activity Dell HPC Solution Centre Industrial HPC Service Cambridge HPC Service Commodity HPC Centre of Excellence Promoting uptake of HPC by UK Industry Driving Discovery Advancing development and application of HPC HPC R& D

5 Cape Town registered users from 31 departments 856 Dell Servers TF sustained DP performance 128 node Westmere (1536 cores) (16 TF) 600 node (9600 core) full non blocking Mellanox FDR IB 2,6 GHz sandy bridge (200 TF) one of the fastest Intel clusters in he UK SKA GPU test bed -128 node 256 card NVIDIA K20 GPU Fastest GPU system in UK 250 TF Designed for maximum I/O throughput and message rate Full non blocking Dual rail Mellanox FDR Connect IB Design for maximum energy efficiency 2 in Green500 Most efficient air cooled supercomputer in the world 4 PB storage – Lustre parallel file system 50GB/s Run as a cost centre – charges our users – 20% income from industry Cambridge HPC vital statistics

6 Cape Town 2013 CORE – Industrial HPC service & consultancy

7 Cape Town 2013 Dell | Cambridge HPC Solution Centre The Solution Centre is a Dell Cambridge joint funded HPC centre of excellence, provide leading edge commodity open source HPC solutions.

8 Cape Town 2013 SA CHPC collaboration HPCS has a long term strategic partnership with CHPC HPCS has been working closely with CHPC for last 6 years Technology strategy, system design procurement HPC system stack development SKA platform development

9 Cape Town 2013 Next generation radio telescope Large multi national Project 100 x more sensitive X faster 5 square km of dish over 3000 km The next big science project Currently the worlds most ambitious IT Project First real exascale ready application Largest global big-data challenge Square Kilometre Array - SKA

10 Cape Town 2013 SKA location Needs a radio-quiet site Very low population density Large amount of space Two sites: Western Australia Karoo Desert RSA A Continental sized Radio Telescope

11 Cape Town 2013 SKA phase 1 implementation SKA1_Low SKA1_Mid incl MeerKAT SKA ElementLocation Dish ArraySKA1_MidRSA Low Frequency Aperture ArraySKA1_LowANZ Survey InstrumentSKA1_AIP_SurveyANZ SKA1_AIP_Survey incl ASKAP +

12 Cape Town 2013 SKA phase 2 implementation SKA2_Low SKA2_Mid_Dish SKA2_AIP_AA SKA ElementLocation Low Frequency Aperture ArraySKA2_LowANZ Mid Frequency Dish ArraySKA2_Mid_DishRSA Mid Frequency Aperture ArraySKA2_Mid_AARSA

13 Cape Town 2013 What is radio astronomy XXXXXX SKY Image Detect & amplify Digitise & delay Correlate Process Calibrate, grid, FFT Integrate s B 12 Astronomical signal (EM wave)

14 Cape Town 2013 SKA – Key scientific drivers Cradle of life Cosmic Magnetism Evolution of galaxies Pulsar survey gravity waves Exploring the dark ages

15 Cape Town 2013 SKA is a cosmic time machine

16 Cape Town 2013 But…… Most importantly the SKA will investigate phenomena we have not even imagined yet Most importantly the SKA will investigate phenomena we have not even imagined yet

17 Cape Town 2013 SKA timeline 2022Operations SKA : Operations SKA Construction of Full SKA, SKA 2 €2 B % SKA construction, SKA 1 €650M 2012Site selection Pre-Construction: 1 yr Detailed design€90M PEP 3 yr Production Readiness System design and refinement of specification Initial concepts stage Preliminary ideas and R&D

18 Cape Town 2013 SKA project structure SKA Board Director General Work Package Consortium 1 Work Package Consortium n Advisory Committees (Science, Engineering, Finance, Funding …) Advisory Committees (Science, Engineering, Finance, Funding …) … … Project Office (OSKAO) Locally funded

19 Cape Town 2013 Work package breakdown UK (lead), AU (CSIRO…), NL (ASTRON…) South Africa SKA, Industry (Intel, IBM…) UK (lead), AU (CSIRO…), NL (ASTRON…) South Africa SKA, Industry (Intel, IBM…) 1.System 2.Science 3.Maintenance and support /Operations Plan 4.Site preparation 5.Dishes 6.Aperture arrays 7.Signal transport 8.Data networks 9.Signal processing 10.Science Data Processor 11.Monitor and Control 12. Power SPO

20 Cape Town 2013 SKA = Streaming data processor Challenge The SDP consortium led by Paul Alexander University of Cambridge 3 year design phase has now started (as of November 2013) To deliver SKA ICT infrastructure need a strong multi-disciplinary team Radio astronomy expertise HPC expertise (scalable software implementations; management) HPC hardware (heterogeneous processors; interconnects; storage) Delivery of data to users (cloud; UI …) Building a broad global consortium: 11 countries: UK, USA, AUS, NZ, Canada, NL, Germany, China, France, Spain, South Korea Radio astronomy observatories; HPC centres; Multi-national ICT companies; sub-contractors

21 Cape Town 2013 SDP consortium members Management GroupingsWorkshare (%) University of Cambridge (Astrophysics & HPFCS) 9.15 Netherlands Institute for Radio Astronomy 9.25 International Centre for Radio Astronomy Research 8.35 SKA South Africa / CHPC 8.15 STFC Laboratories 4.05 Non-Imaging Processing Team 6.95 University of Manchester Max-Planck-Institut für Radioastronomie University of Oxford (Physics) University of Oxford (OeRC) 4.85 Chinese Universities Collaboration 5.85 New Zealand Universities Collaboration 3.55 Canadian Collaboration Forschungszentrum Jülich 2.95 Centre for High Performance Computing South Africa 3.95 iVEC Australia (Pawsey) 1.85 Centro Nacional de Supercomputación 2.25 Fundación Centro de Supercomputación de Castilla y León 1.85 Instituto de Telecomunicações 3.95 University of Southampton 2.35 University College London 2.35 University of Melbourne 1.85 French Universities Collaboration 1.85 Universidad de Chile 1.85

22 Cape Town 2013 SDP –strong industrial partnership Discussions under way with DelI, NVIDIA, Intel, HP IBM, SGI, l, ARM, Microsoft Research Xyratex, Mellanox, Cray, DDN NAG, Cambridge Consultants, Parallel Scientific Amazon, Bull, AMD, Altera, Solar flare, Geomerics, Samsung, CISCO Apologies to those I’ve forgotten to list

23 Cape Town 2013 SDP work packages

24 Cape Town 2013 SKA data rates 16 Tb/s4 Pb/s 24 Tb/s 20 Gb/s 1000Tb/s

25 Cape Town 2013 SKA conceptual data flow

26 Cape Town 2013 SKA conceptual data flow

27 Cape Town 2013 Science data processor pipeline 10 Pflop 1 Eflop 100 Pflop Software complexity 10 Tb/s 200 Pflop 10 Eflop … Incoming Data from collectors Switch Buffer store Switch Buffer store Bulk Store Correlator Beamformer UV Processor Imaging: Non-Imaging: Corner Turning Course Delays Fine F-step/ Correlation Visibility Steering Observation Buffer Gridding Visibilities Imaging Image Storage Corner Turning Course Delays Beamforming/ De-dispersion Beam Steering Observation Buffer Time-series Searching Search analysis Object/timing Storage HPC science processing Image Processor 1000Tb/s 1 Eflop 10 EB/y SKA 2 SKA 1 1 EB/y 10 Tb/s 50 PB 10/1 TB/s

28 Cape Town 2013 SDP processing rack – feasibility model Host processor Multi-core X86 M-Core - >10TFLOP/s To rack switches Disk 1 ≥1TB 56Gb/s PCI Bus Disk 2 ≥1TB Disk 3 ≥1TB Disk 4 ≥1TB Processing blade 1 Processing blade 2 Processing blade 3 Processing blade 4 Processing blade 5 Processing blade 6 Processing blade 7 Processing blade 8 Processing blade 9 Processing blade 10 Processing blade 11 Processing blade 12 Processing blade 13 Processing blade 14 Processing blade 15 Processing blade 16 Processing blade 17 Processing blade 18 Processing blade 19 Processing blade 20 Leaf Switch-1 56Gb/s Leaf Switch-2 56Gb/s 42U Rack Processing Blade: GGPU, MIC,…? 20 TFlop 2x56 Gb/s comms 4 TB storage <1kW power Capable host (dual Xeon) Programmable Significant RAM 20 TFlop 2x56 Gb/s comms 4 TB storage <1kW power Capable host (dual Xeon) Programmable Significant RAM Blade Specification

29 Cape Town 2013 SKA feasibility model

30 Cape Town 2013 SKA conceptual software stack

31 Cape Town 2013 HPC development and prototyping lab for SKA Coordinated out of Cambridge and run jointly by HPCS and CHPC Will work closely with COMP to test and design various potential compute, networking, storage and HPC system / application software components Rigorous system engineering approach, which describes a formalised design and prototyping loop Provides a managed, global lab for the whole of the SDP consortium Provide touch stone and practical place of work for interaction with vendors First major test bed in the form of a Dell / Mellanox / NVIDIA GPU cluster has been deployed in the lab last month and will be used by consortium to drive design R&D SKA Open Architecture Lab

32 Cape Town 2013 The SKA SDP compute facility will be at the time of deployment one of the largest HPC systems in existence Operational management of large HPC systems is challenging at the best of times - When HPC systems are housed in well established research centres with good IT logistics and experienced Linux HPC staff The SKA SDP could be housed in a desert location with little surrounding IT infrastructure, with poor IT logistics and little prior HPC history at the site Potential SKA SDP exascale systems are likely to consist of 100,000 nodes occupy 800 cabinets and consume 30 MW. This is very large – around 5 times the size of one today largest supercomputer –Titan Cray at Oakridge national labs. The SKA SDP HPC operations will be very challenging SKA Exascale computing in the desert

33 Cape Town 2013 Although the operational aspects of the SKA SDP exacscale facility are challenging they are tractable if dealt with systematically and in collaboration with the HPC community. The challenge is tractable

34 Cape Town 2013 We can describe the operational aspects by functional element Machine room requirements ** SDP data connectivity requirements SDP workflow requirements System service level requirements System management software requirements** Commissioning & acceptance test procedures System administration procedure User access procedures Security procedure Maintenance & logistical procedures ** Refresh procedure System staffing & training procedures ** SKA HPC operations – functional elements

35 Cape Town 2013 Machine room infrastructure for exascale HPC facilities is challenging 800 racks, 1600M squared 30MW IT load ~40 Kw of heat per rack Cooling efficiency and heat density management is vital Machine infrastructure at this scale is both costly and time comsuming The power cost alone at todays cost is £30M per year Desert location presents particular problems for data centre Hot ambient temperature - difficult for compressor less cooling Lack of water- difficult for compressor less cooling Very dry air- difficult for humidification Remote location- difficult for DC maintenance Machine room requirements

36 Cape Town 2013 System management software is the vital element in HPC operations System management software today does not scale to exascale Worldwide coordinated effort to develop system management software for exascale Elements of system management software stack:- Power management Network management Storage management Workflow management OS Runtime environment Security management System resilience System monitoring System data analytics Development tool System management software

37 Cape Town 2013 Current HPC technology MBTF for hardware and system software result in failure rates of ~ 2 nodes per week on a cluster a ~600 nodes. It is expected that SKA exascale systems could contain ~100,000 nodes Thus expected failure rates of 300 nodes per week could be realistic During system commissioning this will be 3 or 4 X Fixing nodes quickly is vital otherwise the system will soon degrade into a non functional state The manual engineering processes for fault detection and diagnosis on 600 will not scale to 100,000 nodes. This needs to be automated by the system software layer Vendor hardware replacement logistics need to cope with high turn around rates Maintenance logistics

38 Cape Town 2013 Providing functional staffing levels and experience at remote desert location will be challenging Its hard enough finding good HPC staff to run small scale HPC systems in Cambridge – finding orders of magnitude more staff to run much more complicated systems in a remote desert location will be very Challenging Operational procedures using a combination of remote system administration staff and DC smart hands will be needed. HPC training programmes need to be implemented to skill up way in advance Staffing levels and training

39 Cape Town 2013 Early Cambridge SKA solution - EDSAC 1 Maurice Wilkes


Download ppt "Cape Town 2013 Dr Paul Calleja Director Cambridge HPC Service SKA The worlds largest Radio Telescope streaming data processor."

Similar presentations


Ads by Google