Lattice QCD and GPU-s Robert Edwards, Theory Group Chip Watson, HPC & CIO Jie Chen & Balint Joo, HPC Jefferson Lab TexPoint fonts used in EMF. Read the.

Slides:



Advertisements
Similar presentations
Introduction Glasgow’s NPE research Group uses high precision electromagnetic probes to study the subatomic structure of matter. Alongside this we are.
Advertisements

Excited State Spectroscopy from Lattice QCD
Nuclear Physics in the SciDAC Era Robert Edwards Jefferson Lab SciDAC 2009 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this.
Dynamical Anisotropic-Clover Lattice Production for Hadronic Physics C. Morningstar, CMU K. Orginos, College W&M J. Dudek, R. Edwards, B. Joo, D. Richards,
GPU System Architecture Alan Gray EPCC The University of Edinburgh.
JLab Status & 2016 Planning April 2015 All Hands Meeting Chip Watson Jefferson Lab Outline Operations Status FY15 File System Upgrade 2016 Planning for.
Lattice QCD in Nuclear Physics Robert Edwards Jefferson Lab NERSC 2011 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Algorithms for Lattice Field Theory at Extreme Scales Rich Brower 1*, Ron Babich 1, James Brannick 2, Mike Clark 3, Saul Cohen 1, Balint Joo 4, Tony Kennedy.
1 AppliedMicro X-Gene ® ARM Processors Optimized Scale-Out Solutions for Supercomputing.
Scientific Computing at Jefferson Lab Petabytes, Petaflops and GPUs Chip Watson Scientific Computing Group Jefferson Lab Presented at CLAS12 Workshop,
Jefferson Lab Status Hall A collaboration Dec. 16, 2013 R. D. McKeown Deputy Director For Science.
Exploiting Disruptive Technology: GPUs for Physics Chip Watson Scientific Computing Group Jefferson Lab Presented at GlueX Collaboration Meeting, May 11,
Dynamical Anisotropic-Clover Lattice Production for Hadronic Physics J. Foley, C. Morningstar, CMU K. Orginos, College W&M J. Dudek, R. Edwards, B. Joo,
The Lattice Initiative at Jefferson Lab Robert Edwards Jefferson Lab JLab in a close partnership with MIT has formed the Lattice Hadron Physics Collaboration.
Lattice QCD and GPU-s Robert Edwards, Theory Group Chip Watson, HPC & CIO Jie Chen & Balint Joo, HPC Jefferson Lab TexPoint fonts used in EMF. Read the.
Simulating Quarks and Gluons with Quantum Chromodynamics February 10, CS635 Parallel Computer Architecture. Mahantesh Halappanavar.
Lattice QCD in Nuclear Physics Robert Edwards Jefferson Lab CCP 2011 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Physics Steven Gottlieb, NCSA/Indiana University Lattice QCD: focus on one area I understand well. A central aim of calculations using lattice QCD is to.
R. Ryne, NUG mtg: Page 1 High Energy Physics Greenbook Presentation Robert D. Ryne Lawrence Berkeley National Laboratory NERSC User Group Meeting.
Excited State Spectroscopy using GPUs Robert Edwards Jefferson Lab TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A.
Scientific Computing Experimental Physics Lattice QCD Sandy Philpott May 20, 2011 IT Internal Review 12GeV Readiness.
Baryon Resonances from Lattice QCD Robert Edwards Jefferson Lab N high Q 2, 2011 TexPoint fonts used in EMF. Read the TexPoint manual before you delete.
Kab/SC-PAC Operated by the Jefferson Science Associates for the U.S. Department of Energy Thomas Jefferson National Accelerator Facility SURA.
Hadron Spectroscopy from Lattice QCD
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
Computational Requirements for NP Robert Edwards Jefferson Lab TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAAAAAAAAAAA.
Contractor Assurance System Peer Review April Page 1 NSTAR 2011 May 16, 2011 Jefferson Lab Hugh Montgomery.
Excited baryon spectrum using Lattice QCD Robert Edwards Jefferson Lab JLab Users Group Meeting 2011 TexPoint fonts used in EMF. Read the TexPoint manual.
Baryon Resonance Determination using LQCD Robert Edwards Jefferson Lab Baryons 2013 TexPoint fonts used in EMF. Read the TexPoint manual before you delete.
May 25-26, 2006 LQCD Computing Review1 Jefferson Lab 2006 LQCD Analysis Cluster Chip Watson Jefferson Lab, High Performance Computing.
SciDAC Software Infrastructure for Lattice Gauge Theory Richard C. Brower QCD Project Review May 24-25, 2005 Code distribution see
LQCD Workflow: Gauge Generation, Prop Calcs, Analysis Robert Edwards Jefferson Lab HackFest 2014.
Robert Edwards Jefferson Lab Creutz-Fest 2014 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAAAAAAAAAAA 1983 HADRONS.
SURA BOT 11/5/02 Lattice QCD Stephen J Wallace. SURA BOT 11/5/02 Lattice.
General Discussion some general remarks some questions.
Excited State Spectroscopy from Lattice QCD Robert Edwards Jefferson Lab CERN 2010 TexPoint fonts used in EMF. Read the TexPoint manual before you delete.
1 QCD Thermodynamics at High Temperature Peter Petreczky Large Scale Computing and Storage Requirements for Nuclear Physics (NP), Bethesda MD, April 29-30,
Electromagnetic probes MAMI, Jefferson Lab & MAX-Lab Daniel Watts University of Edinburgh.
Site Report on Physics Plans and ILDG Usage for US Balint Joo Jefferson Lab.
Excited State Spectroscopy from Lattice QCD Robert Edwards Jefferson Lab MENU 2010 TexPoint fonts used in EMF. Read the TexPoint manual before you delete.
What is QCD? Quantum ChromoDynamics is the theory of the strong force
Daniel S. Carman Page 1 Hadron Sep , 2015 Daniel S. Carman Jefferson Laboratory N* Spectrum & Structure Analysis of CLAS Data  CLAS12 N*
HEP and NP SciDAC projects: Key ideas presented in the SciDAC II white papers Robert D. Ryne.
ComPASS Summary, Budgets & Discussion Panagiotis Spentzouris, Fermilab ComPASS PI.
U.S. Department of Energy’s Office of Science Midrange Scientific Computing Requirements Jefferson Lab Robert Edwards October 21, 2008.
Baryons (and Mesons) on the Lattice Robert Edwards Jefferson Lab EBAC May 2010 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this.
1 Melynda Brooks, “Understanding the Origin of the Nucleon Spin” Understanding the Origin of the Nucleon Spin Andi Klein, P-23, Melynda Brooks P-25, Pat.
Dynamical Anisotropic-Clover Lattice Production for Hadronic Physics A. Lichtl, BNL J. Bulava, J. Foley, C. Morningstar, CMU C. Aubin, K. Orginos, W&M.
July 10, 2006TAPS 2006 Experimental Hall-D and the GlueX Experiment at Jefferson Lab Dr. David Lawrence Jefferson Lab Dr. David Lawrence Jefferson Lab.
Fermi National Accelerator Laboratory & Thomas Jefferson National Accelerator Facility SciDAC LQCD Software The Department of Energy (DOE) Office of Science.
Baryon Resonances from Lattice QCD Robert Edwards Jefferson Lab GHP 2011 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Jun Doi IBM Research – Tokyo Early Performance Evaluation of Lattice QCD on POWER+GPU Cluster 17 July 2015.
LQCD Computing Project Overview
EIC NAS review Charge-2 What are the capabilities of other facilities, existing and planned, domestic and abroad, to address the science opportunities.
Project Management – Part I
Welcome Stuart Henderson May 22, 2017.
Computational Requirements
Baryons on the Lattice Robert Edwards Jefferson Lab Hadron 09
Electron Ion Collider New aspects of EIC experiment instrumentation and computing, as well as their possible impact on and context in society (B) COMPUTING.
Future Trends in Nuclear Physics Computing Workshop
Nucleon Resonances from Lattice QCD
Samples of Hall B Results with Strong Italian Impact
Scientific Computing At Jefferson Lab
Baryon Spectroscopy and Resonances
Excited State Spectroscopy from Lattice QCD
A Phenomenology of the Baryon Spectrum from Lattice QCD
Excited State Spectroscopy from Lattice QCD
Excited State Spectroscopy from Lattice QCD
Excited state meson and baryon spectroscopy from Lattice QCD
Baryon Resonances from Lattice QCD
Presentation transcript:

Lattice QCD and GPU-s Robert Edwards, Theory Group Chip Watson, HPC & CIO Jie Chen & Balint Joo, HPC Jefferson Lab TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A AA A A A

Outline Will describe how: Capability computing + Capacity computing + SciDAC –Deliver science & NP milestones Collaborative efforts involve USQCD + JLab & DOE+NSF user communities 2

Hadronic & Nuclear Physics with LQCD Hadronic spectroscopy –Hadron resonance determinations –Exotic meson spectrum (JLab 12GeV ) Hadronic structure –3-D picture of hadrons from gluon & quark spin+flavor distributions –Ground & excited E&M transition form-factors (JLab 6GeV+12GeV+Mainz) –E&M polarizabilities of hadrons (Duke+CERN+Lund) Nuclear interactions –Nuclear processes relevant for stellar evolution –Hyperon-hyperon scattering –3 & 4 nucleon interaction properties [Collab. w/LLNL] (JLab+LLNL) Beyond the Standard Model –Neutron decay constraints on BSM from Ultra Cold Neutron source (LANL) 3

Bridges in Nuclear Physics NP Exascale 4

Spectroscopy Spectroscopy reveals fundamental aspects of hadronic physics –Essential degrees of freedom? –Gluonic excitations in mesons - exotic states of matter? Status –Can extract excited hadron energies & identify spins, –Pursuing full QCD calculations with realistic quark masses. New spectroscopy programs world-wide –E.g., BES III (Beijing), GSI/Panda (Darmstadt) –Crucial complement to 12 GeV program at JLab. Excited nucleon spectroscopy (JLab) JLab GlueX: search for gluonic excitations. 5

USQCD National Effort US Lattice QCD effort: Jefferson Laboratory, BNL and FNAL FNAL Weak matrix elements BNL RHIC Physics JLAB Hadronic Physics SciDAC – R&D Vehicle Software R&D INCITE resources (~20 TF-yr) + USQCD cluster facilities (17 TF-yr): Impact on DOE ’ s High Energy & Nuclear Physics Program 6

Gauge Generation: Cost Scaling Cost: reasonable statistics, box size and “physical” pion mass Extrapolate in lattice spacings: 10 ~ 100 PF-yr PF-years State-of-Art Today, 10TF-yr 2011 (100TF-yr) 7

Computational Requirements Gauge generation : Analysis Current calculations Weak matrix elements: 1 : 1 Baryon spectroscopy: 1 : 10 Nuclear structure: 1 : 4 Computational Requirements: Gauge Generation : Analysis 10 : 1 (2005) 1 : 3 (2010) Core work: Dirac inverters - use GPU-s 8

SciDAC Impact Software development –QCD friendly API’s and libraries: enables high user productivity –Allows rapid prototyping & optimization –Significant software effort for GPU-s Algorithm improvements –Operators & contractions: clusters (Distillation: PRL (2009)) –Mixed-precision Dirac-solvers: INCITE+clusters+GPU-s, 2-3X –Adaptive multi-grid solvers: clusters, ~8X (?) Hardware development via USQCD Facilities –Adding support for new hardware –GPU-s 9

Modern GPU Characteristics Hundreds of simple cores: high flop rate SIMD architecture (single instruction, multiple data) Complex (high bandwidth) memory hierarchy Gaming cards: no memory Error-Correction (ECC) – reliability issue I/O bandwidth << Memory bandwidth Commodity Processorsx86 CPUNVIDIA GT200New Fermi GPU #cores Clock speed3.2 GHz1.4 GHz Main memory bandwidth20 GB/s160 GB/s (gaming card) 180 GB/s (gaming card) I/O bandwidth7 GB/s (dual QDR IB) 3 GB/s 4 GB/s Power80 watts200 watts250 watts 10

Inverter Strong Scaling: V=32 3 x256 Local volume on GPU too small (I/O bottleneck) 3 Tflops 11

Science / Dollar for (Some) LQCD Capacity Apps 12

Hardware: ARRA GPU Clusters GPU clusters: ~530 cards Quads 2.4 GHz Nehalem 48 GB memory / node 117 nodes x 4 GPUs -> 468 GPUs Singles 2.4 GHz Nehalem 24 GB memory / node 64 nodes x 1 GPU -> 64 GPUs

530 GPUs at Jefferson Lab (July)  200,000 cores (1,600 million core hours / year)  600 Tflops peak single precision  100 Tflops aggregate sustained in the inverter, (mixed half / single precision)  Significant increase in dedicated USQCD resources All this for only $1M with hosts, networking, etc. Disclaimer: To exploit this performance, code has to be run on the GPUs, not the CPU (Amdahl’s Law problem). SciDAC-2 (& 3) software effort: move more inverters & other code to gpu A Large Capacity Resource 14

New Science Reach in QCD Spectrum Gauge generation: (next dataset) –INCITE: Crays&BG/P-s, ~ 16K – 24K cores –Double precision Analysis (existing dataset): two-classes –Propagators (Dirac matrix inversions) Few GPU level Single + half precision No memory error-correction –Contractions: Clusters: few cores Double precision + large memory footprint Cost (TF-yr) New: 10 TF-yr Old: 1 TF-yr 10 TF-yr 1 TF-yr 15

Isovector Meson Spectrum 16

Isovector Meson Spectrum

Exotic matter Exotics: world summary 18

Exotic matter Suggests (many) exotics within range of JLab Hall D Previous work: photo- production rates high Current GPU work: (strong) decays - important experimental input Exotics: first GPU results 19

Nucleon & Delta Spectrum First results from GPU-s < 2% error bars [ 56,2 + ] D-wave [ 70,1 - ] P-wave [ 70,1 - ] P-wave [ 56,2 + ] D-wave Discern structure: wave-function overlaps Change at light quark mass? Decays! Suggests spectrum at least as dense as quark model 20

Extending science reach USQCD: –Next calculations: physical quark masses: 100 TF – 1 PF-yr –New INCITE+Early Science application (ANL+ORNL+NERSC) –NSF Blue Waters Petascale (PRAC) Need SciDAC-3 –Significant software effort for next generation GPU-s & heterogeneous environments –Participate in emerging ASCR Exascale initiatives INCITE + LQCD synergy: –ARRA GPU system well matched to current leadership facilities 21

Path to Exascale Enabled by some hybrid GPU system? –Cray + Nvidia ?? NSF GaTech: Tier 2 (experimental facility) –Phase 1: HP cluster+GPU (Nvidia Tesla) –Phase 2: hybrid GPU+ ASCR Exascale facility –Case studies for Science, Software+Runtime, Hardware ASCR call proposals: Exascale Co-Design Center Exascale capacity resource will be needed 22

Summary Capability + Capacity + SciDAC –Deliver science & HEP+NP milestones Petascale (leadership) + Petascale (capacity)+SciDAC-3 Spectrum + decays First contact with experimental resolution Exascale (leadership) + Exascale (capacity)+SciDAC-3 Full resolution Spectrum + transitions Nuclear structure Collaborative efforts: USQCD + JLab user communities 23

Backup slides The end 24

JLab ARRA: Phase 1 25

JLab ARRA: Phase 2 26