U.S. Department of Energy’s Office of Science Midrange Scientific Computing Requirements Jefferson Lab Robert Edwards October 21, 2008.

Slides:



Advertisements
Similar presentations
The Physics of Nuclei, Nuclear Matter and Nucleosynthesis Report of the Nuclear Physics Advisory Panel.
Advertisements

First results from the ATLAS experiment at the LHC
O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Center for Computational Sciences Cray X1 and Black Widow at ORNL Center for Computational.
Excited State Spectroscopy from Lattice QCD
Nuclear Physics in the SciDAC Era Robert Edwards Jefferson Lab SciDAC 2009 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this.
JLab Status & 2016 Planning April 2015 All Hands Meeting Chip Watson Jefferson Lab Outline Operations Status FY15 File System Upgrade 2016 Planning for.
Lattice QCD in Nuclear Physics Robert Edwards Jefferson Lab NERSC 2011 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
Jefferson Lab Status Hall A collaboration Dec. 16, 2013 R. D. McKeown Deputy Director For Science.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
Lattice QCD and GPU-s Robert Edwards, Theory Group Chip Watson, HPC & CIO Jie Chen & Balint Joo, HPC Jefferson Lab TexPoint fonts used in EMF. Read the.
Simulating Quarks and Gluons with Quantum Chromodynamics February 10, CS635 Parallel Computer Architecture. Mahantesh Halappanavar.
Jefferson Lab Strategic Planning Divisional Town Meeting Mission Statement of your Division – What is or should be the mission statement of your division?
QCD Project Overview Ying Zhang September 26, 2005.
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
Lattice QCD in Nuclear Physics Robert Edwards Jefferson Lab CCP 2011 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Outline IT Organization SciComp Update CNI Update
R. Ryne, NUG mtg: Page 1 High Energy Physics Greenbook Presentation Robert D. Ryne Lawrence Berkeley National Laboratory NERSC User Group Meeting.
Excited State Spectroscopy using GPUs Robert Edwards Jefferson Lab TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A.
NLIT May 26, 2010 Page 1 Computing Jefferson Lab Users Group Meeting 8 June 2010 Roy Whitney CIO & CTO.
Scientific Computing Experimental Physics Lattice QCD Sandy Philpott May 20, 2011 IT Internal Review 12GeV Readiness.
Office of Science U.S. Department of Energy RHIC Users Meeting BNL; June 8, 2006 Gulshan Rai RHIC/AGS Users Meeting Gulshan Rai Program Manager for Heavy.
Kab/SC-PAC Operated by the Jefferson Science Associates for the U.S. Department of Energy Thomas Jefferson National Accelerator Facility SURA.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Computer Science Research and Development Department Computing Sciences Directorate, L B N L 1 Storage Management and Data Mining in High Energy Physics.
Computational Requirements for NP Robert Edwards Jefferson Lab TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAAAAAAAAAAA.
Jefferson Lab Status APEX Meeting April 22, 2014 R. D. McKeown Deputy Director For Science.
Contractor Assurance System Peer Review April Page 1 NSTAR 2011 May 16, 2011 Jefferson Lab Hugh Montgomery.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
Proposed 2007 Acquisition Don Holmgren LQCD Project Progress Review May 25-26, 2006 Fermilab.
Office of Science U.S. Department of Energy The Medium Energy Program Outline: Program Overview Some Recent Results FY 2008 Funding Summary.
Lattice QCD and GPU-s Robert Edwards, Theory Group Chip Watson, HPC & CIO Jie Chen & Balint Joo, HPC Jefferson Lab TexPoint fonts used in EMF. Read the.
Jefferson Lab Update R. D. McKeown Jefferson Lab HPS Meeting June 16, 2014.
SciDAC Software Infrastructure for Lattice Gauge Theory Richard C. Brower QCD Project Review May 24-25, 2005 Code distribution see
 Advanced Accelerator Simulation Panagiotis Spentzouris Fermilab Computing Division (member of the SciDAC AST project)
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory Review of U.S. LHC Software and Computing Projects Fermi National Laboratory November.
Brent Gorda LBNL – SOS7 3/5/03 1 Planned Machines: BluePlanet SOS7 March 5, 2003 Brent Gorda Future Technologies Group Lawrence Berkeley.
SURA BOT 11/5/02 Lattice QCD Stephen J Wallace. SURA BOT 11/5/02 Lattice.
1 Cluster Development at Fermilab Don Holmgren All-Hands Meeting Jefferson Lab June 1-2, 2005.
Site Report on Physics Plans and ILDG Usage for US Balint Joo Jefferson Lab.
A QCD Grid: 5 Easy Pieces? Richard Kenway University of Edinburgh.
The 12 GeV Physics Program at Jefferson Lab R. D. McKeown Jefferson Lab College of William and Mary PTSP 2013 – Charlottesville, VA September 9, 2013.
Status and plans at KEK Shoji Hashimoto Workshop on LQCD Software for Blue Gene/L, Boston University, Jan. 27, 2006.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
HEP and NP SciDAC projects: Key ideas presented in the SciDAC II white papers Robert D. Ryne.
ComPASS Summary, Budgets & Discussion Panagiotis Spentzouris, Fermilab ComPASS PI.
Fermi National Accelerator Laboratory & Thomas Jefferson National Accelerator Facility SciDAC LQCD Software The Department of Energy (DOE) Office of Science.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
High Performance Computing Activities at Fermilab James Amundson Breakout Session 5C: Computing February 11, 2015.
Hall D Computing Facilities Ian Bird 16 March 2001.
Particle Physics Sector Young-Kee Kim / Greg Bock Leadership Team Strategic Planning Winter Workshop January 29, 2013.
Baryon Resonances from Lattice QCD Robert Edwards Jefferson Lab GHP 2011 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
LQCD Computing Project Overview
Project Management – Part I
Jefferson Lab Overview
Welcome Stuart Henderson May 22, 2017.
Computational Requirements
Baryons on the Lattice Robert Edwards Jefferson Lab Hadron 09
Electron Ion Collider New aspects of EIC experiment instrumentation and computing, as well as their possible impact on and context in society (B) COMPUTING.
Future Trends in Nuclear Physics Computing Workshop
LQCD Computing Operations
Samples of Hall B Results with Strong Italian Impact
Scientific Computing At Jefferson Lab
Introduction to Jefferson Lab
Excited State Spectroscopy from Lattice QCD
Nuclear Physics Data Management Needs Bruce G. Gibbard
View from Lab Management HPS Collaboration Meeting
Excited State Spectroscopy from Lattice QCD
SOLID Collaboration Meeting
Presentation transcript:

U.S. Department of Energy’s Office of Science Midrange Scientific Computing Requirements Jefferson Lab Robert Edwards October 21, 2008

Office of Science U.S. Department of Energy Jefferson Lab  Lab doubling beam energy  Adding new experimental Hall CD-3 JLab Receives DOE Approval to Start Construction of $310 Million Upgrade

Office of Science U.S. Department of Energy Jefferson Lab Science Goals  Understanding structure, spectroscopy and interactions of hadrons from QCD is the central challenge of nuclear physics  How are charge, current and spin distributed in the nucleon?  What are the effective degrees of freedom describing the low-energy spectrum of the theory?  How does the nucleon-nucleon and hadron-hadron interaction arise from QCD?

Office of Science U.S. Department of Energy Jefferson Lab NP Milestones in Hadron Physics  NP2009 and NP2012 : measurement & determination of mass & electromagnetic properties of low lying baryons (hadrons)  New Hall D – seek information about exotic measurements. Flagship of JLab upgrade  NP2014 : determination of Generalized Parton Distributions within nucleons. These characterize how quarks interact with nucleons

Office of Science U.S. Department of Energy Jefferson Lab Computing efforts in support of mission  Experimental physics data acquisition, storage, and analysis (farm computing)  Lattice QCD theory calculations of fundamental quantities

Office of Science U.S. Department of Energy Jefferson Lab Theory context: USQCD Collaboration  Consists of nearly all high energy and nuclear physicists in the US involved in lattice QCD. Formed nine years ago to develop infrastructure for these studies  Research directly in support of DOE experimental facilities at BNL, FNAL, JLab  SciDAC I & II: software development for lattice QCD  FY06-11: ~$2.0M/yr HEP+NP, $0.4M/yr JLab  USQCD Facilites: LQCD I – Midrange computing  FY06-09: ~$2.5M/yr HEP+NP  Access to USQCD dedicated hardware is allocated within peer-reviewed process  LQCD II: FY10-14 – proposal submitted & reviewed

Office of Science U.S. Department of Energy Jefferson Lab Example: Spectroscopy  Experimental and ab initio N* and Exotic- meson programs aim at discovering effective degrees of freedom of QCD,  NP2009 and NP2012 milestones  Excited Baryon Analysis Center (EBAC) at Jefferson Lab  Spectroscopy of Exotic Mesons is a flagship component of ½ + 3/2 + 5/2 + ½ - 3/2 - 5/ Nucleon spectrum (MeV)

Office of Science U.S. Department of Energy Jefferson Lab Nature of the LQCD calculations Calculations are performed in two steps:  Monte Carlo methods are used to gauge configurations with a probability proportional to their weight in the Feynman path integrals that define QCD  These configurations are stored, and used to calculate a wide variety of physical observables

Office of Science U.S. Department of Energy Jefferson Lab Nature of the LQCD calculations (2)  Leadership machines: gauge generation uses large core count, single sequence of computing, requires multi TF sustained on one job  Midrange machines: analysis jobs typically smaller core count, each configuration an independent job, typically a few 100 GF sustained  Fine grained parallelism; regular hypercubic problems  Computations and communication equally important; low latency required

Office of Science U.S. Department of Energy Jefferson Lab Leadership Machines  USQCD aggressively pursuing national resources:  INCITE 07: ORNL: Cray XT4 - 10M-hrs (largest allocation)  INCITE 08-10: yearly allocations, will increase  ORNL: Cray XT4 - 7M-hrs  ANL: BG/P – 20M-hrs  ESP 08: ANL: BG/P - 250M-hrs [11 TF-yr]  ESP 09: ORNL: XT-5 ~ 60M-hrs ??? [7 TF-yr]  Groups (JLab) also pursuing NSF + other resources  2007: 1 TF-yr (PSC+SDSC)  2008: 2 TF-yr (PSC+SDSC+TACC)  2008: 1 TF-yr (LANL/NNSA)

Office of Science U.S. Department of Energy Jefferson Lab LQCD Computing Project Hardware (Midrange) YearComputerSiteNodes Performance (TF/s) 2002QCDFNAL gJLab PionFNAL QCDOCBNL nJLab KaonFNAL nJLab J/PsiFNAL w/ OASCR SciDAC I LQCD I

Office of Science U.S. Department of Energy Jefferson Lab Distribution of Jobs  Leadership machines: few K to 128K cores  Clusters: currently up to 1K cores. Will grow as lattice sizes scale up on leadership machines

Office of Science U.S. Department of Energy Jefferson Lab Future Hardware goals (BNL+FNAL+JLab) by fiscal year  Initial year based upon initial INCITE and leadership class awards. Moore’s law growth thereafter  Dedicated hardware is for all resources running that year; assumes clusters with 3.5 year life

Office of Science U.S. Department of Energy Jefferson Lab Distribution of Resources ( )

Office of Science U.S. Department of Energy Jefferson Lab LQCD II budget request

Office of Science U.S. Department of Energy Jefferson Lab LQCD II review: Jan Charges to the collaboration by DOE:  Why is a new project needed if OASCR is providing access to Leadership Class machines? In particular, is dedicated hardware, such as additional clusters, essential and cost effective in such an environment? What is the optimal mix of machines, given realistic budget constraints?  What are the plans at FNAL, TJNAF, & BNL for LQCD computing? How are these plans incorporated into your plans for LQCD II ?

Office of Science U.S. Department of Energy Jefferson Lab LQCD II review: Jan Review findings:  The 1-1 mix of Leadership and clusters advocated by LQCD II is the most suitable hardware mix for this project.  The review committee advocates full finding for LQCD II at the level described in their proposal. The scenario in which the funding for LQCD II would be flat with LQCD I would cause the project to miss opportunities that would otherwise enhance several other fields that the Office of Science supports, such as computer hardware and software, nuclear and astrophysics.

Office of Science U.S. Department of Energy Jefferson Lab Characteristics of computing:  Dominant part of calculation: large sparse matrix system solve; application of matrix on vector  Regular grid per compute core: moving to hybrid threaded + MPI model.  SciDAC II software development effort

Office of Science U.S. Department of Energy Jefferson Lab Opportunities for Optimization  Clusters configured for a single application (LQCD) can be better optimized than those serving a dozen applications  Memory and disk lean  Pruned fat tree network (e.g. Infiniband) due to highly local communications pattern  Lower aggregate bandwidth to disk compared to check- point intensive simulations  Overall impact compared to generic clusters: 50% more computing capacity/dollar  2x-5x more cost effective for analysis jobs than leadership class machines

Mflops / $ QCDSP Performance/$ for LQCD Applications Commodity compute nodes (leverage marketplace & Moore’s law) Low latency, high bandwidth network to exploit full I/O capability Supercomputers, leadership class machines JLab SciDAC Prototype Clusters QCDOC /9 cluster at FNAL Japanese Earth Simulator JLab clusters BlueGene/L BlueGene/P

Office of Science U.S. Department of Energy Jefferson Lab NP Approved Missions Data analysis and support for 12 GeV:  Why: Needed to support the current 6 GeV and future 12 GeV programs (detector simulation)  How it’s done today: small cluster / farm, plus all necessary infrastructure (tape library, cache disk, …)  Status: Constrained by tight budgets, but currently keeping up with requirements (barely)  Barriers: Future: multi-threading to prevent memory requirement blow-up on many-core architectures  Special Features: integer intensive and i/o intensive

Office of Science U.S. Department of Energy Jefferson Lab NP Approved Missions Lattice QCD:  Why: Theory calculations in support of JLab experimental physics program  How it’s done today: NSF+INCITE computers + midrange computing within USQCD  Status: need more runs of larger size (more cores), plus more statistics (longer runs)  Barriers: multi-threading to reduce comms. Need LQCD II.  Special Features: floating point intensive + balanced communications – fine grained parallelism

Office of Science U.S. Department of Energy Jefferson Lab NP Proposed Missions or Initiatives  No new proposed missions  Currently fulfilling approved missions, with funding requested to continue those approved missions