1 Metrics for the Office of Science HPC Centers Jonathan Carter User Services Group Lead NERSC User Group Meeting June 12, 2006.

Slides:



Advertisements
Similar presentations
Performance Assessment
Advertisements

Management Plans: A Roadmap to Successful Implementation
1 Performance Assessment An NSF Perspective MJ Suiter Budget, Finance and Awards NSF.
Performance Management Review FAQs
U.S. Department of Energy Office of Science Advanced Scientific Computing Research Program NERSC Users Group Meeting Department of Energy Update June 12,
IDEA What it is and How to Implement the System Texas A & M, February 2013 Shelley A. Chapman, PhD Senior Educational Consultant.
W5HH Principle As applied to Software Projects
Developing Performance Goals That Work For You and ANR Performance Training and Workshop for County Directors November 1, 2013.
Performance Appraisal System Update
Peer Assessment of 5-year Performance ARS National Program 301: Plant, Microbial and Insect Genetic Resources, Genomics and Genetic Improvement Summary.
Peer assessment of group work using WebPA Neil Gordon Symposium on the Benefits of eLearning Technologies University of Manchester, in conjunction with.
Effective Project Management: Traditional, Agile, Extreme
Implementing the new Workload Policy Heads of School Workshop April 2010.
Project Title Project Investigators Project Duration: (E.g., 3 years; currently in year 2, or x months if this is a better representation of project time)
Talbert House Project PASS Goals and Outcomes.
By Saurabh Sardesai October 2014.
Senior Review Evaluations (1 of 5) Proposals due: 6 March 2015 Panel evaluations: Week of 22 April 2015 Performance factors to be evaluated will include.
Contra Costa County Office of Education Local Control Funding Formula Updates and Ideas Presented by: Jannelle Kubinec.
Annual Review Process Georgi Lowe UWSA Office of Human Resources & Workforce Diversity.
Performance Evaluation Process June 19th and June 26 th.
Presented by TS Hamilton. Five Core Competencies We link our courses to CalSTRS core competencies.
Info-Tech Research Group1 Improving Business Satisfaction Moving from Measurement to Action.
Performance Management Dan Robbins. Overview Define performance management Describe the process of developing a performance management system Discuss.
PERFORMANCE AUDIT REPORT ON MANAGEMENT OF PRIMARY HEALTH CARE (A CASE STUDY ON HEALTH CENTERS) 8/16/20151 Dr. Anna Nswilla CDHSMoHSW.
APPRAISAL OF THE HEADTEACHER GOVERNORS’ BRIEFING
Organization Mission Organizations That Use Evaluative Thinking Will Develop mission statements specific enough to provide a basis for goals and.
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
McLean & Company1 Improving Business Satisfaction Moving from Measurement to Action.
Source Selection. What is Source Selection? Source Selection is the process of conducting competitive negotiations. Source Selection allows the Government.
Dear User, This presentation has been designed for you by the Hearts and Minds Support Team. It provides a template for presenting the results of the SAFE.
NASA Earth Observing System Data and Information Systems
Staff Performance Evaluation Process
2010 Results. Today’s Agenda Results Summary 2010 CQS Strengths and Opportunities CQS Benchmarks Demographics Next Steps.
BUGANDO MEDICAL CENTRE PRESENTATION ON OPRAS OVERVIEW
U.S. Department of Energy Office of Science Advanced Scientific Computing Research Program NERSC Users Group Meeting Department of Energy Update September.
December 14, 2011/Office of the NIH CIO Operational Analysis – What Does It Mean To The Project Manager? NIH Project Management Community of Excellence.
Performance Assessment Assessment of Organizational Excellence NSF Advisory Committee for Business and Operations May 5-6, 2005.
Systems Studies Program Peer Review Meeting Albert L. Opdenaker III DOE Program Manager Holiday Inn Express Germantown, Maryland August 29, 2013.
DOE Annual Review of SLAC HEP Research Program June 14-16, 2005 SLAC Charge to Committee Issues Procedures.
March 7, 2013 Texas Education Agency | Office of Assessment and Accountability Division of Performance Reporting Accountability Policy Advisory Committee.
BESAC Dec Outline of the Report I. A Confluence of Scientific Opportunities: Why Invest Now in Theory and Computation in the Basic Energy Sciences?
Office of Acquisition and Property Management Proposed Changes to Attachment G FY 2011 – FY 2015 Five Year Plan.
ASCAC-BERAC Joint Panel on Accelerating Progress Toward GTL Goals Some concerns that were expressed by ASCAC members.
APPRAISAL OF THE HEADTEACHER GOVERNORS’ BRIEFING.
Template This is a template to help, not constrain, you. Modify as appropriate. Move bullet points to additional slides as needed. Don’t cram onto a single.
2005 Customer Satisfaction Study September 2005 NASA Earth Observing System Data and Information Systems.
Exploiting Group Recommendation Functions for Flexible Preferences.
What is Title I and How Can I be Involved? Annual Parent Meeting Pierce Elementary
Office of Science (SC) Overview John Yates Office of Operations Program Management, SC-33 Office of Science Briefing for DOE FIMS Training at ANL – May.
What is Title I and How Can I be Involved? Annual Parent Meeting Pine Hill Middle School September 17, :00 PM.
Template This is a template to help, not constrain, you. Modify as appropriate. Move bullet points to additional slides as needed. Don’t cram onto a single.
BER Long Term Measures As discussed at a previous BERAC meeting with Joel Parriott (OMB) and Bill Valdez (DOE/SC) BERAC is on the hook for evaluating BER’s.
R 0 G125 B 177 R 78 G 47 B 145 R 185 G 50 B 147 R 245 G132 B 107 R 255 G234 B 83 R 123 G193 B 67 R149 G169 B 202 Goal Setting Guide 2015.
Internal Auditing Effectiveness
Module 1: Writing Your Functional Competency Assessment East Carolina University Department of Human Resources Classification and Compensation.
The NPF & Scotland Performs: Analytical Underpinning and Challenges Mairi Spowage Office of the Chief Statistician 23 rd March 2009.
New Framework for Strategic Goal Assessment NSF Advisory Committee for Business and Operations November 8-9, 2006 Tom Cooley, Director, BFA.
Health Management Dr. Sireen Alkhaldi, DrPH Community Medicine Faculty of Medicine, The University of Jordan First Semester 2015 / 2016.
Measuring Turnaround Success October 29 th, 2015 Jeanette P. Cornier, Ph.D.
The NPF & Scotland Performs: Analytical Underpinning and Challenges Mairi Spowage Office of the Chief Statistician 9 th June 2009.
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
Session 5: Selecting and Operationalizing Indicators.
Impact Research 1 Optimizing Your Help Desk: Summary Document.
Assessment Background September 2014 – New National Curriculum introduced into schools Years 1 and 2 (KS1), Years 3 and 4 (Lower KS2), Years 5 and 6 (Upper.
Building PetaScale Applications and Tools on the TeraGrid Workshop December 11-12, 2007 Scott Lathrop and Sergiu Sanielevici.
Selection Criteria and Invitational Priorities School Leadership Program U.S. Department of Education 2005.
APPRAISAL OF THE HEADTEACHER GOVERNORS’ BRIEFING.
Stages of Research and Development
TOPS TRAINING.
Monitoring and Evaluation using the
Presentation transcript:

1 Metrics for the Office of Science HPC Centers Jonathan Carter User Services Group Lead NERSC User Group Meeting June 12, 2006

2 Goals Informational –Metrics Panel –Draft proposal Solicit Feedback –Are proposed metrics reasonable? –Fine tuning ‘capability job’ metrics

3 Office of Science “Metrics Panel” ASCR has asked a panel for recommendations about metrics Panel is headed by Gordon Bell from Microsoft Its goals: –performance measurement and assessment at Office of Science (SC) HPC facilities –appropriateness and comprehensiveness of the measures –science accomplishments and their effects on SC’s science programs –provide input for the Office of Management and Budget (OMB) –evaluation of ASCR progress towards the long-term goals specified in the OMB Program Assessment Rating Tool (PART) NERSC, ORNL and ANL have provided input

4 Current OMB PART Metrics 1.Acquisitions should be no more than 10% more than planned cost and schedule. This metric is reasonable. 2.40% of the computational time is used by jobs with a concurrency of 1/8 or more of the maximum usable compute CPUs. Meeting this metric has positive and negative effects: motivated increased scaling of user codes; not related to the quantity, quality, or productivity of the science. 3.Every year several selected science applications are expected to increase efficiency by at least 50%. This metric was motivated by the desire to increase the percent of peak performance in large science applications, which now has less merit. Should be replaced by a scaling metric.

5 5 Three PART metrics are sufficient to demonstrate DOE Office of Science’s progress in advancing the state of high performance computing. Cost-efficient and timely acquisitions clearly important –Metric #1 retained but slightly modified (scoring). Primary interest of OMB is whether the computational resources in the Office of Science are facilitating scientific discovery: the PART metrics should reflect this interest. –Metrics #2 and #3 should be changed Suggestions for PART Metrics

6 Scientific Discovery is hard to measure in near term Propose using following sets of metrics to assess two factors that are highly influential on scientific discovery –how well the computational facilities provision resources and services (Facility Metrics), and –how well computational scientists use these resources to produce science (Computational Science Metrics) Some combination of these metrics should replace PART #2 and #3

7 Metrics Terminology Goal: the behavior being motivated Metric: what is being measured Value: the value for the metric that must be achieved

8 Facility Metrics How well the computational facilities provision resources and services Specifics of goals and metrics impacts your experience running at NERSC

9 Facility Metrics: User Satisfaction Goal #1: User Satisfaction Meeting the metric means that the users are satisfied with how well the facility provides resources and services. Metric #1.1: Users find the systems and services of a facility useful and helpful. Value #1.1: The overall satisfaction of an annual user survey is 5.25 or better (out of 7). Metric #1.2: Facility responsiveness to user feedback Value #1.2: There is an improved user rating in areas where previous user ratings had fallen below 5.25 (out of 7).

10 Facility Metrics: System Availability Goal #2: Office of Science systems are ready and able to process the user workload. Meeting this metric means the machines are up and available most of the time. Availability has real meaning to users. Metric #2.1: Scheduled availability Scheduled availability is the percentage of time a system is available for users, accounting for any scheduled downtime for maintenance and upgrades. Value #2.1: Within 18 months of delivery and thereafter, scheduled availability is > 95%

11 Facility Metrics: Effective Assistance Goal #3: Facilities provide timely and effective assistance Helping users effectively use complex systems is a key service that leading computational facilities supply. Users desire their inquiry is heard and is being worked. Users also need to have their problems answered properly in a timely manner. Metric #3.1: Problems are recorded and acknowledged Value #3.1: 99% of user problems are acknowledged within 4 working hours. Metric #3.2: Most problems are solved within a reasonable time Value #3.2: 80% of user problems are addressed within 3 working days, either by resolving them or (for longer term problems) by informing the user of a longer term plan and providing periodic updates

12 Facility Metrics: Facilitating Capability Jobs Goal #4: Facilitate running capability jobs Major computational facilities have to run capability jobs. The definition of a capability job needs to be defined by agreement between the Program Office and the Facility. The number of processors that define a capability job is a function of the number of available processors, the number and kind of projects or users that the facility supports. This function has not yet been determined. Metric #4.1: The majority of computational time goes to capability jobs. Value #4.1: T% of all computational time will go to jobs that use more than N CPUs (or x% of the available processors) Metric #4.2: Capability jobs are provided excellent turnaround Value #4.2: For capability jobs, the expansion factor is X or less.

13 Discussion: What is a Capability Job? A job using 1/8 of the processors? A job using 1/10 of the processors? A project that received ≥ 3% of the DOE allocation (3 such projects at NERSC)? A project that received ≥ 2% of the total allocation (12 projects)? A project that received ≥ 1% of the total allocation (25 projects)? A function of both the number of processors and the number of projects at a facility? E.g. 10 * max procs / num projects: NERSC: 10 * 6080 procs / 300 projects = 202 processors Leadership: 10* 10,000 procs / 20 projects = 5,000 processors

14 Discussion: Should we have a Target Expansion Factor? Relationship between Expansion Factor and Allocations: –Inverse relationship between the expected expansion factor and the percentage of resource that is allocated –the more that gets allocated the longer the wait times and the higher the expansion factor For which class of jobs should an Expansion Factor metric apply? –Capability jobs only? –All regular charge jobs? –Other? For which machines should an Expansion Factor metric apply? –Only the largest machine at a facility? –All machines, each weighted by their contribution to the total allocation?

15 Discussion: What should the Target Expansion Factor Be? Traditional Expansion Factor: E(job) = (wait_time + run_time) / run time Proposed Formula (only request time can influence scheduling decisions): E(job) = (wait time + request time) / request time Weight to use in computing the Expansion Factor for a class of jobs: –Simple average –Request time Request time * number of processors (this gives more weight to capability jobs) When to start counting wait time? –On Seaborg and Bassi: when the job enters Idle state –On Jacquard: when the job was submitted (this will change with Maui scheduler)

16 Past NERSC Expansion Factors for Regular Charge Class Quarter Allocation Pressure Seaborg EF Bassi EF Jacquard EF NERSC EF FY05 Q3 Over- allocated FY05 Q4 Over- allocated FY06 Q1Mixed FY06 Q2Very Low FY06 Q3 thru 6/5 Low

17 Past Seaborg Expansion Factors for Regular Charge Class Year procs procs procs ,008 procs 1,024- 2,032 procs 2,048 + procs All FY05 Q FY05 Q FY05 Q FY06 Q FY06 Q FY06 Q3 thru 6/

18 Computational Science Metrics Ability of projects to use facility resources for science

19 Computational Science Metrics: Science Progress CS Goal #1: Science Progress While there are many laudable science goals, it is vital that significant computational progress is made against the Nation’s science challenges and questions. Metric #CS1.1: Progress is demonstrated toward the scientific milestones in the top 20 projects at each facility based on the simulation results planned and promised in their project proposals. Value #CS1.1: For the top 20 projects at each facility, an assessment is made by the related program office regarding how well scientific milestones were met or exceeded relative to plans determined during the review period.

20 Computational Science Metrics: Code Scalability CS Goal #2: Scalability of Computational Science Applications The major challenge facing computational science during the next five to ten years is the increased parallelism needed to use more computational resources. Multi-core chips accelerate the need to respond to this challenge. Metric #CS2.1: Science applications should increase in scalability. Value #CS2.1: The scalability of selected applications increase by a factor of 2 every three years. The definition of scalability (strong, weak, etc.) might be domain- and/or code-specific.