ISACA’s COBIT® Assessment Programme (based on COBIT® 5)

Slides:



Advertisements
Similar presentations
Basic Principles of GMP
Advertisements

Requirements Engineering Processes – 2
Chapter 5 Transfer of Training
Chapter 24 Quality Management.
Copyright © 2003 Pearson Education, Inc. Slide 1 Computer Systems Organization & Architecture Chapters 8-12 John D. Carpinelli.
Chapter 1 The Study of Body Function Image PowerPoint
Copyright © 2011, Elsevier Inc. All rights reserved. Chapter 6 Author: Julia Richards and R. Scott Hawley.
Author: Julia Richards and R. Scott Hawley
Key Stage 3 National Strategy
Science Subject Leader Training
Overview of Performance Measurement. Learning Objectives By the end of the module, you will be able to: Describe what performance measurement is, and.
Knowledge Dietary Managers Association 1 DMA Certification Exam Blueprint and Curriculum Development.
1 Introduction to Safety Management April Objective The objective of this presentation is to highlight some of the basic elements of Safety Management.
for Cabin Safety Inspectors
1 Welcome Safety Regulatory Function Handbook April 2006.
1 Assessing Health Needs Gilbert Burnham, MD, PhD Johns Hopkins University.
The Implementation Structure DG AGRI, October 2005
THE CERTIFYING AUTHORITY
The Managing Authority –Keystone of the Control System
Human Performance Improvement Process
1 DPAS II Process and Procedures for Teachers Developed by: Delaware Department of Education.
Program Goals, Objectives and Performance Indicators A guide for grant and program development 3/2/2014 | Illinois Criminal Justice Information Authority.
NPA: Business Improvement Techniques Contributing to Effective Team Working.
Supporting managers: assessment and the learner journey
Introduction to Auditing
Comparing COBIT 4.1 and COBIT 5
1 European benchmarking with the CAF ROME 17-18th of November 2003.
1 SESSION 5- RECORDING AND REPORTING IN GRADES R-12 Computer Applications Technology Information Technology.
Trustee Gough Ward 3 Forum TDSB/Provincial Report Card
Presenter: Beresford Riley, Government of
EMS Checklist (ISO model)
Qualifications Update: Fashion and Textile Technology Dunblane Hydro 27 November 2013 Qualifications Update: Fashion and Textile Technology Dunblane Hydro.
Fact-finding Techniques Transparencies
Chapter 5 – Enterprise Analysis
The Roles of a Sports Coach
Effectively applying ISO9001:2000 clauses 6 and 7.
1 Quality Indicators for Device Demonstrations April 21, 2009 Lisa Kosh Diana Carl.
Quality Manual for Interoperability Testing
Developing and Implementing a Monitoring & Evaluation Plan
A Process to Identify the Enduring Skills, Processes, & Concepts for your Content Area 1.
COBIT 5 for Information Security Introduction
What is Pay & Performance?
Quality Manual for Interoperability Testing Morten Bruun-Rasmussen Presented by Milan Zoric, ETSI.
© 2012 National Heart Foundation of Australia. Slide 2.
The Aged Care Standards and Accreditation Agency Ltd Continuous Improvement in Residential Aged Care.
©Ian Sommerville 2004Software Engineering, 7th edition. Chapter 27 Slide 1 Quality Management.
ISACA’s COBIT® Assessment Programme
TASSAB – The Tasmanian Secondary Assessment Board au/www/tassab/htm.
Mechanics of FLEX, Planning, Delivery & Quality 20 th & 21 st August 2014 Jude Conlon – Head of Operations & Quality 1.
RTI Implementer Webinar Series: Establishing a Screening Process
1 © 2006 Curriculum K-12 Directorate, NSW Department of Education and Training English K-6 Syllabus Using the syllabus for consistency of assessment.
Module 12 WSP quality assurance tool 1. Module 12 WSP quality assurance tool Session structure Introduction About the tool Using the tool Supporting materials.
©Brooks/Cole, 2001 Chapter 12 Derived Types-- Enumerated, Structure and Union.
1 Phase III: Planning Action Developing Improvement Plans.
PSSA Preparation.
Organization Theory and Health Services Management
COBIT 5 Introduction Presented by.
1 Literacy PERKS Standard 1: Aligned Curriculum. 2 PERKS Essential Elements Academic Performance 1. Aligned Curriculum 2. Multiple Assessments 3. Instruction.
MIS (Management Information System)
WM Software Process & Quality Generic Processes - Slide #1  P. Sorenson SPiCE Reference Model - how to read Chapter 5 Capability Levels (process.
Quality evaluation and improvement for Internal Audit
Demystifying the Business Analysis Body of Knowledge Central Iowa IIBA Chapter December 7, 2005.
Quality Management.  Quality management is becoming increasingly important to the leadership and management of all organisations. I  t is necessary.
Understanding Meaning and Importance of Competency Based Assessment
Assessment Validation. MORE THAN YOU IMAGINE ASQA (Australian Skills Quality Authority) New National Regulator ASQA as of 1 July, 2011.
National PE Cycle of Analysis. Fitness Assessment + Gathering Data Why do we need to asses our fitness levels?? * Strengths + Weeknesses -> Develop Performance.
Chap 2 – Internal Audit Modelling with PAM Assesment and Controlling
ISACA’s COBIT® Assessment Programme (based on COBIT® 5)
Presentation transcript:

ISACA’s COBIT® Assessment Programme (based on COBIT® 5) Presented by:

Session Objectives Understanding the COBIT Assessment Programme, its ISOIEC 15504 base and the use of COBIT 5 content in it Understanding the relationship to ISO/IEC 15504 and why ISACA selected this standard and approach Understanding the COBIT Assessment Programme materials and support from ISACA Copyright ISACA 2014 All rights reserved Slide 2

What is A Process Assessment? ISO/IEC 15504-4 identifies process assessment as an activity that can be performed either as part of a process improvement initiative or as part of a capability determination approach. The purpose of process improvement is to continually improve the enterprise’s effectiveness and efficiency. The purpose of process capability determination is to identify the strengths, weaknesses and risk of selected processes with respect to a particular specified requirement through the processes used and their alignment with the business need. It provides an understandable, logical, repeatable, reliable and robust methodology for assessing the capability of IT processes. Defined in ISO 15504-2 Copyright ISACA 2014 All rights reserved Slide 3

What is the COBIT Assessment Programme? The COBIT Assessment Programme brings together two proven heavyweights in the IT arena, ISO and ISACA. The process assessment standard from ISO, ISO/IEC 15504 is combined with the process model from COBIT 5 to provide an understandable, logical, repeatable, reliable and robust methodology for assessing the capability of IT processes. REVEAL – ISACA’s new COBIT Assessment Process brings COBIT together with ISO15504 – a reference model for assessing process capability (consisting of capability levels which in turn consist of the process attributes and further consist of generic practices). REVEAL – ISACA publications to support the COBIT Assessment Programme include the Process Assessment Model (or PAM); a guide for Certified Assessors (and we will talk more about the concept of “certified assessors” a little later); and a “self assessment” guide for enterprises that would like a less formal assessment using the same basic approach. REVEAL – The PAM – the key reference source for an assessment basically re-states much of the COBIT 4.1 content into an ISO15504 compliant process assessment model for use in assessing IT process capability. Note that PAM is a Process Improvement Process, an analysis of gaps, it is not a risk-based methodology. Copyright ISACA 2014 All rights reserved Slide 4

Programme support The COBIT Assessment Programme www.isaca.org/Knowledge-Center/cobit/Pages/COBIT-Assessment-Programme.aspx products include: COBIT Self Assessment Guide: Using COBIT 5 A self-assessment tool kit COBIT Assessor Guide: Using COBIT 5 COBIT Process Assessment Model (PAM): Using COBIT 5 In addition, Accredited Training Organizations (ATOs) deliver the COBIT Assurance training course to candidates who have obtained the COBIT 5 Foundation certification. ISACA has established a Certified COBIT Assessor certification, to allow appropriately trained and experienced assessors to be able to demonstrate their competence to assessment project sponsors, www.isaca.org/COBIT/Pages/COBIT-5-Certified-Assessor-Program.aspx. Copyright ISACA 2014 All rights reserved Slide 5

Self-assessment approach Simple, stand alone guidance (10 pages plus short appendices and a supporting tool kit) has been developed in a Self-assessment Guide to support completion of a simplified assessment approach. This approach can be used to perform a less rigorous status assessment, perhaps to determine problem or issue areas for internal discussion or for targeting a formal future 15504 compliant assessment. This approach is aligned with the formal approach but does not require evidence collection. It is a good way to learn initially about the programme. Copyright ISACA 2014 All rights reserved Slide 6

Process Assessment Model Assessment Overview Process Assessment Model Assessment Process This figure is reproduced from ISO 15504-2:2003 with the permission of ISO at www.iso.org. Copyright remains with ISO. Copyright ISACA 2014 All rights reserved Slide 7

Process Reference Model (PRM) The COBIT process reference model (PRM) is defined in the Process Assessment Model publication. The PRM content is directly based on COBIT 5: Enabling Processes content, with adjustments only made to reflect ISO/IEC 15504 terminology as necessary. Process domain and scope, purpose and outcomes are defined for each of the 37 COBIT 5 processes. Copyright ISACA 2014 All rights reserved Slide 8

Assessment Overview This figure is reproduced from ISO 15504-2:2003 with the permission of ISO at www.iso.org. Copyright remains with ISO. Copyright ISACA 2014. All rights reserved Slide 9

Measurement Framework The COBIT assessment process measures the extent to which a given process achieves specific attributes relative to that process— ‘process attributes.’ The COBIT assessment process defines nine process attributes (based on ISO/IEC 15504-2) PA 1.1 Process performance PA 2.1 Performance management PA 2.2 Work product management PA 3.1 Process definition PA 3.2 Process deployment PA 4.1 Process measurement PA 4.2 Process control PA 5.1 Process innovation PA 5.2 Continuous optimisation **Note that the Process Reference Model (PRM) in the COBIT PAM refers ONLY to Level 1 – PA1.1. All other levels and attributes PA2.1 to PA5.2 deal with generic outcomes. The COBIT Assessment Programme approach measures the extent to which a given process achieves specific attributes of a process. Those attributes are: REVEAL Process results or performance Management of work products of the process Management of the process performance Definition of the process Deployment of the process Measurement and control of the process Innovation and optimisation of the process Lets take a look at a couple of these in a little more detail so you can get a sense for what they mean. Copyright ISACA 2014. All rights reserved Slide 10

Process Capability Levels Level 5 Optimizing process PA 5.1 Process innovation attribute PA 5.2 Process optimization attribute Optimizing The process is continuously improved to meet relevant current and projected business goals. Level 4 Predictable process PA 4.1 Process measurement attribute PA 4.2 Process control attribute Predictable The process is enacted consistently within defined limits. Level 3 Established process PA 3.1 Process definition attribute PA 3.2 Process deployment attribute Established A defined process is used based on a standard process. Level 2 Managed process PA 2.1 Performance management attribute PA 2.2 Work product management attribute Managed The process is managed and work products are established, controlled and maintained. Level 1 Performed process PA 1.1 Process performance attribute Performed The process is implemented and achieves its process purpose. Level 0 Incomplete process Incomplete The process is not implemented or fails to achieve its purpose. Copyright ISACA 2014. All rights reserved Slide 11

Process Attributes (example) PA 1.1 Process performance The process performance attribute is a measure of the extent to which the process purpose is achieved. As a result of full achievement of this attribute, the process achieves its defined outcomes. The first process attribute – relates to the results or performance of the SPECIFIC PRM process: It’s a measure of the extent to which the process achieves its purpose – what it is designed to do. This attribute is fully achieved when the process achieves its defined outcomes. On this slide and the next one – walk through an example of process attributes PA1 and PA2. Copyright ISACA 2014 All rights reserved Slide 12

Process Attributes (example) PA 2.1 Performance management A measure of the extent to which the performance of the process is managed. As a result of full achievement of this attribute: Objectives for the performance of the process are identified. Performance of the process is planned and monitored. Performance of the process is adjusted to meet plans. Responsibilities and authorities for performing the process are defined, assigned and communicated. Resources and information necessary for performing the process are identified, made available, allocated and used. Interfaces between the involved parties are managed to ensure effective communication and clear assignment of responsibility. PA 2.2 Work product management A measure of the extent to which the work products produced by the process are appropriately managed. As a result of full achievement of this attribute: Requirements for the work products of the process are defined. Requirements for documentation and control of the work products are defined. Work products are appropriately identified, documented and controlled. Work products are reviewed in accordance with planned arrangements and adjusted as necessary to meet requirements. The next attributes relate to management of the process and associated work products: REVEAL - PA 2.1 is a measure of the extent to which the performance of the process is ‘managed’ and is fully achieved when: Process objectives have been defined. The process performance is planned and monitored. Process performance is adjusted to meet plans. Responsibilities and authorities are defined, assigned and communicated. Resource and information requirements are identified, allocated and used. There is effective communication between parties and clear assignment of responsibilities. REVEAL – PA2.2 is a measure of the extent to which the work products produced by the process are managed and is fully achieved when: Requirements for the work products have been defined. Requirements for documentation and control of the work products have been defined. The work products are identified, documented and controlled consistent with the definitions. Work products are reviewed and adjusted as necessary to meet the requirements. We will walk through an example of these shortly. Copyright ISACA 2014. All rights reserved Slide 13

Process Attribute Rating Scale The COBIT assessment process measures the extent to which a given process achieves the ‘process attributes:’ N Not achieved—0 to 15% achievement There is little or no evidence of achievement of the defined attribute in the assessed process. P Partially achieved—> 15% to 50% achievement There is some evidence of an approach to, and some achievement of, the defined attribute in the assessed process. Some aspects of achievement of the attribute may be unpredictable. L Largely achieved—> 50% to 85% achievement There is evidence of a systematic approach to, and significant achievement of, the defined attribute in the assessed process. Some weakness related to this attribute may exist in the assessed process. F Fully achieved—> 85% to 100% achievement There is evidence of a complete and systematic approach to, and full achievement of, the defined attribute in the assessed process. No significant weaknesses related to this attribute exist in the assessed process. But first let us have a look at the ‘rating scale’ . which measures the extent to which a given process achieves each of the process attributes. The PAM (consistent with ISO15504) defines 4 levels of ‘achievement’: Not Achieved – where there is little or no evidence of achievement of the attribute in the process Partially Achieved – some of the elements have been achieved Largely Achieved – evidence of systematic approach to and achievement of the attribute elements – although some weaknesses/improvement opportunities may still exist Fully Achieved – evidence of full achievement of the attribute elements and no significant weaknesses exist Copyright ISACA 2014. All rights reserved Slide 14

Process Attribute Ratings and Capability Levels 1 L / F 2 L / F F 3 L / 4 L / F L / F 5 PA 5.1 Innovation PA 5.2 Optimization Level 5 - Optimizing PA 4.2 Control PA 4.1 Measurement Level 4 - Predictable PA 3.2 Deployment PA 3.1 Definition Level 3 - Established PA 2.2 Work product management PA 2.1 Performance management Level 2 - Managed The Process Attributes are organized into logical ‘levels’ representing the various process capability levels – REVEAL and refer briefly to each of the process attributes and the capability levels. Achievement of a given ‘Process Capability level’ requires the attributes for that level to have been ‘Fully’ or ‘Largely’ achieved – and the attributes for all lower levels to be ‘Fully’ achieved. REVEAL - For example – achieving level 1 capability requires Attribute PA 1.1 to be fully or largely achieved. REVEAL – achieving level 2 requires both PA2.1 and PA2.2 to be fully or largely achieved and PA1.1 to be fully achieved. REVEAL – achieving level 3 requires both PA 3.1 and PA3.2 to be fully or largely achieved and PA1.1, 2.1 and 2.2 to be fully achieved. REVEAL/REVEAL – and so on for capability levels 4 and 5. PA 1.1 Process performance Level 1 - Performed Level 0 - Incomplete L/F = Largely or Fully F= Fully This figure is reproduced from ISO 15504-2:2003 with the permission of ISO at www.iso.org. Copyright remains with ISO. Copyright ISACA 2014. All rights reserved Slide 15

COBIT Assessment Process Overview This figure is reproduced from ISO 15504-2:2003 with the permission of ISO at www.iso.org. Copyright remains with ISO. Copyright ISACA 2014. All rights reserved Slide 16

Process Capability Levels and Attributes ISO The COBIT Process Assessment Model (PAM) uses the PRM and the Measurement Framework to define an assessment model for each of the COBIT 5 processes. The Assessment Model defines ‘indicators’ that support achievement of the 9 process attributes. COBIT 5 This figure is reproduced from ISO 15504-2:2003 with the permission of ISO at www.iso.org. Copyright remains with ISO. Slide 17 17

Process Attribute Rating Assessment indicators in the PAM are used to support the assessors’ judgement in rating process attributes: Provide the basis for repeatability across assessments A rating is assigned based on objective, validated evidence for each process attribute. Traceability needs to be maintained between an attribute rating and the objective evidence used in determining that rating. As implied by their name, indicators do not represent requirements of a process. They represent a common starting point for assessment, which increases the consistency of assessor judgment and enhances the repeatability of the results. The indicators provide a framework for assessment that helps to ensure that:  Assessors have the ability to interpret the organisational unit's instantiation of a process consistently against the Process Assessment Model(s).  The information is captured for subsequent analysis.  The information needed for the Organizational Unit to plan and perform process improvement is captured.  Assessment results are representative, reliable and repeatable. The assignment of a rating for a given Process Attribute needs to be supported by objective, validated evidence. The traceability of the rating and the supporting evidence needs to be maintained. Copyright ISACA 2014. All rights reserved Slide 18

Overview I now want to turn our attention to the Assessment Process itself, to walk through the assessment process at a very high level. Detailed discussion of the process for a 15504 compliant assessment is provided in the Assessor Guide. As a reminder, simplified guidance has been developed in a Self-assessment Guide to support completion a simplified assessment approach for those wanting to perform a simple, judgement but NOT EVIDENCE based self assessment – perhaps to determine problem or issue areas for discussion or for targeting a more formal 15504 compliant assessment. This is a transition slide to indicate we have completed discussion of PAM and will now be moving on to discussion of principle activities in the COBIT Assessment Process. This figure is reproduced from ISO 15504-2:2003 with the permission of ISO at www.iso.org. Copyright remains with ISO. Copyright ISACA 2014 All rights reserved Slide 19

Assessment Process Activities Initiation Planning the assessment Briefing Data collection Data validation Process attributes rating 7 Reporting the results The activities associated with performing a 15504-3 compliant COBIT Assessment consist of these items identified on the slide. We will quickly review the key elements of each of these activities. Copyright ISACA 2014. All rights reserved Slide 20

1. Initiation Identify the sponsor and define the purpose of the assessment: Why it is being carried out? Define the scope of the assessment: Which processes are being assessed? What constraints, if any, apply to the assessment? Identify any additional information that needs to be gathered. Select the assessment participants, the assessment team and define the roles of team members. Define assessment inputs and outputs: Have them approved by the sponsor. The objective of the initiation phase is to ensure that there is a common understanding with the sponsor on the purpose and scope of the assessment, and to identify the individuals with the appropriate competencies to ensure a successful assessment. For example – is the purpose of the assessment to benchmark current performance and identify improvement opportunities – or is the objective to demonstrate contractual/regulatory compliance? Recall, it is highly unlikely an enterprise would assess all 37 COBIT 5 processes, so a simple scoping tool kit has been provided by ISACA. Copyright ISACA 2014. All rights reserved Slide 21 21

2. Planning the Assessment An assessment plan describing all activities performed in conducting the assessment: Is developed Is documented Contains an assessment schedule Identify the project scope. Secure the necessary resources to perform the assessment. Determine the method of collating, reviewing, validating and documenting the information required for the assessment. Co-ordinate assessment activities with the organisational unit being assessed. The Assessment Planning phase includes such things as: Determine the assessment activities. -may be tailored as necessary. Determine the necessary resources and schedule for the assessment. Define how the assessment data will be collected, recorded, stored, analysed and presented with reference to the assessment tool. Define the planned outputs of the assessment. Assessment outputs desired by the sponsor in addition to those required as part of the assessment record are identified and described. Verify conformance to requirements. Detail how the assessment will meet all the requirements in the standard. Manage risks. Potential risk factors and mitigation strategies are documented, prioritised and tracked through assessment planning. All identified risks will be monitored throughout the assessment. Potential risk may include changes to the assessment team, organisational changes, changes to the assessment purpose/scope, lack of resources for assessment, confidentiality, priority of the data, base practices and criticality of indicators and availability of key work products such as documents. Co-ordinate assessment logistics with the Local Assessment Co-ordinator. -compatibility and the availability of technical equipment, identified workspace and scheduling requirements will be met. Review and obtain acceptance of the plan. The sponsor identifies who will approve the assessment plan. The plan, including the assessment schedule and logistics for site visits is reviewed and approved. Confirm the sponsor’s commitment to proceed with the assessment. Copyright ISACA 2014. All rights reserved Slide 22 22

3. Briefing The assessment team leader ensures that the assessment team understands the assessment: Input Process Output Brief the organisational unit on the performance of the assessment: PAM, assessment scope, scheduling, constraints, roles and responsibilities, resource requirements, etc. The Assessor Guide uses Annex A Part 3 A.4 Briefing A.4.1 Overview from ISO 15504-3. Brief the assessment team. Ensure that the team understands the approach defined in the documented process, the assessment inputs and outputs, and is proficient in using the assessment tool. Brief the organisational unit. Explain the assessment purpose, scope, constraints, and model. Stress the confidentiality policy and the benefit of assessment outputs. Present the assessment schedule. Ensure that the staff members understand what is being undertaken and their role in the process. Answer any questions or concerns that they may have. Potential participants and anyone who will see the presentation of the final results should be present at the briefing session. Copyright ISACA 2014. All rights reserved Slide 23 23

4. Data Collection The assessor obtains (and documents) an understanding of the process(es) including process purpose, inputs, outputs and work products, sufficient to enable and support the assessment. Data required for evaluating the processes within the scope of the assessment are collected in a systematic manner. The strategy and techniques for the selection, collection and analysis of data and justification of the ratings are explicitly identified and demonstrable. Each process identified in the assessment scope is assessed on the basis of objective evidence: The objective evidence gathered for each attribute of each process assessed must be sufficient to meet the assessment purpose and scope. Objective evidence that supports the assessors’ judgement of process attribute ratings is recorded and maintained in the assessment record: This record provides evidence to substantiate the ratings and to verify compliance with the requirements. See Assessor Guide. Collect evidence of process performance for each process within the scope. Evidence includes observation of work products and their characteristics, testimony from the process performers, and observation of the infrastructure established for the performance of the process. Collect evidence of process capability for each process within the scope. Evidence of process capability may be more abstract than evidence of process performance. In some cases, the evidence of process performance may be used as evidence of process capability. Record and maintain the references to the evidence that supports the assessors’ judgement of process attribute ratings. Verify the completeness of the data. Ensure that for each process assessed, sufficient evidence exists to meet the assessment purpose and scope. Copyright ISACA 2014. All rights reserved Slide 24 24

5. Data Validation Actions are taken to ensure that the data are accurate and sufficiently cover the assessment scope, including: Seeking information from firsthand, independent sources Using past assessment results Holding feedback sessions to validate the information collected Some data validation may occur as the data is being collected. Assemble and consolidate the data. For each process, relate the evidence to defined process indicators. Validate the data. Ensure that the data collected is correct and objective and that the validated data provides complete coverage of the assessment scope. Copyright ISACA 2014. All rights reserved Slide 25 25

6. Process Attribute Rating For each process assessed, a rating is assigned for each process attribute up to and including the highest capability level defined in the assessment scope. The rating is based on data validated in the previous activity. Traceability must be maintained between the objective evidence collected and the process attribute ratings assigned. For each process attribute rated, the relationship between the indicators and the objective evidence is recorded. Establish and document the decision-making process used to reach agreement on the ratings (e.g., consensus of the assessment team or majority vote). For each process assessed, assign a rating to each process attribute. Use the defined set of assessment indicators in the Process Assessment Model to support the assessors’ judgement. Record the set of process attribute ratings as the process profile and calculate the capability level rating for each process using the Capability Level Ratings criteria. Copyright ISACA 2014. All rights reserved Slide 26 26

7. Reporting the Results The results of the assessment are analysed and presented in a report . The report also covers any key issues raised during the assessment such as: Observed areas of strength and weakness Findings of high risk, i.e., magnitude of gap between assessed capability and desired/required capability See Assessor Guide. Tasks: Assemble and consolidate the data. For each process, relate the evidence to defined process indicators. Validate the data. Ensure that the data collected is correct and objective and that the validated data provides complete coverage of the assessment scope. Process attribute rating: For each process assessed, a rating is assigned for each process attribute up to and including the highest capability level defined in the assessment scope. The rating is based on data validated in the previous activity. Traceability shall be maintained between the objective evidence collected and the process attribute ratings assigned. For each process attribute rated, the relationship between the indicators and the objective evidence shall be recorded. Establish and document the decision-making process used to reach agreement on the ratings (e.g., consensus of the assessment team or majority vote). For each process assessed, assign a rating to each process attribute. Use the defined set of assessment indicators in the Process Assessment Model to support the assessors’ judgement. Record the set of process attribute ratings as the process profile and calculate the capability level rating for each process using the Capability Level Ratings criteria. Reporting the Results Overview During this phase, the results of the assessment are analysed and presented in a report. The report also covers any key issues raised during the assessment such as observed areas of strength and weakness and findings of high risk. Prepare the assessment report. Summarise the findings of the assessment, highlighting the process profiles, key results, observed strengths and weaknesses, identified risk factors, and potential improvement actions (if within the scope of the assessment). Present the assessment results to the participants. Focus the presentation on defining the capability of the processes assessed. Present the assessment results to the sponsor. The assessment results will also be shared with any parties (e.g., organisational unit management, practitioners) specified by the sponsor. Finalise the assessment report and distribute to the relevant parties. Verify and document that the assessment was performed according to requirements. Assemble the Assessment Record. Provide the Assessment Record to the sponsor for retention and storage. Prepare and approve assessor records. For each assessor, records to prove the participation in the assessment are produced. The sponsor or the sponsor’s delegated authority approves the records. Provide feedback from the assessment as a means to improve the assessment process. Copyright ISACA 2014. All rights reserved Slide 27 27

Target Process Capabilities (example) Level 1 Level 2 Level 3 PA 1.1 PA 2.1 PA 2.2 PA 3.1 PA 3.2 Process A Target Capability L Assessed Process B Target Capability Assessed L F Process C Target Capability Assessed L F Depending on the ‘purpose’ of the assessment – it may be appropriate to compare ‘assessed’ capabilities with a ‘target’ or ‘desired’ capability This slide can be used to review nature of ‘required’ or desired process capabilities by process, and as a way of comparing actual assessed capability against the ‘target’ capability. Copyright ISACA 2014 All rights reserved Slide 28

Consequence of Capability Gaps Consequence of Gaps at Various Capability Levels This slide summarises the nature of any gaps that may exist within a given process capability level. Perhaps the easiest way to think about this would be: What is the consequence of NOT being able to achieve the capability level denoted in the first column. This figure is reproduced from ISO 15504-4 2006 with the permission of ISO at www.iso.org. Copyright remains with ISO. Copyright ISACA 2014. All rights reserved Slide 29

Capability Gaps and Risk Risk Associated With Each Capability Level This presents similar information and may be best interpreted as: What is the relative risk if the gap in assessed capability at each maturity level is Substantial, Significant or Slight, e.g., if the gap between your assessed capability and the requirements to achieve Level 2 capability is Substantial – the result would represent a ‘High’ risk to the enterprise. An interpretation could be that: A ‘slight’ gap at Level 1 would be a High Risk – still not achieving Level 1 capability A ‘significant’ gap at level 2 would be a High risk. A ‘slight’ gap at Level 3 would be a Medium risk. This figure is reproduced from ISO 15504-4 2006 with the permission of ISO at www.iso.org. Copyright remains with ISO. Copyright ISACA 2014. All rights reserved Slide 30

Assessor roles and competencies COBIT process assessment roles: Lead assessor—a ‘competent’ assessor responsible for overseeing the assessment activities Assessor—an individual, developing assessor competencies, who performs the assessment activities Assessor competencies: Knowledge, skills and experience: With the process reference model; process assessment model, methods and tools; and rating processes With the processes/domains being assessed Personal attributes that contribute to effective performance Copyright ISACA 2014. All rights reserved Slide 31

Assessor training and certification opportunities Accredited training organizations (ATOs) deliver the COBIT Assurance training course to candidates who have obtained the COBIT 5 Foundation certification. ISACA has established a Certified COBIT Assessor certification, to allow appropriately trained and experienced assessors to be able to demonstrate their competence to assessment project sponsors, www.isaca.org/COBIT/Pages/COBIT-5-Certified-Assessor-Program.aspx Copyright ISACA 2014. All rights reserved Slide 32

Questions? Goodbye and thank you . . . COBIT Assessment Programme: www.isaca.org/cobit-assessment-programme Contact Information: research@isaca.org Questions? Copyright ISACA 2014. All rights reserved Slide 33