Presentation is loading. Please wait.

Presentation is loading. Please wait.

ISACA’s COBIT® Assessment Programme (based on COBIT® 5)

Similar presentations


Presentation on theme: "ISACA’s COBIT® Assessment Programme (based on COBIT® 5)"— Presentation transcript:

1 ISACA’s COBIT® Assessment Programme (based on COBIT® 5)
Presented by:

2 Session Objectives Understanding the COBIT Assessment Programme, its ISOIEC base and the use of COBIT 5 content in it Understanding the relationship to ISO/IEC and why ISACA selected this standard and approach Understanding the COBIT Assessment Programme materials and support from ISACA Copyright ISACA All rights reserved Slide 2

3 What is A Process Assessment?
ISO/IEC identifies process assessment as an activity that can be performed either as part of a process improvement initiative or as part of a capability determination approach. The purpose of process improvement is to continually improve the enterprise’s effectiveness and efficiency. The purpose of process capability determination is to identify the strengths, weaknesses and risk of selected processes with respect to a particular specified requirement through the processes used and their alignment with the business need. It provides an understandable, logical, repeatable, reliable and robust methodology for assessing the capability of IT processes. Defined in ISO Copyright ISACA All rights reserved Slide 3

4 What is the COBIT Assessment Programme?
The COBIT Assessment Programme brings together two proven heavyweights in the IT arena, ISO and ISACA. The process assessment standard from ISO, ISO/IEC is combined with the process model from COBIT 5 to provide an understandable, logical, repeatable, reliable and robust methodology for assessing the capability of IT processes. REVEAL – ISACA’s new COBIT Assessment Process brings COBIT together with ISO15504 – a reference model for assessing process capability (consisting of capability levels which in turn consist of the process attributes and further consist of generic practices). REVEAL – ISACA publications to support the COBIT Assessment Programme include the Process Assessment Model (or PAM); a guide for Certified Assessors (and we will talk more about the concept of “certified assessors” a little later); and a “self assessment” guide for enterprises that would like a less formal assessment using the same basic approach. REVEAL – The PAM – the key reference source for an assessment basically re-states much of the COBIT 4.1 content into an ISO15504 compliant process assessment model for use in assessing IT process capability. Note that PAM is a Process Improvement Process, an analysis of gaps, it is not a risk-based methodology. Copyright ISACA All rights reserved Slide 4

5 Programme support The COBIT Assessment Programme products include: COBIT Self Assessment Guide: Using COBIT 5 A self-assessment tool kit COBIT Assessor Guide: Using COBIT 5 COBIT Process Assessment Model (PAM): Using COBIT 5 In addition, Accredited Training Organizations (ATOs) deliver the COBIT Assurance training course to candidates who have obtained the COBIT 5 Foundation certification. ISACA has established a Certified COBIT Assessor certification, to allow appropriately trained and experienced assessors to be able to demonstrate their competence to assessment project sponsors, Copyright ISACA All rights reserved Slide 5

6 Self-assessment approach
Simple, stand alone guidance (10 pages plus short appendices and a supporting tool kit) has been developed in a Self-assessment Guide to support completion of a simplified assessment approach. This approach can be used to perform a less rigorous status assessment, perhaps to determine problem or issue areas for internal discussion or for targeting a formal future compliant assessment. This approach is aligned with the formal approach but does not require evidence collection. It is a good way to learn initially about the programme. Copyright ISACA All rights reserved Slide 6

7 Process Assessment Model
Assessment Overview Process Assessment Model Assessment Process This figure is reproduced from ISO :2003 with the permission of ISO at Copyright remains with ISO. Copyright ISACA All rights reserved Slide 7

8 Process Reference Model (PRM)
The COBIT process reference model (PRM) is defined in the Process Assessment Model publication. The PRM content is directly based on COBIT 5: Enabling Processes content, with adjustments only made to reflect ISO/IEC terminology as necessary. Process domain and scope, purpose and outcomes are defined for each of the 37 COBIT 5 processes. Copyright ISACA All rights reserved Slide 8

9 Assessment Overview This figure is reproduced from ISO :2003 with the permission of ISO at Copyright remains with ISO. Copyright ISACA All rights reserved Slide 9

10 Measurement Framework
The COBIT assessment process measures the extent to which a given process achieves specific attributes relative to that process— ‘process attributes.’ The COBIT assessment process defines nine process attributes (based on ISO/IEC ) PA 1.1 Process performance PA 2.1 Performance management PA 2.2 Work product management PA 3.1 Process definition PA 3.2 Process deployment PA 4.1 Process measurement PA 4.2 Process control PA 5.1 Process innovation PA 5.2 Continuous optimisation **Note that the Process Reference Model (PRM) in the COBIT PAM refers ONLY to Level 1 – PA1.1. All other levels and attributes PA2.1 to PA5.2 deal with generic outcomes. The COBIT Assessment Programme approach measures the extent to which a given process achieves specific attributes of a process. Those attributes are: REVEAL Process results or performance Management of work products of the process Management of the process performance Definition of the process Deployment of the process Measurement and control of the process Innovation and optimisation of the process Lets take a look at a couple of these in a little more detail so you can get a sense for what they mean. Copyright ISACA All rights reserved Slide 10

11 Process Capability Levels
Level 5 Optimizing process PA 5.1 Process innovation attribute PA 5.2 Process optimization attribute Optimizing The process is continuously improved to meet relevant current and projected business goals. Level 4 Predictable process PA 4.1 Process measurement attribute PA 4.2 Process control attribute Predictable The process is enacted consistently within defined limits. Level 3 Established process PA 3.1 Process definition attribute PA 3.2 Process deployment attribute Established A defined process is used based on a standard process. Level 2 Managed process PA 2.1 Performance management attribute PA 2.2 Work product management attribute Managed The process is managed and work products are established, controlled and maintained. Level 1 Performed process PA 1.1 Process performance attribute Performed The process is implemented and achieves its process purpose. Level 0 Incomplete process Incomplete The process is not implemented or fails to achieve its purpose. Copyright ISACA All rights reserved Slide 11

12 Process Attributes (example)
PA 1.1 Process performance The process performance attribute is a measure of the extent to which the process purpose is achieved. As a result of full achievement of this attribute, the process achieves its defined outcomes. The first process attribute – relates to the results or performance of the SPECIFIC PRM process: It’s a measure of the extent to which the process achieves its purpose – what it is designed to do. This attribute is fully achieved when the process achieves its defined outcomes. On this slide and the next one – walk through an example of process attributes PA1 and PA2. Copyright ISACA All rights reserved Slide 12

13 Process Attributes (example)
PA 2.1 Performance management A measure of the extent to which the performance of the process is managed. As a result of full achievement of this attribute: Objectives for the performance of the process are identified. Performance of the process is planned and monitored. Performance of the process is adjusted to meet plans. Responsibilities and authorities for performing the process are defined, assigned and communicated. Resources and information necessary for performing the process are identified, made available, allocated and used. Interfaces between the involved parties are managed to ensure effective communication and clear assignment of responsibility. PA 2.2 Work product management A measure of the extent to which the work products produced by the process are appropriately managed. As a result of full achievement of this attribute: Requirements for the work products of the process are defined. Requirements for documentation and control of the work products are defined. Work products are appropriately identified, documented and controlled. Work products are reviewed in accordance with planned arrangements and adjusted as necessary to meet requirements. The next attributes relate to management of the process and associated work products: REVEAL - PA 2.1 is a measure of the extent to which the performance of the process is ‘managed’ and is fully achieved when: Process objectives have been defined. The process performance is planned and monitored. Process performance is adjusted to meet plans. Responsibilities and authorities are defined, assigned and communicated. Resource and information requirements are identified, allocated and used. There is effective communication between parties and clear assignment of responsibilities. REVEAL – PA2.2 is a measure of the extent to which the work products produced by the process are managed and is fully achieved when: Requirements for the work products have been defined. Requirements for documentation and control of the work products have been defined. The work products are identified, documented and controlled consistent with the definitions. Work products are reviewed and adjusted as necessary to meet the requirements. We will walk through an example of these shortly. Copyright ISACA All rights reserved Slide 13

14 Process Attribute Rating Scale
The COBIT assessment process measures the extent to which a given process achieves the ‘process attributes:’ N Not achieved—0 to 15% achievement There is little or no evidence of achievement of the defined attribute in the assessed process. P Partially achieved—> 15% to 50% achievement There is some evidence of an approach to, and some achievement of, the defined attribute in the assessed process. Some aspects of achievement of the attribute may be unpredictable. L Largely achieved—> 50% to 85% achievement There is evidence of a systematic approach to, and significant achievement of, the defined attribute in the assessed process. Some weakness related to this attribute may exist in the assessed process. F Fully achieved—> 85% to 100% achievement There is evidence of a complete and systematic approach to, and full achievement of, the defined attribute in the assessed process. No significant weaknesses related to this attribute exist in the assessed process. But first let us have a look at the ‘rating scale’ . which measures the extent to which a given process achieves each of the process attributes. The PAM (consistent with ISO15504) defines 4 levels of ‘achievement’: Not Achieved – where there is little or no evidence of achievement of the attribute in the process Partially Achieved – some of the elements have been achieved Largely Achieved – evidence of systematic approach to and achievement of the attribute elements – although some weaknesses/improvement opportunities may still exist Fully Achieved – evidence of full achievement of the attribute elements and no significant weaknesses exist Copyright ISACA All rights reserved Slide 14

15 Process Attribute Ratings and Capability Levels
1 L / F 2 L / F F 3 L / 4 L / F L / F 5 PA 5.1 Innovation PA 5.2 Optimization Level 5 - Optimizing PA 4.2 Control PA 4.1 Measurement Level 4 - Predictable PA 3.2 Deployment PA 3.1 Definition Level 3 - Established PA 2.2 Work product management PA 2.1 Performance management Level 2 - Managed The Process Attributes are organized into logical ‘levels’ representing the various process capability levels – REVEAL and refer briefly to each of the process attributes and the capability levels. Achievement of a given ‘Process Capability level’ requires the attributes for that level to have been ‘Fully’ or ‘Largely’ achieved – and the attributes for all lower levels to be ‘Fully’ achieved. REVEAL - For example – achieving level 1 capability requires Attribute PA 1.1 to be fully or largely achieved. REVEAL – achieving level 2 requires both PA2.1 and PA2.2 to be fully or largely achieved and PA1.1 to be fully achieved. REVEAL – achieving level 3 requires both PA 3.1 and PA3.2 to be fully or largely achieved and PA1.1, 2.1 and 2.2 to be fully achieved. REVEAL/REVEAL – and so on for capability levels 4 and 5. PA 1.1 Process performance Level 1 - Performed Level 0 - Incomplete L/F = Largely or Fully F= Fully This figure is reproduced from ISO :2003 with the permission of ISO at Copyright remains with ISO. Copyright ISACA All rights reserved Slide 15

16 COBIT Assessment Process Overview
This figure is reproduced from ISO :2003 with the permission of ISO at Copyright remains with ISO. Copyright ISACA All rights reserved Slide 16

17 Process Capability Levels and Attributes
ISO The COBIT Process Assessment Model (PAM) uses the PRM and the Measurement Framework to define an assessment model for each of the COBIT 5 processes. The Assessment Model defines ‘indicators’ that support achievement of the 9 process attributes. COBIT 5 This figure is reproduced from ISO :2003 with the permission of ISO at Copyright remains with ISO. Slide 17 17

18 Process Attribute Rating
Assessment indicators in the PAM are used to support the assessors’ judgement in rating process attributes: Provide the basis for repeatability across assessments A rating is assigned based on objective, validated evidence for each process attribute. Traceability needs to be maintained between an attribute rating and the objective evidence used in determining that rating. As implied by their name, indicators do not represent requirements of a process. They represent a common starting point for assessment, which increases the consistency of assessor judgment and enhances the repeatability of the results. The indicators provide a framework for assessment that helps to ensure that:  Assessors have the ability to interpret the organisational unit's instantiation of a process consistently against the Process Assessment Model(s).  The information is captured for subsequent analysis.  The information needed for the Organizational Unit to plan and perform process improvement is captured.  Assessment results are representative, reliable and repeatable. The assignment of a rating for a given Process Attribute needs to be supported by objective, validated evidence. The traceability of the rating and the supporting evidence needs to be maintained. Copyright ISACA All rights reserved Slide 18

19 Overview I now want to turn our attention to the Assessment Process itself, to walk through the assessment process at a very high level. Detailed discussion of the process for a compliant assessment is provided in the Assessor Guide. As a reminder, simplified guidance has been developed in a Self-assessment Guide to support completion a simplified assessment approach for those wanting to perform a simple, judgement but NOT EVIDENCE based self assessment – perhaps to determine problem or issue areas for discussion or for targeting a more formal compliant assessment. This is a transition slide to indicate we have completed discussion of PAM and will now be moving on to discussion of principle activities in the COBIT Assessment Process. This figure is reproduced from ISO :2003 with the permission of ISO at Copyright remains with ISO. Copyright ISACA All rights reserved Slide 19

20 Assessment Process Activities
Initiation Planning the assessment Briefing Data collection Data validation Process attributes rating 7 Reporting the results The activities associated with performing a compliant COBIT Assessment consist of these items identified on the slide. We will quickly review the key elements of each of these activities. Copyright ISACA All rights reserved Slide 20

21 1. Initiation Identify the sponsor and define the purpose of the assessment: Why it is being carried out? Define the scope of the assessment: Which processes are being assessed? What constraints, if any, apply to the assessment? Identify any additional information that needs to be gathered. Select the assessment participants, the assessment team and define the roles of team members. Define assessment inputs and outputs: Have them approved by the sponsor. The objective of the initiation phase is to ensure that there is a common understanding with the sponsor on the purpose and scope of the assessment, and to identify the individuals with the appropriate competencies to ensure a successful assessment. For example – is the purpose of the assessment to benchmark current performance and identify improvement opportunities – or is the objective to demonstrate contractual/regulatory compliance? Recall, it is highly unlikely an enterprise would assess all 37 COBIT 5 processes, so a simple scoping tool kit has been provided by ISACA. Copyright ISACA All rights reserved Slide 21 21

22 2. Planning the Assessment
An assessment plan describing all activities performed in conducting the assessment: Is developed Is documented Contains an assessment schedule Identify the project scope. Secure the necessary resources to perform the assessment. Determine the method of collating, reviewing, validating and documenting the information required for the assessment. Co-ordinate assessment activities with the organisational unit being assessed. The Assessment Planning phase includes such things as: Determine the assessment activities. -may be tailored as necessary. Determine the necessary resources and schedule for the assessment. Define how the assessment data will be collected, recorded, stored, analysed and presented with reference to the assessment tool. Define the planned outputs of the assessment. Assessment outputs desired by the sponsor in addition to those required as part of the assessment record are identified and described. Verify conformance to requirements. Detail how the assessment will meet all the requirements in the standard. Manage risks. Potential risk factors and mitigation strategies are documented, prioritised and tracked through assessment planning. All identified risks will be monitored throughout the assessment. Potential risk may include changes to the assessment team, organisational changes, changes to the assessment purpose/scope, lack of resources for assessment, confidentiality, priority of the data, base practices and criticality of indicators and availability of key work products such as documents. Co-ordinate assessment logistics with the Local Assessment Co-ordinator. -compatibility and the availability of technical equipment, identified workspace and scheduling requirements will be met. Review and obtain acceptance of the plan. The sponsor identifies who will approve the assessment plan. The plan, including the assessment schedule and logistics for site visits is reviewed and approved. Confirm the sponsor’s commitment to proceed with the assessment. Copyright ISACA All rights reserved Slide 22 22

23 3. Briefing The assessment team leader ensures that the assessment team understands the assessment: Input Process Output Brief the organisational unit on the performance of the assessment: PAM, assessment scope, scheduling, constraints, roles and responsibilities, resource requirements, etc. The Assessor Guide uses Annex A Part 3 A.4 Briefing A.4.1 Overview from ISO Brief the assessment team. Ensure that the team understands the approach defined in the documented process, the assessment inputs and outputs, and is proficient in using the assessment tool. Brief the organisational unit. Explain the assessment purpose, scope, constraints, and model. Stress the confidentiality policy and the benefit of assessment outputs. Present the assessment schedule. Ensure that the staff members understand what is being undertaken and their role in the process. Answer any questions or concerns that they may have. Potential participants and anyone who will see the presentation of the final results should be present at the briefing session. Copyright ISACA All rights reserved Slide 23 23

24 4. Data Collection The assessor obtains (and documents) an understanding of the process(es) including process purpose, inputs, outputs and work products, sufficient to enable and support the assessment. Data required for evaluating the processes within the scope of the assessment are collected in a systematic manner. The strategy and techniques for the selection, collection and analysis of data and justification of the ratings are explicitly identified and demonstrable. Each process identified in the assessment scope is assessed on the basis of objective evidence: The objective evidence gathered for each attribute of each process assessed must be sufficient to meet the assessment purpose and scope. Objective evidence that supports the assessors’ judgement of process attribute ratings is recorded and maintained in the assessment record: This record provides evidence to substantiate the ratings and to verify compliance with the requirements. See Assessor Guide. Collect evidence of process performance for each process within the scope. Evidence includes observation of work products and their characteristics, testimony from the process performers, and observation of the infrastructure established for the performance of the process. Collect evidence of process capability for each process within the scope. Evidence of process capability may be more abstract than evidence of process performance. In some cases, the evidence of process performance may be used as evidence of process capability. Record and maintain the references to the evidence that supports the assessors’ judgement of process attribute ratings. Verify the completeness of the data. Ensure that for each process assessed, sufficient evidence exists to meet the assessment purpose and scope. Copyright ISACA All rights reserved Slide 24 24

25 5. Data Validation Actions are taken to ensure that the data are accurate and sufficiently cover the assessment scope, including: Seeking information from firsthand, independent sources Using past assessment results Holding feedback sessions to validate the information collected Some data validation may occur as the data is being collected. Assemble and consolidate the data. For each process, relate the evidence to defined process indicators. Validate the data. Ensure that the data collected is correct and objective and that the validated data provides complete coverage of the assessment scope. Copyright ISACA All rights reserved Slide 25 25

26 6. Process Attribute Rating
For each process assessed, a rating is assigned for each process attribute up to and including the highest capability level defined in the assessment scope. The rating is based on data validated in the previous activity. Traceability must be maintained between the objective evidence collected and the process attribute ratings assigned. For each process attribute rated, the relationship between the indicators and the objective evidence is recorded. Establish and document the decision-making process used to reach agreement on the ratings (e.g., consensus of the assessment team or majority vote). For each process assessed, assign a rating to each process attribute. Use the defined set of assessment indicators in the Process Assessment Model to support the assessors’ judgement. Record the set of process attribute ratings as the process profile and calculate the capability level rating for each process using the Capability Level Ratings criteria. Copyright ISACA All rights reserved Slide 26 26

27 7. Reporting the Results The results of the assessment are analysed and presented in a report . The report also covers any key issues raised during the assessment such as: Observed areas of strength and weakness Findings of high risk, i.e., magnitude of gap between assessed capability and desired/required capability See Assessor Guide. Tasks: Assemble and consolidate the data. For each process, relate the evidence to defined process indicators. Validate the data. Ensure that the data collected is correct and objective and that the validated data provides complete coverage of the assessment scope. Process attribute rating: For each process assessed, a rating is assigned for each process attribute up to and including the highest capability level defined in the assessment scope. The rating is based on data validated in the previous activity. Traceability shall be maintained between the objective evidence collected and the process attribute ratings assigned. For each process attribute rated, the relationship between the indicators and the objective evidence shall be recorded. Establish and document the decision-making process used to reach agreement on the ratings (e.g., consensus of the assessment team or majority vote). For each process assessed, assign a rating to each process attribute. Use the defined set of assessment indicators in the Process Assessment Model to support the assessors’ judgement. Record the set of process attribute ratings as the process profile and calculate the capability level rating for each process using the Capability Level Ratings criteria. Reporting the Results Overview During this phase, the results of the assessment are analysed and presented in a report. The report also covers any key issues raised during the assessment such as observed areas of strength and weakness and findings of high risk. Prepare the assessment report. Summarise the findings of the assessment, highlighting the process profiles, key results, observed strengths and weaknesses, identified risk factors, and potential improvement actions (if within the scope of the assessment). Present the assessment results to the participants. Focus the presentation on defining the capability of the processes assessed. Present the assessment results to the sponsor. The assessment results will also be shared with any parties (e.g., organisational unit management, practitioners) specified by the sponsor. Finalise the assessment report and distribute to the relevant parties. Verify and document that the assessment was performed according to requirements. Assemble the Assessment Record. Provide the Assessment Record to the sponsor for retention and storage. Prepare and approve assessor records. For each assessor, records to prove the participation in the assessment are produced. The sponsor or the sponsor’s delegated authority approves the records. Provide feedback from the assessment as a means to improve the assessment process. Copyright ISACA All rights reserved Slide 27 27

28 Target Process Capabilities (example)
Level 1 Level 2 Level 3 PA 1.1 PA 2.1 PA 2.2 PA 3.1 PA 3.2 Process A Target Capability L Assessed Process B Target Capability Assessed L F Process C Target Capability Assessed L F Depending on the ‘purpose’ of the assessment – it may be appropriate to compare ‘assessed’ capabilities with a ‘target’ or ‘desired’ capability This slide can be used to review nature of ‘required’ or desired process capabilities by process, and as a way of comparing actual assessed capability against the ‘target’ capability. Copyright ISACA All rights reserved Slide 28

29 Consequence of Capability Gaps
Consequence of Gaps at Various Capability Levels This slide summarises the nature of any gaps that may exist within a given process capability level. Perhaps the easiest way to think about this would be: What is the consequence of NOT being able to achieve the capability level denoted in the first column. This figure is reproduced from ISO with the permission of ISO at Copyright remains with ISO. Copyright ISACA All rights reserved Slide 29

30 Capability Gaps and Risk
Risk Associated With Each Capability Level This presents similar information and may be best interpreted as: What is the relative risk if the gap in assessed capability at each maturity level is Substantial, Significant or Slight, e.g., if the gap between your assessed capability and the requirements to achieve Level 2 capability is Substantial – the result would represent a ‘High’ risk to the enterprise. An interpretation could be that: A ‘slight’ gap at Level 1 would be a High Risk – still not achieving Level 1 capability A ‘significant’ gap at level 2 would be a High risk. A ‘slight’ gap at Level 3 would be a Medium risk. This figure is reproduced from ISO with the permission of ISO at Copyright remains with ISO. Copyright ISACA All rights reserved Slide 30

31 Assessor roles and competencies
COBIT process assessment roles: Lead assessor—a ‘competent’ assessor responsible for overseeing the assessment activities Assessor—an individual, developing assessor competencies, who performs the assessment activities Assessor competencies: Knowledge, skills and experience: With the process reference model; process assessment model, methods and tools; and rating processes With the processes/domains being assessed Personal attributes that contribute to effective performance Copyright ISACA All rights reserved Slide 31

32 Assessor training and certification opportunities
Accredited training organizations (ATOs) deliver the COBIT Assurance training course to candidates who have obtained the COBIT 5 Foundation certification. ISACA has established a Certified COBIT Assessor certification, to allow appropriately trained and experienced assessors to be able to demonstrate their competence to assessment project sponsors, Copyright ISACA All rights reserved Slide 32

33 Questions? Goodbye and thank you . . .
COBIT Assessment Programme: Contact Information: Questions? Copyright ISACA All rights reserved Slide 33


Download ppt "ISACA’s COBIT® Assessment Programme (based on COBIT® 5)"

Similar presentations


Ads by Google