NTTS Conference, Brussels, 18-20 February 20091 OECD/JRC Handbook on constructing composite indicators- Putting theory into practice Michela Nardo & Michaela.

Slides:



Advertisements
Similar presentations
COMP 110: Introduction to Programming Tyler Johnson Feb 18, 2009 MWF 11:00AM-12:15PM Sitterson 014.
Advertisements

COMP 110: Introduction to Programming Tyler Johnson Mar 16, 2009 MWF 11:00AM-12:15PM Sitterson 014.
COMP 110: Introduction to Programming Tyler Johnson Mar 25, 2009 MWF 11:00AM-12:15PM Sitterson 014.
«Productivity growth in services and problems measuring services price indices » EUKLEMS Consortium meeting Brussels, March 2007 Anita Wölfl, CEPII,
The Productivity Gap between Europe and the US: Trends and Causes Marcel P. Timmer Groningen Growth and Development Centre The EU KLEMS project is funded.
1 Alternative measures of well-being Joint work by ECO/ELSA/STD.
Further Analysis MICS3 Regional Workshop on Data Archiving and Dissemination Alexandria, Egypt 3-7 March, 2007.
Statistics for Improving the Efficiency of Public Administration Daniel Peña Universidad Carlos III Madrid, Spain NTTS 2009 Brussels.
Producing monthly estimates of labour market indicators exploiting the longitudinal dimension of the LFS microdata R. Gatto, S. Loriga, A. Spizzichino.
Linking regions and central governments: Indicators for performance-based regional development policy 6 th EUROPEAN CONFERENCE ON EVALUATION OF COHESION.
SOME PRACTICAL ISSUES IN COMPOSITE INDEX CONSTRUCTION Lino Briguglio and Nadia Farrugia Department of Economics, University of Malta Prepared for the INTERNATIONAL.
Wouter Noordkamp The assessment of new platforms on operational performance and manning concepts.
1 Alberto Montanari University of Bologna Basic Principles of Water Resources Management.
6th EC Framework Programme for Research Scientific Support to Policies: “Integrating and Strengthening the European Research Area” Area (i) Policy-oriented.
The “Knowledge sharing platform” in the 6th EC FP Initial training for National Contact Points Brussels 30th October 2002 by Belmiro MARTINS,
Donald T. Simeon Caribbean Health Research Council
Chapter 12 Analyzing Semistructured Decision Support Systems Systems Analysis and Design Kendall and Kendall Fifth Edition.
Design of Experiments Lecture I
DECISION MODELING WITH Multi-Objective Decision Making
Transformations & Data Cleaning
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
IREG-4, Astana, 16 June Rickety Numbers Volatility of international rankings of higher education and implications for policy making ANDREA Saltelli.
See ( OECD-JRC handbook on CI The ‘pros’: Can summarise complex or multi-dimensional issues in view of supporting decision-makers.
Testing the validity of indicators in the field of education The experience of CRELL Rome, October 3-5, 2012, Improving Education through Accountability.
CONCEPTUAL ISSUES IN CONSTRUCTING COMPOSITE INDICES Nadia Farrugia Department of Economics, University of Malta Paper prepared for the INTERNATIONAL CONFERENCE.
CONFERENCE ON SMALL STATES AND RESILIENCE BUILDING Malta, APRIL 2007 " Weighting Procedures for Composite Indicators " Giuseppe Munda European Commission,
1 Constructing Composite Indicators: From Theory to Practice ECFIN, November 11-12, 2010 Andrea Saltelli Ranking and rating: Woodo or Science? Andrea Saltelli,
Copyright © 2006 Pearson Education Canada Inc Course Arrangement !!! Nov. 22,Tuesday Last Class Nov. 23,WednesdayQuiz 5 Nov. 25, FridayTutorial 5.
Chapter 7 Correlational Research Gay, Mills, and Airasian
Michaela Saisana Second Conference on Measuring Human Progress New York, 4-5 March “Reflections on the Human Development Index” (paper by J. Foster)
OECD Short-Term Economic Statistics Working PartyJune Analysis of revisions for short-term economic statistics Richard McKenzie OECD OECD Short.
Measuring Social Life Ch. 5, pp
Scales and Indices While trying to capture the complexity of a phenomenon We try to seek multiple indicators, regardless of the methodology we use: Qualitative.
Presented by Johanna Lind and Anna Schurba Facility Location Planning using the Analytic Hierarchy Process Specialisation Seminar „Facility Location Planning“
ANCOVA Lecture 9 Andrew Ainsworth. What is ANCOVA?
Robustness in assessment of strategic transport projects The 21st International Conference on Multiple Criteria Decision Making Jyväskylä June
Determining Sample Size
Next Colin Clarke-Hill and Ismo Kuhanen 1 Analysing Quantitative Data 1 Forming the Hypothesis Inferential Methods - an overview Research Methods Analysing.
Access to Medicine Index Problem Statement Long-standing debate about: What is the role of the pharmaceutical industry in access to medicines? Where are.
C M Clarke-Hill1 Analysing Quantitative Data Forming the Hypothesis Inferential Methods - an overview Research Methods.
An overview of multi-criteria analysis techniques The main role of the techniques is to deal with the difficulties that human decision-makers have been.
Multi-Criteria Analysis - preference weighting. Defining weights for criteria Purpose: to express the importance of each criterion relative to other criteria.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
Adjusted from slides attributed to Andrew Ainsworth
European Commission DG Joint Research Centre Formal and informal approaches to the quality of information in integrated.
1 Impact of methodological choices on road safety ranking SAMO conference: 20/06/07 Elke Hermans: Transportation.
Eurostat – Unit D5 Key indicators for European policies Third International Seminar on Early Warning and Business Cycle Indicators Annotated outline of.
Module III Multivariate Analysis Techniques- Framework, Factor Analysis, Cluster Analysis and Conjoint Analysis Research Report.
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
Scales and Indices While trying to capture the complexity of a phenomenon We try to seek multiple indicators, regardless of the methodology we use: Qualitative.
METHODS OF SPATIAL ECONOMIC ANALYSIS LECTURE 06 Δρ. Μαρί-Νοέλ Ντυκέν, Αναπληρώτρια Καθηγήτρια, Τηλ Γραφείο Γ.6 UNIVERSITY.
Tutorial I: Missing Value Analysis
Framework and assessment methodology for policy coherence for development: Draft Report for OECD 16 th June, Paris Nick Bozeat.
Growth Diagnostics in Practice Applied Inclusive Growth Analytics Course June 29, 2009 Susanna Lundstrom, PRMED.
Table 3. Merits and Demerits of Selected Water Quality Indices Shweta Tyagi et al. Water Quality Assessment in Terms of Water Quality Index. American Journal.
Methods of multivariate analysis Ing. Jozef Palkovič, PhD.
Simon Deakin CBR, University of Cambridge
Missing data: Why you should care about it and what to do about it
New Techniques and Technologies for Statistics Brussels, March 2017
Open Access and Knowledge Production: ‘Leximetric’ Data Coding
Classroom Assessment A Practical Guide for Educators by Craig A
CHAPTER 6, INDEXES, SCALES, AND TYPOLOGIES
Giuseppe Munda Universitat Autonoma de Barcelona
Measuring Social Life: How Many? How Much? What Type?
Multidimensional Scaling and Correspondence Analysis
Imputation in UNECE Statistical Databases: Principles and Practices
New Techniques and Technologies for Statistics Brussels, March 2017
Sustainable buildings
Towards a Work Programme for the Common Implementation Strategy for the Water Framework Directive (2000/60/EC) Water Directors Meeting 28 November.
Presentation transcript:

NTTS Conference, Brussels, February OECD/JRC Handbook on constructing composite indicators- Putting theory into practice Michela Nardo & Michaela Saisana & JRC-IPSC- G09 Econometric and Applied Statistics Unit NTTS (New Techniques and Technologies for Statistics) Seminar Brussels, February 2009

NTTS Conference, Brussels, February Step 1. Developing a theoretical framework Step 2. Selecting indicators Step 3. Imputation of missing data Step 4. Multivariate analysis Step 5. Normalisation of data Step 6. Weighting and aggregation Step 7. Robustness and sensitivity Step 8. Back to the details (indicators) Step 9. Association with other variables Step 10. Presentation and dissemination The ten recommended steps JRC/OECD Handbook on composite indicators [… and ranking systems] 2 rounds of consultation with OECD high level statistical committee Finally endorsed in March 2008

NTTS Conference, Brussels, February What is badly defined is likely to be badly measured…. This means : To give a definition of the phenomenon to describe To define the number and identity of indicators To describe the nested structure of these indicators Step 1. Developing a Theoretical/conceptual framework

NTTS Conference, Brussels, February Step 1. Developing a Theoretical/conceptual framework

NTTS Conference, Brussels, February Framework: a way of encoding a natural/social system into a formal system No clear way Definition of the phenomenon Systems are complex A framework is valid only within a given information space Formal coherence but also reflexivity Step 1. Developing a Theoretical/conceptual framework

NTTS Conference, Brussels, February Problems found in practice… Go from an ideal to a workable framework Recognize that a framework is only a possible way of representing the reality Recognize that a framework is shaped by the objective of the analysis – implies the definition of selection criteria (e.g. input, process, output for measuring innovation activity) Define the direction of an indicator (i.e. define the objective of the CI). E.g. social cohesion – OECD lists 35 indicator among which number of marriages and divorces, the fertility rate, number of foreigners,.. BUT Does a large group of foreign-born population decrease social cohesion? Select sound variables (merge different source of information, soft data, different years, …) Step 1. Developing a Theoretical/conceptual framework

NTTS Conference, Brussels, February A composite indicator is above all the sum of its parts… Step 2. Selecting variables

NTTS Conference, Brussels, February Dempster and Rubin (1983), The idea of imputation is both seductive and dangerous Step 3. Imputation of missing data

NTTS Conference, Brussels, February Rule of thumb: (Little Rubin 87) If a variable has more than 5% missing values, cases are not deleted, and many researchers are much more stringent than this. Case deletion Pros: no artificial assumptions Cons: higher standard error (less info is used) Risk: think that the dataset is complete Step 3. Imputation of missing data

NTTS Conference, Brussels, February Imputation with MAR and MCAR assumptions Single imputation: one value for each missing data implicit modeling explicit modeling Multiple imputation: more than 1 value for each missing data (Markov Chain Monte Carlo algorithm) Step 3. Imputation of missing data

NTTS Conference, Brussels, February Analysing the underlying structure of the data is still an art … Step 4. Multivariate Analysis

NTTS Conference, Brussels, February Grouping information on individual indicators Is the nested structure of the composite indicator well-defined? Is the set of available individual indicators sufficient/ appropriate to describe the phenomenon? Expert opinion and Multivariate statistics Revision of the indicators or the structure Step 4. Multivariate Analysis

NTTS Conference, Brussels, February Correlation one may see a high correlation among sub- indicators as something to correct for, e.g. by making the weights inversely proportional to the strength of the overall correlation for a given sub-indicator, e.g. Index of Relative Intensity of Regional Problems in the EU (Commission of the European Communities, 1984). schools of thought Practitioners of multicriteria decision analysis would instead tend to consider the existence of correlations as a feature of the problem, not to be corrected for, as correlated sub-indicators may indeed reflect non-compensable different features of the problem. e.g. A cars speed and beauty are likely correlated with one another, but this does not imply that we are willing to trade speed for design. Step 4. Multivariate Analysis

NTTS Conference, Brussels, February Step 4. Multivariate Analysis Principal Component Analysis

NTTS Conference, Brussels, February Grouping information on countries/indicators Cluster analysis serves as: a purely statistical method of aggregation of the indicators, a diagnostic tool for exploring the impact of the methodological choices made during the construction phase of the composite indicator, a method of disseminating information on the composite indicator without losing that on the dimensions of the individual indicators, and a method for selecting groups of countries to impute missing data with a view to decreasing the variance of the imputed values. Cluster analysis Step 4. Multivariate Analysis

NTTS Conference, Brussels, February Cluster analysis: group information on countries based on their similarity on different individual indicators. Step 4. Multivariate Analysis

NTTS Conference, Brussels, February Avoid adding up apples and oranges … Step 5. Normalisation of data

NTTS Conference, Brussels, February Convert variables expressed in different measurement units (e.g. no. of people, money per capita, % of total exports) into a single (dimensionless) unit Issue: which is the most suitable normalization robustness Step 5. Normalisation of data

NTTS Conference, Brussels, February Pro: simple, indep. to outliers Cons: levels lost Pro: no distortion from mean Cons: outliers Pro: simple Cons: outliers Pro: benchmarking Cons: max be outliers Pro: reduce skewness Cons: values near to 0 Pro: indep on slight changes in def. Cons: no track slight improvements Step 5. Normalisation of data

NTTS Conference, Brussels, February Pro: simple, no outliers Cons: p arbitrary, no absolute levels Pro: minimize cycles Cons: penalize irregular indic. Pro: based on opinions Cons: no absolute levels Pro: no levels but changes Cons: low levels are equal to high levels Step 5. Normalisation of data

NTTS Conference, Brussels, February The relative importance of the indicators is a source of contention … 6. Weighting and aggregation should be done along the lines of the underlying theoretical framework. To select appropriate weighting and aggregation procedure(s) that respect both the theoretical framework and the data properties. To discuss whether correlation issues among indicators should be accounted for. To discuss whether compensability among indicators should be allowed. Step 6. Weighting and aggregation

NTTS Conference, Brussels, February No general consensus A composite cannot be objective Weighting implies a subjective choice Soundness and transparency Step 6. Weighting and aggregation

NTTS Conference, Brussels, February Weights based on statistical models Factor analysis Benefit of the doubt Regression Unobserved components model Weights based on opinions Budget allocation Public opinion Analytic hierarchy process Conjoint analysis

NTTS Conference, Brussels, February Pros: deals with correlation Cons: correlation does not mean redundancy only used when correlation sensitive to factor extraction and rotation Factor analysis Step 6. Weighting and aggregation

NTTS Conference, Brussels, February The performance indicator is the ratio of the distance between the origin and the actual observed point and that of the projected point in the frontier:. Step 6. Weighting and aggregation Benefit of the doubt Data Envelopment Analysis (DEA) employs linear programming tools to estimate an efficiency frontier that would be used as a benchmark to measure the relative performance of countries

NTTS Conference, Brussels, February Benefit of the doubt Pros: * the composite will be sensible to national priorities * real benchmark * any other weighting scheme gives a lower index * reveal policy priorities Cons: * weights are country specifics (no comparability) * a priori constraints on weights * weights depend on the benchmark Step 6. Weighting and aggregation But the cross-efficient DEA allows ranking (and comparability)

NTTS Conference, Brussels, February Budget allocation The budget allocation process (BAP) has four different phases: Selection of experts for the valuation; Allocation of budget to the individual indicators; Calculation of the weights; Pros: weights not based on statistical manipulation experts tends to increase legitimacy and acceptance Cons: reliability (weights could reflect local conditions) not suited for more than 10 indicators at the time Step 6. Weighting and aggregation

NTTS Conference, Brussels, February Analytic hierarchy process The core of AHP (Saaty 70s) is an ordinal pair-wise comparison of attributes. The comparisons are made per pairs of individual indicators: Which of the two is the more important? And, by how much? The preference is expressed on a semantic scale of 1 to 9. A preference of 1 indicates equality between two individual indicators, while a preference of 9 indicates that the individual indicator is 9 times more important than the other one. The results are represented in a comparison matrix (Table 20), where Aii = 1 and Aij = 1 / Aji. Step 6. Weighting and aggregation

NTTS Conference, Brussels, February Analytic hierarchy process Pros: suitable for qualitative and quantitative data Cons: high number of pairwise comparisons depends on the setting of the experiment Step 6. Weighting and aggregation

NTTS Conference, Brussels, February Aggregation rules: Linear aggregation implies full (and constant) compensability rewards sub-indicators proportionally to the weights Geometric mean entails partial (non constant) compensability rewards more those countries with higher scores The absence of synergy or conflict effects among the indicators is a necessary condition to admit either linear or geometric aggregation & weights express trade-offs between indicators Multi-criteria analysis different goals are equally legitimate and important (e.g. social, economic dimensions), thus a non-compensatory logic Step 6. Weighting and aggregation

NTTS Conference, Brussels, February When different goals are equally legitimate and important, then a non compensatory logic may be necessary. Example: physical, social and economic figures must be aggregated. If the analyst decides that an increase in economic performance can not compensate a loss in social cohesion or a worsening in environmental sustainability, then neither the linear nor the geometric aggregation are suitable. Instead, a non-compensatory multicriteria approach will assure non compensability by formalizing the idea of finding a compromise between two or more legitimate goals. + does not reward outliers + different goals are equally legitimate and important + no normalisation is required BUT - computational cost when the number of countries is high Multi-criteria type of aggregation Step 6. Weighting and aggregation

NTTS Conference, Brussels, February Step 7. Robustness and sensitivity Cynics say that models can be built to conclude anything provided that suitable assumptions are fed into them. [The Economist, 1998] many indices rarely have adequate scientific foundations to support precise rankings: […] typical practice is to acknowledge uncertainty in the text of the report and then to present a table with unambiguous rankings [Andrews et al. (2004: 1323)] Step 7. Robustness and Sensitivity

NTTS Conference, Brussels, February Uncertainty + Sensitivity Analysis

NTTS Conference, Brussels, February stakeholder involvement Steps in the Development of an Index Step 1. Developing a theoretical framework Step 2. Selecting indicators Step 3. Imputation of missing data Step 4. Multivariate analysis Step 5. Normalisation of data Step 6. Weighting and aggregation Step 7. Robustness and sensitivity Step 8. Links to other variables Step 9. Back to the details Step 10. Presentation and dissemination Sources of uncertainty

NTTS Conference, Brussels, February As a result, an uncertainty analysis should naturally include a careful mapping of all these uncertainties (/assumptions) onto the space of the output (/inferences). Two things can happen: The latter outcome calls in turn for a revision of the ranking system, or a further collection of indicators. The space of the inference is still narrow enough, so as to be meaningful. The space of the inference is too wide to be meaningful.

NTTS Conference, Brussels, February JRC/OECD Handbook on composite indicators [… and ranking systems] Why the ten steps? The three pillars of a well-designed CI: 1.Solid conceptual framework, 2.good quality data, 3.sound methodology (+ robustness assessment) Composite indicators between analysis and advocacy, Saltelli, 2007, Social Indicators Research, 81:65-77 If the 3 conditions are met then CIs can be used for policy-making

NTTS Conference, Brussels, February Some recent CIs developed in collaboration with the JRC using the philosophy of the Handbook 2008 European Lifelong Learning Index (Bertelsmann Foundation, CCL) 2008/2006 Environmental Performance Index (Yale & Columbia University) 2007 Alcohol Policy Index (New York Medical College) 2007 Composite Learning Index (Canadian Council on Learning) 2002/2005 Environmental Sustainability Index (Yale & Columbia University) JRC/OECD Handbook on composite indicators [… and ranking systems]

NTTS Conference, Brussels, February The Composite Learning Index (Canadian Council on Learning) Results Policy messageSensitivity analysis Composite Learning Index (Canadian Council on Learning) Framework

NTTS Conference, Brussels, February Framework (WHO report) Results Policy message Sensitivity analysis The Alcohol Policy Index (New York Medical College) Comparative Analysis of Alcohol Control Policies in 30 Countries Brand, Saisana, Rynn, Pennoni, Lowenfels, 2007, PLoS Medicine, 4(4):

NTTS Conference, Brussels, February The Alcohol Policy Index (New York Medical College) Maybe an example of a country where strong laws are poorly enforced. high estimated amount of unrecorded consumption (>50% of the total). predominantly Islamic population

NTTS Conference, Brussels, February Our Tools for Sensitivity Analysis free software Two fractional variance measures First order effect – one factor influence by itself Total effect - influence inclusive of all interactions with other factors

NTTS Conference, Brussels, February Saltelli A., M. Ratto, T. Andres, F. Campolongo, J. Cariboni, D. Gatelli, M. Saisana, S. Tarantola, 2008, Global sensitivity analysis. The Primer, John Wiley & Sons, England. OECD/JRC, 2008, Handbook on Constructing Composite Indicators. Methodology and user Guide, OECD Publishing, ISBN Authored by Nardo M., Saisana M., Saltelli A., Tarantola S., Hoffman A., Giovannini E. Selected References Books

NTTS Conference, Brussels, February Brand, D. A., Saisana, M., Rynn, L. A., Pennoni, F., Lowenfels, A. B., 2007, Comparative Analysis of Alcohol Control Policies in 30 Countries, PLoS Medicine 4(4): Cherchye, L., Moesen, W., Rogge, N., van Puyenbroeck, T., Saisana, M., Saltelli, A., Liska, R., Tarantola, S., 2008, Creating Composite Indicators with DEA and Robustness Analysis: the case of the Technology Achievement Index, Journal of Operational Research Society 59: Saisana, M., Saltelli, A., Tarantola, S., 2005, Uncertainty and sensitivity analysis techniques as tools for the analysis and validation of composite indicators, Journal of the Royal Statistical Society A 168(2): Saltelli, A., M. Ratto, S. Tarantola and F. Campolongo (2005) Sensitivity Analysis for Chemical Models, Chemical Reviews, 105(7) pp 2811 – Munda G., Nardo M., Saisana M., 2009, Measuring uncertainties in composite indicators of sustainability. Int. J. Environmental Technology and Management (forthcoming). Selected References Journal articles

NTTS Conference, Brussels, February Munda G., Saisana M., 2009, Methodological Considerations on Regional Sustainability Assessment based on Multicriteria and Sensitivity Analysis. Regional Studies (forthcoming). Tarantola S., Nardo M., Saisana M., Gatelli D., 2006, A new estimator for sensitivity analysis of model output: An application to the e-business readiness composite indicator. Reliability Engineering & System Safety 91(10-11), Selected References Journal articles

NTTS Conference, Brussels, February Saisana, M., 2008, The 2007 Composite Learning Index: Robustness Issues and Critical Assessment, EUR Report EN, European Commission, JRC-IPSC, Italy. Saisana M., DHombres B., 2008, Higher Education Rankings: Robustness Issues and Critical Assessment, EUR Report EN, European Commission, JRC-IPSC, Italy. Saisana M., Munda G., 2008, Knowledge Economy: measures and drivers, EUR Report EN, European Commission, JRC-IPSC. Saisana M., Saltelli, A., 2008, Sensitivity Analysis for the 2008 Environmental Performance Index, EUR Report EN, European Commission, JRC-IPSC. Selected References EUR reports

NTTS Conference, Brussels, February Munda G. and Nardo M. (2005), Constructing Consistent Composite Indicators: the Issue of Weights, EUR EN, Joint Research Centre, Ispra. Nardo M., Tarantola S., Saltelli A., Andropoulos C., Buescher R., Karageorgos G., Latvala A. and Noel F. (2004), The e-business readiness composite indicator for 2003: a pilot study, EUR Saisana M., Tarantola S., 2002, State-of-the-art Report on Current Methodologies and Practices for Composite Indicator Development, EUR Report EN, European Commission, JRC-IPSC, Italy. Selected References EUR reports