1 / 46

OECD/JRC Handbook on constructing composite indicators- Putting theory into practice Michela Nardo & Michaela Sai

OECD/JRC Handbook on constructing composite indicators- Putting theory into practice Michela Nardo & Michaela Saisana michela.nardo@jrc.it & michaela.saisana@jrc.it JRC-IPSC- G09 Econometric and Applied Statistics Unit NTTS (New Techniques and Technologies for Statistics) Seminar

henry
Télécharger la présentation

OECD/JRC Handbook on constructing composite indicators- Putting theory into practice Michela Nardo & Michaela Sai

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. OECD/JRC Handbook on constructing composite indicators- Putting theory into practice Michela Nardo & Michaela Saisana michela.nardo@jrc.it & michaela.saisana@jrc.it JRC-IPSC- G09 Econometric and Applied Statistics Unit NTTS (New Techniques and Technologies for Statistics) Seminar Brussels, 18-20 February 2009

  2. JRC/OECD Handbook on composite indicators [… and ranking systems] The ten recommended steps Step 1. Developing a theoretical framework Step 2. Selecting indicators Step 3. Imputation of missing data Step 4. Multivariate analysis Step 5. Normalisation of data Step 6. Weighting and aggregation Step 7. Robustness and sensitivity Step 8. Back to the details (indicators) Step 9. Association with other variables Step 10. Presentation and dissemination • 2 rounds of consultation with OECD high level statistical committee • Finally endorsed in March 2008 http://composite-indicators.jrc.ec.europa.eu/

  3. Step 1. Developing a Theoretical/conceptual framework What is badly defined is likely to be badly measured…. • This means : • To give a definition of the phenomenon to describe • To define the number and identity of indicators • To describe the nested structure of these indicators

  4. Step 1. Developing a Theoretical/conceptual framework

  5. Step 1. Developing a Theoretical/conceptual framework Definition of the phenomenon Framework: a way of encoding a natural/social system into a formal system No clear way A framework is valid only within a given information space Systems are complex Formal coherence but also reflexivity

  6. Step 1. Developing a Theoretical/conceptual framework Problems found in practice… • Go from an ideal to a workable framework • Recognize that a framework is only a possible way of • representing the reality • Recognize that a framework is shaped by the • objective of the analysis – implies the definition of • selection criteria (e.g. input, process, output for • measuring innovation activity) • Define the direction of an indicator (i.e. define the objective of the CI).E.g. social cohesion – OECD lists 35 indicator among which number of marriages and divorces, the fertility rate, number of foreigners,.. BUT Does a large group of foreign-born population decrease social cohesion? • Select sound variables (merge different source of information, soft data, different years, …)

  7. Step 2. Selecting variables A composite indicator is above all the sum of its parts…

  8. Step 3. Imputation of missing data Dempster and Rubin (1983), “The idea of imputation is both seductive and dangerous”

  9. Step 3. Imputation of missing data Case deletion Rule of thumb: (Little Rubin ‘87) If a variable has more than 5% missing values, cases are not deleted, and many researchers are much more stringent than this. Pros: no artificial assumptions Cons: higher standard error (less info is used) Risk: think that the dataset is complete

  10. Step 3. Imputation of missing data Imputation with MAR and MCAR assumptions Single imputation: one value for each missing data implicit modeling explicit modeling Multiple imputation: more than 1 value for each missing data (Markov Chain Monte Carlo algorithm)

  11. Step 4. Multivariate Analysis Analysing the underlying structure of the data is still an art …

  12. Is the nested structure of the composite indicator well-defined? • Is the set of available individual indicators sufficient/ appropriate to describe the phenomenon? Revision of the indicators or the structure Expert opinion and Multivariate statistics Step 4. Multivariate Analysis Grouping information on individual indicators

  13. Step 4. Multivariate Analysis one may see a high correlation among sub-indicators as something to correct for, e.g. by making the weights inversely proportional to the strength of the overall correlation for a given sub-indicator, e.g. Index of Relative Intensity of Regional Problems in the EU (Commission of the European Communities, 1984). Correlation schools of thought • Practitioners of multicriteria decision analysis would instead tend to consider the existence of correlations as a feature of the problem, not to be corrected for, as correlated sub-indicators may indeed reflect non-compensable different features of the problem. • e.g. A car’s speed and beauty are likely correlated with one another, but this does not imply that we are willing to trade speed for design.

  14. Step 4. Multivariate Analysis Principal Component Analysis

  15. Step 4. Multivariate Analysis Cluster analysis • Grouping information on countries/indicators • Cluster analysis serves as: • a purely statistical method of aggregation of the indicators, • a diagnostic tool for exploring the impact of the methodological choices made during the construction phase of the composite indicator, • a method of disseminating information on the composite indicator without losing that on the dimensions of the individual indicators, and • a method for selecting groups of countries to impute missing data with a view to decreasing the variance of the imputed values.

  16. Step 4. Multivariate Analysis Cluster analysis: group information on countries based on their similarity on different individual indicators.

  17. Step 5. Normalisation of data • Avoid adding up apples and oranges …

  18. Step 5. Normalisation of data Convert variables expressed in different measurement units (e.g. no. of people, money per capita, % of total exports) into a single (dimensionless) unit Issue: which is the most suitable normalization robustness

  19. Pro: simple, indep. to outliers Cons: levels lost Pro: no distortion from mean Cons: outliers Pro: simple Cons: outliers Pro: benchmarking Cons: max be outliers Pro: reduce skewness Cons: values near to 0 Pro: indep on slight changes in def. Cons: no track slight improvements Step 5. Normalisation of data

  20. Pro: simple, no outliers Cons: p arbitrary, no absolute levels Pro: minimize cycles Cons: penalize irregular indic. Pro: based on opinions Cons: no absolute levels Pro: no levels but changes Cons: low levels are equal to high levels Step 5. Normalisation of data

  21. Step 6. Weighting and aggregation The relative importance of the indicators is a source of contention …

  22. Step 6. Weighting and aggregation No general consensus Weighting implies a subjective choice A composite cannot be “objective” Soundness and transparency

  23. Weights based on statistical models Factor analysis Benefit of the doubt Regression Unobserved components model Weights based on opinions Budget allocation Public opinion Analytic hierarchy process Conjoint analysis

  24. Step 6. Weighting and aggregation Factor analysis • Pros: deals with correlation • Cons: • correlation does not mean redundancy • only used when correlation • sensitive to factor extraction and rotation

  25. Step 6. Weighting and aggregation Benefit of the doubt Data Envelopment Analysis (DEA) employs linear programming tools to estimate an efficiency frontier that would be used as a benchmark to measure the relative performance of countries The performance indicator is the ratio of the distance between the origin and the actual observed point and that of the projected point in the frontier: .

  26. Step 6. Weighting and aggregation Benefit of the doubt Pros: * the composite will be sensible to national priorities * real benchmark * any other weighting scheme gives a lower index * reveal policy priorities Cons: * weights are country specifics (no comparability) * a priori constraints on weights * weights depend on the benchmark But the cross-efficient DEA allows ranking (and comparability)

  27. Step 6. Weighting and aggregation Budget allocation • The budget allocation process (BAP) has four different phases: • Selection of experts for the valuation; • Allocation of budget to the individual indicators; • Calculation of the weights; Pros: weights not based on statistical manipulation experts tends to increase legitimacy and acceptance Cons: reliability (weights could reflect local conditions) not suited for more than 10 indicators at the time

  28. Step 6. Weighting and aggregation Analytic hierarchy process The core of AHP (Saaty 70s) is an ordinal pair-wise comparison of attributes. The comparisons are made per pairs of individual indicators: Which of the two is the more important? And, by how much? The preference is expressed on a semantic scale of 1 to 9. A preference of 1 indicates equality between two individual indicators, while a preference of 9 indicates that the individual indicator is 9 times more important than the other one. The results are represented in a comparison matrix (Table 20), where Aii = 1 and Aij = 1 / Aji.

  29. Step 6. Weighting and aggregation Analytic hierarchy process Pros: suitable for qualitative and quantitative data Cons: high number of pairwise comparisons depends on the setting of the experiment

  30. Step 6. Weightingand aggregation Aggregation rules: • Linear aggregation implies full (and constant) compensability rewards sub-indicators proportionally to the weights • Geometric mean entails partial (non constant) compensability rewards more those countries with higher scores The absence of synergy or conflict effects among the indicators is a necessary condition to admit either linear or geometric aggregation & weights express trade-offs between indicators • Multi-criteria analysis • different goals are equally legitimate and important (e.g. social, economic dimensions), thus a non-compensatory logic

  31. Step 6. Weightingand aggregation • When different goals are equally legitimate and important, then a non compensatory logic may be necessary. • Example: physical, social and economic figures must be aggregated. If the analyst decides that an increase in economic performance can not compensate a loss in social cohesion or a worsening in environmental sustainability, then neither the linear nor the geometric aggregation are suitable. • Instead, a non-compensatory multicriteria approach will assure non compensability by formalizing the idea of finding a compromise between two or more legitimate goals. • + does not reward outliers • + different goals are equally legitimate and important • + no normalisation is required • BUT • - computational cost when the number of countries is high Multi-criteria type of aggregation

  32. Step 7. Robustness and Sensitivity “Cynics say that models can be built to conclude anything provided that suitable assumptions are fed into them.” [The Economist, 1998] Step 7. Robustness and sensitivity “many indices rarely have adequate scientific foundations to support precise rankings: […] typical practice is to acknowledge uncertainty in the text of the report and then to present a table with unambiguous rankings” [Andrews et al. (2004: 1323)]

  33. Uncertainty + Sensitivity Analysis

  34. Steps in the Development of an Index stakeholder involvement Step 1. Developing a theoretical framework Step 2. Selecting indicators Step 3. Imputation of missing data Step 4. Multivariate analysis Step 5. Normalisation of data Step 6. Weighting and aggregation Step 7. Robustness and sensitivity Step 8. Links to other variables Step 9. Back to the details Step 10. Presentation and dissemination Sources of uncertainty

  35. As a result, an uncertainty analysis should naturally include a careful mapping of all these uncertainties (/assumptions) onto the space of the output (/inferences). • Two things can happen: • The latter outcome calls in turn for a revision of the ranking system, or a further collection of indicators. The space of the inference is still narrow enough, so as to be meaningful. The space of the inference is too wide to be meaningful.

  36. JRC/OECD Handbook on composite indicators [… and ranking systems] Why the ten steps? • The three pillars of a well-designed CI: • Solid conceptual framework, • good quality data, • sound methodology (+ robustness assessment) If the 3 conditions are met then CIs can be used for policy-making Composite indicators between analysis and advocacy, Saltelli, 2007, Social Indicators Research, 81:65-77

  37. JRC/OECD Handbook on composite indicators [… and ranking systems] Some recent CIs developed in collaboration with the JRC using the philosophy of the Handbook 2008 European Lifelong Learning Index (Bertelsmann Foundation, CCL) 2008/2006 Environmental Performance Index (Yale & Columbia University) 2007 Alcohol Policy Index (New York Medical College) 2007 Composite Learning Index (Canadian Council on Learning) 2002/2005 Environmental Sustainability Index(Yale & Columbia University) http://composite-indicators.jrc.ec.europa.eu/

  38. Composite Learning Index (Canadian Council on Learning) Framework Sensitivity analysis Policy message Results The Composite Learning Index (Canadian Council on Learning)

  39. Framework (WHO report) Comparative Analysis of Alcohol Control Policies in 30 Countries Brand, Saisana, Rynn, Pennoni, Lowenfels, 2007, PLoS Medicine, 4(4):752-759 Results Sensitivity analysis Policy message The Alcohol Policy Index (New York Medical College)

  40. Maybe an example of a country where strong laws are poorly enforced. high estimated amount of unrecorded consumption (>50% of the total). predominantly Islamic population The Alcohol Policy Index (New York Medical College)

  41. Our Tools for Sensitivity Analysis Two fractional variance measures First order effect – one factor influence by itself Total effect - influence inclusive of all interactions with other factors free software

  42. Selected References Books • Saltelli A., M. Ratto, T. Andres, F. Campolongo, J. Cariboni, D. Gatelli, M. Saisana, S. Tarantola, 2008, Global sensitivity analysis. The Primer, John Wiley & Sons, England. • OECD/JRC, 2008, Handbook on Constructing Composite Indicators. Methodology and user Guide, OECD Publishing, ISBN 978-92-64-04345-9. Authored by Nardo M., Saisana M., Saltelli A., Tarantola S., Hoffman A., Giovannini E.

  43. Selected References Journal articles • Brand, D. A., Saisana, M., Rynn, L. A., Pennoni, F., Lowenfels, A. B., 2007, Comparative Analysis of Alcohol Control Policies in 30 Countries, PLoS Medicine 4(4):752-759. • Cherchye, L., Moesen, W., Rogge, N., van Puyenbroeck, T., Saisana, M., Saltelli, A., Liska, R., Tarantola, S., 2008, Creating Composite Indicators with DEA and Robustness Analysis: the case of the Technology Achievement Index, Journal of Operational Research Society 59:239-251. • Saisana, M., Saltelli, A., Tarantola, S., 2005, Uncertainty and sensitivity analysis techniques as tools for the analysis and validation of composite indicators, Journal of the Royal Statistical Society A 168(2):307-323. • Saltelli, A., M. Ratto, S. Tarantola and F. Campolongo (2005) Sensitivity Analysis for Chemical Models, Chemical Reviews, 105(7) pp 2811 – 2828. • Munda G., Nardo M., Saisana M., 2009, Measuring uncertainties in composite indicators of sustainability. Int. J. Environmental Technology and Management (forthcoming).

  44. Selected References Journal articles • Munda G., Saisana M., 2009, Methodological Considerations on Regional Sustainability Assessment based on Multicriteria and Sensitivity Analysis. Regional Studies (forthcoming). • Tarantola S., Nardo M., Saisana M., Gatelli D., 2006, A new estimator for sensitivity analysis of model output: An application to the e-business readiness composite indicator. Reliability Engineering & System Safety 91(10-11), 1135-1141.

  45. Selected References EUR reports • Saisana, M., 2008, The 2007 Composite Learning Index: Robustness Issues and Critical Assessment, EUR Report 23274 EN, European Commission, JRC-IPSC, Italy. • Saisana M., D’Hombres B., 2008, Higher Education Rankings: Robustness Issues and Critical Assessment, EUR Report 23487 EN, European Commission, JRC-IPSC, Italy. • Saisana M., Munda G., 2008, Knowledge Economy: measures and drivers, EUR Report 23486 EN, European Commission, JRC-IPSC. • Saisana M., Saltelli, A., 2008, Sensitivity Analysis for the 2008 Environmental Performance Index, EUR Report 23485 EN, European Commission, JRC-IPSC.

  46. Selected References EUR reports • Munda G. and Nardo M. (2005), Constructing Consistent Composite Indicators: the Issue of Weights, EUR 21834 EN, Joint Research Centre, Ispra. • Nardo M., Tarantola S., Saltelli A., Andropoulos C., Buescher R. , Karageorgos G. , Latvala A. and Noel F. (2004), The e-business readiness composite indicator for 2003: a pilot study, EUR 21294. • Saisana M., Tarantola S., 2002, State-of-the-art Report on Current Methodologies and Practices for Composite Indicator Development, EUR Report 20408 EN, European Commission, JRC-IPSC, Italy.

More Related