Difference between revisions of "Evaluating performance of environmental health assessments"

From Testiwiki
Jump to: navigation, search
m
m
 
(3 intermediate revisions by 2 users not shown)
Line 1: Line 1:
<accesscontrol>Members of projects,,Workshop2008,,beneris,,Erac,,Heimtsa,,Hiwate,,Intarese</accesscontrol>
+
[[heande:Evaluating performance of environmental health assessments]]
 +
[[Category:Open policy practice]]
 +
[[Category:Decision analysis and risk management]]
 +
{{nugget
 +
|authors= [[User:Mikko Pohjola|Mikko V. Pohjola]], Marko Tainio, [[User:Jouni|Jouni T. Tuomisto]]
 +
|reference=  Pohjola, M.V.; Pohjola, P.; Tainio, M.; Tuomisto, J.T. Perspectives to Performance of Environment and Health Assessments and Models—From Outputs to Outcomes? Int. J. Environ. Res. Public Health 2013, 10:2621-2642. http://dx.doi.org/10.3390/ijerph10072621
 +
|urn= 
 +
|ethics= 
 +
|journal=
 +
}}
  
[[Category:Manuscript]]
+
:''This page is the final manuscript of'' Pohjola, M.V.; Pohjola, P.; Tainio, M.; Tuomisto, J.T. Perspectives to Performance of Environment and Health Assessments and Models—From Outputs to Outcomes? Int. J. Environ. Res. Public Health 2013, 10:2621-2642. http://dx.doi.org/10.3390/ijerph10072621. ''Originally, the text was written on page [[:heande:Evaluating performance of environmental health assessments]].
  
This is a manuscript on evaluating performance of environmental health assessments. It discusses currently prevailing perspectives to evaluating the goodness of assessments and proposes a new more comprehensive approach based on properties of good assessments.
+
'''Keywords
  
== Title ==
+
Evaluation, EMS, Effectiveness, Open Assessment, REA, Outcome
  
'''The properties of good assessments - a new approach to evaluating performance of environmental health assessments
+
'''Highlights
 +
* Conventional evaluation of models and assessments focuses on processes and outputs
 +
* Recent evaluation approaches also emphasize societal model/assessment outcomes
 +
* Model/assessment effectiveness is the likelihood of delivering intended outcomes
 +
* An outcome-oriented turn is taking place in modelling, assessment and evaluation
 +
* New views combine design, making and evaluation of models, assessments and practices
  
 
== Abstract ==
 
== Abstract ==
  
Environmental health assessments produce descriptions of reality as answers to questions concerning relations between environmental phenomena and human health and well-being. This information can be used by various kinds of actors in multiple societal contexts in order to create understanding about these phenomena and to decide upon possible actions of dealing with them and their effects. The quality of the descriptions and the ways they are produced and provided for use significantly influence the effectiveness of the information. Also efficiency in producing and using the information is important with regard to the goodness of environmental health assessments. Understanding the factors that constitute the overall performance of environmental health assessments is thus crucial in evaluating assessments and especially in designing and executing good assessments.
+
This paper reviews different perspectives to evaluation and management of model and assessment performance. While there is an increasing need to evaluate the success of environment and health relevant models and assessments according to their societal outcomes, most perspectives are primarily focused on the outputs of modelling and assessment as well as the procedures of their making. The practical application of outputs and the consequential societal outcomes, are most often left without explicit consideration. Only quite recently, the approaches to evaluation and management of model and assessment performance have recognized the outcomes as an essential performance criterion.
  
Various perspectives to consider performance of assessments have been taken building on the bases of different fields of study, but a fully comprehensive approach that would sufficiently address all aspects of this complex issue in a concise manner has not yet evolved. The two main views of how assessment performance has been considered in recent scientific literature on fields related to environmental health can be characterized as (1) the quality assurance approach, attempting to provide procedural guidelines for good assessment practice, and (2) the uncertainty analysis approach, attempting to identify and categorize types of uncertainties that tend to reside in the information produced in assessments.
+
The perspectives to model and assessment performance can be categorized according to their foci as i) quality assurance/control, ii) uncertainty analysis, iii) technical assessment of models, iv) effectiveness and v) other perspectives. In practice many approaches found in the literature combine characteristics of different perspectives, but the categorization, however, illustrates which aspects are primarily emphasized.
  
We propose a new approach for considering the performance of environmental health assessments that builds on considering assessments as trialogical processes of creating collective belief systems. It describes the general properties of good assessments which can be categorized as properties related to (i) quality of information content, (ii) applicability of information and (iii) efficiency of assessment. The properties of good assessments can be used as assessment design and execution principles as well as a framework for evaluating past assessments.
+
It seems that the existing approaches and perspectives do not sufficiently serve all needs of evaluation and management of model and assessment performance. We need a more comprehensive approach that covers a) the making of models and assessments, b) their outputs, c) their use and other interaction with the societal context, as well as d) the decisions and actions influenced by the models and assessments, and e) the consequential outcomes. Achieving this necessitates a thorough account of the mechanisms of collective knowledge creation and the relations between knowledge and action - aspects that are often not recognized as relevant to evaluation and management of model and assessment performance. Some attempts to combine certain collaborative and pragmatic methods, frameworks and tools into a comprehensive methodology for design, making and evaluation of models, assessments as well as practices are being made, and the early results look promising.
  
The proposed new approach is capable of incorporating the main aspects of both the quality assurance approach and the uncertainty assessment aproach, but provides a more comprehensive and coherent framework that addresses the assessment process, assessment products as well as the use processes as a whole. It considers the assessment product as the mediating shared object of activity and scrutinizes it as the central object bridging the assessment process with the use process and thereby enables explication of all relevant aspects of assessment performance in a single framework.
+
== Introduction ==
 +
 
 +
In this paper we review different perspectives to evaluation and management of model and assessment performance. While there is an increasing need to evaluate the success of models, modelling, and assessment upon issues of environment and health according to their societal outcomes, most evaluation approaches do not seem to support this need. In a recent thematic issue on the assessment and evaluation of environmental models and software (McIntosh et al., 2011a), Matthews et al. (2011) opened up a debate on how to best respond to the increasing desire to evaluate the success of environmental modelling and software projects in terms of their outcomes, i.e changes to values, attitudes, and behaviour outside the walls of the research organization, rather than mere outputs. Until now, there has been limited appreciation within the environmental modelling and software community regarding the challenges of shifting the focus of evaluation from outputs to outcomes (Matthews et al., 2011).
 +
 
 +
The current situation in the different overlapping communities of environment and health related assessment, such as integrated assessment (e.g. van der Sluijs, 2002), health impact assessment (e.g. WHO, 1999), risk assessment (e.g. NRC, 1983, 1996, 2009), chemical safety assessment (ECHA, 2008), environmental impact assessment (e.g. Wood, 1995), and integrated environmental health impact assessment (Briggs, 2008), appears to be somewhat similar. For example, a recent study on the state of the art in environmental health assessment revealed that despite the fact that most assessment approaches explicitly state influencing of society as their aim, this is rarely manifested in the principles and practices of evaluating assessment performance (Pohjola et al., 2012).
 +
 
 +
The emphasis in the scientific discourses on evaluating models and assessments has been on rather scientific and technical aspects of evaluation within the research domain, and perspectives that also duly address the impacts of modelling and assessment in broader societal contexts have emerged only quite recently and are still relatively rare (cf. McIntosh et al., 2011b). Such evaluations are qualitatively different (Matthews et al., 2011), which invokes a need for reconsidering the criteria for evaluating modelling and assessment endeavours, as well as the frameworks within which they are applied.
 +
 
 +
Evaluation of models and assessments is not only a matter of judging the goodness of a model or assessment, but has a more profound influence on their use as well as making. On the one hand, evaluation informs users e.g. regarding the possible and appropriate uses of models and assessment or model outputs. On the other hand, the criteria for evaluations also tend to direct the design and execution of modelling and assessment endeavours to meet those criteria, particularly if they are explicitly known in advance (cf. What you measure is what you get (WYMIWYG) by Hummel and Huitt, 1994).
 +
 
 +
This paper contributes to the debate by first describing a framework within which model and assessment performance can be considered, then reviewing different perspectives to model and assessment performance, and finally discussing the capability of existing approaches to address the challenges of what could be called the "outcome-oriented turn in evaluation and management of models and assessments". Some current attempts to develop a comprehensive approach to design, making and evaluation of models, assessments as well as practices are also briefly presented.
 +
 
 +
== Models, assessments and performance ==
 +
 
 +
For several years, the authors have been engaged with developing methods and tools for environmental health modelling and assessment. Most notably this work has resulted in the open assessment method and Opasnet web-workspace for science-based collaborative knowledge creation and policy support upon virtually any topic (Pohjola et al. 2011, 2012, Tuomisto and Pohjola 2007, http://en.opasnet.org). These developments have required thorough familiarization of and several inquiries into the kinds of existing modelling and assessment means and practices, and particularly the factors that determine how well they perform. This review is based on the extensive information and knowledge base on model and assessment performance that has been obtained by these inquiries.
 +
 
 +
Here modelling and assessment are considered as instances of fundamentally the same issue of science-based support to decision making regarding environment and health. In fact, assessments virtually always involve modelling of some kind, at least implicit conceptual models. Conversely, modelling is also often identified with assessment (Jakeman and Letcher, 2003). In addition, decision support systems, information support tools, integrated modelling frameworks and other software tools and information systems to assist in developing, running, and analyzing models and assessments are perceived as integral parts of modelling and assessment (e.g. Rizzoli et al., 2008).
 +
 
 +
Models and assessments can be considered e.g. as diagnostic, prognostic, or summative according to the kinds of questions they address (Briggs, 2008), ex-ante or ex-post according to their timing in relation to the activities being assessed (Pope et al., 2004), and regulatory or academic according to the contexts of their development and application (Pohjola et al., 2012). They can also be developed, executed, and applied by many kinds of actors, e.g. consultants, federal agencies or academic researchers. Modelling and assessment should be, however, clearly distinguished from purely curiosity-driven research, ad hoc assessments, and models or assessments made only to justify predetermined decisions. In addition, it should be noted that "assessment of models", often interchangeable with "evaluation of models", is also excluded from the concept of assessment adopted here. Instead, technical assessment of models is considered in the review below as one of the common perspectives to evaluating model and assessment performance. Altogether, modelling and assessment can be considered as fundamentally having two purposes: i) describing reality, and ii) serving the needs of practical decision-making.
 +
 
 +
Inspired by the conceptual framework presented by Matthews et al. (2011), and influenced by the works of e.g. Blackstock et al. (2007), Bina (2008), Leviton (2003), and Patton (2002), Figure 1 describes the essential aspects in the interaction of modelling and assessment with their broader societal context.
 +
 
 +
[[File:Context Process and product.PNG|thumb|center|700px|Figure 1. The essential aspects in the interaction of modelling and assessment with their societal context.]]
 +
 
 +
The endeavours of modelling and assessment are here broken down into 1) process, the procedures and practices of developing and executing models and assessments, 2) output, the products or results that come out of these modelling and assessment processes, and 3) use, the practical application of the outputs. The surrounding context provides enablers for modelling and assessment endeavours to take place e.g. in the form of funding, facilities and education, but also as acceptance or demand for certain kinds of modelling and assessment endeavours. On the other hand, the context also poses constraints e.g. in the form of prioritization of scarce resources, non-acceptance of some modelling and assessment endeavours, or lacking of capability for applying their outputs. The societal context is also the medium where the outcomes of modelling and assessment, i.e. the changes in values, attitudes and behaviour among others than assessors and modellers themselves are (or are not) realized. It is important to distinguish the outcomes from what Matthews et al. (2011) termed process effects, i.e. the changes in the capacity of those engaged in the modelling and assessment endeavours, which does not directly result in impacts in the broader societal context.
 +
 
 +
In Figure 1, use is intentionally located on the boundary between the modelling and assessment domain and the context in order to emphasize that the practical application of outputs should be considered as a shared responsibility between the modellers or assessors and the intended users outside the modelling and assessment domain. This implies that the responsibility of modellers and assessors should not end in completion of the output. On the other hand, the users of the outputs should not expect realization of the outcomes without having to invest any effort themselves in their delivery. The boundary between the modelling and assessment domain and its context is marked with a dashed line, indicating that the domain should not be perceived as a strictly separate entity, but in continuous interaction with its context.
 +
 
 +
The case-specific uses of modelling and assessment outputs in varying contexts are critical for achieving the intended outcomes of models and assessments. Therefore, in order for models or assessments to perform well, it is not sufficient to only provide good descriptions of reality. They must also fulfil their instrumental purposes of serving the needs of their intended uses, and thereby resulting in the desired outcomes. However, this aspect of model and assessment performance is often given little attention, as will turn out when the existing perspectives to model and assessment performance are reviewed and discussed in light of the framework described in Figure 1.
  
== Introduction ==
+
== Perspectives to model and assessment performance ==
 +
 
 +
Within the abundance of contributions in the scientific literature addressing model and assessment performance, certain commonalities between contributions can be identified. These contributions can be categorized as representing different perspectives according to their primary focus: i) quality assurance/control, ii) uncertainty analysis, iii) technical assessment of models, iv) effectiveness, or v) other perspectives. These perspectives provide the structure for the review. As the number of different relevant contributions is so large that it is not be possible to include them all in the review, we have instead attempted to compile representative and comprehensive collections of examples, grouped according to certain subcategories, for each perspective. Recent contributions are emphasized, but some important or illustrative examples that were published before 2000 have been included as well. Also the relevant contributions stemming from the method and tool development by the authors are also included in the review.
 +
 
 +
=== Quality assurance/control ===
 +
 
 +
One of the major themes in modelling and assessment performance related literature is what can be referred to as quality assurance/control (QA/QC) perspective. The focus in this perspective is primarily on determining how the processes of modelling or assessment are to be conducted in order to assure the quality of the output. The QA/QC perspective appears commonly in both modelling and assessment literature, sometimes extending to also consider the processes of decision-making.
 +
 +
There are multiple alternative definitions for quality (see e.g. Reeves and Bednar, 1994). However, as regards modelling and assessment, the interpretation is mostly analogous with the perception in the ISO-9000 framework, i.e. as the organisational structures, responsibilities, procedures, processes, and resources to assure and improve quality (Harteloh, 2003). Also the hierarchy of evidence, commonly applied in the field of medicine, is a quality assurance/control perspective, which ranks types of evidence strictly according to the procedure by which they were obtained (Guyatt et al., 1995). However, as pointed out by Cartwright (2007) with regard to randomized controlled trials, the procedure alone cannot guarantee delivery of useful information in practical contexts.
 +
 
 +
One common variation of this perspective is stepwise procedural guidance. Such guidance provides often relatively strict and detailed descriptions of the steps or phases of a modelling or assessment process that are to be executed in a more or less defined order. Faithful execution of the procedure is assumed to lead to good outputs. A somewhat similar, but perhaps less rigorous, variation of the QA/QC perspective is checklist-type guidance emphasizing issues that need to be taken account of in the modelling or assessment process or their evaluation. The checklists can be more or less detailed and they usually do not strictly define the order or sequence of execution.
 +
 
 +
Also the accounts that address evaluation of input quality can be considered as manifestations of the QA/QC perspective. However, the primary focus in QA/QC is often on the outputs, and the input quality evaluations typically complement uncertainty analyses or technical assessments of models. For example, model parameter uncertainty analysis can be considered as an example of evaluation of input quality, but in practice it is most often considered as an aspect of either uncertainty analysis or technical assessment of models.
 +
 
 +
{|{{prettytable}}
 +
|+ '''Table 1. Examples of quality assurance/control perspective to modelling and assessment performance.
 +
|-----
 +
! Type
 +
! Author(s)
 +
! Description
 +
|-----
 +
| rowspan="7"| Stepwise procedural guidance
 +
| Jakeman et al. 2006
 +
| Ten iterative steps in development and evaluation of environmental models
 +
|-----
 +
| Refsgaard et al. 2005
 +
| HarmoniQuA guidance for quality assurance in multidisciplinary model-based water management
 +
|-----
 +
| van Delden et al. 2011
 +
| Methodology for the design and development of integrated models for policy support
 +
|-----
 +
| Briggs 2008
 +
| Framework for integrated environmental health impact assessment
 +
|-----
 +
| Hoekstra et al. 2010
 +
| BRAFO tiered approach for benefit-risk assessment of foods
 +
|-----
 +
| Liu et al. 2008
 +
| Generic framework for effective decision support through integrated modelling and scenario analysis
 +
|-----
 +
| Mahmoud et al. 2009
 +
| Formal framework for scenario development in support of environmental decision making
 +
|-----
 +
| rowspan="4"| Checklist guidance
 +
| Granger Morgan and Dowlatadabi 1996
 +
| Seven attributes of good integrated assessment of climate change
 +
|-----
 +
| Risbey et al. 1996
 +
| Listing of end use independent process based considerations for integrated assessment
 +
|-----
 +
| Forristal et al. 2008
 +
| QA/QC performance measurement scheme for risk assessment in Canada
 +
|-----
 +
| Risbey et al. 2005
 +
| Checklist for quality assistance in environmental modelling
 +
|-----
 +
| rowspan="3"| Evaluation of input quality
 +
| van der Sluijs et al. 2005
 +
| Pedigree analysis in model-based environmental assessment
 +
|-----
 +
| Brown et al. 2005
 +
| Methodology for recording uncertainties about environmental data
 +
|-----
 +
| Kloprogge et al. 2011
 +
| Method for analyzing assumptions in model-based environmental assessments
 +
|}
 +
 
 +
Characteristic for stepwise guidance is that it attempts to predetermine a procedure in order to guarantee good quality of outputs. As such, it takes a proactive approach to managing performance. Checklist-type guidance and evaluation of input quality can also be applied proactively, but the examples found in literature mostly represent a reactive approach of evaluating already executed modelling and assessment processes.
 +
 
 +
=== Uncertainty analysis ===
 +
 
 +
Another major theme in the modelling and assessment performance literature can be referred to as the uncertainty analysis perspective. The contributions within this theme vary significantly, ranging from descriptions of single methods to overarching frameworks, but the common main idea is characterization of certain properties of the outputs of models and assessments. Fundamentally the perspective builds on quantitative statistical methods based on probability calculus (O'Hagan, 2011), but also other than probability-based approaches to uncertainty have been presented (e.g. Colyvan, 2008). Many manifestations of this perspective in the context of environment and health modelling and assessment also extend to consider qualitative properties of the outputs. The uncertainty analysis perspective appears to be more commonly represented in the domain of assessment than of modelling. In addition to model and assessment outputs, models themselves could be considered as outputs whose properties are considered and evaluated in a similar fashion. However, this issue is addressed as a separate perspective in the next section, technical assessment of models.
 +
 
 +
One variation of the uncertainty analysis perspective is identification and typifying of the kinds and sources of uncertainty that characterize the modelling or assessment outputs and their quality. Some uncertainties are often considered as being primarily expressible in quantitative, while others in qualitative terms. The sources of uncertainty can also extend to include aspects of the modelling and assessment processes. In some cases also intended or possible uses and use contexts of the outputs are acknowledged.
 +
 
 +
Also different types of guidance on how to assess or deal with different kinds of uncertainties exist. Such frameworks usually combine both qualitative and quantitative aspects of uncertainty deriving from various sources. Consequently, aspects of the processes of producing the outputs, e.g. input quality, and acknowledgment of the intended or possible uses and use contexts of the outputs, e.g. in terms of acceptance, are often also included in the frameworks. The primary focus is, however, on the characteristics of the model and assessment outputs.
 +
 
 +
In addition to the broad characterizations and frameworks, also numerous more or less explicit methods, means and practices to analyze uncertainties of model and assessment outputs exist. For example, sensitivity, importance, and value of information analysis and Bayesian modelling, in addition to the standard statistical characterization, are essential in the context of environment and health modelling and assessment. Such methods are dominantly quantitative. The standard means for statistical analysis are considered common knowledge and not discussed here in any more detail.
  
Environmental health assessments produce descriptions of reality as answers to questions concerning relations between environmental phenomena and human health [VIITE]. As environment can be considered to cover about the whole reality that surrounds us and health can be considered to cover a whole spectrum of different diseases as well as aspects of well-being, it can be seen that the scope of environmental health assessment is a very wide-ranging and complex one, and it can be broken down into several more detailed or strictly bound disciplines. The term environmental health assessment is used in this paper in a very broad sense as an overarching meta-discipline that can include all kinds of systematic inquiries that provide information relevant to environment and health. Such activities take place e.g. under the labels of integrated assessment and modeling, risk assessment, health impact assessment, environmental impact assessment, environmental modeling and water resources modeling and management to name a few.
+
{|{{prettytable}}
 +
|+ '''Table 2. Examples of uncertainty analysis perspective to model and assessment performance.
 +
|-----
 +
! Type
 +
! Author(s)
 +
! Description
 +
|-----
 +
| rowspan="3"| Identification of kinds of uncertainty
 +
| Walker et al. 2003
 +
| Conceptual basis for uncertainty management in model-based decision support
 +
|-----
 +
| Briggs et al. 2008
 +
| Uncertainty in epidemiology and health risk and impact assessment
 +
|-----
 +
| van Asselt and Rotmans 2002
 +
| Uncertainty in integrated assessment modelling
 +
|-----
 +
| rowspan="3"| Guidance on dealing with uncertainties
 +
| van der Sluijs et al. 2008
 +
| Knowledge quality assessment for complex policy decisions
 +
|-----
 +
| Blind and Refsgaard 2007
 +
| Operationalising uncertainty in integrated water resource management
 +
|-----
 +
| Refsgaard et al. 2007
 +
| Framework and guidance for dealing with uncertainty in environmental modelling
 +
|-----
 +
| rowspan="4"| Methods for uncertainty analysis
 +
| Kann and Weyant 2000
 +
| Approaches for performing uncertainty analysis in large-scale energy/economic policy models
 +
|-----
 +
| Brouwer and De Blois 2008
 +
| Modelling of risk and uncertainty underlying the cost and effectiveness of water quality measures
 +
|-----
 +
| Basson and Petrie 2007
 +
| Consideration of uncertainty in decision making supported by Life Cycle Assessment
 +
|-----
 +
| Borgonovo 2008
 +
| Sensitivity analysis of model outputs with input constraints
 +
|}
  
The information produced in environmental health assessments can be used by various kinds of actors in multiple societal contexts in order to create understanding about these phenomena and to decide upon possible actions of dealing with them and their effects [VIITE]. It can be considered as providing input to practical decision-making in political processes, but also in e.g industrial and commercial organizations as well as on the level of everyday activities of an individual citizen [VIITE]. In addition to providing aids for finding solutions to specific decision making problems the information provided by environmental health assessments can function in increasing the general level of awareness and understanding about important issues relevant to environment and health [VIITE]. This kind of societal learning can also take place on several levels, e.g. among political and industrial decision makers, NGOs and other stakeholder organizations as well as public at large, but also within the community of scientific experts on the particular or related fields [VIITE]. The general intentionality of environmental health assessments can thus be characterized as being three-fold as:
+
For the examples of uncertainty analysis perspective, it appears characteristic that the issue of uncertainty is typically approached from an external observer’s point of view. The evaluation of performance is thus mainly considered as a separate, typically ex-post, activity taking place in addition to the actual modelling or assessment process, not as its integral proactive part.
  
#Answering to specific needs of decision making
+
=== Technical assessment of models ===
#Contributing to societal collective learning about environment and health
 
#Contributing to scientific research on environment and health
 
  
The simultaneous parallel goals of environmental health assessment and the range of potential receivers of the information produced in them bring about plural requirements for the quality of the information and the ways of producing and providing it. It appears obvious that there are several simultaneously affecting factors arising from different aspects of an environmental health assessment that can significantly influence the effectiveness of its products. For example credibility of data sources, means of data synthesis, framing of  assessment problem, clarity of information representation or choice of media for disseminating assessment results can all be crucial with regard to achieving the intended outcomes of assessment [VIITE]. Also efficiency of the processes of producing and using the information is important when considering the goodness of assessments in the world with ever-present scarcity of resources [VIITE].
+
If the uncertainty analysis perspective is more commonly adopted in assessment literature, the technical assessment of models is a grand theme particularly in the modelling literature. In contrast to the focus on model or assessment outputs in the uncertainty analysis approach, the focus in this perspective is on the models and their characteristics. In addition to models, the perspective also considers different kinds of software tools that are applied in developing, running, and analyzing the models.
  
It is necessary to understand the factors that constitute the overall performance of environmental health assessments in order to be able to reasonably evaluate the goodness of assessments [VIITE]. Furthermore, and even more importantly, it is essential in being able to design and execute good assessments [VIITE?]. Useful conceptual tools for addressing this issue are thus needed. Various perspectives to consider the performance of assessments have been taken building on the bases of different fields of study relevant to environmental health assessment, but a fully comprehensive approach that would sufficiently address all aspects of this complex issue in a concise manner has not yet evolved [VIITE].
+
Particularly the object of interest in technical assessment of models is development and application of formal methods for testing and evaluating models within defined domains of application. Generally, model evaluation and performance is considered to cover structural features of models, representativeness of model results in relation to a certain part of reality, as well as usefulness with regard to a designated task (cf. Beck, 2002). However, if usefulness is considered at all, it mainly addresses expert use of models, which corresponds mostly to the so-called process effects rather than outcomes. Most commonly technical assessment of models takes place in terms of validation and verification by comparing models and their results against each other or measurement data.
  
In the following chapter we briefly review some recently published contributions to evaluating performance of environmental health assessments and make an attempt to describe the general characteristics of the most commonly held views. Then we propose a new framework for evaluating the performance of environmental health assessments, which attempts to address the issues in a more comprehensive, yet concise, manner. In the end of the paper we discuss the differences in perspectives and the capabilities of the currently prevailing approaches and the new approach in helping to achieve the general goals of environmental health assessment as explained above.
+
A variation of this perspective, more common for the discourses in assessment literature, is analysis of model uncertainty. Here the aim typically is to characterize the properties of a model in order to be able to correctly interpret or evaluate its outputs. Model uncertainty is often considered as one aspect of a broader uncertainty concept.
  
== Approaches to assessment performance ==
+
{|{{prettytable}}
 +
|+ '''Table 3. Examples of technical assessment of models perspective to model and assessment performance.
 +
|-----
 +
! Type
 +
! Author(s)
 +
! Description
 +
|-----
 +
| rowspan="8"| Means for model and software evaluation
 +
| Alexandrov et al. 2011
 +
| Realistic criteria for environmental model and software evaluation
 +
|-----
 +
| Refsgaard et al. 2004
 +
| Terminology and methodological framework for modelling and model evaluation
 +
|-----
 +
| Matthews et al. 2011
 +
| Evaluation methods of environmental modelling and software in a comprehensive conceptual framework
 +
|-----
 +
| Bai et al. 2009
 +
| Top-down framework for watershed model evaluation and selection
 +
|-----
 +
| Wyat Appel et al. 2011
 +
| Overview of atmospheric model evaluation tool (AMET)
 +
|-----
 +
| Xu et al. 2007
 +
| Appropriateness framework for the Dutch Meuse decision support system
 +
|-----
 +
| Sojda 2007
 +
| Empirical evaluation of decision support systems
 +
|-----
 +
| Wagener and Kollat 2007
 +
| Numerical and visual evaluation of hydrological and environmental models
 +
|-----
 +
| rowspan="4"| Evaluation of models
 +
| Mo et al. 2011
 +
| Evaluation of an ecosystem model for wheat-maize cropping system in North China
 +
|-----
 +
| Pollino et al. 2007
 +
| Parameterisation and evaluation of a Bayesian network for use in an ecological risk assessment
 +
|-----
 +
| Sonneveld et al. 2011
 +
| Evaluating quantitative and qualitative models for water erosion assessment in Ethiopia
 +
|-----
 +
| Aertsen et al. 2011
 +
| Evaluation of modelling techniques for forest site productivity prediction using SMAA
 +
|-----
 +
| rowspan="3"| Analysis of model uncertainty
 +
| Nilsen and Aven 2003
 +
| Model uncertainty in the context of risk analysis
 +
|-----
 +
| Moschandreas 2002
 +
| Scenario, model and parameter uncertainty in risk assessment
 +
|-----
 +
| Refsgaard et al. 2006
 +
| Framework for dealing with uncertainty due to model structure error
 +
|}
  
*uncertainty analysis
+
The technical assessment of models is predominantly a reactive perspective to evaluate models, as it requires an existing model or software system that can be tested and analyzed. The evaluation, however, is usually considered as being an integral part of the model development, not a separate entity, which also enables application of technical assessment of models in different developmental stages within the modelling or assessment process. On the other hand, the common practice of self evaluation of models may also lead e.g. to limited usability, credibility and acceptability unless interaction with the context is otherwise realized.
**statistical (trad.) + broadened, more intangible, uncertainty (e.g. pedigree)
 
**stresses the product (as such)
 
  
*quality assurance
+
=== Effectiveness ===
**origins in engineering and industrial production
 
**stresses production process
 
  
It seems that the scientific discourse regarding issues relevant to performance of environmental health assessment or related disciplines has somewhat shifted broadened the last 15 years from discussions on evaluating or reducing quantitative uncertainty within assessment (or model) outputs mainly by statistical means towards discussions that also address the more qualitative aspects of assessment performance arising not only from the products of assessments, but also from the processes of producing, and sometimes also using, the assessment outputs. This has resulted to a variety of different perspectives being presented and applied in attempts to evaluate the goodness of assessment outputs or finding the factors that tend to improve the goodness of assessments. For example Refsgaard et al. (2007) identified 14 different partly complementary methods to commonly used in uncertainty assessment and characterization in their study on uncertainty in environmental modeling.
+
Whereas the three former perspectives have been subjects of discussion regarding model and assessment performance for longer times, emphasizing of model and assessment effectiveness can be considered to have become a major topic only quite recently. Contributions addressing effectiveness appear more common in the assessment literature, but the topic has recently been addressed also in the context of modelling and software.
  
We have identified two main views of considering assessment performance in recent scientific literature on fields related to environmental health. They can be characterized as (1) the quality assurance approach, attempting to provide procedural guidelines for good assessment (or modeling) practice (e.g. Forristal et al. 2008, Jakeman 2006, Guimarães Pereira and Funtowicz 2005, Refsgaard et al. 2005; 2004, Risbey et al. 1996), and (2) uncertainty analysis approach, attempting to identify and categorize types and sources of uncertainties that tend to reside in the information produced in assessments (or by models) (e.g. Blind and Refsgaard 2007, van der Sluijs 2005, Brown et al. 2005, Walker et al. 2003, van Asselt and Rotmans 2002, Kann and Weyant 2000). These views are however not exclusive, but rather represent alternative perspectives to basically the same issue which can generally be called assessment performance. In fact some studies have applied both approaches in order to characterize uncertainties and then draw procedural guidelines for reducing them (e.g. Refsgaard et al. 2007; 2006, van der Sluijs et al. 2008, Risbey et al. 2005, Janssen et al. 2005). Also other relevant contributions to scientific discourse on assessment performance which do not fall neatly into either of the above-mentioned categories have been given, e.g. the seven attributes of a good integrated assessment of climate change (Morgan and Dowlatadabi 1996), important features of integrated assessment modeling (Parker et al. 2002) and insights from the UPEM conference (van der Sluijs 2006).
+
In the effectiveness perspective, the aim of modelling and assessment is generally seen as to promote changes in values, attitudes, and behaviour outside the walls of the research community (Matthews et al., 2011) by maximizing the likelihood of an assessment process to achieve the desired results and the goals set for it (Hokkanen and Kojo, 2003). In principle, performance of models and assessments is thus fundamentally characterized in terms of the impacts delivered into the broader societal context. However, due to the complexity of reality, evaluation of outcomes is often perceived as very difficult, if not impossible (Kauppinen et al., 2006), and possibly even leading to incorrect conclusions regarding effectiveness (cf. Ekboir, 2003). Consequently, the effectiveness criteria and frameworks often address various aspects of process and output, as well as contextual enablers and constraints, rather than outcomes, as factors of effectiveness. Some contributions also make a distinction between (immediate) impacts and (indirect) outcomes. As a result, although the aim is to address outcomes, some approaches to effectiveness in the end turn out somewhat similar to those considered as checklist guidance in quality assurance/control (in Table 1).
  
*MNP guidance
+
In addition, the approaches emphasizing the use of models, tools and their outputs can be considered as a manifestation of the effectiveness perspective. They can generally be characterized as attempts to operationalise the interaction of modelling and assessment with the practical uses of their outputs. Most of the contributions are, however, relatively tool-centred (cf. evaluation of decision support systems as models), and most often little attention is given to the cognitive processes involved in the delivery and reception of information produced by models and assessments.
*EPA modeling / modeling QA guidance
 
*IRGC framework: assessment typification
 
  
Let us now look into some general characteristics as well as similarities and differences of these two main approaches. In the uncertainty analysis approach the focus is primarily on the information produced in an assessment or the model used in the assessment. The uncertainty analysis approach attempts to identify and characterize the types and sources of uncertainties that the information might contain. Also guidance on how to estimate uncertainties is sometimes provided. In the uncertainty analysis methods that identify and address also qualitative aspects of uncertainty, some uncertainty types actually stem from either the assessment process (e.g. choice and reliability of data sources, model validation), the way the information is provided for use (e.g. representation, chosen media), or the match between the information content and its use (e.g. fitness for purpose, robustness) not directly from the information content as such. Even so, the perspective to the process is "through the product". The uncertainty analysis approach explicitly considers the evaluation of performance as a separate process taking place in practice often in the late phases of the assessment or after the assessment has finished, when the information product is more or less ready.
+
{|{{prettytable}}
 +
|+ '''Table 4. Examples of effectiveness perspective to model and assessment performance.
 +
|-----
 +
! Type
 +
! Author(s)
 +
! Description
 +
|-----
 +
| rowspan="7"|Frameworks and criteria for effectiveness
 +
| Kauppinen et al. 2006
 +
| Framework for the effectiveness of prospective human impact assessment
 +
|-----
 +
| Quigley and Taylor 2004
 +
| Process, impact and outcome indicators for evaluating health impact assessment
 +
|-----
 +
| Clark and Majone 1985
 +
| Criteria for appraisal of scientific inquiries with policy implications
 +
|-----
 +
| Hildén et al. 2004
 +
| Necessary conditions and facilitating factors for effectiveness in strategic environmental assessment
 +
|-----
 +
| Baker and McLelland 2003
 +
| Components of policy effectiveness in participatory environmental assessment
 +
|-----
 +
| Pohjola and Tuomisto 2011
 +
| Dimensions of openness for analyzing the potential for effectiveness in participatory policy support
 +
|-----
 +
| Tuomisto and Pohjola 2007
 +
| Properties of good assessment for evaluating effectiveness of assessments (updated description at: http://en.opasnet.org/w/Properties_of_good_assessment)
 +
|-----
 +
| rowspan="5"|Effectiveness evaluations
 +
| Wismar et al. 2007
 +
| Several cases of evaluating effectiveness of health impact assessment in Europe
 +
|-----
 +
| Fischer and Gazzola 2006
 +
| General effectiveness criteria for strategic environmental assessment and their adaptation for Italy
 +
|-----
 +
| Leu et al. 1996
 +
| Environmental impact assessment evaluation model and its application in Taiwan
 +
|-----
 +
| Pölönen et al. 2010
 +
| Effectiveness of the Finnish environmental impact assessment system
 +
|-----
 +
| Matthews et al. 2011
 +
| Example of outcome evaluation for environmental modelling and software
 +
|-----
 +
| rowspan="8"| Use of models, tools and assessments
 +
| Larocque et al. 2011
 +
| Framework to assist decision makers in the use of ecosystem model predictions
 +
|-----
 +
| Sterk et al. 2011
 +
| Analysis of contribution of land-use modelling to societal problem solving
 +
|-----
 +
| Diez and McIntosh 2011
 +
| Use of decision and information support tools in desertification policy and management
 +
|-----
 +
| McIntosh et al. 2008.
 +
| Developing tools to support environmental management and policy
 +
|-----
 +
| Siebenhüner and Barth 2005
 +
| Role of computer modelling in participatory integrated assessments
 +
|-----
 +
| Inman et al. 2011
 +
| Usage and perceived effectiveness of decision support systems in participatory planning
 +
|-----
 +
| Dewar et al. 1996
 +
| Credible uses of the distributed interactive simulation (DIS) system
 +
|-----
 +
| Pohjola et al. 2012
 +
| Analysis of interaction between environmental health assessment and policy making
 +
|}
  
In the quality assurance approach the focus is primarily on the assessment (or modeling) process. The quality assurance approach attempts to provide procedural guidelines whose purpose is to lead the assessors to do every important task during an assessment, and thereby guarantee the goodness of the output. The guidelines may be either in the form of assessment/modeling steps to follow, as highlighting certain procedural issues that deserve more or special attention or as guidelines to conducting a separate quality assurance/control process alongside the assessment/modeling process. Among the quality assurance guidelines there is a varying degree of emphasis on the interaction between the assessors/modelers and the intended users of assessment outputs as well as on recognizing the purpose of the assessment or modeling study.
+
The approaches to effectiveness range from external ex-post evaluations to support for development and management of modelling, assessment as well as decision making practices. All approaches, however, explicitly acknowledge the role of use in delivering the effects of knowledge provided by models and assessment, despite that the criteria against which effectiveness is considered may vary.
  
*similarities/differences between QA/UA
+
=== Other perspectives ===
*QA & UA combined = sufficient?
 
*complementary characteristics of other approaches?
 
  
*Two basic lines of thinking: uncertainty assessment & quality assurance
+
Although many contributions to modelling and assessment performance in relevant literature can be quite comfortably located within the four perspectives above, there are also some other aspects that deserve to be mentioned. Those to be brought up here address credibility and acceptability, information quality, and communication.
**QA: procedural guidelines (internal, public, public interactive); process &rarr; product
 
**UA: content (and wrapping) of the product and the process; product &rarr; process
 
*QA:
 
**CEPA (Forrestal et al.) 2008, principle-based approach to QA (risk assessment)
 
**Jakeman et al. 2006, 10 iterative steps in model development and evaluation (environmental modeling)
 
**Risbey et al. 1996 procedural issues to address (integrated assessment)
 
**Guimarães & Funtowicz 2005, QA by extended peer-review (groundwater resources, post-normal science)
 
**EPA 2002: guidance for QA &rarr; guidance for modeling
 
**Refsgaard et al. 2004 & 2005, modeling guidelines new QA guidelines (model based water management)
 
*UA:
 
**van Asselt & Rotmans 2002, typology of uncertainty in IAM & pluralistic uncertainty management (integrated assessment modeling)
 
**van der Sluijs et al. 2005, NUSAP (model-based environmental assessment)
 
**Walker at al. 2003, conceptual basis for uncertainty management (model-based decision support)
 
**Brown et al. 2005, data uncertainty recording (environmental data)
 
**Blind & Refsgaard 2007, uncertainty in data and models (water resources management)
 
**Kann & Weyant 2002, Decision making under uncertainty (energy/economic policy models)
 
QA/UA combined:
 
**MNP, Risbey et al. 2005, Janssen et al. 2005, van der Sluijs 2008, guidance for QA/UA & communication (environmental management)
 
**Refsgaard et al. 2006 & 2007, guidelines to dealing with uncertainty (environmental modeling)
 
*other:
 
**Morgan & Dowlatadabi 1996, 7 attributes of good climate change IA's
 
**Parker et al. 2002, Progress in IAM
 
**van der Sluijs 2007, uncertainty and precaution / UPEM (environmental management)
 
  
== Properties of good assessments ==
+
Credibility is often considered necessary for acceptance of modelling and assessment endeavours and their outputs. It can be obtained more or less formally or informally e.g. through peer review, extended peer-review (Funtowicz and Ravetz, 1990) or reputation. Credibility and acceptability are often considered as aspects of broader performance concepts.
  
We propose a new approach for considering the performance of environmental health assessments that builds on considering assessments as trialogical processes of creating collective belief systems. It describes the general properties of good assessments which can be categorized as properties related to (i) quality of information content, (ii) applicability of information and (iii) efficiency of assessment. The properties of good assessments can be used as assessment design and execution principles as well as a framework for evaluating past assessments.
+
Modelling and assessment are essentially processes of producing information. Therefore, the contributions regarding information quality also outside the domains of modelling and assessment are of relevance here. This perspective resembles the uncertainty analysis perspective as they both focus into certain properties of an information product. Similarly, the variation among contributions addressing information quality is big.
  
*multi-perspective approach to performance
+
Also communication of results, e.g. in terms of communicating uncertainties and risk information, is linked to performance of models and assessments. However, the issues of communication are not necessarily considered as integral parts of modelling and assessments endeavours. For example, risk assessment, risk management and risk communication are traditionally considered as separate, yet interrelated, entities, each having their own aims, practices, and practitioners (e.g. WHO/FAO, 2006).
**effectiveness
 
***quality of content
 
***applicability
 
**efficiency
 
  
*Evaluation process
+
{|{{prettytable}}
**a priori and/or a posteriori view
+
|+ '''Table 5. Examples of other perspectives to model and assessment performance.
**identification of purpose
+
! Characterization
**evaluation of quality of content (uncertainty + relevance)
+
! Author(s)
***in principle reality, but in practice golden standard as reference point {{disclink|What if a golden standard does not exist?}}
+
! Description
**evaluation of applicability
+
|-----
**evaluation of efficiency (effort expenditure)
+
| rowspan="2"| Acceptance and credibility
**overall performance
+
| Alexandrov et al. 2011
***potential for effectiveness/effort given purpose
+
| Obtaining model credibility through peer-reviewed publication process
****can be further evaluated retrospectively against realized effectiveness (possibly against redefined purpose)
+
|-----
 +
| Aumann 2009
 +
| Model credibility in the context of policy appraisal
 +
|-----
 +
| rowspan="4"| Information quality
 +
| Wang and Strong 1996
 +
| A conceptual framework of data quality
 +
|----
 +
| Moody and Walsh 1999
 +
| An asset valuation approach to value of information
 +
|-----
 +
| Skyrme 1994
 +
| Ten aspects that add value to information
 +
|-----
 +
| Tongchuay and Praneetpolgrang 2010
 +
| Knowledge quality in knowledge management systems
 +
|-----
 +
| rowspan="4"| Communication
 +
| Wardekker et al. 2008
 +
| Uncertainty communication in environmental assessments
 +
|-----
 +
| Janssen et al. 2005
 +
| Checklist for assessing and communicating uncertainties
 +
|-----
 +
| Covello et al. 2001
 +
| Communication challenges posed by a release of a pathogen in an urban setting
 +
|-----
 +
| Bischof and Eppler 2011
 +
| Clarity in knowledge communication
 +
|}
  
 
== Discussion ==
 
== Discussion ==
  
The proposed new approach is capable of incorporating the main aspects of both the quality assurance view and the uncertainty assessment view, but provides a more comprehensive and coherent framework that addresses the assessment process, assessment products as well as the use processes as a whole. It considers the assessment product as the mediating shared object of activity and scrutinizes it as the central object bridging the assessment process with the use process and thereby enables explication of all relevant aspects of assessment performance in a single framework.
+
By considering the reviewed contributions in light of the framework described in Figure 1, it seems that none of the perspectives nor any individual contributions alone sufficiently serve all the needs of evaluation and management of model and assessment performance. In most of the contributions, the main emphasis is on the processes and outputs of modelling and assessment while contextual aspects, outcomes, as well as use are addressed to a lesser extent, although more frequently in recent literature. The accounts of interaction of modelling and assessment with their societal context also appear to be vaguer in comparison to the commonly applied perspectives focusing on processes and outputs. No fundamental differences in perspectives to performance between domains of modelling and assessment can be seen.
 +
 
 +
Many approaches to performance seem to perceive evaluation as a separate entity that most often takes place only after the development of a model or assessment, and often is considered a responsibility of others than modellers and assessors themselves. Some major exceptions to this are the essentially proactive stepwise guidance in the quality assurance/control perspective, and the technical assessment of models perspective, in which the evaluation is often integrated in the model development. In addition, some of the effectiveness frameworks are explicitly intended as means to support design and execution, not only evaluation, of models and assessments.
 +
 
 +
The emphasis on processes and outputs in evaluation and management of model and assessment performance is in line with the fact that e.g. the issues of effectiveness and policy-relevance have become major topics also in modelling and assessment only during the last decades. As modellers, assessors, and researchers more generally, have been lacking requirements and incentives for effectiveness and policy-relevance (cf. Harris, 2002), correspondingly the practices, principles and methods of performance management and evaluation have not developed to address these issues. Instead, the impacts of modelling and assessment have mostly been considered mainly in terms of their process effects (cf. Matthews et al., 2011) within the communities of modellers and assessors, rather than outcomes in the broader societal context. Virtually all modelling and assessment endeavors in the fields of environment and health are, however, at least nominally, motivated by the aim to influence societal decision-making. The societal outcomes should thus be considered as the ultimate criterion for model and assessment performance. This has also been recognized in many new approaches to modeling, assessment as well as their evaluation (e.g. Tijhuis et al., 2012; Matthews et al., 2011). However, the complexity of addressing the outcomes remains a challenge. In the eyes of the evaluators, the relative simplicity of considering only processes, outputs or direct impacts in tightly bound settings of expert activities may appear inviting in comparison to attempting to account for complex indirect impacts within the broader social context. Unfortunately, this may not be adequate for serving the purposes of modelling, assessment and their evaluation.
 +
 
 +
It appears that a more comprehensive approach that covers all aspects of modelling and assessment in their societal context, as described in Figure 1, is needed to support model and assessment evaluation and management. In practice, this requires taking account of the making of models and assessments, their outputs, their use and other interaction with the societal context, as well as the decisions and actions influenced by the models and assessments, and the consequential outcomes. Such an approach would need to combine the essential characteristics of the different perspectives into one methodology, framework, or tool. However, a mere compilation of features taken from different perspectives would probably not be sufficient. A more thorough account of the mechanisms of collective knowledge creation and the relations between knowledge and action in a societal context is needed in order to truly bridge models and assessments with their outcomes (Pohjola et al., 2011). Unfortunately these aspects are barely even recognized in most current approaches to model and assessment performance.
 +
 
 +
Despite that the contributions within the effectiveness perspective were above characterized as often being somewhat vague, it seems likely that the effectiveness perspective provides the best basis for a comprehensive approach as the concept brings together the making, content, use as well as outcomes of models and assessments. Attempts to combine methods, frameworks and tools into a comprehensive methodology for design, making and evaluation of models, assessments as well as practices are being made, and the early results look promising. For example, some of the above mentioned effectiveness frameworks have been jointly applied for evaluating the effectiveness of two open assessments on alternative biofuel sources (Sandström et al. manuscript, http://en.opasnet.org/w/Biofuel_assessments) as well as analysing how different health, safety and environment assessment approaches serve the needs of decision making in public policy as well as manufacturing industry (Pohjola manuscript). A more noteworthy effort in this aspect is the ongoing TEKAISU-project (http://en.opasnet.org/w/Tekaisu) in which methods and tools for evaluation and management of knowledge-based city-level decision making are developed and applied. This development combines four above mentioned effectiveness frameworks by Kauppinen et al. (2006), Tuomisto and Pohjola (2007), Pohjola and Tuomisto (2011), and Pohjola et al. (2012) with a theoretical framework and practical tool for evaluating and managing practices in the social and health field by Koivisto and Pohjola (2012, https://pilotointi.innokyla.fi). All these combined frameworks and tools essentially build on collaboration and pragmatism, and thus emphasize the interrelations between knowledge and action in a social context. Consequently, the approach does not focus primarily on any separate aspect, process, output, use or outcome, of model and assessment performance, but considers and addresses the knowledge that is created, transferred and applied within the intertwined processes of modelling, assessment and decision making. This, we believe, is the key to making the whole chain from design of models and assessments to societal outcomes of decisions and actions both evaluable and manageable.
 +
 
 +
The emphasis in evaluation and management of models and assessments has been on rather scientific and technical aspects within the research domain, and perspectives addressing the impacts of models and assessments in terms of societal outcomes have emerged only quite recently and are still relatively rare. Still, it can be said that an outcome-oriented turn is taking place in evaluation and management of models and assessments, at least in development of theories and related academic discourses. However, a lot of work still remains in developing the frameworks, methods, tools and ultimately the common practices of modelling and assessment to sufficiently serve the needs of practical decision making for sustainable and healthy future. Succeeding in this will require better recognition of the mechanisms how models and assessments influence the knowledge, decisions and actions that deliver the consequences.
 +
 
 +
==Acknowledgements==
 +
 
 +
This review builds on research done in several projects receiving funding from various sources. Most importantly the authors would like to mention the EU projects INTARESE (Integrated Assessment of Health Risks of Environmental Stressors in Europe, 2005-2011, GOCE-CT-2005-018385), and BENERIS (Benefit–Risk Asessment of Food: An iterative Value-of-Information approach, 2006–2009, FOOD-CT-2006-022936), Safefoodera project BEPRARIBEAN (Best Practices for Risk-Benefit Analysis of Foods, project ID 08192), Academy of Finland (Grants 218114 and 126532), and SYTYKE doctoral programme in environmental health at the University of Eastern Finland.
 +
 
 +
== References ==
 +
 
 +
Aertsen, W., Kint, V., van Orshoven, J., Muys, B., 2011. Evaluation of modelling techniques for forest site productivity prediction in contrasting ecoregions using stochastic multicriteria acceptability analysis (SMAA). Environmental Modelling & Software 26, 929-937.
 +
 
 +
Alexandrov, G.A., Ames, D., Bellochi, G., Bruen, M., Crout, N., Erechtchoukova, M., Hildebrandt, A., Hoffman, F., Jackisch, C., Khaiter, P., Mannina, G., Matsunaga, T., Purucker, S.T., Rivington, M., Samaniego, L., 2011. Technical assessment and evaluation of environmental models and software: Letter to the editor. Environmental Modelling & Software 26, 328-336.
 +
 
 +
van Asselt, M.B.A., Rotmans, J., 2002. Uncertainty in integrated assessment modelling: From positivism to Pluralism. Climatic Change 54, 75-105.
 +
 
 +
Aumann, C.A., 2011. Constructing model credibility in the context of policy appraisal. Environmental Modelling & Software 26, 258-265.
 +
 
 +
Bai, Y., Wagener, T., Reed, P., 2009. A top-down framework for watershed model evaluation and selection under uncertainty. Environmental Modelling & Software 24, 901-916.
 +
 
 +
Baker, D.C., McLelland, J.N., 2003. Evaluating the effectiveness of British Columbia's environmental assessment process for first nations' participation in mining development. Environmental Impact Assessment Review 23, 581-603.
 +
 
 +
Beck, B., 2002. Model evaluation and performance, in: El-Shaarawi, A.H., Piegorsch, W.W. (Eds.), Encyclopedia of Environmetrics. John Wiley & Sons Ltd, Chichester, Volume 3, pp. 1275-1279.
 +
 
 +
Basson, L., Petrie, J.G., 2007. An integrated approach for the consideration of uncertainty in decision making supported by Life Cycle Assessment. Environmental Modelling & Software 22, 167-176.
 +
 
 +
Bina, O., 2008. Context and systems: Thinking more broadly about effectiveness in strategic environmental assessments in China. Environmental Management 42, 717-733.
 +
 
 +
Bischof, N., Eppler, M.J., 2011. Caring for clarity in knowledge communication. Journal of Universal Computer Science 17, 1455-1473.
 +
 
 +
Blackstock, K.L., Kelly, G.J., Horsey, B.L., 2007. Developing and applying a framework to evaluate participatory research for sustainability. Ecological Economics 60, 726-742.
 +
 
 +
Blind, M.W., Refsgaard, J.C., 2007. Operationalising uncertainty in data and models for integrated water resource management. Water, Science & Technology 56, 1-12.
 +
 
 +
Borgonovo, E., 2008. Sensitivuty analysis of model output with input constraints: A generalized rationale for local methods. Risk Analysis 28, 667-680.
 +
 
 +
Briggs, D.J., 2008. A framework for integrated environmental health impact assessment of systemic risks. Environmental Health, 7:61. doi:10.1186/1476-069X-7-61
 +
 
 +
Briggs, D.J., Sable, C.E., Lee, K., 2009. Uncertainty in epidemiology and health risk and impact assessment. Environmental Geochemistry and Health 31, 189-203.
 +
 
 +
Brouwer, R., De Blois, C., 2008. Integrated modelling of risk and uncertainty underlying the cost and effectiveness of water quality measures. Environmental Modelling & Software 23, 922-937.
 +
 
 +
Brown, J.D., Heuvelink, G.B., Refsgaard, J.C., 2005. An integrated methodology for recording uncertainties about environmental data. Water, Science & Technology 52, 153-160.
 +
 
 +
Cartwright, N., 2007. Are RCTs the gold standard?. BioSocieties 2, 11-20. doi:10.1017/S1745855207005029
 +
 
 +
Clark, W.C., Majone, G., 1985. The critical appraisal of scientific inquiries with policy implications. Science, Technology & Human Values 10, 6-19.
 +
 
 +
Colyvan, M., 2008. Is probability the only coherent approach to uncertainty?. Risk Analysis 28, 645652
 +
 
 +
Covello, V.T., Peters, R.G., Wojtecki, J.G., Hyde, R.C., 2001. Risk communication, the west nile virus epidemic, and bioterrorism: Responding to the communication challenges posed by the intentional or unintentional release of a pathogen in an urban setting. Journal of Urban Health: Bulletin of the New York Academy of Medicine 78, 382-391.
 +
 
 +
van Delden, H., Seppelt, R., White, R., Jakeman, A.J., 2011. A methodology for the design and development of integrated models for policy support. Environmental Modelling & Software 26, 266-279.
 +
 
 +
Dewar, J.A., Bankes, S.C., Hodges, J.S., Lucas, T., Saunders-Newton, D.K., Vye, P., 1996. Credible uses of the distributed interactive simulation (DIS) system. RAND, Santa Monica, CA.
 +
 
 +
Diez, E., McIntosh, B.S., 2011. Organisational drivers for, constraints on and impacts of decision and information support tool use in desertification policy and management. Environmental Modelling & Software 26, 317-327.
 +
 
 +
ECHA, 2008. Guidance on Information Requirements and Chemical Safety Assessment. Guidance for the implementation of REACH. European Chemicals Agency.
 +
 
 +
Ekboir, J., 2003. Why impact analysis should not be used for research evaluation and what the alternatives are. Agricultural Systems 78, 166-184.
 +
 
 +
Fischer, T.B., Gazzola, P., 2006. SEA effectiveness criteria - equally valid in all countries? The case of Italy. Environmental Impact Assessment Review 26, 396-409.
 +
 
 +
Forristal, P.M., Wilke, D.L., McCarthy, L.S., 2008. Improving the quality of risk assessments in Canada using a principle-based apporach. Regulatory Toxicology and Pharmacology 50, 336-344.
 +
 
 +
Funtowicz, S.O. and J.R. Ravetz 1990. Uncertainty and Quality in Science for Policy. Kluwer Academic Publishers, the Netherlands.
 +
 
 +
Granger Morgan, M., Dowlatadabi, H., 1996. Learning from integrated assessment of climate change. Climatic change 34, 337-368.
 +
 
 +
Guyatt, G.H., Sackett, D.L., Sinclair, J.C., Hayward, R., Cook, D.J., Cook, R.J., 1995. Users' guides to the medical literature. IX. A method for grading health care recommendations. JAMA 274, 1800–1804.
 +
 
 +
Harris, G., 2002. Integrated assessment and modelling: an essential way of doing science. Environmental Modelling & Software 17, 201-207.
 +
 
 +
Harteloh, P.P.M., 2002. Quality systems in health care: a sociotechnical approach. Health Policy 64, 391-398.
 +
 
 +
Hildén, M., Furman, E., Kaljonen, M., 2004. Views on planning and expectations of SEA: the case of transport planning. Environmental Impact Assessment Review 24, 519-536.
 +
 
 +
Hoekstra, J., Hart, A., Boobis, A., Claupein, E., Cockburn, A., Hunt, A., Knudsen, I., Richardson, D., Schilter, B., Schütte, K., Torgerson, P.R., Verhagen, H., Watzl, B., Chiodini, A., 2010. BRAFO tiered approach for benefit-risk assessment of foods. Food and Chemical Toxicology, In press. doi:10.1016/j.fct.2010.05.049
 +
 
 +
Hokkanen, P., Kojo, M., 2003. How environmental impact assessment influences decision-making [in Finnish]. Ympäristöministeriö, Helsinki.
 +
 
 +
Hummel, J., Huitt, W., 1994. What you measure is what you get. GaASCD Newsletter: The Reporter, 10-11. Retrieved 13.10.2011 from http://www.edpsycinteractive.org/papers/wymiwyg.html
 +
 
 +
Inman, D., Blind, M., Ribarova, I., Krause, A., Roosenschoon, O., Kassahun, A., Scholten, H., Arampatzis, G., Abrami, G., McIntosh, B., Jeffrey, P., 2011. Perceived effectiveness of environmental decision support systems in participatory planing: Evidence from small groups of end-users. Environmental Modelling & Software 26, 302-309.
 +
 
 +
Jakeman, A.J., Letcher, R.A., 2003. Integrated assessment and modelling: features, principles and examples for catchment management. Environmental Modelling & Software 18, 491-501.
 +
 
 +
Jakeman, A.J., Letcher, R.A., Norton, J.P., 2006. Ten iterative steps in development and evaluation of environmental models. Environmental Modelling & Software 21, 602-614.
 +
 
 +
Janssen, P.H.M., Petersen, A.C., van der Sluijs, J.P., Risbey, J.S., Ravetz, J.R., 2005. A guidance for assessing and communicating uncertainties. Water Science & Technology 52, 125-131.
 +
 
 +
Kann, A., Weyant, J.P., 2000. Approaches for performing uncertainty analysis in large-scale energy/economic policy models. Environmental Modeling and Assessment 5, 29-46.
 +
 
 +
Kauppinen, T., Nelimarkka, K., Perttilä, K., 2006. The effectiveness of human impact assessment in the Finnish Health Cities Network. Public Health 120, 1033-1041.
 +
 
 +
Kloprogge, P., van der Sluijs, J.P., Petersen, A.C., 2011. A method for the analysis of assumptions in model-based environmental assessments. Environmental Modelling & Software 26, 289-301.
 +
 
 +
Koivisto, J., Pohjola, P., 2012. Practices, modifications and generativity - REA: a practical tool for managing the innovation processes of practices. Systems, Signs & Actions 5(1), 100-116. http://www.sysiac.org/uploads/SySiAc2011-Koivisto-Pohjola.pdf
 +
 
 +
Larocque, G.R., Bhatti, J.S., Ascough II, J.C., Liu, J., Luckai, N., Mailly, D., Archambault, L., Gordon, A.M., 2011. An analytical framework to assist decision makers in the use of forest ecosystem model predictions. Environmental Modelling & Software 26, 280-288.
 +
 
 +
Leu, W., Williams, W.P., Bark, A.W., 1996. Development of an environmental impact assessment evaluation model and its application: Taiwan case study. Environmental Impact Assessment Review 16, 115-133.
 +
 
 +
Leviton, L.C., 2003. Evaluation use: Advances, challenges and applications. American Journal of Evaluation 24, 525-535.
 +
 
 +
Liu, Y., Gupta, H., Springer, E., Wagener, T., 2008. Linking science with environmental decision making: Experiences from an integrated modelling approach to supporting sustainable water resources management. Environmental Modeling & Software 23, 846-858.
 +
 
 +
Mahmoud, M., Liu, Y., Hartmann, H., Stewart, S., Wagener, T., Semmens, D., Stewart, R., Gupta, H., Dominguez, D., Dominguez, F., Hulse, D., Letcher, R., Rashleigh, B., Smith, C., Street, R., Ticehurst, J., Twery, M., van Delden, H., Waldick, R., White, D., Winter, L., 2009. A formal framework for scenario development in support of environmental decision-making. Environmental Modelling & Software 24, 798-808.
 +
 
 +
Matthews, K.B., Rivington, M., Blackstock, K.L., McCrum, G., Buchan, K., Miller, D.G., 2011. Raising the bar? - The challenges of evaluating the outcomes of environmental modelling and software. Environmental Modelling & Software 26 (3), 247-257.
 +
 
 +
McIntosh, B.S., Giupponi, C., Voinov, A.A., Smith, C., Matthews, K.B., Monticino, M., Kolkman, M.J., Crossman, N., van Ittersum, M., Haase, D., Haase, A., Mysiak, J., Groot, J.C.J., Sieber, S., Verweij, P., Quinn, N., Waeger, P., Gaber, N., Hepting, D., Scholten, H., Sulis, A., van Delden, H., Gaddis, E., Assaf, H., 2008. Bridging the gaps between design and use: Developing tools to support environmental management and policy, in: Jakeman, A.J., Voinov, A.A., Rizzoli, A.E., Chen, S.H. (Eds.): Environmental Modelling, Software and Decision Support. Elsevier, Amsterdam, pp. 33-48.
 +
 
 +
McIntosh, B.S., Alexandrov, G., Matthews, K., Mysiak, J., van Ittersum, M. (Eds.), 2011a. Thematic issue on the assessment and evaluation of environmental models and software. Environmental Modelling & Software 26, 245-336.
 +
 
 +
McIntosh, B.S., Alexandrov, G., Matthews, K., Mysiak, J., van Ittersum, M., 2011b. Preface: Thematic issue on the assessment and evaluation of environmental models and software. Environmental Modelling & Software 26, 245-246.
 +
 
 +
Mo, X., Liu, S., Lin, Z., 2011. Evaluation of an ecosystem model for a wheat-maize double cropping system over the North China Plain. Environmental Modelling & Software, In press. doi:10.1016/j.envsoft.2011.07.002
 +
 
 +
Moody, D., Walsh, P., 1999. Measuring the value of information: An asset valuation approach. Seventh European Conference on Information System (ECIS’99), Copenhagen Business School, Frederiksberg, Denmark, 23-25 June, 1999. Available: http://wwwinfo.deis.unical.it/zumpano/2004-2005/PSI/lezione2/ValueOfInformation.pdf Accessed 13.10.2011.
 +
 
 +
Moschandreas, D.J., Karuchit, S., 2002. Scenario-model-parameter: a new method of cumulative risk uncertainty analysis. Environment International 28, 247-261.
 +
 
 +
Nilsen, T., Aven, T., 2003. Models and model uncertainty in the context of risk analysis. Reliability Engineering & System Safety 79, 309-317.
 +
 
 +
NRC, 1983. Risk Assessment in the Federal Government: Managing the Progress. The National Research Council. National Academy Press, Washington D.C.
 +
 
 +
NRC, 1996. Understanding Risk: Informing Decisions in a Democratic Society. The National Research Council. National Academy Press, Washington D.C.
 +
 
 +
NRC, 2009. Science and Decisions: Advancing Risk Assessment. National Research Council. National Academy Press, Washington D.C.
 +
 
 +
O'Hagan, A., 2011. Probabilistic uncertainty specification: Overview, elaboration techniques and their application to a mechanistic model of carbon flux. Environmental Modelling & Software, In press. doi:10.1016/j.envsoft.2011.03.003
 +
 
 +
Patton, M.Q., 2008: Utilization-focused evaluation, fourth edition. SAGE Publications, Inc. MN.
 +
 
 +
Pohjola, M.V., Leino, O., Kollanus, V., Tuomisto, J.T., Gunnlaugsdόttir, H., Holm, F., Kalogeras, N., Luteijn, J.M., Magnusson, S.H., Odekerken, G., Tijhuis, M.J., Ueland, Ø., White, B.C., Verhagen, H., 2012. State of the art in benefit–risk analysis: environmental health. Food Chem. Toxicol. 2012, 50:40-55. http://dx.doi.org/10.1016/j.fct.2011.06.004
 +
 
 +
Pohjola, M.V., Pohjola, P., Paavola, S., Bauters, M., Tuomisto, J.T., 2011. Pragmatic Knowledge Services. Journal of Universal Computer Science 17:472-497. http://dx.doi.org/10.3217/jucs-017-03-0472
 +
 
 +
Pohjola, M.V., Tuomisto, J.T., 2011. Openness in participation, assessment, and policy making upon issues of environment and environmental health: a review of literature and recent project results. Environmental Health 10, 58. doi:10.1186/1476-069X-10-58
 +
 
 +
Pohjola, M.V. Assessment of impacts to health, safety, and environment in the context of materials processing and related public policy. Manuscript.
 +
 
 +
Pollino, C.A., Woodberry, O., Nicholson, A., Korb, K., Hart, B.T., 2007. Parameterisation and evaluation of a Bayesian network for use in an ecological risk assessment.
 +
 
 +
Pope, J., Annandale, D., Morrison-Saunders, A., 2004. Conceptualising sustainability assessment. Environmental Impact Assessment Review 24, 595-616.
 +
 
 +
Pölönen, I., Hokkanen, P., Jalava, K., 2011. The effectiveness of the Finnish EIA system - What works, what doesn't, and what could be improved? Environmental Impact Assessment Review 31, 120-128.
 +
 
 +
Quigley, R.J., Taylor, L.C., 2004. Evaluating health impact assessment. Public Health 118, 544-552.
  
*general shortcomings of QA & UA:
+
Reeves, C.A., Bednar, D.A., 1994. Defining quality: Alternatives and implications. Academy of Management Review 19, 419-445.
**use process often not explicitly included and use purpose not considered
 
**societal aspects often neglected
 
**all aspects of performance not covered in any single approach
 
**evaluation as a separate process, often only after assessment
 
  
*properties of good assessments:
+
Refsgaard, J.C., Henriksen, H.J., 2004. Modelling guidelines - terminology and guiding principles. Advances in Water Resources 27, 71-82.
**focus on the mediator of the overall intellectual process - the shared object of activity
 
***product|use purpose, assessment process|product
 
***trialogical process, knowledge creation metaphor
 
****not only collection and use of existing information or learning to deal with new situation, but also creation of new knowledge
 
**capable of incorporating the goods from other perspectives into a more comprehensive and coherent approach
 
**evaluation of performance can only be done meaningfully against purpose
 
***various intentions need to be identified, explicated and prioritized in all assessments
 
  
*Assessment products as belief systems (Hilpinen)
+
Refsgaard, J.C., Henriksen, H.J., Harrar, W.G., Scholten, H., Kassahun, A., 2005. Quality assurance in model-based water management - review of existing practice and outline of new approaches. Environmental Modelling & Software 20, 1201-1215.
*Trialogue: the knowledge creation metaphor, belief systems as mediators
 
*Assessment participation as innovatiove knowledge communities (Paavola, Hakkarainen)
 
  
*an integral part of assessment process &rarr; applicable also in design and execution (a priori), not only evaluation (a posteriori)
+
Refsgaard, J.C., van der Sluijs, J.P., Brown, J., van der Keur, P., 2006. A framework for dealing with uncertainty due to model structure error. Advances in Water Resources 29, 1586-1597.
  
*data about hypothesis
+
Refsgaard, J.C., van der Sluijs, J.P., Lajer Højberg, A., Vanrolleghem, P.A., 2007. Uncertainty in the environmental modelling process - A framework and guidance. Environmental Modelling & Software 22, 1543-1556.
  
*an assessment is always case- and context-specific, i.e. evaluable against a particular purpose (or purposes) only, but the components of an assessment may (should!) be more generally usable (unlike Beck 2002 says)
+
Risbey, J., Kandlikar, M., Patwardhan, A., 1996. Assessing integrated assessment. Climatic Change 34, 369-395.
  
== Conclusions ==
+
Risbey, J. van der Sluijs, J.P., Kloprogge, P., Ravetz, J., Funtowicz, S., Corral Quintana, S., 2005. Application of a checklist for quality assistance in environmental modelling to an energy model. Environmental Modeling and assessment 10, 63-79.
  
*There is more to assessment performance than just statistical uncertainty and data source reliability
+
Rizzoli, A.E., Leavesley, G., Ascough II, J.C., Argent, R.M., Athanasiadis, I.N., Brilhante, V., Claeys, F.H.A., David, O., Donatelli, M., Gijsbergs, P., Havlik, D., Kassahun, A., Krause, P., Quinn, N.W.T., Scholten, H., Sojda, R.S., Villa, F., 2008. Integrated modelling frameworks for environmental assessment decision support, in: Jakeman, A.J., Voinov, A.A., Rizzoli, A.E., Chen, S.H. (Eds.): Environmental Modelling, Software and Decision Support. Elsevier, Amsterdam, pp. 101-118.
*Overall performance of assessment can be evaluated systematically and explicitly
 
**requires consistent information structure
 
**a priori evaluation should be made an inherent part of assessment process
 
  
== Competing interests ==
+
Sandström, V., Tuomisto, J.T., Majaniemi, S., Rintala, T., Pohjola, M.V. Evaluating effectiveness of open assessments on alternative biofuel sources. Manuscript.
  
== Authors' contributions ==
+
Siebenhüner, B., Barth, V., 2005. The role of computer modelling in participatory integrated assessments. Environmental Impact Assessment Review 25, 367-389.
  
== Acknowledgements ==
+
Skyrme, D.J., 1994. Ten ways to add value to your business. Managing Information 1, 20-25. Available: http://www.skyrme.com/pubs/tenways.htm (accessed 13.10.2011)
  
== References ==
+
Sterk, B., van Ittersum, M.K., Leeuwis, C., 2011. How, when, and for what reasons does land use modelling contribute to societal problem solving?. Environmental Modelling & Software 26, 310-316.
 +
 
 +
van der Sluijs, J.P., 2002. Integrated Assessment, in: Tolba, M.K. (Ed.), Volume 4, Responding to Global Environmental Change, pp. 250-253, in: Munn, T. (Ed.), Encyclopedia of Global Environmental Change. John Wiley & Sons, Ltd, Chichester.
 +
 
 +
van der Sluijs, J.P., Craye, M., Funtowicz, S., Kloprogge, P., Ravetz, J., Risbey, J., 2005. Combining quantitative and qualitative measures of uncertainty in model-based environmental assessment: The NUSAP system. Risk Analysis 25, 481-492.
 +
 
 +
van der Sluijs, J.P., Petersen, A.C., Janssen, P.H.M., Risbey, J.S., Ravetz, J.R., 2008: Exploring the quality of evidence for complex and contested policy decisions. Environmental Research Letters 3, 024008. doi:10.1088/1748-9326/3/2/024008
 +
 
 +
Sojda, R.S., 2007. Empirical evaluation of decision support systems: Needs, definitions, potential methods, and an example pertaining to waterfowl management. Environmental Modelling & Software 22, 269-277.
 +
 
 +
Sonneveld, B.G.J.S., Keyzer, M.A., Stroosnijder, L., 2011. Evaluating quantitative and qualitative models: An application for nationwide water erosion assessment in Ethiopia. Environmental Modelling & Software 26, 1161-1170.
 +
 
 +
Tijhuis, M.J., de Jong, N., Pohjola, M.V., Gunnlaugsdόttir, H., Hendriksen, M., Hoekstra, J., Holm, F., Kalogeras, N. Leino, O., van Leeuwen, F.X., Luteijn, J.M., Magnússon, S.H., Odekerken, G., Rompelberg, C., Tuomisto, J.T., Ueland, Ø., White, B.C., Verhagen, H., 2012. State of the art in benefit–risk analysis: food and nutrition. Food Chem. Toxicol. 2012, 50:5-25. http://dx.doi.org/10.1016/j.fct.2011.06.010
 +
 
 +
Tongchuay, C., Praneetpolgrang, P., 2010. Knowledge quality and quality metrics in knowledge management systems. Proceedings of the Fifth International Conference on eLearning for Knowledge-Based Society. December 11-12, 2008. Bangkok Metro, Thailand.
 +
 
 +
Tuomisto, J.T., Pohjola, M.V., 2007. Open Risk Assessment - A new way of providing information for decision-making. Publications of the National Public Health Institute B18/2007. KTL - National Public Health Institute, Kuopio.
 +
 
 +
Wagener, T., Kollat, J., 2007. Numerical and visual evaluation of hydrological and environmental models using the Monte Carlo analysis toolbox. Environmental Modelling & Software 22, 1021-1033.
 +
 
 +
Walker, W.E., Harremoës, P., Rotmans, J., van der Sluijs, J.P., van Asselt, M.B.A., Janssen, P., Krayer von Krauss, M.P., 2003. Defining uncertainty: A conceptual basis for uncertainty management in model-based decision support. Integrated Assessment 4, 5-17.
 +
 
 +
Wang, R.Y., Strong, D.M., 1996. Beyond accuracy: What data quality means to data consumers. Journal of Management Information Systems 12, 5-34.
 +
 
 +
Wardekker, J.A., van der Sluijs, J.P., Janssen, P.H.M., Kloprogge, P., Petersen, A.C., 2008. Uncertainty communication in environmental assessments: views from the Dutch science-policy interface. Environmental Science & Policy 11, 627-641.
 +
 
 +
Wismar, M., Blau, J., Ernst, K., Figueras, J. (Eds.), 2007. The Effectiveness of Health Impact Assessment.: Scope and limitations of supporting decision-making in Europe. WHO, Copenhagen.
 +
 
 +
WHO, 1999. Health Impact Assessment (HIA), main concepts and suggested approach. Gothenburg consensus paper. World Health Organization, Brussels.
 +
 
 +
WHO/FAO, 2006. Food safety risk analysis: A guide for national food safety authorities. World Health Organization, Food and Agriculture Organization of the United Nations, Rome.
  
== Figures ==
+
Wood, C., 1995. Environmental impact assessment: a comparative review. Longman Scientific & Technical, New York.
  
== Tables ==
+
Wyat Appel, K., Gilliam, R.C., Davis, N., Howard, S.C., 2011. Overview of the atmospheric model evaluation tool (AMET) v1.1 for evaluating meteorological and air quality models. Environmental Modelling & Software 26, 434-443.
  
== Additional files ==
+
Xu, Y., Booij, M.J., Mynett, A.E., 2007. An appropriateness framework for the Dutch Meuse decision support system. Environmental Modelling & Software 22, 1667-1678.

Latest revision as of 16:20, 4 March 2015



This page is the final manuscript of Pohjola, M.V.; Pohjola, P.; Tainio, M.; Tuomisto, J.T. Perspectives to Performance of Environment and Health Assessments and Models—From Outputs to Outcomes? Int. J. Environ. Res. Public Health 2013, 10:2621-2642. http://dx.doi.org/10.3390/ijerph10072621. Originally, the text was written on page heande:Evaluating performance of environmental health assessments.

Keywords

Evaluation, EMS, Effectiveness, Open Assessment, REA, Outcome

Highlights

  • Conventional evaluation of models and assessments focuses on processes and outputs
  • Recent evaluation approaches also emphasize societal model/assessment outcomes
  • Model/assessment effectiveness is the likelihood of delivering intended outcomes
  • An outcome-oriented turn is taking place in modelling, assessment and evaluation
  • New views combine design, making and evaluation of models, assessments and practices

Abstract

This paper reviews different perspectives to evaluation and management of model and assessment performance. While there is an increasing need to evaluate the success of environment and health relevant models and assessments according to their societal outcomes, most perspectives are primarily focused on the outputs of modelling and assessment as well as the procedures of their making. The practical application of outputs and the consequential societal outcomes, are most often left without explicit consideration. Only quite recently, the approaches to evaluation and management of model and assessment performance have recognized the outcomes as an essential performance criterion.

The perspectives to model and assessment performance can be categorized according to their foci as i) quality assurance/control, ii) uncertainty analysis, iii) technical assessment of models, iv) effectiveness and v) other perspectives. In practice many approaches found in the literature combine characteristics of different perspectives, but the categorization, however, illustrates which aspects are primarily emphasized.

It seems that the existing approaches and perspectives do not sufficiently serve all needs of evaluation and management of model and assessment performance. We need a more comprehensive approach that covers a) the making of models and assessments, b) their outputs, c) their use and other interaction with the societal context, as well as d) the decisions and actions influenced by the models and assessments, and e) the consequential outcomes. Achieving this necessitates a thorough account of the mechanisms of collective knowledge creation and the relations between knowledge and action - aspects that are often not recognized as relevant to evaluation and management of model and assessment performance. Some attempts to combine certain collaborative and pragmatic methods, frameworks and tools into a comprehensive methodology for design, making and evaluation of models, assessments as well as practices are being made, and the early results look promising.

Introduction

In this paper we review different perspectives to evaluation and management of model and assessment performance. While there is an increasing need to evaluate the success of models, modelling, and assessment upon issues of environment and health according to their societal outcomes, most evaluation approaches do not seem to support this need. In a recent thematic issue on the assessment and evaluation of environmental models and software (McIntosh et al., 2011a), Matthews et al. (2011) opened up a debate on how to best respond to the increasing desire to evaluate the success of environmental modelling and software projects in terms of their outcomes, i.e changes to values, attitudes, and behaviour outside the walls of the research organization, rather than mere outputs. Until now, there has been limited appreciation within the environmental modelling and software community regarding the challenges of shifting the focus of evaluation from outputs to outcomes (Matthews et al., 2011).

The current situation in the different overlapping communities of environment and health related assessment, such as integrated assessment (e.g. van der Sluijs, 2002), health impact assessment (e.g. WHO, 1999), risk assessment (e.g. NRC, 1983, 1996, 2009), chemical safety assessment (ECHA, 2008), environmental impact assessment (e.g. Wood, 1995), and integrated environmental health impact assessment (Briggs, 2008), appears to be somewhat similar. For example, a recent study on the state of the art in environmental health assessment revealed that despite the fact that most assessment approaches explicitly state influencing of society as their aim, this is rarely manifested in the principles and practices of evaluating assessment performance (Pohjola et al., 2012).

The emphasis in the scientific discourses on evaluating models and assessments has been on rather scientific and technical aspects of evaluation within the research domain, and perspectives that also duly address the impacts of modelling and assessment in broader societal contexts have emerged only quite recently and are still relatively rare (cf. McIntosh et al., 2011b). Such evaluations are qualitatively different (Matthews et al., 2011), which invokes a need for reconsidering the criteria for evaluating modelling and assessment endeavours, as well as the frameworks within which they are applied.

Evaluation of models and assessments is not only a matter of judging the goodness of a model or assessment, but has a more profound influence on their use as well as making. On the one hand, evaluation informs users e.g. regarding the possible and appropriate uses of models and assessment or model outputs. On the other hand, the criteria for evaluations also tend to direct the design and execution of modelling and assessment endeavours to meet those criteria, particularly if they are explicitly known in advance (cf. What you measure is what you get (WYMIWYG) by Hummel and Huitt, 1994).

This paper contributes to the debate by first describing a framework within which model and assessment performance can be considered, then reviewing different perspectives to model and assessment performance, and finally discussing the capability of existing approaches to address the challenges of what could be called the "outcome-oriented turn in evaluation and management of models and assessments". Some current attempts to develop a comprehensive approach to design, making and evaluation of models, assessments as well as practices are also briefly presented.

Models, assessments and performance

For several years, the authors have been engaged with developing methods and tools for environmental health modelling and assessment. Most notably this work has resulted in the open assessment method and Opasnet web-workspace for science-based collaborative knowledge creation and policy support upon virtually any topic (Pohjola et al. 2011, 2012, Tuomisto and Pohjola 2007, http://en.opasnet.org). These developments have required thorough familiarization of and several inquiries into the kinds of existing modelling and assessment means and practices, and particularly the factors that determine how well they perform. This review is based on the extensive information and knowledge base on model and assessment performance that has been obtained by these inquiries.

Here modelling and assessment are considered as instances of fundamentally the same issue of science-based support to decision making regarding environment and health. In fact, assessments virtually always involve modelling of some kind, at least implicit conceptual models. Conversely, modelling is also often identified with assessment (Jakeman and Letcher, 2003). In addition, decision support systems, information support tools, integrated modelling frameworks and other software tools and information systems to assist in developing, running, and analyzing models and assessments are perceived as integral parts of modelling and assessment (e.g. Rizzoli et al., 2008).

Models and assessments can be considered e.g. as diagnostic, prognostic, or summative according to the kinds of questions they address (Briggs, 2008), ex-ante or ex-post according to their timing in relation to the activities being assessed (Pope et al., 2004), and regulatory or academic according to the contexts of their development and application (Pohjola et al., 2012). They can also be developed, executed, and applied by many kinds of actors, e.g. consultants, federal agencies or academic researchers. Modelling and assessment should be, however, clearly distinguished from purely curiosity-driven research, ad hoc assessments, and models or assessments made only to justify predetermined decisions. In addition, it should be noted that "assessment of models", often interchangeable with "evaluation of models", is also excluded from the concept of assessment adopted here. Instead, technical assessment of models is considered in the review below as one of the common perspectives to evaluating model and assessment performance. Altogether, modelling and assessment can be considered as fundamentally having two purposes: i) describing reality, and ii) serving the needs of practical decision-making.

Inspired by the conceptual framework presented by Matthews et al. (2011), and influenced by the works of e.g. Blackstock et al. (2007), Bina (2008), Leviton (2003), and Patton (2002), Figure 1 describes the essential aspects in the interaction of modelling and assessment with their broader societal context.

Error creating thumbnail: Unable to save thumbnail to destination
Figure 1. The essential aspects in the interaction of modelling and assessment with their societal context.

The endeavours of modelling and assessment are here broken down into 1) process, the procedures and practices of developing and executing models and assessments, 2) output, the products or results that come out of these modelling and assessment processes, and 3) use, the practical application of the outputs. The surrounding context provides enablers for modelling and assessment endeavours to take place e.g. in the form of funding, facilities and education, but also as acceptance or demand for certain kinds of modelling and assessment endeavours. On the other hand, the context also poses constraints e.g. in the form of prioritization of scarce resources, non-acceptance of some modelling and assessment endeavours, or lacking of capability for applying their outputs. The societal context is also the medium where the outcomes of modelling and assessment, i.e. the changes in values, attitudes and behaviour among others than assessors and modellers themselves are (or are not) realized. It is important to distinguish the outcomes from what Matthews et al. (2011) termed process effects, i.e. the changes in the capacity of those engaged in the modelling and assessment endeavours, which does not directly result in impacts in the broader societal context.

In Figure 1, use is intentionally located on the boundary between the modelling and assessment domain and the context in order to emphasize that the practical application of outputs should be considered as a shared responsibility between the modellers or assessors and the intended users outside the modelling and assessment domain. This implies that the responsibility of modellers and assessors should not end in completion of the output. On the other hand, the users of the outputs should not expect realization of the outcomes without having to invest any effort themselves in their delivery. The boundary between the modelling and assessment domain and its context is marked with a dashed line, indicating that the domain should not be perceived as a strictly separate entity, but in continuous interaction with its context.

The case-specific uses of modelling and assessment outputs in varying contexts are critical for achieving the intended outcomes of models and assessments. Therefore, in order for models or assessments to perform well, it is not sufficient to only provide good descriptions of reality. They must also fulfil their instrumental purposes of serving the needs of their intended uses, and thereby resulting in the desired outcomes. However, this aspect of model and assessment performance is often given little attention, as will turn out when the existing perspectives to model and assessment performance are reviewed and discussed in light of the framework described in Figure 1.

Perspectives to model and assessment performance

Within the abundance of contributions in the scientific literature addressing model and assessment performance, certain commonalities between contributions can be identified. These contributions can be categorized as representing different perspectives according to their primary focus: i) quality assurance/control, ii) uncertainty analysis, iii) technical assessment of models, iv) effectiveness, or v) other perspectives. These perspectives provide the structure for the review. As the number of different relevant contributions is so large that it is not be possible to include them all in the review, we have instead attempted to compile representative and comprehensive collections of examples, grouped according to certain subcategories, for each perspective. Recent contributions are emphasized, but some important or illustrative examples that were published before 2000 have been included as well. Also the relevant contributions stemming from the method and tool development by the authors are also included in the review.

Quality assurance/control

One of the major themes in modelling and assessment performance related literature is what can be referred to as quality assurance/control (QA/QC) perspective. The focus in this perspective is primarily on determining how the processes of modelling or assessment are to be conducted in order to assure the quality of the output. The QA/QC perspective appears commonly in both modelling and assessment literature, sometimes extending to also consider the processes of decision-making.

There are multiple alternative definitions for quality (see e.g. Reeves and Bednar, 1994). However, as regards modelling and assessment, the interpretation is mostly analogous with the perception in the ISO-9000 framework, i.e. as the organisational structures, responsibilities, procedures, processes, and resources to assure and improve quality (Harteloh, 2003). Also the hierarchy of evidence, commonly applied in the field of medicine, is a quality assurance/control perspective, which ranks types of evidence strictly according to the procedure by which they were obtained (Guyatt et al., 1995). However, as pointed out by Cartwright (2007) with regard to randomized controlled trials, the procedure alone cannot guarantee delivery of useful information in practical contexts.

One common variation of this perspective is stepwise procedural guidance. Such guidance provides often relatively strict and detailed descriptions of the steps or phases of a modelling or assessment process that are to be executed in a more or less defined order. Faithful execution of the procedure is assumed to lead to good outputs. A somewhat similar, but perhaps less rigorous, variation of the QA/QC perspective is checklist-type guidance emphasizing issues that need to be taken account of in the modelling or assessment process or their evaluation. The checklists can be more or less detailed and they usually do not strictly define the order or sequence of execution.

Also the accounts that address evaluation of input quality can be considered as manifestations of the QA/QC perspective. However, the primary focus in QA/QC is often on the outputs, and the input quality evaluations typically complement uncertainty analyses or technical assessments of models. For example, model parameter uncertainty analysis can be considered as an example of evaluation of input quality, but in practice it is most often considered as an aspect of either uncertainty analysis or technical assessment of models.

Table 1. Examples of quality assurance/control perspective to modelling and assessment performance.
Type Author(s) Description
Stepwise procedural guidance Jakeman et al. 2006 Ten iterative steps in development and evaluation of environmental models
Refsgaard et al. 2005 HarmoniQuA guidance for quality assurance in multidisciplinary model-based water management
van Delden et al. 2011 Methodology for the design and development of integrated models for policy support
Briggs 2008 Framework for integrated environmental health impact assessment
Hoekstra et al. 2010 BRAFO tiered approach for benefit-risk assessment of foods
Liu et al. 2008 Generic framework for effective decision support through integrated modelling and scenario analysis
Mahmoud et al. 2009 Formal framework for scenario development in support of environmental decision making
Checklist guidance Granger Morgan and Dowlatadabi 1996 Seven attributes of good integrated assessment of climate change
Risbey et al. 1996 Listing of end use independent process based considerations for integrated assessment
Forristal et al. 2008 QA/QC performance measurement scheme for risk assessment in Canada
Risbey et al. 2005 Checklist for quality assistance in environmental modelling
Evaluation of input quality van der Sluijs et al. 2005 Pedigree analysis in model-based environmental assessment
Brown et al. 2005 Methodology for recording uncertainties about environmental data
Kloprogge et al. 2011 Method for analyzing assumptions in model-based environmental assessments

Characteristic for stepwise guidance is that it attempts to predetermine a procedure in order to guarantee good quality of outputs. As such, it takes a proactive approach to managing performance. Checklist-type guidance and evaluation of input quality can also be applied proactively, but the examples found in literature mostly represent a reactive approach of evaluating already executed modelling and assessment processes.

Uncertainty analysis

Another major theme in the modelling and assessment performance literature can be referred to as the uncertainty analysis perspective. The contributions within this theme vary significantly, ranging from descriptions of single methods to overarching frameworks, but the common main idea is characterization of certain properties of the outputs of models and assessments. Fundamentally the perspective builds on quantitative statistical methods based on probability calculus (O'Hagan, 2011), but also other than probability-based approaches to uncertainty have been presented (e.g. Colyvan, 2008). Many manifestations of this perspective in the context of environment and health modelling and assessment also extend to consider qualitative properties of the outputs. The uncertainty analysis perspective appears to be more commonly represented in the domain of assessment than of modelling. In addition to model and assessment outputs, models themselves could be considered as outputs whose properties are considered and evaluated in a similar fashion. However, this issue is addressed as a separate perspective in the next section, technical assessment of models.

One variation of the uncertainty analysis perspective is identification and typifying of the kinds and sources of uncertainty that characterize the modelling or assessment outputs and their quality. Some uncertainties are often considered as being primarily expressible in quantitative, while others in qualitative terms. The sources of uncertainty can also extend to include aspects of the modelling and assessment processes. In some cases also intended or possible uses and use contexts of the outputs are acknowledged.

Also different types of guidance on how to assess or deal with different kinds of uncertainties exist. Such frameworks usually combine both qualitative and quantitative aspects of uncertainty deriving from various sources. Consequently, aspects of the processes of producing the outputs, e.g. input quality, and acknowledgment of the intended or possible uses and use contexts of the outputs, e.g. in terms of acceptance, are often also included in the frameworks. The primary focus is, however, on the characteristics of the model and assessment outputs.

In addition to the broad characterizations and frameworks, also numerous more or less explicit methods, means and practices to analyze uncertainties of model and assessment outputs exist. For example, sensitivity, importance, and value of information analysis and Bayesian modelling, in addition to the standard statistical characterization, are essential in the context of environment and health modelling and assessment. Such methods are dominantly quantitative. The standard means for statistical analysis are considered common knowledge and not discussed here in any more detail.

Table 2. Examples of uncertainty analysis perspective to model and assessment performance.
Type Author(s) Description
Identification of kinds of uncertainty Walker et al. 2003 Conceptual basis for uncertainty management in model-based decision support
Briggs et al. 2008 Uncertainty in epidemiology and health risk and impact assessment
van Asselt and Rotmans 2002 Uncertainty in integrated assessment modelling
Guidance on dealing with uncertainties van der Sluijs et al. 2008 Knowledge quality assessment for complex policy decisions
Blind and Refsgaard 2007 Operationalising uncertainty in integrated water resource management
Refsgaard et al. 2007 Framework and guidance for dealing with uncertainty in environmental modelling
Methods for uncertainty analysis Kann and Weyant 2000 Approaches for performing uncertainty analysis in large-scale energy/economic policy models
Brouwer and De Blois 2008 Modelling of risk and uncertainty underlying the cost and effectiveness of water quality measures
Basson and Petrie 2007 Consideration of uncertainty in decision making supported by Life Cycle Assessment
Borgonovo 2008 Sensitivity analysis of model outputs with input constraints

For the examples of uncertainty analysis perspective, it appears characteristic that the issue of uncertainty is typically approached from an external observer’s point of view. The evaluation of performance is thus mainly considered as a separate, typically ex-post, activity taking place in addition to the actual modelling or assessment process, not as its integral proactive part.

Technical assessment of models

If the uncertainty analysis perspective is more commonly adopted in assessment literature, the technical assessment of models is a grand theme particularly in the modelling literature. In contrast to the focus on model or assessment outputs in the uncertainty analysis approach, the focus in this perspective is on the models and their characteristics. In addition to models, the perspective also considers different kinds of software tools that are applied in developing, running, and analyzing the models.

Particularly the object of interest in technical assessment of models is development and application of formal methods for testing and evaluating models within defined domains of application. Generally, model evaluation and performance is considered to cover structural features of models, representativeness of model results in relation to a certain part of reality, as well as usefulness with regard to a designated task (cf. Beck, 2002). However, if usefulness is considered at all, it mainly addresses expert use of models, which corresponds mostly to the so-called process effects rather than outcomes. Most commonly technical assessment of models takes place in terms of validation and verification by comparing models and their results against each other or measurement data.

A variation of this perspective, more common for the discourses in assessment literature, is analysis of model uncertainty. Here the aim typically is to characterize the properties of a model in order to be able to correctly interpret or evaluate its outputs. Model uncertainty is often considered as one aspect of a broader uncertainty concept.

Table 3. Examples of technical assessment of models perspective to model and assessment performance.
Type Author(s) Description
Means for model and software evaluation Alexandrov et al. 2011 Realistic criteria for environmental model and software evaluation
Refsgaard et al. 2004 Terminology and methodological framework for modelling and model evaluation
Matthews et al. 2011 Evaluation methods of environmental modelling and software in a comprehensive conceptual framework
Bai et al. 2009 Top-down framework for watershed model evaluation and selection
Wyat Appel et al. 2011 Overview of atmospheric model evaluation tool (AMET)
Xu et al. 2007 Appropriateness framework for the Dutch Meuse decision support system
Sojda 2007 Empirical evaluation of decision support systems
Wagener and Kollat 2007 Numerical and visual evaluation of hydrological and environmental models
Evaluation of models Mo et al. 2011 Evaluation of an ecosystem model for wheat-maize cropping system in North China
Pollino et al. 2007 Parameterisation and evaluation of a Bayesian network for use in an ecological risk assessment
Sonneveld et al. 2011 Evaluating quantitative and qualitative models for water erosion assessment in Ethiopia
Aertsen et al. 2011 Evaluation of modelling techniques for forest site productivity prediction using SMAA
Analysis of model uncertainty Nilsen and Aven 2003 Model uncertainty in the context of risk analysis
Moschandreas 2002 Scenario, model and parameter uncertainty in risk assessment
Refsgaard et al. 2006 Framework for dealing with uncertainty due to model structure error

The technical assessment of models is predominantly a reactive perspective to evaluate models, as it requires an existing model or software system that can be tested and analyzed. The evaluation, however, is usually considered as being an integral part of the model development, not a separate entity, which also enables application of technical assessment of models in different developmental stages within the modelling or assessment process. On the other hand, the common practice of self evaluation of models may also lead e.g. to limited usability, credibility and acceptability unless interaction with the context is otherwise realized.

Effectiveness

Whereas the three former perspectives have been subjects of discussion regarding model and assessment performance for longer times, emphasizing of model and assessment effectiveness can be considered to have become a major topic only quite recently. Contributions addressing effectiveness appear more common in the assessment literature, but the topic has recently been addressed also in the context of modelling and software.

In the effectiveness perspective, the aim of modelling and assessment is generally seen as to promote changes in values, attitudes, and behaviour outside the walls of the research community (Matthews et al., 2011) by maximizing the likelihood of an assessment process to achieve the desired results and the goals set for it (Hokkanen and Kojo, 2003). In principle, performance of models and assessments is thus fundamentally characterized in terms of the impacts delivered into the broader societal context. However, due to the complexity of reality, evaluation of outcomes is often perceived as very difficult, if not impossible (Kauppinen et al., 2006), and possibly even leading to incorrect conclusions regarding effectiveness (cf. Ekboir, 2003). Consequently, the effectiveness criteria and frameworks often address various aspects of process and output, as well as contextual enablers and constraints, rather than outcomes, as factors of effectiveness. Some contributions also make a distinction between (immediate) impacts and (indirect) outcomes. As a result, although the aim is to address outcomes, some approaches to effectiveness in the end turn out somewhat similar to those considered as checklist guidance in quality assurance/control (in Table 1).

In addition, the approaches emphasizing the use of models, tools and their outputs can be considered as a manifestation of the effectiveness perspective. They can generally be characterized as attempts to operationalise the interaction of modelling and assessment with the practical uses of their outputs. Most of the contributions are, however, relatively tool-centred (cf. evaluation of decision support systems as models), and most often little attention is given to the cognitive processes involved in the delivery and reception of information produced by models and assessments.

Table 4. Examples of effectiveness perspective to model and assessment performance.
Type Author(s) Description
Frameworks and criteria for effectiveness Kauppinen et al. 2006 Framework for the effectiveness of prospective human impact assessment
Quigley and Taylor 2004 Process, impact and outcome indicators for evaluating health impact assessment
Clark and Majone 1985 Criteria for appraisal of scientific inquiries with policy implications
Hildén et al. 2004 Necessary conditions and facilitating factors for effectiveness in strategic environmental assessment
Baker and McLelland 2003 Components of policy effectiveness in participatory environmental assessment
Pohjola and Tuomisto 2011 Dimensions of openness for analyzing the potential for effectiveness in participatory policy support
Tuomisto and Pohjola 2007 Properties of good assessment for evaluating effectiveness of assessments (updated description at: http://en.opasnet.org/w/Properties_of_good_assessment)
Effectiveness evaluations Wismar et al. 2007 Several cases of evaluating effectiveness of health impact assessment in Europe
Fischer and Gazzola 2006 General effectiveness criteria for strategic environmental assessment and their adaptation for Italy
Leu et al. 1996 Environmental impact assessment evaluation model and its application in Taiwan
Pölönen et al. 2010 Effectiveness of the Finnish environmental impact assessment system
Matthews et al. 2011 Example of outcome evaluation for environmental modelling and software
Use of models, tools and assessments Larocque et al. 2011 Framework to assist decision makers in the use of ecosystem model predictions
Sterk et al. 2011 Analysis of contribution of land-use modelling to societal problem solving
Diez and McIntosh 2011 Use of decision and information support tools in desertification policy and management
McIntosh et al. 2008. Developing tools to support environmental management and policy
Siebenhüner and Barth 2005 Role of computer modelling in participatory integrated assessments
Inman et al. 2011 Usage and perceived effectiveness of decision support systems in participatory planning
Dewar et al. 1996 Credible uses of the distributed interactive simulation (DIS) system
Pohjola et al. 2012 Analysis of interaction between environmental health assessment and policy making

The approaches to effectiveness range from external ex-post evaluations to support for development and management of modelling, assessment as well as decision making practices. All approaches, however, explicitly acknowledge the role of use in delivering the effects of knowledge provided by models and assessment, despite that the criteria against which effectiveness is considered may vary.

Other perspectives

Although many contributions to modelling and assessment performance in relevant literature can be quite comfortably located within the four perspectives above, there are also some other aspects that deserve to be mentioned. Those to be brought up here address credibility and acceptability, information quality, and communication.

Credibility is often considered necessary for acceptance of modelling and assessment endeavours and their outputs. It can be obtained more or less formally or informally e.g. through peer review, extended peer-review (Funtowicz and Ravetz, 1990) or reputation. Credibility and acceptability are often considered as aspects of broader performance concepts.

Modelling and assessment are essentially processes of producing information. Therefore, the contributions regarding information quality also outside the domains of modelling and assessment are of relevance here. This perspective resembles the uncertainty analysis perspective as they both focus into certain properties of an information product. Similarly, the variation among contributions addressing information quality is big.

Also communication of results, e.g. in terms of communicating uncertainties and risk information, is linked to performance of models and assessments. However, the issues of communication are not necessarily considered as integral parts of modelling and assessments endeavours. For example, risk assessment, risk management and risk communication are traditionally considered as separate, yet interrelated, entities, each having their own aims, practices, and practitioners (e.g. WHO/FAO, 2006).

Table 5. Examples of other perspectives to model and assessment performance.
Characterization Author(s) Description
Acceptance and credibility Alexandrov et al. 2011 Obtaining model credibility through peer-reviewed publication process
Aumann 2009 Model credibility in the context of policy appraisal
Information quality Wang and Strong 1996 A conceptual framework of data quality
Moody and Walsh 1999 An asset valuation approach to value of information
Skyrme 1994 Ten aspects that add value to information
Tongchuay and Praneetpolgrang 2010 Knowledge quality in knowledge management systems
Communication Wardekker et al. 2008 Uncertainty communication in environmental assessments
Janssen et al. 2005 Checklist for assessing and communicating uncertainties
Covello et al. 2001 Communication challenges posed by a release of a pathogen in an urban setting
Bischof and Eppler 2011 Clarity in knowledge communication

Discussion

By considering the reviewed contributions in light of the framework described in Figure 1, it seems that none of the perspectives nor any individual contributions alone sufficiently serve all the needs of evaluation and management of model and assessment performance. In most of the contributions, the main emphasis is on the processes and outputs of modelling and assessment while contextual aspects, outcomes, as well as use are addressed to a lesser extent, although more frequently in recent literature. The accounts of interaction of modelling and assessment with their societal context also appear to be vaguer in comparison to the commonly applied perspectives focusing on processes and outputs. No fundamental differences in perspectives to performance between domains of modelling and assessment can be seen.

Many approaches to performance seem to perceive evaluation as a separate entity that most often takes place only after the development of a model or assessment, and often is considered a responsibility of others than modellers and assessors themselves. Some major exceptions to this are the essentially proactive stepwise guidance in the quality assurance/control perspective, and the technical assessment of models perspective, in which the evaluation is often integrated in the model development. In addition, some of the effectiveness frameworks are explicitly intended as means to support design and execution, not only evaluation, of models and assessments.

The emphasis on processes and outputs in evaluation and management of model and assessment performance is in line with the fact that e.g. the issues of effectiveness and policy-relevance have become major topics also in modelling and assessment only during the last decades. As modellers, assessors, and researchers more generally, have been lacking requirements and incentives for effectiveness and policy-relevance (cf. Harris, 2002), correspondingly the practices, principles and methods of performance management and evaluation have not developed to address these issues. Instead, the impacts of modelling and assessment have mostly been considered mainly in terms of their process effects (cf. Matthews et al., 2011) within the communities of modellers and assessors, rather than outcomes in the broader societal context. Virtually all modelling and assessment endeavors in the fields of environment and health are, however, at least nominally, motivated by the aim to influence societal decision-making. The societal outcomes should thus be considered as the ultimate criterion for model and assessment performance. This has also been recognized in many new approaches to modeling, assessment as well as their evaluation (e.g. Tijhuis et al., 2012; Matthews et al., 2011). However, the complexity of addressing the outcomes remains a challenge. In the eyes of the evaluators, the relative simplicity of considering only processes, outputs or direct impacts in tightly bound settings of expert activities may appear inviting in comparison to attempting to account for complex indirect impacts within the broader social context. Unfortunately, this may not be adequate for serving the purposes of modelling, assessment and their evaluation.

It appears that a more comprehensive approach that covers all aspects of modelling and assessment in their societal context, as described in Figure 1, is needed to support model and assessment evaluation and management. In practice, this requires taking account of the making of models and assessments, their outputs, their use and other interaction with the societal context, as well as the decisions and actions influenced by the models and assessments, and the consequential outcomes. Such an approach would need to combine the essential characteristics of the different perspectives into one methodology, framework, or tool. However, a mere compilation of features taken from different perspectives would probably not be sufficient. A more thorough account of the mechanisms of collective knowledge creation and the relations between knowledge and action in a societal context is needed in order to truly bridge models and assessments with their outcomes (Pohjola et al., 2011). Unfortunately these aspects are barely even recognized in most current approaches to model and assessment performance.

Despite that the contributions within the effectiveness perspective were above characterized as often being somewhat vague, it seems likely that the effectiveness perspective provides the best basis for a comprehensive approach as the concept brings together the making, content, use as well as outcomes of models and assessments. Attempts to combine methods, frameworks and tools into a comprehensive methodology for design, making and evaluation of models, assessments as well as practices are being made, and the early results look promising. For example, some of the above mentioned effectiveness frameworks have been jointly applied for evaluating the effectiveness of two open assessments on alternative biofuel sources (Sandström et al. manuscript, http://en.opasnet.org/w/Biofuel_assessments) as well as analysing how different health, safety and environment assessment approaches serve the needs of decision making in public policy as well as manufacturing industry (Pohjola manuscript). A more noteworthy effort in this aspect is the ongoing TEKAISU-project (http://en.opasnet.org/w/Tekaisu) in which methods and tools for evaluation and management of knowledge-based city-level decision making are developed and applied. This development combines four above mentioned effectiveness frameworks by Kauppinen et al. (2006), Tuomisto and Pohjola (2007), Pohjola and Tuomisto (2011), and Pohjola et al. (2012) with a theoretical framework and practical tool for evaluating and managing practices in the social and health field by Koivisto and Pohjola (2012, https://pilotointi.innokyla.fi). All these combined frameworks and tools essentially build on collaboration and pragmatism, and thus emphasize the interrelations between knowledge and action in a social context. Consequently, the approach does not focus primarily on any separate aspect, process, output, use or outcome, of model and assessment performance, but considers and addresses the knowledge that is created, transferred and applied within the intertwined processes of modelling, assessment and decision making. This, we believe, is the key to making the whole chain from design of models and assessments to societal outcomes of decisions and actions both evaluable and manageable.

The emphasis in evaluation and management of models and assessments has been on rather scientific and technical aspects within the research domain, and perspectives addressing the impacts of models and assessments in terms of societal outcomes have emerged only quite recently and are still relatively rare. Still, it can be said that an outcome-oriented turn is taking place in evaluation and management of models and assessments, at least in development of theories and related academic discourses. However, a lot of work still remains in developing the frameworks, methods, tools and ultimately the common practices of modelling and assessment to sufficiently serve the needs of practical decision making for sustainable and healthy future. Succeeding in this will require better recognition of the mechanisms how models and assessments influence the knowledge, decisions and actions that deliver the consequences.

Acknowledgements

This review builds on research done in several projects receiving funding from various sources. Most importantly the authors would like to mention the EU projects INTARESE (Integrated Assessment of Health Risks of Environmental Stressors in Europe, 2005-2011, GOCE-CT-2005-018385), and BENERIS (Benefit–Risk Asessment of Food: An iterative Value-of-Information approach, 2006–2009, FOOD-CT-2006-022936), Safefoodera project BEPRARIBEAN (Best Practices for Risk-Benefit Analysis of Foods, project ID 08192), Academy of Finland (Grants 218114 and 126532), and SYTYKE doctoral programme in environmental health at the University of Eastern Finland.

References

Aertsen, W., Kint, V., van Orshoven, J., Muys, B., 2011. Evaluation of modelling techniques for forest site productivity prediction in contrasting ecoregions using stochastic multicriteria acceptability analysis (SMAA). Environmental Modelling & Software 26, 929-937.

Alexandrov, G.A., Ames, D., Bellochi, G., Bruen, M., Crout, N., Erechtchoukova, M., Hildebrandt, A., Hoffman, F., Jackisch, C., Khaiter, P., Mannina, G., Matsunaga, T., Purucker, S.T., Rivington, M., Samaniego, L., 2011. Technical assessment and evaluation of environmental models and software: Letter to the editor. Environmental Modelling & Software 26, 328-336.

van Asselt, M.B.A., Rotmans, J., 2002. Uncertainty in integrated assessment modelling: From positivism to Pluralism. Climatic Change 54, 75-105.

Aumann, C.A., 2011. Constructing model credibility in the context of policy appraisal. Environmental Modelling & Software 26, 258-265.

Bai, Y., Wagener, T., Reed, P., 2009. A top-down framework for watershed model evaluation and selection under uncertainty. Environmental Modelling & Software 24, 901-916.

Baker, D.C., McLelland, J.N., 2003. Evaluating the effectiveness of British Columbia's environmental assessment process for first nations' participation in mining development. Environmental Impact Assessment Review 23, 581-603.

Beck, B., 2002. Model evaluation and performance, in: El-Shaarawi, A.H., Piegorsch, W.W. (Eds.), Encyclopedia of Environmetrics. John Wiley & Sons Ltd, Chichester, Volume 3, pp. 1275-1279.

Basson, L., Petrie, J.G., 2007. An integrated approach for the consideration of uncertainty in decision making supported by Life Cycle Assessment. Environmental Modelling & Software 22, 167-176.

Bina, O., 2008. Context and systems: Thinking more broadly about effectiveness in strategic environmental assessments in China. Environmental Management 42, 717-733.

Bischof, N., Eppler, M.J., 2011. Caring for clarity in knowledge communication. Journal of Universal Computer Science 17, 1455-1473.

Blackstock, K.L., Kelly, G.J., Horsey, B.L., 2007. Developing and applying a framework to evaluate participatory research for sustainability. Ecological Economics 60, 726-742.

Blind, M.W., Refsgaard, J.C., 2007. Operationalising uncertainty in data and models for integrated water resource management. Water, Science & Technology 56, 1-12.

Borgonovo, E., 2008. Sensitivuty analysis of model output with input constraints: A generalized rationale for local methods. Risk Analysis 28, 667-680.

Briggs, D.J., 2008. A framework for integrated environmental health impact assessment of systemic risks. Environmental Health, 7:61. doi:10.1186/1476-069X-7-61

Briggs, D.J., Sable, C.E., Lee, K., 2009. Uncertainty in epidemiology and health risk and impact assessment. Environmental Geochemistry and Health 31, 189-203.

Brouwer, R., De Blois, C., 2008. Integrated modelling of risk and uncertainty underlying the cost and effectiveness of water quality measures. Environmental Modelling & Software 23, 922-937.

Brown, J.D., Heuvelink, G.B., Refsgaard, J.C., 2005. An integrated methodology for recording uncertainties about environmental data. Water, Science & Technology 52, 153-160.

Cartwright, N., 2007. Are RCTs the gold standard?. BioSocieties 2, 11-20. doi:10.1017/S1745855207005029

Clark, W.C., Majone, G., 1985. The critical appraisal of scientific inquiries with policy implications. Science, Technology & Human Values 10, 6-19.

Colyvan, M., 2008. Is probability the only coherent approach to uncertainty?. Risk Analysis 28, 645652

Covello, V.T., Peters, R.G., Wojtecki, J.G., Hyde, R.C., 2001. Risk communication, the west nile virus epidemic, and bioterrorism: Responding to the communication challenges posed by the intentional or unintentional release of a pathogen in an urban setting. Journal of Urban Health: Bulletin of the New York Academy of Medicine 78, 382-391.

van Delden, H., Seppelt, R., White, R., Jakeman, A.J., 2011. A methodology for the design and development of integrated models for policy support. Environmental Modelling & Software 26, 266-279.

Dewar, J.A., Bankes, S.C., Hodges, J.S., Lucas, T., Saunders-Newton, D.K., Vye, P., 1996. Credible uses of the distributed interactive simulation (DIS) system. RAND, Santa Monica, CA.

Diez, E., McIntosh, B.S., 2011. Organisational drivers for, constraints on and impacts of decision and information support tool use in desertification policy and management. Environmental Modelling & Software 26, 317-327.

ECHA, 2008. Guidance on Information Requirements and Chemical Safety Assessment. Guidance for the implementation of REACH. European Chemicals Agency.

Ekboir, J., 2003. Why impact analysis should not be used for research evaluation and what the alternatives are. Agricultural Systems 78, 166-184.

Fischer, T.B., Gazzola, P., 2006. SEA effectiveness criteria - equally valid in all countries? The case of Italy. Environmental Impact Assessment Review 26, 396-409.

Forristal, P.M., Wilke, D.L., McCarthy, L.S., 2008. Improving the quality of risk assessments in Canada using a principle-based apporach. Regulatory Toxicology and Pharmacology 50, 336-344.

Funtowicz, S.O. and J.R. Ravetz 1990. Uncertainty and Quality in Science for Policy. Kluwer Academic Publishers, the Netherlands.

Granger Morgan, M., Dowlatadabi, H., 1996. Learning from integrated assessment of climate change. Climatic change 34, 337-368.

Guyatt, G.H., Sackett, D.L., Sinclair, J.C., Hayward, R., Cook, D.J., Cook, R.J., 1995. Users' guides to the medical literature. IX. A method for grading health care recommendations. JAMA 274, 1800–1804.

Harris, G., 2002. Integrated assessment and modelling: an essential way of doing science. Environmental Modelling & Software 17, 201-207.

Harteloh, P.P.M., 2002. Quality systems in health care: a sociotechnical approach. Health Policy 64, 391-398.

Hildén, M., Furman, E., Kaljonen, M., 2004. Views on planning and expectations of SEA: the case of transport planning. Environmental Impact Assessment Review 24, 519-536.

Hoekstra, J., Hart, A., Boobis, A., Claupein, E., Cockburn, A., Hunt, A., Knudsen, I., Richardson, D., Schilter, B., Schütte, K., Torgerson, P.R., Verhagen, H., Watzl, B., Chiodini, A., 2010. BRAFO tiered approach for benefit-risk assessment of foods. Food and Chemical Toxicology, In press. doi:10.1016/j.fct.2010.05.049

Hokkanen, P., Kojo, M., 2003. How environmental impact assessment influences decision-making [in Finnish]. Ympäristöministeriö, Helsinki.

Hummel, J., Huitt, W., 1994. What you measure is what you get. GaASCD Newsletter: The Reporter, 10-11. Retrieved 13.10.2011 from http://www.edpsycinteractive.org/papers/wymiwyg.html

Inman, D., Blind, M., Ribarova, I., Krause, A., Roosenschoon, O., Kassahun, A., Scholten, H., Arampatzis, G., Abrami, G., McIntosh, B., Jeffrey, P., 2011. Perceived effectiveness of environmental decision support systems in participatory planing: Evidence from small groups of end-users. Environmental Modelling & Software 26, 302-309.

Jakeman, A.J., Letcher, R.A., 2003. Integrated assessment and modelling: features, principles and examples for catchment management. Environmental Modelling & Software 18, 491-501.

Jakeman, A.J., Letcher, R.A., Norton, J.P., 2006. Ten iterative steps in development and evaluation of environmental models. Environmental Modelling & Software 21, 602-614.

Janssen, P.H.M., Petersen, A.C., van der Sluijs, J.P., Risbey, J.S., Ravetz, J.R., 2005. A guidance for assessing and communicating uncertainties. Water Science & Technology 52, 125-131.

Kann, A., Weyant, J.P., 2000. Approaches for performing uncertainty analysis in large-scale energy/economic policy models. Environmental Modeling and Assessment 5, 29-46.

Kauppinen, T., Nelimarkka, K., Perttilä, K., 2006. The effectiveness of human impact assessment in the Finnish Health Cities Network. Public Health 120, 1033-1041.

Kloprogge, P., van der Sluijs, J.P., Petersen, A.C., 2011. A method for the analysis of assumptions in model-based environmental assessments. Environmental Modelling & Software 26, 289-301.

Koivisto, J., Pohjola, P., 2012. Practices, modifications and generativity - REA: a practical tool for managing the innovation processes of practices. Systems, Signs & Actions 5(1), 100-116. http://www.sysiac.org/uploads/SySiAc2011-Koivisto-Pohjola.pdf

Larocque, G.R., Bhatti, J.S., Ascough II, J.C., Liu, J., Luckai, N., Mailly, D., Archambault, L., Gordon, A.M., 2011. An analytical framework to assist decision makers in the use of forest ecosystem model predictions. Environmental Modelling & Software 26, 280-288.

Leu, W., Williams, W.P., Bark, A.W., 1996. Development of an environmental impact assessment evaluation model and its application: Taiwan case study. Environmental Impact Assessment Review 16, 115-133.

Leviton, L.C., 2003. Evaluation use: Advances, challenges and applications. American Journal of Evaluation 24, 525-535.

Liu, Y., Gupta, H., Springer, E., Wagener, T., 2008. Linking science with environmental decision making: Experiences from an integrated modelling approach to supporting sustainable water resources management. Environmental Modeling & Software 23, 846-858.

Mahmoud, M., Liu, Y., Hartmann, H., Stewart, S., Wagener, T., Semmens, D., Stewart, R., Gupta, H., Dominguez, D., Dominguez, F., Hulse, D., Letcher, R., Rashleigh, B., Smith, C., Street, R., Ticehurst, J., Twery, M., van Delden, H., Waldick, R., White, D., Winter, L., 2009. A formal framework for scenario development in support of environmental decision-making. Environmental Modelling & Software 24, 798-808.

Matthews, K.B., Rivington, M., Blackstock, K.L., McCrum, G., Buchan, K., Miller, D.G., 2011. Raising the bar? - The challenges of evaluating the outcomes of environmental modelling and software. Environmental Modelling & Software 26 (3), 247-257.

McIntosh, B.S., Giupponi, C., Voinov, A.A., Smith, C., Matthews, K.B., Monticino, M., Kolkman, M.J., Crossman, N., van Ittersum, M., Haase, D., Haase, A., Mysiak, J., Groot, J.C.J., Sieber, S., Verweij, P., Quinn, N., Waeger, P., Gaber, N., Hepting, D., Scholten, H., Sulis, A., van Delden, H., Gaddis, E., Assaf, H., 2008. Bridging the gaps between design and use: Developing tools to support environmental management and policy, in: Jakeman, A.J., Voinov, A.A., Rizzoli, A.E., Chen, S.H. (Eds.): Environmental Modelling, Software and Decision Support. Elsevier, Amsterdam, pp. 33-48.

McIntosh, B.S., Alexandrov, G., Matthews, K., Mysiak, J., van Ittersum, M. (Eds.), 2011a. Thematic issue on the assessment and evaluation of environmental models and software. Environmental Modelling & Software 26, 245-336.

McIntosh, B.S., Alexandrov, G., Matthews, K., Mysiak, J., van Ittersum, M., 2011b. Preface: Thematic issue on the assessment and evaluation of environmental models and software. Environmental Modelling & Software 26, 245-246.

Mo, X., Liu, S., Lin, Z., 2011. Evaluation of an ecosystem model for a wheat-maize double cropping system over the North China Plain. Environmental Modelling & Software, In press. doi:10.1016/j.envsoft.2011.07.002

Moody, D., Walsh, P., 1999. Measuring the value of information: An asset valuation approach. Seventh European Conference on Information System (ECIS’99), Copenhagen Business School, Frederiksberg, Denmark, 23-25 June, 1999. Available: http://wwwinfo.deis.unical.it/zumpano/2004-2005/PSI/lezione2/ValueOfInformation.pdf Accessed 13.10.2011.

Moschandreas, D.J., Karuchit, S., 2002. Scenario-model-parameter: a new method of cumulative risk uncertainty analysis. Environment International 28, 247-261.

Nilsen, T., Aven, T., 2003. Models and model uncertainty in the context of risk analysis. Reliability Engineering & System Safety 79, 309-317.

NRC, 1983. Risk Assessment in the Federal Government: Managing the Progress. The National Research Council. National Academy Press, Washington D.C.

NRC, 1996. Understanding Risk: Informing Decisions in a Democratic Society. The National Research Council. National Academy Press, Washington D.C.

NRC, 2009. Science and Decisions: Advancing Risk Assessment. National Research Council. National Academy Press, Washington D.C.

O'Hagan, A., 2011. Probabilistic uncertainty specification: Overview, elaboration techniques and their application to a mechanistic model of carbon flux. Environmental Modelling & Software, In press. doi:10.1016/j.envsoft.2011.03.003

Patton, M.Q., 2008: Utilization-focused evaluation, fourth edition. SAGE Publications, Inc. MN.

Pohjola, M.V., Leino, O., Kollanus, V., Tuomisto, J.T., Gunnlaugsdόttir, H., Holm, F., Kalogeras, N., Luteijn, J.M., Magnusson, S.H., Odekerken, G., Tijhuis, M.J., Ueland, Ø., White, B.C., Verhagen, H., 2012. State of the art in benefit–risk analysis: environmental health. Food Chem. Toxicol. 2012, 50:40-55. http://dx.doi.org/10.1016/j.fct.2011.06.004

Pohjola, M.V., Pohjola, P., Paavola, S., Bauters, M., Tuomisto, J.T., 2011. Pragmatic Knowledge Services. Journal of Universal Computer Science 17:472-497. http://dx.doi.org/10.3217/jucs-017-03-0472

Pohjola, M.V., Tuomisto, J.T., 2011. Openness in participation, assessment, and policy making upon issues of environment and environmental health: a review of literature and recent project results. Environmental Health 10, 58. doi:10.1186/1476-069X-10-58

Pohjola, M.V. Assessment of impacts to health, safety, and environment in the context of materials processing and related public policy. Manuscript.

Pollino, C.A., Woodberry, O., Nicholson, A., Korb, K., Hart, B.T., 2007. Parameterisation and evaluation of a Bayesian network for use in an ecological risk assessment.

Pope, J., Annandale, D., Morrison-Saunders, A., 2004. Conceptualising sustainability assessment. Environmental Impact Assessment Review 24, 595-616.

Pölönen, I., Hokkanen, P., Jalava, K., 2011. The effectiveness of the Finnish EIA system - What works, what doesn't, and what could be improved? Environmental Impact Assessment Review 31, 120-128.

Quigley, R.J., Taylor, L.C., 2004. Evaluating health impact assessment. Public Health 118, 544-552.

Reeves, C.A., Bednar, D.A., 1994. Defining quality: Alternatives and implications. Academy of Management Review 19, 419-445.

Refsgaard, J.C., Henriksen, H.J., 2004. Modelling guidelines - terminology and guiding principles. Advances in Water Resources 27, 71-82.

Refsgaard, J.C., Henriksen, H.J., Harrar, W.G., Scholten, H., Kassahun, A., 2005. Quality assurance in model-based water management - review of existing practice and outline of new approaches. Environmental Modelling & Software 20, 1201-1215.

Refsgaard, J.C., van der Sluijs, J.P., Brown, J., van der Keur, P., 2006. A framework for dealing with uncertainty due to model structure error. Advances in Water Resources 29, 1586-1597.

Refsgaard, J.C., van der Sluijs, J.P., Lajer Højberg, A., Vanrolleghem, P.A., 2007. Uncertainty in the environmental modelling process - A framework and guidance. Environmental Modelling & Software 22, 1543-1556.

Risbey, J., Kandlikar, M., Patwardhan, A., 1996. Assessing integrated assessment. Climatic Change 34, 369-395.

Risbey, J. van der Sluijs, J.P., Kloprogge, P., Ravetz, J., Funtowicz, S., Corral Quintana, S., 2005. Application of a checklist for quality assistance in environmental modelling to an energy model. Environmental Modeling and assessment 10, 63-79.

Rizzoli, A.E., Leavesley, G., Ascough II, J.C., Argent, R.M., Athanasiadis, I.N., Brilhante, V., Claeys, F.H.A., David, O., Donatelli, M., Gijsbergs, P., Havlik, D., Kassahun, A., Krause, P., Quinn, N.W.T., Scholten, H., Sojda, R.S., Villa, F., 2008. Integrated modelling frameworks for environmental assessment decision support, in: Jakeman, A.J., Voinov, A.A., Rizzoli, A.E., Chen, S.H. (Eds.): Environmental Modelling, Software and Decision Support. Elsevier, Amsterdam, pp. 101-118.

Sandström, V., Tuomisto, J.T., Majaniemi, S., Rintala, T., Pohjola, M.V. Evaluating effectiveness of open assessments on alternative biofuel sources. Manuscript.

Siebenhüner, B., Barth, V., 2005. The role of computer modelling in participatory integrated assessments. Environmental Impact Assessment Review 25, 367-389.

Skyrme, D.J., 1994. Ten ways to add value to your business. Managing Information 1, 20-25. Available: http://www.skyrme.com/pubs/tenways.htm (accessed 13.10.2011)

Sterk, B., van Ittersum, M.K., Leeuwis, C., 2011. How, when, and for what reasons does land use modelling contribute to societal problem solving?. Environmental Modelling & Software 26, 310-316.

van der Sluijs, J.P., 2002. Integrated Assessment, in: Tolba, M.K. (Ed.), Volume 4, Responding to Global Environmental Change, pp. 250-253, in: Munn, T. (Ed.), Encyclopedia of Global Environmental Change. John Wiley & Sons, Ltd, Chichester.

van der Sluijs, J.P., Craye, M., Funtowicz, S., Kloprogge, P., Ravetz, J., Risbey, J., 2005. Combining quantitative and qualitative measures of uncertainty in model-based environmental assessment: The NUSAP system. Risk Analysis 25, 481-492.

van der Sluijs, J.P., Petersen, A.C., Janssen, P.H.M., Risbey, J.S., Ravetz, J.R., 2008: Exploring the quality of evidence for complex and contested policy decisions. Environmental Research Letters 3, 024008. doi:10.1088/1748-9326/3/2/024008

Sojda, R.S., 2007. Empirical evaluation of decision support systems: Needs, definitions, potential methods, and an example pertaining to waterfowl management. Environmental Modelling & Software 22, 269-277.

Sonneveld, B.G.J.S., Keyzer, M.A., Stroosnijder, L., 2011. Evaluating quantitative and qualitative models: An application for nationwide water erosion assessment in Ethiopia. Environmental Modelling & Software 26, 1161-1170.

Tijhuis, M.J., de Jong, N., Pohjola, M.V., Gunnlaugsdόttir, H., Hendriksen, M., Hoekstra, J., Holm, F., Kalogeras, N. Leino, O., van Leeuwen, F.X., Luteijn, J.M., Magnússon, S.H., Odekerken, G., Rompelberg, C., Tuomisto, J.T., Ueland, Ø., White, B.C., Verhagen, H., 2012. State of the art in benefit–risk analysis: food and nutrition. Food Chem. Toxicol. 2012, 50:5-25. http://dx.doi.org/10.1016/j.fct.2011.06.010

Tongchuay, C., Praneetpolgrang, P., 2010. Knowledge quality and quality metrics in knowledge management systems. Proceedings of the Fifth International Conference on eLearning for Knowledge-Based Society. December 11-12, 2008. Bangkok Metro, Thailand.

Tuomisto, J.T., Pohjola, M.V., 2007. Open Risk Assessment - A new way of providing information for decision-making. Publications of the National Public Health Institute B18/2007. KTL - National Public Health Institute, Kuopio.

Wagener, T., Kollat, J., 2007. Numerical and visual evaluation of hydrological and environmental models using the Monte Carlo analysis toolbox. Environmental Modelling & Software 22, 1021-1033.

Walker, W.E., Harremoës, P., Rotmans, J., van der Sluijs, J.P., van Asselt, M.B.A., Janssen, P., Krayer von Krauss, M.P., 2003. Defining uncertainty: A conceptual basis for uncertainty management in model-based decision support. Integrated Assessment 4, 5-17.

Wang, R.Y., Strong, D.M., 1996. Beyond accuracy: What data quality means to data consumers. Journal of Management Information Systems 12, 5-34.

Wardekker, J.A., van der Sluijs, J.P., Janssen, P.H.M., Kloprogge, P., Petersen, A.C., 2008. Uncertainty communication in environmental assessments: views from the Dutch science-policy interface. Environmental Science & Policy 11, 627-641.

Wismar, M., Blau, J., Ernst, K., Figueras, J. (Eds.), 2007. The Effectiveness of Health Impact Assessment.: Scope and limitations of supporting decision-making in Europe. WHO, Copenhagen.

WHO, 1999. Health Impact Assessment (HIA), main concepts and suggested approach. Gothenburg consensus paper. World Health Organization, Brussels.

WHO/FAO, 2006. Food safety risk analysis: A guide for national food safety authorities. World Health Organization, Food and Agriculture Organization of the United Nations, Rome.

Wood, C., 1995. Environmental impact assessment: a comparative review. Longman Scientific & Technical, New York.

Wyat Appel, K., Gilliam, R.C., Davis, N., Howard, S.C., 2011. Overview of the atmospheric model evaluation tool (AMET) v1.1 for evaluating meteorological and air quality models. Environmental Modelling & Software 26, 434-443.

Xu, Y., Booij, M.J., Mynett, A.E., 2007. An appropriateness framework for the Dutch Meuse decision support system. Environmental Modelling & Software 22, 1667-1678.