Uploaded by User46124

Sensitivity analysis - Wikipedia

Sensitivity analysis - Wikipedia
Sensitivity analysis
Sensitivity analysis is the study of how the uncertainty in the output of a mathematical model or system (numerical or otherwise) can be apportioned to different sources of uncertainty in its inputs.[1][2] A related
practice is uncertainty analysis, which has a greater focus on uncertainty quantification and propagation of uncertainty; ideally, uncertainty and sensitivity analysis should be run in tandem.
The process of recalculating outcomes under alternative assumptions to determine the impact of a variable under sensitivity analysis can be useful for a range of purposes,[3] including:
Testing the robustness of the results of a model or system in the presence of uncertainty.
Increased understanding of the relationships between input and output variables in a system or model.
Uncertainty reduction, through the identification of model inputs that cause significant uncertainty in the output and should therefore be the focus of attention in order to increase robustness (perhaps by further
Searching for errors in the model (by encountering unexpected relationships between inputs and outputs).
Model simplification – fixing model inputs that have no effect on the output, or identifying and removing redundant parts of the model structure.
Enhancing communication from modelers to decision makers (e.g. by making recommendations more credible, understandable, compelling or persuasive).
Finding regions in the space of input factors for which the model output is either maximum or minimum or meets some optimum criterion (see optimization and Monte Carlo filtering).
In case of calibrating models with large number of parameters, a primary sensitivity test can ease the calibration stage by focusing on the sensitive parameters. Not knowing the sensitivity of parameters can result in
time being uselessly spent on non-sensitive ones.[4]
To seek to identify important connections between observations, model inputs, and predictions or forecasts, leading to the development of better models.[5][6]
Settings and constraints
Core methodology
One-at-a-time (OAT/OFAT)
Local methods
Scatter plots
Regression analysis
Variance-based methods
Variogram-based methods
Variogram analysis of response surfaces (VARS)
Alternative methods
High-dimensional model representations (HDMR)
Fourier amplitude sensitivity test (FAST)
Other issues
Assumptions vs. inferences
Pitfalls and difficulties
1 of 13
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
Social sciences
In meta-analysis
Multi-criteria decision making
Time-critical decision making
Model calibration and improvement
Sensitivity auditing
Related concepts
See also
Sensitivity analysis events
Further reading
External links
A mathematical model (for example a climate model, an economic model, or a finite element model in engineering etc.) can be highly complex, and as a result its relationships between inputs and outputs may be poorly
understood. In such cases, the model can be viewed as a black box, i.e. the output is an "opaque" function of its inputs.
Quite often, some or all of the model inputs are subject to sources of uncertainty, including errors of measurement, absence of information and poor or partial understanding of the driving forces and mechanisms. This
uncertainty imposes a limit on our confidence in the response or output of the model. Further, models may have to cope with the natural intrinsic variability of the system (aleatory), such as the occurrence of stochastic
Good modeling practice requires that the modeler provide an evaluation of the confidence in the model. This requires, first, a quantification of the uncertainty in any model results (uncertainty analysis); and second, an
evaluation of how much each input is contributing to the output uncertainty. Sensitivity analysis addresses the second of these issues (although uncertainty analysis is usually a necessary precursor), performing the role of
ordering by importance the strength and relevance of the inputs in determining the variation in the output.[2]
In models involving many input variables, sensitivity analysis is an essential ingredient of model building and quality assurance. National and international agencies involved in impact assessment studies have included
sections devoted to sensitivity analysis in their guidelines. Examples are the European Commission (see e.g. the guidelines for impact assessment),[8] the White House Office of Management and Budget, the
Intergovernmental Panel on Climate Change and US Environmental Protection Agency's modelling guidelines.[9]
Settings and constraints
The choice of method of sensitivity analysis is typically dictated by a number of problem constraints or settings. Some of the most common are
Computational expense: Sensitivity analysis is almost always performed by running the model a (possibly large) number of times, i.e. a sampling-based approach.[10] This can be a significant problem when,
A single run of the model takes a significant amount of time (minutes, hours or longer). This is not unusual with very complex models.
The model has a large number of uncertain inputs. Sensitivity analysis is essentially the exploration of the multidimensional input space, which grows exponentially in size with the number of inputs. See the curse
of dimensionality.
Computational expense is a problem in many practical sensitivity analyses. Some methods of reducing computational expense include the use of emulators (for large models),
and screening methods (for reducing the dimensionality of the problem). Another method is to use an event-based sensitivity analysis method for variable selection for timeconstrained applications.[11] This is an input variable selection (IVS) method that assembles together information about the trace of the changes in system inputs and outputs
2 of 13
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
using sensitivity analysis to produce an input/output trigger/event matrix that is designed to map the relationships between input data as causes that trigger events and the output
data that describes the actual events. The cause-effect relationship between the causes of state change i.e. input variables and the effect system output parameters determines
which set of inputs have a genuine impact on a given output. The method has a clear advantage over analytical and computational IVS method since it tries to understand and
interpret system state change in the shortest possible time with minimum computational overhead.[11][12]
Correlated inputs: Most common sensitivity analysis methods assume independence between model inputs, but sometimes inputs can be strongly correlated. This is still an immature field of research and definitive
methods have yet to be established.
Nonlinearity: Some sensitivity analysis approaches, such as those based on linear regression, can inaccurately measure sensitivity when the model response is nonlinear with respect to its inputs. In such cases,
variance-based measures are more appropriate.
Model interactions: Interactions occur when the perturbation of two or more inputs simultaneously causes variation in the output greater than that of varying each of the inputs alone. Such interactions are present in
any model that is non-additive, but will be neglected by methods such as scatterplots and one-at-a-time perturbations.[13] The effect of interactions can be measured by the total-order sensitivity index.
Multiple outputs: Virtually all sensitivity analysis methods consider a single univariate model output, yet many models output a large number of possibly spatially or time-dependent data. Note that this does not
preclude the possibility of performing different sensitivity analyses for each output of interest. However, for models in which the outputs are correlated, the sensitivity measures can be hard to interpret.
Given data: While in many cases the practitioner has access to the model, in some instances a sensitivity analysis must be performed with "given data", i.e. where the sample points (the values of the model inputs for
each run) cannot be chosen by the analyst. This may occur when a sensitivity analysis has to be performed retrospectively, perhaps using data from an optimisation or uncertainty analysis, or when data comes from a
discrete source.[14]
Core methodology
There are a large number of approaches to performing a sensitivity analysis, many of which have been developed to address one
or more of the constraints discussed above.[2] They are also distinguished by the type of sensitivity measure, be it based on (for
example) variance decompositions, partial derivatives or elementary effects. In general, however, most procedures adhere to the
following outline:
1. Quantify the uncertainty in each input (e.g. ranges, probability distributions). Note that this can be difficult and many methods
exist to elicit uncertainty distributions from subjective data.[15]
2. Identify the model output to be analysed (the target of interest should ideally have a direct relation to the problem tackled by
the model).
3. Run the model a number of times using some design of experiments,[16] dictated by the method of choice and the input
4. Using the resulting model outputs, calculate the sensitivity measures of interest.
In some cases this procedure will be repeated, for example in high-dimensional problems where the user has to screen out
unimportant variables before performing a full sensitivity analysis.
The various types of "core methods" (discussed below) are distinguished by the various sensitivity measures which are calculated.
These categories can somehow overlap. Alternative ways of obtaining these measures, under the constraints of the problem, can
be given.
One-at-a-time (OAT/OFAT)
One of the simplest and most common approaches is that of changing one-factor-at-a-time (OFAT or OAT), to see what effect this
produces on the output.[17][18][19] OAT customarily involves
Moving one input variable, keeping others at their baseline (nominal) values, then,
Returning the variable to its nominal value, then repeating for each of the other inputs in the same way.
Ideal scheme of a possibly sampling-based sensitivity analysis. Uncertainty arising
from different sources – errors in the data, parameter estimation procedure,
alternative model structures – are propagated through the model for uncertainty
analysis and their relative importance is quantified via sensitivity analysis.
Sensitivity may then be measured by monitoring changes in the output, e.g. by partial derivatives or linear regression. This
appears a logical approach as any change observed in the output will unambiguously be due to the single variable changed. Furthermore, by changing one variable at a time, one can keep all other variables fixed to their
central or baseline values. This increases the comparability of the results (all 'effects' are computed with reference to the same central point in space) and minimizes the chances of computer programme crashes, more
3 of 13
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
likely when several input factors are changed simultaneously. OAT is frequently preferred by modellers because of practical
reasons. In case of model failure under OAT analysis the modeller immediately knows which is the input factor responsible for the
Despite its simplicity however, this approach does not fully explore the input space, since it does not take into account the
simultaneous variation of input variables. This means that the OAT approach cannot detect the presence of interactions between
input variables.[20]
Local methods
Local methods involve taking the partial derivative of the output Y with respect to an input factor Xi:
where the subscript X0 indicates that the derivative is taken at some fixed point in the space of the input (hence the 'local' in the
name of the class). Adjoint modelling[21][22] and Automated Differentiation[23] are methods in this class. Similar to OAT/OFAT,
local methods do not attempt to fully explore the input space, since they examine small perturbations, typically one variable at a
Sampling-based sensitivity analysis by scatterplots. Y (vertical axis) is a function of
four factors. The points in the four scatterplots are always the same though sorted
differently, i.e. by Z1, Z2, Z3, Z4 in turn. Note that the abscissa is different for each
plot: (−5, +5) for Z1, (−8, +8) for Z2, (−10, +10) for Z3 and Z4. Z4 is most important in
influencing Y as it imparts more 'shape' on Y.
Scatter plots
A simple but useful tool is to plot scatter plots of the output variable against individual input variables, after (randomly) sampling the model over its input distributions. The advantage of this approach is that it can also
deal with "given data", i.e., a set of arbitrarily-placed data points, and gives a direct visual indication of sensitivity. Quantitative measures can also be drawn, for example by measuring the correlation between Y and Xi, or
even by estimating variance-based measures by nonlinear regression.[14]
Regression analysis
Regression analysis, in the context of sensitivity analysis, involves fitting a linear regression to the model response and using standardized regression coefficients as direct measures of sensitivity. The regression is required
to be linear with respect to the data (i.e. a hyperplane, hence with no quadratic terms, etc., as regressors) because otherwise it is difficult to interpret the standardised coefficients. This method is therefore most suitable
when the model response is in fact linear; linearity can be confirmed, for instance, if the coefficient of determination is large. The advantages of regression analysis are that it is simple and has a low computational cost.
Variance-based methods
Variance-based methods[24][25][26] are a class of probabilistic approaches which quantify the input and output uncertainties as probability distributions, and decompose the output variance into parts attributable to input
variables and combinations of variables. The sensitivity of the output to an input variable is therefore measured by the amount of variance in the output caused by that input. These can be expressed as conditional
expectations, i.e., considering a model Y=f(X) for X={X1, X2, ... Xk}, a measure of sensitivity of the ith variable Xi is given as,
where "Var" and "E" denote the variance and expected value operators respectively, and X~i denotes the set of all input variables except Xi. This expression essentially measures the contribution Xi alone to the uncertainty
(variance) in Y (averaged over variations in other variables), and is known as the first-order sensitivity index or main effect index. Importantly, it does not measure the uncertainty caused by interactions with other
variables. A further measure, known as the total effect index, gives the total variance in Y caused by Xi and its interactions with any of the other input variables. Both quantities are typically standardised by dividing by
4 of 13
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
Variance-based methods allow full exploration of the input space, accounting for interactions, and nonlinear responses. For these reasons they are widely used when it is feasible to calculate them. Typically this calculation
involves the use of Monte Carlo methods, but since this can involve many thousands of model runs, other methods (such as emulators) can be used to reduce computational expense when necessary. Note that full variance
decompositions are only meaningful when the input factors are independent from one another.[27]
Variogram-based methods
Variogram analysis of response surfaces (VARS)
One of the major shortcomings of the previous sensitivity analysis methods is that none of them considers the spatially ordered structure of the response surface/output of the model Y=f(X) in the parameter space. By
utilizing the concepts of directional variograms and covariograms, variogram analysis of response surfaces (VARS) addresses this weakness through recognizing a spatially continuous correlation structure to the values of
Y, and hence also to the values of
Basically, the higher the variability the more heterogeneous is the response surface along a particular direction/parameter, at a specific perturbation scale. Accordingly, in the VARS framework, the values of directional
variograms for a given perturbation scale can be considered as a comprehensive illustration of sensitivity information, through linking variogram analysis to both direction and perturbation scale concepts. As a result, the
VARS framework accounts for the fact that sensitivity is a scale-dependent concept, and thus overcomes the scale issue of traditional sensitivity analysis methods.[30] More importantly, VARS is able to provide relatively
stable and statistically robust estimates of parameter sensitivity with much lower computational cost than other strategies (about two orders of magnitude more efficient).[31] Noteworthy, it has been shown that there is a
theoretical link between the VARS framework and the variance-based and derivative-based approaches.
Screening is a particular instance of a sampling-based method. The objective here is rather to identify which input variables are contributing significantly to the output uncertainty in high-dimensionality models, rather
than exactly quantifying sensitivity (i.e. in terms of variance). Screening tends to have a relatively low computational cost when compared to other approaches, and can be used in a preliminary analysis to weed out
uninfluential variables before applying a more informative analysis to the remaining set. One of the most commonly used screening method is the elementary effect method.[32][33]
Alternative methods
A number of methods have been developed to overcome some of the constraints discussed above, which would otherwise make the estimation of sensitivity measures infeasible (most often due to computational expense).
Generally, these methods focus on efficiently calculating variance-based measures of sensitivity.
Emulators (also known as metamodels, surrogate models or response surfaces) are data-modeling/machine learning approaches that involve building a relatively simple mathematical function, known as an emulator, that
approximates the input/output behaviour of the model itself.[34] In other words, it is the concept of "modelling a model" (hence the name "metamodel"). The idea is that, although computer models may be a very complex
series of equations that can take a long time to solve, they can always be regarded as a function of their inputs Y=f(X). By running the model at a number of points in the input space, it may be possible to fit a much simpler
emulator η(X), such that η(X)≈f(X) to within an acceptable margin of error. Then, sensitivity measures can be calculated from the emulator (either with Monte Carlo or analytically), which will have a negligible additional
computational cost. Importantly, the number of model runs required to fit the emulator can be orders of magnitude less than the number of runs required to directly estimate the sensitivity measures from the model.[35]
Clearly the crux of an emulator approach is to find an η (emulator) that is a sufficiently close approximation to the model f. This requires the following steps,
1. Sampling (running) the model at a number of points in its input space. This requires a sample design.
2. Selecting a type of emulator (mathematical function) to use.
3. "Training" the emulator using the sample data from the model – this generally involves adjusting the emulator parameters until the emulator mimics the true model as well as possible.
Sampling the model can often be done with low-discrepancy sequences, such as the Sobol sequence – due to mathematician Ilya M. Sobol or Latin hypercube sampling, although random designs can also be used, at the
loss of some efficiency. The selection of the emulator type and the training are intrinsically linked, since the training method will be dependent on the class of emulator. Some types of emulators that have been used
5 of 13
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
successfully for sensitivity analysis include,
Gaussian processes[35] (also known as kriging), where the any combination of output points is assumed to be distributed as a multivariate Gaussian distribution. Recently, "treed" Gaussian processes have been used
to deal with heteroscedastic and discontinuous responses.[36][37]
Random forests,[34] in which a large number of decision trees are trained, and the result averaged.
Gradient boosting,[34] where a succession of simple regressions are used to weight data points to sequentially reduce error.
Polynomial chaos expansions,[38] which use orthogonal polynomials to approximate the response surface.
Smoothing splines,[39] normally used in conjunction with HDMR truncations (see below).
The use of an emulator introduces a machine learning problem, which can be difficult if the response of the model is highly nonlinear. In all cases it is useful to check the accuracy of the emulator, for example using crossvalidation.
High-dimensional model representations (HDMR)
A high-dimensional model representation (HDMR)[40][41] (the term is due to H. Rabitz[42]) is essentially an emulator approach, which involves decomposing the function output into a linear combination of input terms
and interactions of increasing dimensionality. The HDMR approach exploits the fact that the model can usually be well-approximated by neglecting higher-order interactions (second or third-order and above). The terms
in the truncated series can then each be approximated by e.g. polynomials or splines (REFS) and the response expressed as the sum of the main effects and interactions up to the truncation order. From this perspective,
HDMRs can be seen as emulators which neglect high-order interactions; the advantage being that they are able to emulate models with higher dimensionality than full-order emulators.
Fourier amplitude sensitivity test (FAST)
The Fourier amplitude sensitivity test (FAST) uses the Fourier series to represent a multivariate function (the model) in the frequency domain, using a single frequency variable. Therefore, the integrals required to
calculate sensitivity indices become univariate, resulting in computational savings.
Methods based on Monte Carlo filtering.[43][44] These are also sampling-based and the objective here is to identify regions in the space of the input factors corresponding to particular values (e.g. high or low) of the output.
Other issues
Assumptions vs. inferences
In uncertainty and sensitivity analysis there is a crucial trade off between how scrupulous an analyst is in exploring the input assumptions and how wide the resulting inference may be. The point is well illustrated by the
econometrician Edward E. Leamer:[45][46]
I have proposed a form of organized sensitivity analysis that I call 'global sensitivity analysis' in which a neighborhood of alternative assumptions is selected and the corresponding interval of inferences is
identified. Conclusions are judged to be sturdy only if the neighborhood of assumptions is wide enough to be credible and the corresponding interval of inferences is narrow enough to be useful.
Note Leamer's emphasis is on the need for 'credibility' in the selection of assumptions. The easiest way to invalidate a model is to demonstrate that it is fragile with respect to the uncertainty in the assumptions or to show
that its assumptions have not been taken 'wide enough'. The same concept is expressed by Jerome R. Ravetz, for whom bad modeling is when uncertainties in inputs must be suppressed lest outputs become
Pitfalls and difficulties
6 of 13
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
Some common difficulties in sensitivity analysis include
Too many model inputs to analyse. Screening can be used to reduce dimensionality.
The model takes too long to run. Emulators (including HDMR) can reduce the number of model runs needed.
There is not enough information to build probability distributions for the inputs. Probability distributions can be constructed from expert elicitation, although even then it may be hard to build distributions with great
confidence. The subjectivity of the probability distributions or ranges will strongly affect the sensitivity analysis.
Unclear purpose of the analysis. Different statistical tests and measures are applied to the problem and different factors rankings are obtained. The test should instead be tailored to the purpose of the analysis, e.g.
one uses Monte Carlo filtering if one is interested in which factors are most responsible for generating high/low values of the output.
Too many model outputs are considered. This may be acceptable for quality assurance of sub-models but should be avoided when presenting the results of the overall analysis.
Piecewise sensitivity. This is when one performs sensitivity analysis on one sub-model at a time. This approach is non conservative as it might overlook interactions among factors in different sub-models (Type II
Some examples of sensitivity analyses performed in various disciplines follow here.
Environmental computer models are increasingly used in a wide variety of studies and applications. For example, global climate models are used for both short-term weather forecasts and long-term climate change.
Moreover, computer models are increasingly used for environmental decision-making at a local scale, for example for assessing the impact of a waste water treatment plant on a river flow, or for assessing the behavior and
life-length of bio-filters for contaminated waste water.
In both cases sensitivity analysis may help to understand the contribution of the various sources of uncertainty to the model output uncertainty and the system performance in general. In these cases, depending on model
complexity, different sampling strategies may be advisable and traditional sensitivity indices have to be generalized to cover multiple model outputs,[48] heteroskedastic effects and correlated inputs.[12]
In a decision problem, the analyst may want to identify cost drivers as well as other quantities for which we need to acquire better knowledge in order to make an informed decision. On the other hand, some quantities
have no influence on the predictions, so that we can save resources at no loss in accuracy by relaxing some of the conditions. See Corporate finance: Quantifying uncertainty. Additionally to the general motivations listed
above, sensitivity analysis can help in a variety of other circumstances specific to business:
To identify critical assumptions or compare alternative model structures
To guide future data collections
To optimize the tolerance of manufactured parts in terms of the uncertainty in the parameters
To optimize resources allocation
However, there are also some problems associated with sensitivity analysis in the business context:
Variables are often interdependent (correlated), which makes examining each variable individually unrealistic. E.G. changing one factor such as sales volume, will most likely affect other factors such as the selling
Often the assumptions upon which the analysis is based are made by using past experience/data which may not hold in the future.
Assigning a maximum and minimum (or optimistic and pessimistic) value is open to subjective interpretation. For instance one person's 'optimistic' forecast may be more conservative than that of another person
performing a different part of the analysis. This sort of subjectivity can adversely affect the accuracy and overall objectivity of the analysis.
Social sciences
Sensitivity analysis is common practice in social sciences. A famous early example is Mroz (1987), who analysed econometric models of female labor market participation.[49]
In modern econometrics the use of sensitivity analysis to anticipate criticism is the subject of one of Peter Kennedy's "ten commandments of applied econometrics":[50]
7 of 13
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
Thou shall confess in the presence of sensitivity. Corollary: Thou shall anticipate criticism [•••] When reporting a sensitivity analysis, researchers should explain fully their specification search so that the
readers can judge for themselves how the results may have been affected. This is basically an 'honesty is the best policy' approach, advocated by Leamer, (1978[51]).
Sensitivity analysis can also be used in model-based policy assessment studies.[52] Sensitivity analysis can be used to assess the robustness of composite indicators,[53] also known as indices, such as the Environmental
Performance Index.
Sensitivity analysis is common in many areas of physics and chemistry.[54]
With the accumulation of knowledge about kinetic mechanisms under investigation and with the advance of power of modern computing technologies, detailed complex kinetic models are increasingly used as predictive
tools and as aids for understanding the underlying phenomena. A kinetic model is usually described by a set of differential equations representing the concentration-time relationship. Sensitivity analysis has been proven
to be a powerful tool to investigate a complex kinetic model.[55][56][57]
Kinetic parameters are frequently determined from experimental data via nonlinear estimation. Sensitivity analysis can be used for optimal experimental design, e.g. determining initial conditions, measurement positions,
and sampling time, to generate informative data which are critical to estimation accuracy. A great number of parameters in a complex model can be candidates for estimation but not all are estimable.[57] Sensitivity
analysis can be used to identify the influential parameters which can be determined from available data while screening out the unimportant ones. Sensitivity analysis can also be used to identify the redundant species and
reactions allowing model reduction.
Modern engineering design makes extensive use of computer models to test designs before they are manufactured. Sensitivity analysis allows designers to assess the effects and sources of uncertainties, in the interest of
building robust models.[58]
In meta-analysis
In a meta analysis, a sensitivity analysis tests if the results are sensitive to restrictions on the data included. Common examples are large trials only, higher quality trials only, and more recent trials only. If results are
consistent it provides stronger evidence of an effect and of generalizability.[59]
Multi-criteria decision making
Sometimes a sensitivity analysis may reveal surprising insights about the subject of interest. For instance, the field of multi-criteria decision making (MCDM) studies (among other topics) the problem of how to select the
best alternative among a number of competing alternatives. This is an important task in decision making. In such a setting each alternative is described in terms of a set of evaluative criteria. These criteria are associated
with weights of importance. Intuitively, one may think that the larger the weight for a criterion is, the more critical that criterion should be. However, this may not be the case. It is important to distinguish here the notion
of criticality with that of importance. By critical, we mean that a criterion with small change (as a percentage) in its weight, may cause a significant change of the final solution. It is possible criteria with rather small
weights of importance (i.e., ones that are not so important in that respect) to be much more critical in a given situation than ones with larger weights.[60][61] That is, a sensitivity analysis may shed light into issues not
anticipated at the beginning of a study. This, in turn, may dramatically improve the effectiveness of the initial study and assist in the successful implementation of the final solution.
Time-critical decision making
Producing time-critical accurate knowledge about the state of a system (effect) under computational and data acquisition (cause) constraints is a major challenge, especially if the knowledge required is critical to the
system operation where the safety of operators or integrity of costly equipment is at stake, e.g., during manufacturing or during environment substrate drilling. Understanding and interpreting, a chain of interrelated
events, predicted or unpredicted, that may or may not result in a specific state of the system, is the core challenge of this research. Sensitivity analysis may be used to identify which set of input data signals has a significant
impact on the set of system state information (i.e. output). Through a cause-effect analysis technique, sensitivity can be used to support the filtering of unsolicited data to reduce the communication and computational
8 of 13
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
capabilities of a standard supervisory control and data acquisition system.[12]
Model calibration and improvement
One application of sensitivity analysis addresses the question of "What's important to model or system development?" One can seek to identify important connections between observations, model inputs, and predictions
or forecasts. That is, one can seek to understand what observations (measurements of dependent variables) are most and least important to model inputs (parameters representing system characteristics or excitation),
what model inputs are most and least important to predictions or forecasts, and what observations are most and least important to the predictions and forecasts. Often the results are surprising, lead to finding problems in
the data or model development, and fixing the problems. This leads to better models.[5][6] In biomedical engineering, sensitivity analysis can be used to determine system dynamics in ODE-based kinetic models.
Parameters corresponding to stages of differentiation can be varied to determine which parameter is most influential on cell fate. Therefore, the most limiting step can be identified and the cell state for most advantageous
scale-up and expansion can be determined.[62] Additionally, complex networks in systems biology can be better understood through fitting mass-action kinetic models. Sensitivity analysis on rate coefficients can then be
conducted to determine optimal therapeutic targets within the system of interest.[63]
Sensitivity auditing
It may happen that a sensitivity analysis of a model-based study is meant to underpin an inference, and to certify its robustness, in a context where the inference feeds into a policy or decision making process. In these
cases the framing of the analysis itself, its institutional context, and the motivations of its author may become a matter of great importance, and a pure sensitivity analysis – with its emphasis on parametric uncertainty –
may be seen as insufficient. The emphasis on the framing may derive inter-alia from the relevance of the policy study to different constituencies that are characterized by different norms and values, and hence by a
different story about 'what the problem is' and foremost about 'who is telling the story'. Most often the framing includes more or less implicit assumptions, which could be political (e.g. which group needs to be protected)
all the way to technical (e.g. which variable can be treated as a constant).
In order to take these concerns into due consideration the instruments of SA have been extended to provide an assessment of the entire knowledge and model generating process. This approach has been called 'sensitivity
auditing'. It takes inspiration from NUSAP,[64] a method used to qualify the worth of quantitative information with the generation of `Pedigrees' of numbers. Likewise, sensitivity auditing has been developed to provide
pedigrees of models and model-based inferences.[65] Sensitivity auditing has been especially designed for an adversarial context, where not only the nature of the evidence, but also the degree of certainty and uncertainty
associated to the evidence, will be the subject of partisan interests.
Related concepts
Sensitivity analysis is closely related with uncertainty analysis; while the latter studies the overall uncertainty in the conclusions of the study, sensitivity analysis tries to identify what source of uncertainty weighs more on
the study's conclusions.
The problem setting in sensitivity analysis also has strong similarities with the field of design of experiments. In a design of experiments, one studies the effect of some process or intervention (the 'treatment') on some
objects (the 'experimental units'). In sensitivity analysis one looks at the effect of varying the inputs of a mathematical model on the output of the model itself. In both disciplines one strives to obtain information from the
system with a minimum of physical or numerical experiments.
See also
Elementary effects method
Experimental uncertainty analysis
Fourier amplitude sensitivity testing
Info-gap decision theory
Interval FEM
Perturbation analysis
Probabilistic design
Probability bounds analysis
ROC curve
Uncertainty quantification
Variance-based sensitivity analysis
Sensitivity analysis events
9 of 13
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
PAST EVENT Eighth International Conference on Sensitivity Analysis of Model Output - November 30th to December 3rd 2016, Reunion Island. (http://samo2016.univ-reunion.fr/) A Conference celebrating the 90th
birthday of Ilya_M._Sobol
PAST EVENT Numbers for policy: Practical problems in quantification, Castelldefels (Barcelona), August 26 - September 1, 2018. (https://www.uib.no/en/svt/115575/numbers-policy-practical-problems-quantification)
COMING EVENT Ninth International Conference on Sensitivity Analysis of Model Output, Barcelona, October 28-30 2019 (http://symposium.uoc.edu/23220/detail/ninth-international-conference-on-sensitivity-analysisof-model-output.html)
1. Saltelli, A. (2002). "Sensitivity Analysis for Importance Assessment". Risk Analysis. 22 (3): 1–12.
CiteSeerX (https://citeseerx.ist.psu.edu/viewdoc/summary?doi=
doi:10.1111/0272-4332.00040 (https://doi.org/10.1111%2F0272-4332.00040).
14. Paruolo, P.; Saisana, M.; Saltelli, A. (2013). "Ratings and Rankings: Voodoo or Science?". Journal of the
Royal Statistical Society, Series A. 176 (3): 609–634. arXiv:1104.3009 (https://arxiv.org/abs/1104.3009).
doi:10.1111/j.1467-985X.2012.01059.x (https://doi.org/10.1111%2Fj.1467-985X.2012.01059.x).
2. Saltelli, A.; Ratto, M.; Andres, T.; Campolongo, F.; Cariboni, J.; Gatelli, D.; Saisana, M.; Tarantola, S.
(2008). Global Sensitivity Analysis: The Primer. John Wiley & Sons.
15. O'Hagan, A.; et al. (2006). Uncertain Judgements: Eliciting Experts' Probabilities
(https://books.google.com/books?id=H9KswqPWIDQC). Chichester: Wiley. ISBN 9780470033302.
3. Pannell, D. J. (1997). "Sensitivity Analysis of Normative Economic Models: Theoretical Framework and
Practical Strategies". Agricultural Economics. 16 (2): 139–152. doi:10.1016/S0169-5150(96)01217-0
16. Sacks, J.; Welch, W. J.; Mitchell, T. J.; Wynn, H. P. (1989). "Design and Analysis of Computer
Experiments". Statistical Science. 4 (4): 409–435. doi:10.1214/ss/1177012413 (https://doi.org
4. Bahremand, A.; De Smedt, F. (2008). "Distributed Hydrological Modeling and Sensitivity Analysis in
Torysa Watershed, Slovakia". Water Resources Management. 22 (3): 293–408.
doi:10.1007/s11269-007-9168-x (https://doi.org/10.1007%2Fs11269-007-9168-x).
17. Campbell, J.; et al. (2008). "Photosynthetic Control of Atmospheric Carbonyl Sulfide During the Growing
Season" (http://www.escholarship.org/uc/item/82r9s2x3). Science. 322 (5904): 1085–1088.
Bibcode:2008Sci...322.1085C (http://adsabs.harvard.edu/abs/2008Sci...322.1085C).
doi:10.1126/science.1164015 (https://doi.org/10.1126%2Fscience.1164015). PMID 19008442
5. Hill, M.; Kavetski, D.; Clark, M.; Ye, M.; Arabi, M.; Lu, D.; Foglia, L.; Mehl, S. (2015). "Practical use of
computationally frugal model analysis methods". Groundwater. 54 (2): 159–170. doi:10.1111/gwat.12330
(https://doi.org/10.1111%2Fgwat.12330). PMID 25810333 (https://www.ncbi.nlm.nih.gov/pubmed
6. Hill, M.; Tiedeman, C. (2007). Effective Groundwater Model Calibration, with Analysis of Data,
Sensitivities, Predictions, and Uncertainty. John Wiley & Sons.
18. Bailis, R.; Ezzati, M.; Kammen, D. (2005). "Mortality and Greenhouse Gas Impacts of Biomass and
Petroleum Energy Futures in Africa". Science. 308 (5718): 98–103. Bibcode:2005Sci...308...98B
(http://adsabs.harvard.edu/abs/2005Sci...308...98B). doi:10.1126/science.1106881 (https://doi.org
/10.1126%2Fscience.1106881). PMID 15802601 (https://www.ncbi.nlm.nih.gov/pubmed/15802601).
7. Der Kiureghian, A.; Ditlevsen, O. (2009). "Aleatory or epistemic? Does it matter?" (http://orbit.dtu.dk
Structural Safety. 31 (2): 105–112. doi:10.1016/j.strusafe.2008.06.020 (https://doi.org
19. Murphy, J.; et al. (2004). "Quantification of modelling uncertainties in a large ensemble of climate
change simulations". Nature. 430 (7001): 768–772. Bibcode:2004Natur.430..768M
(http://adsabs.harvard.edu/abs/2004Natur.430..768M). doi:10.1038/nature02771 (https://doi.org
/10.1038%2Fnature02771). PMID 15306806 (https://www.ncbi.nlm.nih.gov/pubmed/15306806).
8. http://ec.europa.eu/governance/impact/commission_guidelines/docs/iag_2009_en.pdf
20. Czitrom (1999). "One-Factor-at-a-Time Versus Designed Experiments" (https://www.questia.com
/googleScholar.qst?docId=5001888588). American Statistician. 53 (2).
9. http://www.epa.gov/CREM/library/cred_guidance_0309.pdf
10. Helton, J. C.; Johnson, J. D.; Salaberry, C. J.; Storlie, C. B. (2006). "Survey of sampling based methods
for uncertainty and sensitivity analysis" (https://digital.library.unt.edu/ark:/67531/metadc891681/).
Reliability Engineering and System Safety. 91 (10–11): 1175–1209. doi:10.1016/j.ress.2005.11.017
11. Tavakoli, Siamak; Mousavi, Alireza (2013). "Event tracking for real-time unaware sensitivity analysis
(EventTracker)" (http://bura.brunel.ac.uk/handle/2438/8690). IEEE Transactions on Knowledge and Data
Engineering. 25 (2): 348–359. doi:10.1109/tkde.2011.240 (https://doi.org/10.1109%2Ftkde.2011.240).
12. Tavakoli, Siamak; Mousavi, Alireza; Poslad, Stefan (2013). "Input variable selection in time-critical
knowledge integration applications: A review, analysis, and recommendation paper"
(http://bura.brunel.ac.uk/handle/2438/8691). Advanced Engineering Informatics. 27 (4): 519–536.
doi:10.1016/j.aei.2013.06.002 (https://doi.org/10.1016%2Fj.aei.2013.06.002).
13. Saltelli, A.; Annoni, P. (2010). "How to avoid a perfunctory sensitivity analysis". Environmental Modeling
and Software. 25 (12): 1508–1517. doi:10.1016/j.envsoft.2010.04.012 (https://doi.org
10 of 13
21. Cacuci, Dan G. Sensitivity and Uncertainty Analysis: Theory. I. Chapman & Hall.
22. Cacuci, Dan G.; Ionescu-Bujor, Mihaela; Navon, Michael (2005). Sensitivity and Uncertainty Analysis:
Applications to Large-Scale Systems. II. Chapman & Hall.
23. Griewank, A. (2000). Evaluating Derivatives, Principles and Techniques of Algorithmic Differentiation.
24. Sobol', I. (1990). Sensitivity estimates for nonlinear mathematical models. Matematicheskoe
Modelirovanie 2, 112–118. in Russian, translated in English in Sobol' , I. (1993). Sensitivity analysis for
non-linear mathematical models. Mathematical Modeling & Computational Experiment (Engl. Transl.),
1993, 1, 407–414.
25. Homma, T.; Saltelli, A. (1996). "Importance measures in global sensitivity analysis of nonlinear models".
Reliability Engineering and System Safety. 52: 1–17. doi:10.1016/0951-8320(96)00002-6 (https://doi.org
26. Saltelli, A., K. Chan, and M. Scott (Eds.) (2000). Sensitivity Analysis. Wiley Series in Probability and
Statistics. New York: John Wiley and Sons.
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
27. Saltelli, A.; Tarantola, S. (2002). "On the relative importance of input factors in mathematical models:
safety assessment for nuclear waste disposal". Journal of the American Statistical Association. 97 (459):
702–709. doi:10.1198/016214502388618447 (https://doi.org/10.1198%2F016214502388618447).
39. Ratto, M.; Pagano, A. (2010). "Using recursive algorithms for the efficient identification of smoothing
spline ANOVA models". AStA Advances in Statistical Analysis. 94 (4): 367–388.
doi:10.1007/s10182-010-0148-8 (https://doi.org/10.1007%2Fs10182-010-0148-8).
28. Razavi, Saman; Gupta, Hoshin V. (1 January 2016). "A new framework for comprehensive, robust, and
efficient global sensitivity analysis: 1. Theory". Water Resources Research. 52 (1): 423–439.
Bibcode:2016WRR....52..423R (http://adsabs.harvard.edu/abs/2016WRR....52..423R).
doi:10.1002/2015WR017558 (https://doi.org/10.1002%2F2015WR017558). ISSN 1944-7973
40. Li, G.; Hu, J.; Wang, S.-W.; Georgopoulos, P.; Schoendorf, J.; Rabitz, H. (2006). "Random SamplingHigh Dimensional Model Representation (RS-HDMR) and orthogonality of its different order component
functions". Journal of Physical Chemistry A. 110 (7): 2474–2485. Bibcode:2006JPCA..110.2474L
(http://adsabs.harvard.edu/abs/2006JPCA..110.2474L). doi:10.1021/jp054148m (https://doi.org
/10.1021%2Fjp054148m). PMID 16480307 (https://www.ncbi.nlm.nih.gov/pubmed/16480307).
29. Razavi, Saman; Gupta, Hoshin V. (1 January 2016). "A new framework for comprehensive, robust, and
efficient global sensitivity analysis: 2. Application". Water Resources Research. 52 (1): 440–455.
Bibcode:2016WRR....52..440R (http://adsabs.harvard.edu/abs/2016WRR....52..440R).
doi:10.1002/2015WR017559 (https://doi.org/10.1002%2F2015WR017559). ISSN 1944-7973
41. Li, G., W. S. W., and R. H. (2002). Practical approaches to construct RS-HDMR component functions.
Journal of Physical Chemistry 106, 8721{8733.
30. Haghnegahdar, Amin; Razavi, Saman (1 September 2017). "Insights into sensitivity analysis of Earth
and environmental systems models: On the impact of parameter perturbation scale"
(https://www.sciencedirect.com/science/article/pii/S1364815216309070). Environmental Modelling &
Software. 95: 115–131. doi:10.1016/j.envsoft.2017.03.031 (https://doi.org
31. Gupta, H; Razavi, S (2016). "Challenges and Future Outlook of Sensitivity Analysis"
/978-0-12-803011-0). In Petropoulos, George; Srivastava, Prashant. Sensitivity Analysis in Earth
Observation Modelling (1st ed.). pp. 397–415. ISBN 9780128030318.
42. Rabitz, H (1989). "System analysis at molecular scale". Science. 246 (4927): 221–226.
Bibcode:1989Sci...246..221R (http://adsabs.harvard.edu/abs/1989Sci...246..221R).
doi:10.1126/science.246.4927.221 (https://doi.org/10.1126%2Fscience.246.4927.221). PMID 17839016
43. Hornberger, G.; Spear, R. (1981). "An approach to the preliminary analysis of environmental systems".
Journal of Environmental Management. 7: 7–18.
44. Saltelli, A.; Tarantola, S.; Campolongo, F.; Ratto, M. (2004). Sensitivity Analysis in Practice: A Guide to
Assessing Scientific Models. John Wiley and Sons.
45. Leamer, Edward E. (1983). "Let's Take the Con Out of Econometrics". American Economic Review. 73
(1): 31–43. JSTOR 1803924 (https://www.jstor.org/stable/1803924).
46. Leamer, Edward E. (1985). "Sensitivity Analyses Would Help". American Economic Review. 75 (3):
308–313. JSTOR 1814801 (https://www.jstor.org/stable/1814801).
32. Morris, M. D. (1991). "Factorial sampling plans for preliminary computational experiments".
Technometrics. 33 (2): 161–174. CiteSeerX (https://citeseerx.ist.psu.edu/viewdoc
/summary?doi= doi:10.2307/1269043 (https://doi.org/10.2307%2F1269043).
JSTOR 1269043 (https://www.jstor.org/stable/1269043).
48. Fassò, Alessandro (2006). "Sensitivity Analysis for Environmental Models and Monitoring Networks"
(http://www.iemss.org/iemss2006/papers/s7/268_Fasso_0.pdf) (PDF). Preprint.
33. Campolongo, F.; Cariboni, J.; Saltelli, A. (2007). "An effective screening design for sensitivity analysis of
large models". Environmental Modelling and Software. 22 (10): 1509–1518.
doi:10.1016/j.envsoft.2006.10.004 (https://doi.org/10.1016%2Fj.envsoft.2006.10.004).
49. Mroz, Thomas A. (1987). "The Sensitivity of an Empirical Model of Married Women's Hours of Work to
Economic and Statistical Assumptions". Econometrica. 55 (4): 765–799. doi:10.2307/1911029
(https://doi.org/10.2307%2F1911029). JSTOR 1911029 (https://www.jstor.org/stable/1911029).
34. Storlie, C.B.; Swiler, L.P.; Helton, J.C.; Sallaberry, C.J. (2009). "Implementation and evaluation of
nonparametric regression procedures for sensitivity analysis of computationally demanding models".
Reliability Engineering & System Safety. 94 (11): 1735–1763. doi:10.1016/j.ress.2009.05.007
50. Kennedy, P. (2007). A Guide to Econometrics (https://books.google.com/books?id=B8I5SP69e4kC)
(Fifth ed.). Blackwell. ISBN 9780262611831.
35. Oakley, J.; O'Hagan, A. (2004). "Probabilistic sensitivity analysis of complex models: a Bayesian
approach". J. Royal Stat. Soc. B. 66 (3): 751–769. CiteSeerX (https://citeseerx.ist.psu.edu
/viewdoc/summary?doi= doi:10.1111/j.1467-9868.2004.05304.x (https://doi.org
52. Saltelli, Andrea (2006) "The critique of modelling and sensitivity analysis in the scientific discourse: An
overview of good practices" (http://www.modeling.uga.edu/tauc/background_material/WashingtonMain.pdf) Archived (https://web.archive.org/web/20110720084018/http://www.modeling.uga.edu
/tauc/background_material/Washington-Main.pdf) 2011-07-20 at the Wayback Machine, Transatlantic
Uncertainty Colloquium (TAUC) Washington, October 10–11
36. Gramacy, R. B.; Taddy, M. A. (2010). "Categorical Inputs, Sensitivity Analysis, Optimization and
Importance Tempering with tgp Version 2, an R Package for Treed Gaussian Process Models"
(https://cran.r-project.org/web/packages/tgp/vignettes/tgp2.pdf) (PDF). Journal of Statistical Software. 33
(6). doi:10.18637/jss.v033.i06 (https://doi.org/10.18637%2Fjss.v033.i06).
37. Becker, W.; Worden, K.; Rowson, J. (2013). "Bayesian sensitivity analysis of bifurcating nonlinear
models". Mechanical Systems and Signal Processing. 34 (1–2): 57–75. Bibcode:2013MSSP...34...57B
(http://adsabs.harvard.edu/abs/2013MSSP...34...57B). doi:10.1016/j.ymssp.2012.05.010 (https://doi.org
47. Ravetz, J.R., 2007, No-Nonsense Guide to Science, New Internationalist Publications Ltd.
51. Leamer, E. (1978). Specification Searches: Ad Hoc Inferences with Nonexperimental Data. John Wiley &
Sons, Ltd, p. vi.
53. Saisana, M.; Saltelli, A.; Tarantola, S. (2005). "Uncertainty and Sensitivity analysis techniques as tools
for the quality assessment of composite indicators". Journal of the Royal Statistical Society, Series A.
168 (2): 307–323. doi:10.1111/j.1467-985x.2005.00350.x (https://doi.org
54. Saltelli, A.; Ratto, M.; Tarantola, S.; Campolongo, F. (2005). "Sensitivity Analysis for Chemical Models".
Chemical Reviews. 105 (7): 2811–2828. doi:10.1021/cr040659d (https://doi.org/10.1021%2Fcr040659d).
PMID 16011325 (https://www.ncbi.nlm.nih.gov/pubmed/16011325).
38. Sudret, B., (2008), Global sensitivity analysis using polynomial chaos expansions}, Reliability
Engineering & System Safety 93(7): 964-979,
11 of 13
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
55. Rabitz, H.; Kramer, M.; Dacol, D. (1983). "Sensitivity Analysis in Chemical Kinetics". Annual Review of
Physical Chemistry. 34: 419–461. Bibcode:1983ARPC...34..419R (http://adsabs.harvard.edu
/abs/1983ARPC...34..419R). doi:10.1146/annurev.pc.34.100183.002223 (https://doi.org
60. Triantaphyllou, E.; A. Sanchez (1997). "A Sensitivity Analysis Approach for Some Deterministic MultiCriteria Decision-Making Methods" (http://www.csc.lsu.edu/trianta/Journal_PAPERS1/SENSIT1.htm).
Decision Sciences. 28 (1): 151–194. doi:10.1111/j.1540-5915.1997.tb01306.x (https://doi.org
/10.1111%2Fj.1540-5915.1997.tb01306.x). Retrieved 2010-06-28.
56. Turanyi, T. (1990). "Sensitivity analysis of complex kinetic systems. Tools and applications". Journal of
Mathematical Chemistry. 5 (3): 203–248. doi:10.1007/BF01166355 (https://doi.org
61. Triantaphyllou, E. (2000). Multi-Criteria Decision Making: A Comparative Study (http://www.csc.lsu.edu
/trianta/Books/DecisionMaking1/Book1.htm). Dordrecht, The Netherlands: Kluwer Academic Publishers
(now Springer). p. 320. ISBN 978-0-7923-6607-2.
57. Komorowski, M.; Costa, M. J.; Rand, D. A.; Stumpf, M. P. H. (2011). "Sensitivity, robustness, and
identifiability in stochastic chemical kinetics models" (https://www.ncbi.nlm.nih.gov/pmc/articles
/PMC3102369). Proc Natl Acad Sci U S A. 108 (21): 8645–50. arXiv:1104.1274 (https://arxiv.org
/abs/1104.1274). Bibcode:2011PNAS..108.8645K (http://adsabs.harvard.edu
/abs/2011PNAS..108.8645K). doi:10.1073/pnas.1015814108 (https://doi.org
/10.1073%2Fpnas.1015814108). PMC 3102369 (https://www.ncbi.nlm.nih.gov/pmc/articles
/PMC3102369). PMID 21551095 (https://www.ncbi.nlm.nih.gov/pubmed/21551095).
62. Selekman JA, Das A, Grundl NJ, Palecek SP. Improving efficiency of human pluripotent stem cell
differentiation platforms using an integrated experimental and computational approach. Biotechnol
Bioeng. 2013;110(11):3024-37.
58. Becker, W.; Rowson, J.; Oakley, J.E.; Yoxall, A.; Manson, G.; Worden, K. (2011). "Bayesian sensitivity
analysis of a model of the aortic valve". Journal of Biomechanics. 44 (8): 1499–506.
doi:10.1016/j.jbiomech.2011.03.008 (https://doi.org/10.1016%2Fj.jbiomech.2011.03.008).
PMID 21481873 (https://www.ncbi.nlm.nih.gov/pubmed/21481873).
59. clinicalevidence.bmj.com > Glossary > sensitivity analysis (http://clinicalevidence.bmj.com/ceweb
/resources/glossary.jsp) Retrieved on June 21, 2010
63. Tian D, Solodin NM, Rajbhandari P, Bjorklund K, Alarid ET, Kreeger PK. A kinetic model identifies
phosphorylated estrogen receptor-α (ERα) as a critical regulator of ERα dynamics in breast cancer.
FASEB J. 2015;29(5):2022-31.
64. Van der Sluijs, JP; Craye, M; Funtowicz, S; Kloprogge, P; Ravetz, J; Risbey, J (2005). "Combining
quantitative and qualitative measures of uncertainty in model based environmental assessment: the
NUSAP system". Risk Analysis. 25 (2): 481–492. doi:10.1111/j.1539-6924.2005.00604.x (https://doi.org
/10.1111%2Fj.1539-6924.2005.00604.x). PMID 15876219 (https://www.ncbi.nlm.nih.gov/pubmed
65. Saltelli, A., van der Sluijs, J., Guimarães Pereira, Â., 2013, Funtowiz, S.O., What do I make of your
Latinorum? Sensitivity auditing of mathematical modelling, International Journal Foresight and
Innovation Policy, 9 (2/3/4), 213–234.
Further reading
Cannavó, F. (2012). "Sensitivity analysis for volcanic source modeling quality assessment and model selection". Computers & Geosciences. 44: 52–59. Bibcode:2012CG.....44...52C (http://adsabs.harvard.edu
/abs/2012CG.....44...52C). doi:10.1016/j.cageo.2012.03.008 (https://doi.org/10.1016%2Fj.cageo.2012.03.008).
Fassò A. (2007) "Statistical sensitivity analysis and water quality". In Wymer L. Ed, Statistical Framework for Water Quality Criteria and Monitoring. Wiley, New York.
Fassò A., Perri P.F. (2002) "Sensitivity Analysis". In Abdel H. El-Shaarawi and Walter W. Piegorsch (eds) Encyclopedia of Environmetrics, Volume 4, pp 1968–1982, Wiley.
Fassò A., Esposito E., Porcu E., Reverberi A.P., Vegliò F. (2003) "Statistical Sensitivity Analysis of Packed Column Reactors for Contaminated Wastewater". Environmetrics. Vol. 14, n.8, 743–759.
Haug, Edward J.; Choi, Kyung K.; Komkov, Vadim (1986) Design sensitivity analysis of structural systems. Mathematics in Science and Engineering, 177. Academic Press, Inc., Orlando, FL.
Pianosi, F.; Beven, K.; Freer, J.; Hall, J.W.; Rougier, J.; Stephenson, D.B.; Wagener, T. (2016). "Sensitivity analysis of environmental models: A systematic review with practical workflow". Environmental Modeling and
Software. 79: 214–232. doi:10.1016/j.envsoft.2016.02.008 (https://doi.org/10.1016%2Fj.envsoft.2016.02.008).
Pilkey, O. H. and L. Pilkey-Jarvis (2007), Useless Arithmetic. Why Environmental Scientists Can't Predict the Future. New York: Columbia University Press.
Santner, T. J.; Williams, B. J.; Notz, W.I. (2003) Design and Analysis of Computer Experiments; Springer-Verlag.
Taleb, N. N., (2007) The Black Swan: The Impact of the Highly Improbable, Random House.
External links
Joseph Hart, Julie Bessac, Emil Constantinescu (2018), "Global sensitivity analysis for statistical model parameters", arXiv:1708.07441
International Journal of Chemical Kinetics – September 2008 (http://www3.interscience.wiley.com/journal/121401979/issue) – Special Issue on Sensitivity Analysis
Reliability Engineering and System Safety (Volume 91, 2006) (http://www.elsevier.com/wps/find/journaldescription.cws_home/405908/description#description) – special issue on sensitivity analysis
web-page on Sensitivity analysis (https://ec.europa.eu/jrc/en/samo) – (Joint Research Centre of the European Commission)
SimLab (https://ec.europa.eu/jrc/en/samo/simlab), the free software for global sensitivity analysis of the Joint Research Centre
Sensitivity Analysis Excel Add-In (http://www.life-cycle-costing.de/sensitivity_analysis/) is a free (for private and commercial use) Excel Add-In that allows for simple sample based sensitivity analysis runs
MUCM Project (http://mucm.ac.uk/index.html) – Extensive resources for uncertainty and sensitivity analysis of computationally-demanding models.
GEM-SA (http://www.tonyohagan.co.uk/academic/GEM/index.html) – a program for performing sensitivity analysis with Gaussian processes.
SALib (https://salib.github.io/SALib/) Sensitivity Analysis Library in Python (Numpy). Contains Sobol, Morris, Fractional Factorial and FAST methods.
12 of 13
3/17/2019, 2:36 PM
Sensitivity analysis - Wikipedia
Retrieved from "https://en.wikipedia.org/w/index.php?title=Sensitivity_analysis&oldid=887843474"
This page was last edited on 15 March 2019, at 05:15 (UTC).
Text is available under the Creative Commons Attribution-ShareAlike License; additional terms may apply. By using this site, you agree to the Terms of Use and Privacy Policy. Wikipedia® is a registered trademark of the
Wikimedia Foundation, Inc., a non-profit organization.
13 of 13
3/17/2019, 2:36 PM
Random flashcards

0 Cards oauth2_google_0810629b-edb6-401f-b28c-674c45d34d87

Rekening Agen Resmi De Nature Indonesia

9 Cards denaturerumahsehat

Nomor Rekening Asli Agen De Nature Indonesia

2 Cards denaturerumahsehat


2 Cards oauth2_google_3524bbcd-25bd-4334-b775-0f11ad568091

English Training Melbourne

2 Cards Einstein College of Australia

Create flashcards