Managing uncertainty in computer models: aircraft engine design and food safety risk assessment
Submitting Institution
University of SheffieldUnit of Assessment
Mathematical SciencesSummary Impact Type
TechnologicalResearch Subject Area(s)
Mathematical Sciences: Statistics
Information and Computing Sciences: Artificial Intelligence and Image Processing
Economics: Econometrics
Summary of the impact
Pratt & Whitney (one of the world's largest makers of aircraft
engines) has developed a process, "Design for Variation" (DFV), that uses
Bayesian methods developed at Sheffield for analysing uncertainty in
computer model predictions within the design, manufacture and service of
aircraft engines. The DFV process significantly improves cost efficiency
by increasing the time an engine stays operational on the wing of an
aircraft, so reducing the time that the aircraft is unavailable due to
engine maintenance. DFV also saves costs by identifying design and process
features that have little impact on engine performance, but are expensive
to maintain. Pratt & Whitney estimate the DFV process to generate
savings, for a large fleet of military aircraft, of [text removed for
publication].
The UK Food and Environment Research Agency (Fera) has used these methods
in their risk analyses, for example in assessing risks of exposure to
pesticides.
Underpinning research
The research is concerned with statistical methods for handling
uncertainty in computer models. By "computer model", we mean a
deterministic mathematical model of a physical system, implemented on a
computer. Uncertainty can arise from not knowing the true values for the
model inputs, and an imperfect model structure; we may not understand the
underlying physics perfectly, and/or it may not be feasible to implement
our best description of the physical process on a computer. Such models
can take a long time to run. Reinman et al. [S1] describe
computationally expensive models used at Pratt & Whitney: finite
element models, as used for heat transfer and mechanical stress modelling,
can take hours to run at one choice of input values, and computational
fluid dynamics models based on Navier-Stokes equations can take days to
run just once.
Kennedy and O'Hagan [R1] is concerned with calibrating computer
models to data. They consider a computer model with two types of inputs:
uncertain, fixed "calibration" inputs, and variable, known "control"
inputs. Physical experiments are conducted at different values of the
control inputs, and the aim is to find values of the calibration inputs so
that the computer model outputs match the physical data as closely as
possible. An important development is the inclusion of a "model
discrepancy" function: a mechanism for learning the error in the model
structure and correcting the model prediction at new input settings. The
authors use a Bayesian framework, and show how to quantify all sources of
uncertainty when predicting with the model.
This research was started at the University of Nottingham by O'Hagan as
Principal Investigator and Kennedy as Research Associate, funded by an
EPSRC grant with support from the National Radiological Protection Board [G1].
O'Hagan and Kennedy moved to the University of Sheffield in January 1999,
where they continued and developed their research, leading to their 2001
publication.
Oakley and O'Hagan [R2] is concerned with identifying the most
influential inputs in a computer model. They consider a computer model
with multiple inputs, the true values of which are uncertain. The aim is
to quantify how each uncertain input contributes to the uncertainty in the
model output. A variance-based approach was taken, which quantifies how
much the output variance can be reduced by learning the true value of an
input.
The key development is to use a Gaussian process emulator to speed up
computation of the variance-based measures of input influence. The
emulator is a statistical approximation of the computer model, which is
constructed from a relatively small number of computer model runs, and can
then be used as a fast surrogate. Previous computational methods required
large numbers of runs of the computer model, and were infeasible for
computationally expensive models. (Faster computers have not solved this
problem. As computing power increases, model users may choose to run their
models at higher resolution, improving accuracy, but requiring more
computational effort).
This research was conducted at the University of Sheffield, funded by an
EPSRC grant with O'Hagan as Principal Investigator and Oakley as Research
Associate (now Lecturer at the University of Sheffield) [G2].
References to the research
Papers:
R1 Kennedy, M.C., O'Hagan, A. (2001). Bayesian calibration of
computer models (with discussion). Journal of the Royal Statistical
Society, Series B 63, 425-64. doi: 10.1111/1467-9868.00294
R2 Oakley, J.E., O'Hagan, A. (2004). Probabilistic sensitivity
analysis of complex models: a Bayesian approach. Journal of the Royal
Statistical Society, Series B 66, 751-69. doi: 10.1111/j.1467-9868.2004.05304.x
Grants:
G1 Engineering and Physical Sciences Research Council, £105,685
(1995-98) Bayesian uncertainty analysis and computer model inadequacy,
with support from the National Radiological Protection Board. PI: Anthony
O'Hagan
G2 Engineering and Physical Sciences Research Council, £86,940
(2000-02). Realising Our Potential Award: Bayesian elicitation of expert
opinion. PI: Anthony O'Hagan.
Details of the impact
The Sheffield research has changed the way Pratt & Whitney designs
and manufactures aircraft engines, and the way the Food and Environment
Research Agency (Fera), part of the Department for Environment, Food and
Rural Affairs (Defra) conducts risk assessments.
Commercial impact
The Design for Variation initiative was led by Grant Reinman, a
statistician at Pratt & Whitney. Pratt & Whitney learned of
Kennedy and O'Hagan's work from a literature search. A key dissemination
route for Oakley & O'Hagan (2004) was the software package GEM-SA,
which implements the methodology in this paper, written by Marc Kennedy
during his time at Sheffield. It was made available for free download, and
Pratt & Whitney have used it in their design processes (though they
have now built on it to develop their own software). Kennedy is now a risk
analyst at Fera, and so disseminated the research within Fera directly.
Pratt & Whitney's Design for Variation process has five steps: (i)
define probabilistic design criteria; (ii) use computer models and
physical experiments to identify causes of performance variation and
uncertainty; (iii) find the optimum design to satisfy the design criteria;
continue data collection to (iv) validate the models; and (v) ensure the
models remain consistent with the real world. Methods in Oakley &
O'Hagan (2004) and Kennedy & O'Hagan (2001) play an essential role in
steps (ii) and (iii), and hence have contributed to what Al Brockett, a
former vice president of engineering module centres at Pratt &
Whitney, describes as a "paradigm shift" and a "high-visibility
strategic priority" in the way they design and manufacture aircraft
engines [S1].
An illustration is given in Reinman et al. [S2]. In the design of
a jet engine turbine airfoil, a computer model predicted the life
expectancy of the airfoil, given its design. There was variability in
airfoil life expectancy due to part-to-part variation, engine-to-engine
variation, and environmental variation, and the designers wanted to know
how to reduce variation in life expectancy. A variance-based sensitivity
analysis was used: the analysis told them how much of the output variance
was caused by each source of input variation. As the model was
computationally expensive, the analysis could not have been done without
Oakley & O'Hagan (2004). The designers used the results to assess the
most cost-effective way of reducing variability in life expectancy, by
targeting the most important sources of input variation (and not wasting
resources by reducing unimportant input variation).
Pratt & Whitney calibrate their computer models to data using Kennedy
& O'Hagan (2001). This method allows them to account for all sources
of uncertainty in their model predictions — in particular, uncertainty due
to a model not representing reality perfectly. Reinman et al. explain the
benefits: "Significant insight can be gained from the calibration
results. In a recent study, assumptions typically made about boundary
conditions near the airfoil surfaces were found to be over 20% higher
than what the calibration process revealed them to be. Part temperatures
were being over-estimated, and correspondingly airfoil life was being
under-estimated" [S1].
To quantify the financial benefits of DFV, Pratt & Whitney did a
Business Case Study to assess the value of quantifying and managing
uncertainty over the entire life cycle of an engine (from design through
to service), using sensitivity analysis and calibration methods within
their DFV process. The published saving in sustainment costs from doing
this, for a large fleet of military aircraft, was approximately [text
removed for publication] [S3]. The company also estimates that its
component-level DFV initiatives "have yielded a 64% to 88% return on
investment by reducing design iterations, improving manufacturability,
increasing reliability, improving on-time deliveries, and providing
other performance benefits" [S1].
Change to professional practice in environmental management
The Sheffield research has also changed the way the Food and
Environmental Research Agency (Fera) conduct probabilistic risk
assessments. Kennedy et al. [S4] report an analysis funded by the
UK Health and Safety Executive's Chemicals Regulation Directorate (Defra
project no. PS2005), investigating risks of exposure to pesticide from the
spray drift of an agricultural boom sprayer. A computer model predicted
the level of exposure to bystanders and residents after a crop-spraying
event. The model had uncertain and variable inputs, such as the height of
the boom, distance of a bystander from the source, wind speed, etc. Using
the sensitivity analysis method of Oakley & O'Hagan (2004),
implemented in GEM-SA, they quantified the contribution of each
uncertain/variable input to the output uncertainty, to give risk managers
information on how best to manage risks by reducing output uncertainty.
Due to the computational expense of the model, this would not have been
feasible without Oakley & O'Hagan (2004). The analysis in this case
suggested reducing boom height and variation in boom height has the
potential to reduce exposure.
Other ongoing projects at Fera are using GEM-SA for contaminated land and
assessing the impact of recycling pesticide containers: Defra research
project PS1010 — Development of Category 4 Screening Levels for
Assessment of Land Affected by Contamination; and Defra research project
PS2808 — Recycling of Home and Garden Pesticide Containers.
Sources to corroborate the impact
S1 ANSYS (2013). ANSYS Advantage, 7, 2, p. 18. Available at http://tinyurl.com/q63h646
S2 Reinman, G. et al. (2012). Design for variation. Quality
Engineering, 24: 317-45 doi: 10.1080/08982112.2012.651973
S3 Statistician,Pratt & Whitney, letter on file corroborating
savings in sustainment costs.
S4 Kennedy M.C., Butler Ellis, M.C., Miller, P.C.H. (2012). BREAM:
A probabilistic bystander and resident exposure assessment model of spray
drift from an agricultural boom sprayer. Computers and Electronics in
Agriculture, 88: 63-71.