Muito Bom Tem A Tebela Da Compração de Métodos de Vv.
Muito Bom Tem A Tebela Da Compração de Métodos de Vv.
Muito Bom Tem A Tebela Da Compração de Métodos de Vv.
SAND2016-5737C
2
Some thoughts on quality of models
and assessments
Do we have the courage to make sure that Sandia’s modeling
capabilities are not oversold?
Expert judgment is not sufficient. Experts can be wrong and
have their own biases.
Just because someone pays you to do something doesn’t
mean it is valid.
My job today is to provide some background on validation as
it is performed in DOE’s Advanced Simulation and Computing
program and outline some problems with using this
framework for complex systems models.
I will emphasize reliance on data and benchmarks, performing
uncertainty analyses and examining worst-case scenarios, and
addressing the risk of using a model for a particular situation. 3
Do you know how good (or bad) your
modeling and simulation is?
“Perform due diligence and communicate frankly about
assumptions, approximations, and limitations affecting
simulation credibility.
Nothing against
ostriches, of course
4
From the world of physics and
engineering….
5
Prediction is hard:
Limited physical data (observational or experimental)
Limited simulations (high computational demands…)
Imperfect computational models (missing physics, etc.)
Under-resolved approximations or numerics
Unknown model parameters and boundary conditions
Imperfect humans
We want to extrapolate to conditions beyond validation
regime…
Model Framework
7
Verification: Are equations solved
correctly?
Verification: a math and computer science issue
Software quality is the bare minimum
Software verification tools/concepts:
Requirements and software change and revision control
Unit and regression tests covering intended use and features, core and edge cases, not
just line coverage (gcov/lcov)
Static and dynamic analysis (memory, performance, red flags)
Software quality engineering processes, design and code reviews
Build in from the start!
• Numerical analysis tells us what properties an algorithm possesses —
symmetry, stability, conservation, convergence, etc. — and under
what conditions.
• In code verification, we test whether the implementation of the algorithm
exhibits these properties with respect to known, exact solutions.
8
Best Practices: Verification Throughout
System Design Physics
Meshing Discretization
Linear solve
“If the test data are shown in blue and the simulation data are
shown in , then all I want to see is green.”
(heard at Los Alamos, October 2005)
10
Is this model (black line) valid?
h Exp. Data
Best calculation
Validation with experimental error,
numerical and parametric uncertainty
Fine
Medium
Coarse
Extrapolated
Exp. Data
h
The Validation Hierarchy
A key construct
Example Validation Projects:
Model to Use Case
Validation of the radiative transfer equation (RTE) used in
Sierra/Thermal-Fluid Dynamics calculations of pool-fire soot
radiation.
Assess the validity of key model assumptions, which include: 1) gray
soot radiation and 2) the size of the soot particles is much less than
the wavelength of the radiation so that scattering is negligible.
Validation experiments performed with lightly-sooting liquid
hydrocarbon fuels that yielded fully turbulent fires 2 m diameter.
Radiation intensities were measured, with supplemental
measurements of air flow and temperature, fuel temperature and
burn rate, and flame surface emissive power, wall heat, and flame
height and width provide a complete set of boundary condition data.
model Model error due to modeling assumptions and model = E – (num + input -
approximations D).
16
Validation Metrics:
We have an interval (E uval) in which model falls:
model [E- uval , E+ uval].
17
Performing validation
Are we ready to estimate model [E- uval , E+ uval] ?
What do we need? An estimate of numerical, parameter, and
experimental uncertainties
2 2
𝑢𝑣𝑎𝑙 = 𝑢𝑛𝑢𝑚 + 𝑢𝑖𝑛𝑝𝑢𝑡 + 𝑢𝐷2
19
Supporting capabilities
Sensitivity Analysis
Identify most important variables and their interactions
Understand code output variations as input factors vary
Often correlation coefficients, scatterplots, or variance-based indices
Uncertainty Quantification
Determine the probability distribution of code outputs, given
uncertainty in input factors
Assess the likelihood of typical or extreme outputs given input
uncertainties: determine mean or median performance, assess
variability in model responses, find probability of failure
Assess how close code predictions are to experimental data (validation)
or performance limits (margins)
Calibration
determine optimal parameter values that yield simulation results which
“best match” the experimental data in some sense
Least-squares methods, Bayesian calibration methods 20
More on UQ/SA
Advances in UQ methods have been tremendous BUT still
severely limited with respect to dimensionality
Most of our UQ/SA methods deal with continuous variables,
not large numbers of discrete entities
Sampling approaches are probably the only mainstream
viable approach currently
Surrogate models have also advanced tremendously the past
15 years, but again, limited to a handful of variables and
responses, mainly continuous
21
Challenge:
general coupled multi-physics
Can we efficiently propagate UQ across scales/disciplines?
Naively wrapping multi-physics with UQ often too costly
Can we invert loops and perform multi-physics analysis on UQ-
enriched simulations (couple individual sampling loops based on
scalar statistics, random fields, stochastic processes)?
Embedded methods: examining GPUs, re-structuring of sampling
loops (Eric Phipps, Sandia)
Instead of N separate samples where each requires a full solve of
residuals/Jacobians, restructure the operations
Form a large block nonlinear problem
Krylov Basis Recycling
Increased vectorization and data locality
Implemented through Stokhos embedded UQ package
22
Lessons from the ASC Program
DATA: Can I design and execute a validation experiment?
What benchmarks are available? Difference between data
used for calibrating the model vs. validation.
Early emphasis on identifying and quantifying epistemic vs.
aleatory uncertainty
More recently, issues of model form uncertainty and model
selection are of interest
Bayesian methods are becoming more popular
Limits of the validation hierarchy
IT IS VERY EXPENSIVE!!!!
Our validation statements are fairly limited
Validation is a process
23
Frontiers of V&V/UQ
Complexity of codes, multi-physics couplings
Few examples of validation across the hierarchy
UQ: Focus is on methods that minimize the number of function
evaluations while maximizing accuracy in response statistics
Calibration: People are tacking more complicated scenarios,
want parameters estimated with uncertainties from a variety of
experimental configurations
People are doing experiment to simulation comparisons, but
characterizing, propagating, and comparing uncertainties
remains a big issue
DATA, DATA, DATA
How to deal with models that are “theory rich and data poor”
vs. models that are “data rich and theory poor”?
24
Recommendations
Focus on Data: what data was used to build the model, what
data was used to calibrate the model, what data was used to
evaluate the model, what data was used for validation? What
benchmark data sets are available?
Use of SA/UQ/Optimization in terms of understanding the
behavior of the model, identifying worst case scenarios,
understanding the spread of possible outcomes and their
likelihood.
Perform risk management on the use of the model: what are
the risks associated with using this model?
25
Thoughts to consider
Hans Meir’s thoughts on validation of complex systems
models:
Naïve practitioners in the softer sciences are pretending to “validate”
their tools based on harder-science verification practices.
We should be enforcing a course-correction that steers the managers
and naïve practitioners away from that pseudo-science.
Instead, we should recognize the deep uncertainty inherent in softer
science challenges, stop pretending to gloss over that uncertainty with
sophisticated modeling and validation tricks, and prevent the
inappropriate use of models for predictive/forecasting purposes.
26
Thoughts to consider
Models don’t forecast….People Do. (Laura McNamara, SAND,
NAS)
29