English Français

Documentation / Guide méthodologique : PDF version

Methodological reference guide for Uranie v4.9.0

Methodological reference guide for Uranie v4.9.0

The Uranie team

CEA DES

Abstract

This documentation is introducing the theoretical basics upon which the Uranie platform (based on Uranie v4.9.0), has been developed at CEA/DES. Since this platform is designed for uncertainty propagation, sensitivity analysis and surrogate model generation, the main methods which have been implemented are introduced and discussed. This document is however not made to give a complete overview of the methodology but more to open the scope of the reader by largely relying on a list of references, without getting to attached on the structure of the Uranie platform itself.


Table of Contents

I. Glossary
II. Basic statistical elements
II.1. Random variable modelisation
II.1.1. The probability distributions
II.2. Statistical treatments and operations
II.2.1. Normalising the variable
II.2.2. Computing the ranking
II.2.3. Computing the elementary statistic
II.2.4. The quantile computation
II.2.5. Correlation matrix
II.3. Combining these aspects: performing PCA
II.3.1. Theoretical introduction
III. The Sampler module
III.1. Introduction
III.2. The Stochastic methods
III.2.1. Introduction
III.2.2. Correlating samples drawn from different marginals
III.2.3. The maximin LHS
III.2.4. The constrained LHS
III.3. QMC method
IV. Generating surrogate models
IV.1. Introduction
IV.1.1. Quality criteria definition
IV.1.2. Adapting the fitting strategy
IV.2. The linear regression
IV.3. Chaos polynomial expansion
IV.3.1. Introduction
IV.3.2. Nisp in a nutshell
IV.4. The artificial neural network
IV.4.1. Introduction to the formal neuron
IV.4.2. The working principle
IV.5. The kriging method
IV.5.1. Theoretical introduction
IV.5.2. Running a kriging
V. Sensitivity analysis
V.1. Brief reminder of theoretical aspects
V.1.1. Theoretical aspects
V.1.2. List of available methods
V.2. The finite differences method
V.2.1. General presentation of finite difference sensitivity indices
V.3. The regression method
V.3.1. General presentation of regression's coefficients
V.3.2. Getting a confidence-interval estimation
V.4. The Morris screening method
V.4.1. Principle of the Morris' method
V.5. The Sobol method
V.5.1. Sobol's sensitivity indices
V.6. Fourier-based methods
V.6.1. Introducing the method
V.6.2. Implementation of methods
V.7. The Johnson relative weight
V.7.1. Introducing the method
V.8. Sensitivity Indices based on HSIC
V.8.1. Introducing the method
VI. Dealing with optimisation issues
VI.1. Introduction
VI.1.1. Single criterion case
VI.1.2. The pareto concept in a nutshell
VI.2. Multicriteria optimisation
VI.2.1. Hitchhiker's guide to genetic algorithms
VI.2.2. General discussion on multi and many criteria problem.
VII. The Calibration module
VII.1. Brief reminder of theoretical aspects
VII.1.1. The distance used to compare observations and model predictions
VII.1.2. Discussing assumptions and theoretical background
VII.2. Using minimisation techniques
VII.3. Analytical linear Bayesian estimation
VII.3.1. Prediction values
VII.4. The Approximation Bayesian Computation techniques (ABC)
VII.4.1. Rejection ABC algorithm
VII.5. The Markov-chain approach
VII.5.1. Markov-chain principle
VII.5.2. The Metropolis-Hasting algorithm
VIII. The Uncertainty modeler module
VIII.1. Introduction
VIII.2. Tests based on the Empirical Distribution Function ("EDF tests")
VIII.3. The Circe method
VIII.3.1. Main principle of the CIRCE method
References

List of Figures

II.1. Principle of the truncated PDF generation (right-hand side) from the orginal one (left-hand side).
II.2. Example of PDF, CDF and inverse CDF for Uniform distribution.
II.3. Example of PDF, CDF and inverse CDF for LogUniform distributions.
II.4. Example of PDF, CDF and inverse CDF for Triangular distributions.
II.5. Example of PDF, CDF and inverse CDF for Logtriangular distributions.
II.6. Example of PDF, CDF and inverse CDF for Normal distributions.
II.7. Example of PDF, CDF and inverse CDF for LogNormal distributions.
II.8. Example of PDF, CDF and inverse CDF for Trapezium distributions.
II.9. Example of PDF, CDF and inverse CDF for UniformByParts distributions.
II.10. Example of PDF, CDF and inverse CDF for Exponential distributions.
II.11. Example of PDF, CDF and inverse CDF for Cauchy distributions.
II.12. Example of PDF, CDF and inverse CDF for GumbelMax distributions.
II.13. Example of PDF, CDF and inverse CDF for Weibull distributions.
II.14. Example of PDF, CDF and inverse CDF for Beta distributions.
II.15. Example of PDF, CDF and inverse CDF for GenPareto distributions.
II.16. Example of PDF, CDF and inverse CDF for Gamma distributions.
II.17. Example of PDF, CDF and inverse CDF for InvGamma distributions.
II.18. Example of PDF, CDF and inverse CDF for Student distribution.
II.19. Example of PDF, CDF and inverse CDF for generalized normal distributions.
II.20. Example of PDF, CDF and inverse CDF for a composed distribution made out of three normal distributions with respective weights.
II.21. Illustration of the results of 100000 quantile determinations, applied to a reduced centered gaussian distribution, comparing the usual and Wilks methods. The number of points in the reduced centered gaussian distribution is varied, as well as the confidence level.
III.1. Schematic view of the input/output relation through a code
III.2. Comparison of the two sampling methods SRS (left) and LHS (right) with samples of size 8.
III.3. Comparison of deterministic design-of-experiments obtained using either SRS (left) or LHS (right) algorithm, when having two independent random variables (uniform and normal one)
III.4. Transformation of a classical LHS (left) to its corresponding maximin LHS (right) when considering a problem with two uniform distributions between 0 and 1.
III.5. Matrix of distribution of three uniformly distributed variables on which three linear constraints are applied. The diagonal are the marginal distributions while the off-diagonal are the two-by-two scatter plots.
III.6. Comparison of both quasi Monte-Carlo sequences with both LHS and SRS sampling when dealing with two uniform variables.
III.7. Comparison of design-of-experiments made with Petras algorithm, using different level values, when dealing with two uniform variables.
IV.1. Sketch of the evolution of the bias, the variance and their sum, as a function of the complexity of the model.
IV.2. Sketches of under-trained (left), over-trained (middle) and properly trained (right) surrogate models, given that the black points show the training database, while the yellow ones show the testing database
IV.3. Evolution of the different kinds of error used to determine when does one start to over-train a model
IV.4. Schematical view of the projection of the original value from the code onto the subspace spanned by the column of H (in blue).
IV.5. Schematic view of the Nisp methodology
IV.6. Schematic description of a formal neuron, as seen in McCulloch and Pitts [McCulloch1943].
IV.7. Example of transfer functions: the hyperbolic tangent (left) and the logistical one (right)
IV.8. Schematic description of the working flow of an artificial neural network as used in Uranie
IV.9. Influence of the variance parameter in the Matern function once fix at 0.5, 1 and 2 (from left to right). The correlation length is set to 1 while the smoothness is set to 3/2.
IV.10. Influence of the correlation length parameter in the Matern function once fix at 0.5, 1 and 2 (from left to right). The variance is set to 1 while the smoothness is set to 3/2.
IV.11. Influence of the smoothness parameter in the Matern function once fix at 0.5, 1.5 and 2.5 (from left to right). Both the variance and the correlation length are set to 1.
IV.12. Evolution of the different covariance functions implemented in Uranie.
IV.13. Example of kriging method applied on a simple uni-dimensional function, with a training site of six points, and tested on a basis of about hundred points, with either a gaussian correlation function (left) or a matern3/2 one (right).
IV.14. Schematic description of the kriging procedure as done within Uranie
V.1. Schematic view of two trajectories drawn randomly in the discretised hyper-volume (with p=6) for two different values of the elementary variation (the optimal one in black and the smallest one in pink, as detailed on the figure itself).
VI.1. Naive example of an imaginary optimisation case relying on two objectives that only depend on a single input variable.
VI.2. Description of the children production process in the Uranie implementation of the genetic algorithm
VI.3. Comparison of two Pareto sets (left) and fronts (right) from vizir (blue) and MOEAD (ref) when the hollow bar case is studied with very low number of points, i.e. about 20 (simulating higher dimensions).
/language/en