1. Introduction
Global sensitivity analysis (GSA) focuses on attributing the uncertainties of model outputs, or related performance indicators, to their inputs, thereby assessing the impact of input uncertainties on outputs or performance indicators [
1,
2]. Various GSA methods have been developed for this purpose [
3,
4]. These methods include the screening method [
5,
6], variance-based methods [
7,
8], moment-independent methods [
9,
10], and derivative-based methods [
11,
12]. Among these, variance-based sensitivity indices, also known as Sobol’ indices [
7,
8], are particularly notable for their mathematical elegance in measuring the individual, interaction, and total contributions of each input to the model output uncertainty, see, e.g. [
13,
14].
In the limit state method, probabilistic reliability analysis is based on the estimation of the failure probability [
15]. Sobol' indices have been widely used in structural reliability analysis to pinpoint variables that significantly influence failure probability [
16,
17]. These sensitivity indices, which focus on failure probability
Pf, are derived from the variance decomposition of a binary function representing failure and success [
16,
17]. Fort et al. [
18] expanded on Sobol's sensitivity indices by introducing a contrast function in place of variance, allowing the indices to be oriented towards variance, probability, and quantile. This approach maintains the non-negative property of the indices and ensures that their sum equals one, as in traditional Sobol sensitivity analysis. Extending GSA to
Pf and design quantile represented a significant advancement in civil engineering, as these quantities are crucial in structural reliability assessments [
19,
20].
The development of GSA methods focused on reliability demands precise estimation of
Pf [
21]. However, the complexity of mathematical models and the difficulty of uncertainty propagation using sampling-based methods such as Monte Carlo (MC) simulations, due to the large number of runs required, present significant challenges [
22]. Nonlinear finite element models are particularly demanding on CPU time [
23], which further complicates the process of structural reliability estimation, see, e.g., [
24,
25]. To ensure the most accurate estimate of
Pf, it is essential to develop methods that provide precise
Pf estimates while minimizing the computational costs associated with repeated calls to the computational model [
26,
27].
The computational burden can be reduced by using metamodels, also known as surrogate models, which approximate the behavior of complex models with simpler ones [
28,
29]. Methods such as polynomial response surface [
30,
31], response surface based multi-fidelity model [
32], polynomial chaos expansion (PCE) [
33,
34], Gaussian process [
35,
36], Kriging [
37,
38], and neural network [
39,
40,
41] are used to the creation of such metamodels. While the use of metamodels significantly reduces computational burden by approximating complex models with simpler ones, there are some critical drawbacks to this approach [
42,
43,
44]. Despite the efficiency of new meta-models in sampling [
45,
46], traditional (quasi-) Monte Carlo methods remain the primary choice for practical sensitivity analysis [
3,
4].
Existing research frequently explores the rate of advancements across various Monte Carlo based reliability applications [
47,
48], but there is a notable lack of studies focusing on the implications of these advancements for enhancing the efficiency of reliability-oriented global sensitivity analyses (GSAs). Ensuring the accurate estimation of failure probability
Pf with the available number of simulations is critical because it directly influences reliability assessments and decision-making processes. More accurate
Pf estimation is advantageous when using both the original model and the metamodel, depending on the available computational resources.
The solution proposed in this article involves adopting an alternative measure of structural reliability based on Cliff's Delta [
48], which can be calculated using double-nested-loop simulations. This approach enhances the precision of
Pf estimation without requiring additional computational effort, offering a robust alternative to traditional metamodel-based GSA. By maximizing the utility of existing simulations, this method ensures that reliability analyses are both more accurate and computationally efficient.
2. Cliff Delta
Cliff delta, denoted as
δC, was initially devised by Norman Cliff, primarily for handling ordinal data [
49]. It serves as a metric to assess the frequency with which values from one distribution exceed those in another distribution. A key feature of
δC is that it does not necessitate any specific assumptions regarding the distributions’ form or variability.
The formula for computing the sample estimate of
δC is expressed as:
In this equation, the two distributions are characterized by sizes n and m, with respective elements yi and yj. Here, the notation [⋅][⋅] refers to the Iverson bracket notation, resulting 1 if the condition within the brackets holds true, and 0 otherwise. This statistical approach allows for an intuitive comparison of two distributions by quantifying the dominance of one distribution over the other.
Building upon the foundational description of
δC, this measure is specifically applied to assess the relationship between resistance
R and load force
F within a framework of structural reliability. In the limit state, a structure is reliable if
R ≥
F, otherwise failure occurs [
19]. Let the difference between
R and
F be denoted as
where
R and
F are statistically independent random variables. In the Monte Carlo simulations, these variables are represented as arrays
R and
F, corresponding to resistance and force, respectively. Cliff's delta,
δC, is utilized to quantitatively evaluate the extent to which values of resistance exceed or fall below those of force, thereby providing fundamental insights into system reliability.
Assuming arrays
R and
F are of equal size, with
n entries each, the formula for calculating
δC is simplified to:
In this expression, Ri and Fj represent the i-th and j-th entries in the R and F arrays, respectively, denoting random realizations of resistance and force. The Iverson brackets, [⋅][⋅], return 1 when the enclosed condition is true and 0 otherwise. This formulation facilitates a direct comparison between resistance and force across all sampled scenarios.
The estimation of δC between the measurements of resistance and force provides a metric quantifying the frequency with which resistance values surpass those of force. Employed as a statistical tool, this measure assesses how frequently resistance can withstand or exceed applied forces.
4. The case study
In the case study, the new sensitivity indices are compared with the results of Sobol sensitivity analysis using a simple example. In Equation (2), the resistance can be considered as:
where
K is constant. The load force is considered as
All input random variables, X1, X2,…,X5, follow a Gaussian probability density function with a mean value of zero and a standard deviation of one.
Table 1 presents the estimated values of
C0 =
, where
is obtained using
n = 12,000 runs of the Latin Hypercube Sampling (LHS) method [
57,
58]. The conditional values of
are estimated using double-nested-loop computation of LHS method. When a single input variable,
Xi, is fixed, it is sampled using
n=12,000 runs, and for each realization,
is calculated again using
n=12,000 runs of the LHS method. This numerical procedure is analogous to Sobol [
7,
8], but the sensitivity measurement is not based on variance but on Cliff's delta. In computing of
Si, the computational complexity of
E(
|
Xi) is
n2=144,000,000.
In the
Table 1, the data in the last column represent a set of values that are highly consistent, with low variance. Most values range between 0.924 and 0.940, with a last exceptions approaching 1.000. In
Table 1, the value 1 is always present on the last row, ensuring that the sum of all indices equals one.
The column with K=0 exhibits values close to zero, indicating the absence of dominance of R over F in the observations of . Conversely, values far from zero suggest a dominance of R over F in the observation of .
In general, the strong influence of an input variable or variables occurs when the mean value of the fixed realizations of
is significantly different from
C0, see Equations (13), (14), and (15). It can be noted that the accuracy of estimation of sensitivity indices is lowest for
K=4, where the conditioned realizations of
in the last column of
Table 1 are very consistent, with low variance, and are minimally different from
C0.
The influences of input variables and their groups, as expressed by sensitivity indices, are displayed in
Figure 1,
Figure 2 and
Figure 3.
In the
Figure 1, all first-order sensitivity indices are zero. The second-order sensitivity indices
S12=0.18 and
S23=0.18 have the same value, which is due to the nature of Equation 10 and the same characteristics of the input random variables. The dominant influence is the interaction effect of variables
X4 and
X5, as indicated by the value
S45=0.26.
Increasing the value of the constant
K distances the random realizations of resistance
R from load action
F and enhances the value of Cliff's Delta. The interaction effects indicated by sensitivity indices
S12,
S23, and
S45 decrease with increasing K, while the proportion of third, fourth, and fifth-order sensitivity indices increases, see
Figure 2 and
Figure 3.
In sensitivity analysis based on Cliff's Delta, the impact of input variables on the observed changes in is quantified using sensitivity indices of the first order and higher orders. To derive meaningful conclusions and categorize input variables into influential, less influential, and non-influential groups, it is essential to assign the effects of each input variable without the complexity of interpreting numerous sensitivity indices.
To achieve this, the concept of the total effect index is employed. This index captures the comprehensive contribution of a factor, Xi, to the changes observed in . Specifically, it encompasses both the first-order effects and all higher-order effects resulting from interactions. The total effect index provides a robust measure of the impact that each input variable has on the , accounting for all potential interactions.
For instance, in a five-factor model, the total effect of the factor
X1 is calculated by summing all terms in Equation (17) where the factor
X1 is included.
This sum accounts for
X1 direct influence on
as well as its synergistic effects with other factors. The total effect measure provides the educated answer to the question, Which factor can be fixed anywhere over its range of variability without affecting the
. The total effect index reflects both the main and the interaction influences of
X1 on the outcome, providing a comprehensive view of its relative importance in the system's reliability, measured by the distance from
F to
R. Total effects for the case study are displayed in
Figure 4.
The sensitivity analysis results show the Total Sensitivity Indices for each input variable (X1, X2, X3, X4, X5) across five distinct constant values (K=0, 1, 2, 3, 4), using Cliff's delta as the sensitivity measure.
As
K increases from 0 to 4, a general trend of increasing total sensitivity indices is observed for
X1,
X2, as depicted in
Figure 4. For
X3, the total sensitivity indices exhibit a slightly convex pattern. An increasing trend for
X3 is observed from
K=1 to
K=4. This suggests that these variables become more influential with higher values of
K, as measured by Cliff's delta. Conversely, the sensitivity indices for
X4 and
X5 decrease with increasing
K.
The variable X2 consistently exhibits the highest total sensitivity index across all tested values of K, indicating its dominant influence on Cliff's delta. This effect is attributable to X2's involvement in both additive terms of the resistance function, as shown in Equation (22). The variables X1 and X3 also demonstrate an increasing influence, although they remain slightly less dominant than X2 but are notably more influential than X4 and X5 as K increases. The influence of X4 and X5, which are involved in the load force equation, decreases especially as K surpasses 1, highlighting their reduced significance in affecting Cliff's delta.
The sensitivity analysis outcomes reflect a decreasing probability
P(
R<
F), which diminishes as
K increases. The Cliff delta-based sensitivity analysis exhibits characteristics of reliability-oriented sensitivity analysis [
19], describing the change in the influence of each variable on Cliff's delta due to
K. The influence of the results of the sensitivity indices by the value of the deterministic quantity
K is the main difference compared to Sobol sensitivity analysis.
The results of the classical Sobol sensitivity analysis can be calculated analytically. Non-zero values of Sobol's sensitivity indices were obtained only for second-order sensitivity indices =1/3, =1/3, =1/3 , other Sobol indices are zero. The total effect Sobol sensitivity indices are =1/3, =2/3, =1/3, =1/3, =1/3. The dominant influence of the input variable X2 confirms the most important conclusions of the newly introduced sensitivity analysis based on Cliff's Delta, however, there are differences. The results of Sobol's sensitivity analysis are independent of the value of the constant K, because Sobol's indices are based on the decomposition of variance, which the deterministic variable K does not affect.
Although classic Sobol's sensitivity analysis of model output is empathetic to the results of reliability-oriented types of sensitivity analyses, it is not directly oriented towards reliability, as Sobol sensitivity indices do not reflect changes in
Pf [
19].
The sensitivity indices results based on Cliff's Delta closely resemble those obtained from the sensitivity analysis based on
Pf . The pie chart on the left side of
Figure 4 is practically identical to the chart in
Figure 1. Moreover, the pie chart on the right side of
Figure 4 is very similar to the chart on the right side of
Figure 5. The results of the sensitivity analysis are the same, but the sensitivity scale is different. Overall, this demonstrates a high degree of similarity between Cliff's Delta and
Pf, from which useful conclusions can be drawn.
The sensitivity indices based on Cliff's Delta are derived from the calculation of
Pf using Cliff's Delta. While it may seem that estimating Cliff's Delta is more demanding compared to
Pf, using double-loop simulations to estimate Cliff's Delta leads to a more accurate numerical estimate of
Pf and extracts more information from the available simulations. It can be noted that a similar double-loop simulation is used for estimating
Pf based on the numerical integration of the distributions of the random variables
R and
F , see e.g. [
59].
The advantage of estimating Cliff's Delta is that it does not require knowledge of the distributions or approximation approaches of the random variables R and F. This flexibility allows for a more robust analysis in practical scenarios where precise distribution functions may not be available or easily determined.
5. Comparative Analysis of Pf Estimations Using Cliff's Delta and Basic Definition
It can be showed that the estimation of Pf according to Equation (9) is more accurate compared to the basic definition in Equation (4) when Cliff's delta is calculated using a double nested-loop simulation according to Equation (3).
Let
Pf denote the failure probability estimated from a binomial distribution, and let
n represent the number of simulations or experiments.
The standard error
SE for the failure probability estimation is provided by the binomial distribution
The estimate of the failure probability Pf using Cliff’s Delta in Equation (9) is more accurate because it utilizes an increased number of simulations through a double-nested-loop process. However, the standard error SE2(Pf) for this estimate cannot be straightforwardly determined using an analytical formula similar to SE1(Pf). Instead, empirical observations indicate that the standard error for Pf estimated with Cliff’s Delta is smaller. This improved accuracy is attributed to the increased statistical information gained from the double-nested-loop simulations. The accuracy of both estimates can be demonstrated in the following case study using the Monte Carlo method.
The limit state of a rod made of elastic material, subjected to axial tension, is being studied. Let resistance
R have a Gaussian probability density function with a mean value
μR=9 kN and standard deviation
σR=0.9 kN, and let
F have a Gaussian probability density function with a mean value
μF=4 kN and standard deviation
σF=1.218887 kN. Under these assumptions, Z is a random variable with a Gaussian probability density function with a mean value
μZ=9−4=5 kN and standard deviation
σZ =(0.9
2+1.218887
2)
0.5=1.515 kN. The failure probability
Pf =0.000483477 is evaluated by numerical integration of the Gaussian probability density function of
Z from negative infinity to zero. According to EN1990 [
15], a structure designed with this failure probability is classified in reliability class RC1 for reference periods of 1 and 50 years.
The aim of the case study is to compare the accuracy of
Pf estimation according to the basic definition in Equation (4) and the alternative formula in Equation (9). The procedure is as follows: In the first step, the random variables
R and
F are simulated for 10,000 runs using the Monte Carlo (MC) method. The estimation of
Pf from Equation (4) is plotted in
Figure 6 as the first blue point from the left. The estimation of
Pf from Equation (9) is plotted in
Figure 7 as the first blue point from the left. Next, another 10,000 runs of MC are generated, and the procedure is repeated. In total, 100 estimates of
Pf according to Equation (4) are plotted in the left quarter of
Figure 6, and 100 estimates of
Pf according to Equation (9) are plotted in the left quarter of
Figure 7. This procedure is further repeated for 100,000, 1 million, and 10 million runs of the Monte Carlo (MC) method, as shown in the subsequent quarters in
Figure 6 and
Figure 7.
The numerical study revealed that the failure probability
Pf estimation using Cliff's Delta
δC, see Equation (9), is more accurate compared to the basic definition, see Equation (4). As illustrated in
Figure 6 and
Figure 7, the alternative method resulted in noticeably lower standard deviations σ
Pf (approximately three times smaller) across varying Monte Carlo run counts, indicating improved consistency and precision. In contrast, the basic definition exhibited higher variability, especially with fewer Monte Carlo runs. These findings underscore the efficacy of the alternative formula for more reliable
Pf estimation in structural reliability assessments.
6. Discussion
In civil engineering, estimating the resistance
R using nonlinear FEM calculations is very time-consuming, and studies are limited by the number of computationally expensive Monte Carlo or LHS method runs required to accurately simulate real structural behavior, see e.g., [
60,
61,
62,
63]. In recent years, the role of mathematical models that offer quick analytical solutions has become increasingly significant due to their efficiency in providing rapid responses [
64,
65]. For each model, it is useful to utilize all realized simulations to estimate failure probability as efficiently as possible.
Equation (9) offers an efficient alternative for estimating
Pf with high utilization of information from the realized runs of the computational model. Although estimating Cliff's Delta is more numerically demand than direct estimation of failure probability using Equation (4), the same number of model runs (the same number of random realizations of
R and
F) ensures a more accurate estimation of failure probability
Pf, see
Figure 6 compared to
Figure 7. These advantages highlight the utility of Cliff's Delta in both reliability and sensitivity analysis, making it an innovative tool for engineers and researchers to gain deeper insights into the reliability and performance of structural systems.
The computational complexity of estimating Cliff's Delta can be effectively reduced by optimizing computer algorithms. The following function formulates the algorithm for calculating Cliff's Delta using ten thousand random realizations of R and F in arrays AR and AF. In the Delphi programming language, the function to calculate Cliff's Delta can be written as follows:
`pascal
const max=10000;
var
AR,AF: array[1..max] of Single;
AB : array[1..max] of Boolean;
Function Cliffd:extended;
var i1,i2 :integer;
Dominant, Dominated : Int64;
begin
Dominant:=0;
Dominated:=0;
for i1:=1 to max do
begin
if AB[i1] then Dominant:=Dominant+max
else
for i2:=1 to max do
begin
if (AR[i2]>AF[i1]) then Inc(Dominant) else
if (AR[i2]<AF[i1]) then Inc(Dominated);
end;
end;
Cliffd:=(Dominant-Dominated)/(max*max);
end;
`
Before calling the function, the array AB is populated with random realizations for which the inner loop runs are unnecessary.
`pascal
for i1 := 1 to max do if AF[i1] < 5.49836 then AB[i1] := true else AB[i1] := false;
`
The constant 5.49836 is the smallest value of the random variable R generated by the Monte Carlo method. If a random realization of F is less than 5.49836, then all random realizations of R are less than 5.49836, and the inner loop does not need to be executed; it is sufficient to assign Dominant:=Dominant+max. In the case study, using max=10000, the estimation of Cliff's Delta was accelerated by approximately two times.
The algorithm efficiently reduces the double-loop computational burden especially for Cliff's Delta estimates close to one (very small
Pf values), where the vast majority of random realizations of
R are much higher than
F. According to the EN1990 [
15] standard, common building structures are designed with a
Pf around 7.2 ·10
-5. For estimating the standard error of the
Pf using Equation (25), one needs to perform 1,388,889 Monte Carlo simulation runs to achieve a coefficient of variation of 0.1. This results in a standard error of
SE1=7.2⋅10−6. In practice, this requires conducting Monte Carlo runs until 100 failure events are observed, i.e., 100 runs where
Z < 0. Subsequently, using Equation (9) instead of Equation (4) will improve the accuracy of the
Pf estimate.
Conducting sensitivity analysis based on Cliff's Delta offers functionalities that are identical to sensitivity analysis based on failure probability focused on reliability. Cliff's Delta provides a robust measure for evaluating the extent to which one distribution dominates another without making specific assumptions about the distributions' form or variability. This flexibility is particularly valuable in reliability studies where input variables may not follow normal distributions or exhibit significant variability. By focusing on the frequency with which resistance values exceed load values, Cliff's Delta directly aligns with the fundamental concepts of reliability engineering. The ability to estimate failure probabilities using Cliff's Delta enhances the robustness of reliability and sensitivity analysis, especially in cases where performing additional resistance or load simulations is numerically difficult or impossible.