1. Introduction
The issue of student dropout in online education has become a critical concern for educators and institutions alike. Online education offers unparalleled flexibility and accessibility, allowing students from diverse backgrounds to pursue their academic goals. However, the flexibility that makes online learning appealing also presents unique challenges, particularly regarding student engagement and retention [
1,
2]. High dropout rates in online courses not only affect individual students’ academic progress but also undermine the efficacy of educational programs and the reputation of institutions offering these courses. With the rapid expansion of online learning platforms, understanding the factors that influence student retention and dropout rates has become crucial for designing effective interventions aimed at mitigating dropout rates and enhancing student success [
1,
2,
3,
4].
Traditional models of student retention, such as those used in face-to-face learning environments [
5,
6,
7,
8], often fail to capture the complex and dynamic patterns observed in online dropout data. These models typically assume linear relationships and do not account for the myriad factors that can influence a student’s decision to drop out [
9]. Online learning environments introduce variables such as technological issues, varying levels of digital literacy, and the need for self-regulation and motivation, all of which can significantly impact student retention [
9]. Consequently, there is a pressing need for the development of more sophisticated statistical tools that can accurately model these complexities and provide coverage for dropout time. While most of the existing studies on modelling dropout focuses of the use of machine learning classification techniques [
8,
10,
11,
12,
13], only few [
2] has utilized survival analysis and probability modelling to analyze students dropout time. Thus, this study introduces the Exponentiated Weibull Weibull (EWW) distribution, a five-parameter model specifically designed to better represent dropout times in online learning environments. The EWW distribution extends traditional survival analysis models by incorporating additional parameters that allow for greater flexibility in modeling the hazard function. The hazard function, which represents the instantaneous risk of dropout at any given time, can take on various shapes, including increasing, decreasing, and bathtub-shaped hazard rates. This flexibility is particularly important for capturing the uncommon patterns of student engagement and disengagement in online courses.
The motivation for modelling student dropout times by the EWW distribution comes from the requirement to include the unusual patterns of online learning engagement and disengagement. Earlier studies have applied different statistical techniques in the analysis of dropout rates, including logistic regression and Cox proportional hazards models [
2,
14,
15,
16,
17]. Moreover, these models are often inadequate to properly predict the non-monotonic hazard rates of dropout that appear in educational data, where the risk of student leaving his/her course is lower at the beginning and then increases thereafter. Using the EWW distribution in this setting allows us to investigate the dropout dynamics in a more fine-grained manner. EWW distribution provides a better fit for the data by introducing parameters which is able to capture both the initial decrease, as well as long-term increase in dropout hazard rates. A closer fit is important in generating appropriate interventions that directly target dropout and retention rate for online learning courses [
2].
The EWW distribution is built on the foundational principles of the Weibull distribution, a widely used model in reliability engineering and survival analysis due to its flexibility in modelling various types of data [
18]. The Exponentiated Weibull (EW) distribution are known for characterization of both increasing and decreasing hazard rates, making it suitable for diverse applications [
18]. The EWW distribution extends this versatility by incorporating additional parameters, allowing it to accommodate more complex hazard functions often observed in real-world data, such as the bathtub-shaped hazard function which maybe observed in dropout studies [
2,
19,
20]. This enhancement is important for capturing the unusual patterns of student engagement and disengagement, providing a more accurate and detailed understanding of dropout dynamics in online learning environments.
Bayesian estimation has gained popularity in recent years due to its ability to incorporate prior information and provide more accurate parameter estimates, especially in small sample sizes or complex models [
18,
21,
22,
23,
24,
25,
26]. Unlike the traditional Maximum Likelihood Estimation (MLE), which relies solely on the observed data, Bayesian methods combine prior distributions with the likelihood of the observed data to generate posterior distributions. This approach can yield more robust estimates, particularly in scenarios with high variability or limited data [
27,
28].
In this paper, I used both MLE and Bayesian estimation methods to estimate the parameters of the EWW distribution. The comparison of these methods is essential to determine the most effective approach for parameter estimation in this context. Simulation studies and real-life data analysis are conducted to evaluate the accuracy and precision of these estimates.
3. Five-Parameter Exponentiated Weibull Weibull Distribution
The cumulative distribution function (CDF) of the exponentiated Weibull-generated family [
32] for a random variable
is given by
where
and
are the two shape parameters, and
is the rate parameter. This CDF provides a broader family of continuous distributions. The corresponding probability density function (PDF) is given by
In this study, we propose a novel five-parameter model that serves as a competitive extension to the Weibull distribution by utilizing the EW-G distribution. By replacing
in equation (
6) with the cumulative distribution function of the Weibull distribution from equation (
4), we obtain the following result.
Simplifying the expression inside the exponential:
Therefore, the cumulative distribution function (CDF) of the five-parameter Exponentiated-Weibull Weibull (EWW) distribution for a random variable
is given by:
where
,
and
are the three shape parameters, and
and
are the rate parameters.
Correspondingly, substituting equations (
3) and (
4) in (
7), we have the pdf of the five-parameter exponentiated-Weibull Weibull distribution given by
The survival, hazard and cumulative hazard functions follow from the cumulative density function as follows respectively:
In addition, the median survival time is given by:
Note that the derivations in equations (
8) to (
13) are different from the one presented by [
32] in terms of parameterization of the Weibull base distribution. While the approach of [
32] is applicable in medicine, our own parameterization is applicable for modelling failure times [
33,
34] which is relevant for the specific situation addressed in this paper.
Figure 1 shows that when are all equal to 1, we have an increasing hazard and when are less than 1, we have a decreasing hazard. Similarly, when are less than 1 but at least 1, we have a bathtub hazard and is less than 1 but and are greater than 1, we have a unimodal hazard.
6. Simulation study and results
To illustrate the EWW dropout model using both the maximum likelihood estimator and Bayesian estimator, we simulated random samples that corresponds to the four hazard types presented in
Figure 1. Specifically, the following simulation parameters were used for the various hazard types.
Monotone increasing: .
Monotone decreasing: .
Unimodal: .
Bathtub-shaped: .
The dropout (
) time was generated using the following equation:
where
U is the standardized uniform random variable with parameters
. To achieve censoring, a random censoring scheme was adopted. In this scheme, a random variable
Z that follows a Bernoulli distribution with parameter
, denoted as Bernoulli
, is defined. If
, then
corresponds to the dropout time of student
i. Otherwise,
corresponds to the completion time of student
i. To study the effect of censoring on the estimators, we set
which corresponds to
censoring percentages. In addition, the estimators were evaluated using bias and root mean square error. Bias and root mean square error for each parameter set are given below:
The sample size
n was fixed at 200 throughout the experiment. To achieve stability, each simulation run was replicated 1000 times [
35]. For the Bayesian estimation,
posterior samples were generated with a thinning value set to 10 (i.e., every 10th estimate was saved). The initial burn-in period, during which samples were discarded, was set at 2000.
In
Table 1, we present the parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE), biases, and root mean squared errors (RMSE) for both Bayesian and MLE methods. The table summarizes these statistics for five different parameters:
,
,
,
, and
, evaluated at different censoring levels (
) of 10%, 20%, 30%, 40%, and 50% for scenario 1: increasing hazard type. The results show that Bayesian estimates consistently outperform Maximum Likelihood Estimates (MLE) across all parameters. For
, Bayesian estimates are close to 1 the true value with low standard deviations and root mean squared errors (RMSE), whereas MLEs exhibit higher variability and larger biases, particularly at the 10% quantile. For
, Bayesian estimates also display consistency and low biases, while MLEs vary significantly with large biases and higher RMSE values.
Similarly
Table 1 shows that the Bayesian estimates for
are accurate with small positive biases and low RMSE values, whereas MLEs show greater variability and higher biases, notably at the 20% quantile. For
, Bayesian estimates are close to the true value with low biases and RMSE, while MLEs are more variable with larger biases and higher RMSE values. Lastly, Bayesian estimates for
are accurate with small biases and low RMSE, in contrast to MLEs which exhibit higher variability, larger biases, and higher RMSE values. Overall, Bayesian estimation demonstrates higher accuracy and precision compared to MLE for all parameters, indicating its robustness in this context.
Table 2 presents the parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE), biases, and root mean squared errors (RMSE) for both Bayesian and MLE methods evaluated at different censoring levels (
) of 10%, 20%, 30%, 40%, and 50% for a decreasing hazard type. The results show that Bayesian estimates consistently exhibit lower biases and RMSE values compared to MLEs across all parameters and censoring levels. Bayesian estimates for
,
,
,
, and
are more accurate and precise, with smaller SDs and biases. In contrast, MLEs show higher variability, larger biases, and greater RMSE values, indicating less reliable estimates. For instance, the bias for
in Bayesian estimates remains small and close to zero, while MLEs show a significant bias, especially at the 10% censoring level. Similarly, Bayesian estimates for
and
demonstrate consistent accuracy, whereas MLEs vary widely with substantial biases and higher RMSE values.
Table 3 presents parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE), biases, and root mean squared errors (RMSE) for both Bayesian and MLE methods, evaluated at different censoring levels (
) of 10%, 20%, 30%, 40%, and 50% for a unimodal hazard type. For parameter
, Bayesian estimates are consistently close to 4 with low SD and RMSE, whereas MLEs show significant variability and larger biases, especially at the 10% censoring level. Bayesian estimates for
remain close to the true value of 0.3, displaying small biases and low RMSEs, while MLEs vary considerably with large biases and higher RMSEs. For
, Bayesian estimates are accurate with low biases and RMSEs, but MLEs exhibit higher variability and significant biases. In the case of
, Bayesian estimates are near the true value of 3 with low biases and RMSEs, while MLEs show greater variability and higher biases. Lastly, Bayesian estimates for
demonstrate accuracy with small biases and low RMSEs, in contrast to MLEs which display high variability, larger biases, and elevated RMSEs.
The results presented in
Table 4 indicate that Bayesian estimates consistently outperform Maximum Likelihood Estimates (MLE) across various parameters and censoring levels for the bathtub hazard type. Bayesian estimates for all parameters (
,
,
,
,
) show low standard deviations (SD) and root mean squared errors (RMSE), with biases generally close to zero. In contrast, MLEs exhibit higher variability, larger biases, and higher RMSE values across the board. For example, Bayesian estimates of
at the 10% censoring level have a low bias of 0.024 and an RMSE of 0.029, whereas MLEs show a high bias of 0.152 and an RMSE of 0.729. Similarly, Bayesian estimates of
are close to the true value with low biases and RMSEs across all censoring levels, while MLEs are more variable with larger biases and higher RMSE values.
Overall across the four hazard types and censoring levels, the Bayesian estimates demonstrate higher accuracy and precision compared to the MLEs for all parameters, as indicated by lower biases and RMSE values. The standard deviations for Bayesian estimates are also consistently smaller than the standard errors for MLEs, further highlighting the robustness of the Bayesian approach in this context. This simulation underscores the advantages of Bayesian estimation in providing more reliable parameter estimates for the Exponentiated Weibull Weibull (EWW) distribution.
7. Application to Modelling Time to Students’ Dropout from an Online Class Presentation
The dataset used in this study was extracted from the student behaviour and performance dataset for 23,344 students in the United Kingdom, accessible at [
36]. This dataset includes various data points such as assessment results and extensive logs of student interactions with the Virtual Learning Environment (VLE). The final result of each student is categorized as Distinction, Pass, Withdrawn, or Fail, along with the submission date measured in days since the start of the module presentation. For the analysis in this study, a strategic sample of 500 students was selected to reflect the actual distribution of the final results. The goodness of fit for the proposed EWW distribution, alongside the Exponential Weibull (EW), Weibull Weibull (WW), and Weibull Exponential (WE) distributions [
32], was assessed using the log-likelihood, Akaike Information Criterion (AIC), and Bayesian Information Criterion (BIC). These three competing distributions were selected because they are special cases of the EWW distribution. The WW can be derived from the EWW if we set
. The EWW reduces to the WE if
and
. Similarly, the EWW reduces to the EW if
and
.
Figure 2 shows the plots of , , for the five-parameter exponentiated-Weibull Weibull (EWW) alongside the Exponential Weibull (EW), Weibull Weibull (WW), and Weibull Exponential (WE) distributions for the student dropout dataset based on the Bayesian parameter estimates. The plot shows that the most appropriate distribution for the student dropout dataset is EWW.
The results presented in
Table 5 highlight the superior performance of Bayesian estimates over Maximum Likelihood Estimates (MLE) for various parameters for the student dropout dataset. The Bayesian estimates for
,
,
,
, and
exhibit lower standard deviations (SD), indicating higher precision. For instance, the Bayesian estimate of
is 0.233 with an SD of 0.008, compared to the MLE estimate of 0.536 with a standard error (SE) of 0.137. Similarly, Bayesian estimates for
and
are close to their true values with minimal SDs (0.002), while MLE estimates show greater variability and higher SEs (0.149 and 0.154, respectively). For the parameter
, the Bayesian method yields an estimate of 0.635 with an SD of 0.005, significantly more precise than the MLE estimate of 0.759 with an SE of 0.129. Lastly, the Bayesian estimate for
is 0.358 with an SD of 0.003, whereas the MLE estimate is 1.190 with a considerably higher SE of 0.196. Overall, these findings underscore the robustness and precision of Bayesian estimation in parameter estimation compared to the MLE method for the student dropout dataset.
The goodness-of-fit results presented in
Table 6 for the five-parameter Exponentiated Weibull Weibull (EWW) model, alongside the Exponential Weibull (EW), Weibull Weibull (WW), and Weibull Exponential (WE) models for the student dropout dataset, indicate that the EWW model provides the best fit. This is evidenced by the EWW model having the highest LogLikelihood (-750.75) and the lowest values for the Akaike Information Criterion (AIC = 1511.5) and the Bayesian Information Criterion (BIC = 1532.57). In contrast, the WW, WE, and EW models exhibit significantly lower LogLikelihood values (-1799.22, -1859.54, and -1029.02 respectively) and higher AIC (3606.44, 3725.09, and 2064.04 respectively) and BIC (3623.3, 3737.73, and 2076.68 respectively) values. These findings underscore the superior performance of the EWW model in capturing the underlying characteristics of the student dropout data, making it the most suitable choice among the compared models.
Table 7 shows the estimate of the dropout times for the student from the online class module using the four distributions based on the Bayesian parameter estimate. The result showed that median dropout time using the EWW distribution from the online class presentation is 0.8 months (24 days) from the day the class commenced. On the other hand, if any of the WW, WE, EW distributions is assumed, the median dropout time from the online class presentation is approximately 2 months (60 days) from the day the class commenced. The median estimate using any of WW, WE, EW distribution is significantly larger than EWW and also unrealistic given that most students dropout from online class modules in the first few days after the commencement of the online presentation [
11,
15].
Figure 1.
Plot of , , survival and hazard functions for the five-parameter exponentiated-Weibull Weibull Distribution for increasing (black), decreasing (red), bathtub (green) and unimodal (blue) hazard types.
Figure 1.
Plot of , , survival and hazard functions for the five-parameter exponentiated-Weibull Weibull Distribution for increasing (black), decreasing (red), bathtub (green) and unimodal (blue) hazard types.
Figure 2.
Plot of , , for the five-parameter exponentiated-Weibull Weibull (EWW) alongside the Exponential Weibull (EW), Weibull Weibull (WW), and Weibull Exponential (WE) distributions for the student dropout dataset based on the Bayesian parameter estimates.
Figure 2.
Plot of , , for the five-parameter exponentiated-Weibull Weibull (EWW) alongside the Exponential Weibull (EW), Weibull Weibull (WW), and Weibull Exponential (WE) distributions for the student dropout dataset based on the Bayesian parameter estimates.
Table 1.
Parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE), biases, and root mean squared errors (RMSE) for both Bayesian and MLE methods were evaluated at different censoring levels () of 10%, 20%, 30%, 40%, and 50% for an increasing hazard type.
Table 1.
Parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE), biases, and root mean squared errors (RMSE) for both Bayesian and MLE methods were evaluated at different censoring levels () of 10%, 20%, 30%, 40%, and 50% for an increasing hazard type.
|
|
Bayesian |
MLE |
Parameter |
|
Estimate |
SD |
Bias |
RMSE |
Estimate |
SE |
Bias |
RMSE |
|
|
0.991 |
0.014 |
-0.009 |
0.017 |
1.910 |
4.459 |
0.910 |
4.551 |
|
|
0.946 |
0.025 |
-0.055 |
0.060 |
2.001 |
1.871 |
1.001 |
2.122 |
|
10% |
1.033 |
0.030 |
0.033 |
0.045 |
1.741 |
0.570 |
0.741 |
0.935 |
|
|
1.013 |
0.010 |
0.013 |
0.017 |
0.450 |
0.920 |
-0.550 |
1.072 |
|
|
1.030 |
0.022 |
0.030 |
0.037 |
0.398 |
0.347 |
-0.602 |
0.695 |
|
|
|
|
0.003 |
0.035 |
|
|
0.300 |
1.875 |
|
|
0.977 |
0.016 |
-0.023 |
0.028 |
1.187 |
0.645 |
0.187 |
0.671 |
|
|
0.959 |
0.025 |
-0.041 |
0.048 |
0.434 |
0.348 |
-0.566 |
0.664 |
|
20% |
1.043 |
0.033 |
0.043 |
0.055 |
2.284 |
3.407 |
1.284 |
3.641 |
|
|
0.996 |
0.011 |
-0.004 |
0.011 |
1.385 |
1.658 |
0.385 |
1.702 |
|
|
1.047 |
0.026 |
0.047 |
0.054 |
1.444 |
1.067 |
0.444 |
1.155 |
|
|
|
|
0.004 |
0.039 |
|
|
0.347 |
1.567 |
|
|
0.961 |
0.021 |
-0.039 |
0.044 |
0.457 |
0.861 |
-0.543 |
1.018 |
|
|
0.947 |
0.026 |
-0.053 |
0.059 |
1.335 |
5.281 |
0.335 |
5.291 |
|
30% |
1.049 |
0.031 |
0.049 |
0.058 |
1.079 |
3.170 |
0.079 |
3.171 |
|
|
0.980 |
0.016 |
-0.021 |
0.026 |
1.204 |
3.943 |
0.204 |
3.948 |
|
|
1.044 |
0.026 |
0.044 |
0.051 |
0.753 |
1.150 |
-0.247 |
1.176 |
|
|
|
|
-0.004 |
0.048 |
|
|
-0.034 |
2.921 |
|
|
0.952 |
0.024 |
-0.048 |
0.054 |
0.831 |
1.252 |
-0.169 |
1.263 |
|
|
0.934 |
0.028 |
-0.066 |
0.072 |
1.210 |
1.368 |
0.210 |
1.384 |
|
40% |
1.044 |
0.030 |
0.044 |
0.053 |
0.754 |
1.096 |
-0.246 |
1.123 |
|
|
0.958 |
0.020 |
-0.042 |
0.046 |
0.733 |
1.172 |
-0.267 |
1.202 |
|
|
1.040 |
0.023 |
0.040 |
0.046 |
1.190 |
1.216 |
0.190 |
1.231 |
|
|
|
|
-0.014 |
0.054 |
|
|
-0.056 |
1.241 |
|
|
0.921 |
0.030 |
-0.079 |
0.085 |
0.457 |
1.036 |
0.636 |
1.216 |
|
|
0.949 |
0.024 |
-0.051 |
0.056 |
1.335 |
0.348 |
-0.385 |
0.519 |
|
50% |
1.068 |
0.034 |
0.068 |
0.077 |
1.079 |
0.131 |
1.076 |
1.084 |
|
|
0.926 |
0.029 |
-0.074 |
0.079 |
1.204 |
0.489 |
-0.297 |
0.572 |
|
|
1.052 |
0.028 |
0.052 |
0.059 |
0.753 |
0.736 |
0.466 |
0.871 |
|
|
|
|
-0.017 |
0.071 |
|
|
0.299 |
0.853 |
Table 2.
Parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE), biases, and root mean squared errors (RMSE) for both Bayesian and MLE methods were evaluated at different censoring levels () of 10%, 20%, 30%, 40%, and 50% for decreasing hazard type.
Table 2.
Parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE), biases, and root mean squared errors (RMSE) for both Bayesian and MLE methods were evaluated at different censoring levels () of 10%, 20%, 30%, 40%, and 50% for decreasing hazard type.
|
|
Bayesian |
MLE |
Parameter |
|
Estimate |
SD |
Bias |
RMSE |
Estimate |
SE |
Bias |
RMSE |
|
|
0.703 |
0.014 |
0.003 |
0.015 |
0.442 |
0.436 |
-0.258 |
0.506 |
|
|
0.591 |
0.036 |
-0.109 |
0.115 |
0.991 |
1.085 |
0.291 |
1.124 |
|
10% |
0.704 |
0.024 |
0.004 |
0.024 |
0.812 |
0.826 |
0.112 |
0.833 |
|
|
1.014 |
0.011 |
0.014 |
0.018 |
1.623 |
3.481 |
0.623 |
3.537 |
|
|
0.982 |
0.013 |
-0.019 |
0.023 |
0.472 |
0.257 |
-0.528 |
0.588 |
|
|
|
|
-0.021 |
0.039 |
|
|
0.048 |
1.318 |
|
|
0.722 |
0.017 |
0.022 |
0.027 |
0.945 |
0.548 |
0.245 |
0.601 |
|
|
0.607 |
0.031 |
-0.093 |
0.098 |
0.373 |
0.246 |
-0.328 |
0.410 |
|
20% |
0.685 |
0.020 |
-0.015 |
0.025 |
1.926 |
1.649 |
1.226 |
2.055 |
|
|
1.027 |
0.013 |
0.027 |
0.030 |
3.613 |
3.452 |
2.613 |
4.330 |
|
|
0.970 |
0.012 |
-0.030 |
0.032 |
0.627 |
0.193 |
-0.374 |
0.420 |
|
|
|
|
-0.018 |
0.042 |
|
|
0.677 |
1.563 |
|
|
0.660 |
0.020 |
-0.040 |
0.044 |
0.442 |
1.913 |
0.528 |
1.985 |
|
|
0.606 |
0.038 |
-0.094 |
0.101 |
0.991 |
0.813 |
-0.319 |
0.873 |
|
30% |
0.752 |
0.034 |
0.052 |
0.062 |
0.812 |
7.484 |
1.799 |
7.697 |
|
|
0.983 |
0.013 |
-0.017 |
0.021 |
1.623 |
1.883 |
0.192 |
1.893 |
|
|
0.999 |
0.017 |
-0.001 |
0.017 |
0.472 |
0.394 |
-0.286 |
0.487 |
|
|
|
|
-0.020 |
0.049 |
|
|
0.383 |
2.587 |
|
|
0.740 |
0.019 |
0.040 |
0.044 |
0.304 |
0.645 |
-0.396 |
0.757 |
|
|
0.595 |
0.029 |
-0.105 |
0.109 |
2.514 |
3.627 |
1.814 |
4.055 |
|
40% |
0.664 |
0.019 |
-0.036 |
0.041 |
1.167 |
0.752 |
0.467 |
0.885 |
|
|
1.021 |
0.013 |
0.021 |
0.024 |
2.512 |
7.925 |
1.512 |
8.068 |
|
|
0.954 |
0.013 |
-0.046 |
0.047 |
0.126 |
0.141 |
-0.874 |
0.885 |
|
|
|
|
-0.025 |
0.053 |
|
|
0.505 |
2.930 |
|
|
0.687 |
0.016 |
-0.013 |
0.021 |
0.444 |
1.294 |
-0.256 |
1.320 |
|
|
0.579 |
0.041 |
-0.121 |
0.128 |
0.729 |
3.649 |
0.029 |
3.650 |
|
50% |
0.716 |
0.027 |
0.016 |
0.031 |
0.827 |
4.394 |
0.127 |
4.395 |
|
|
1.001 |
0.009 |
0.001 |
0.010 |
1.609 |
10.079 |
0.609 |
10.098 |
|
|
0.980 |
0.014 |
-0.021 |
0.025 |
0.608 |
0.297 |
-0.392 |
0.492 |
|
|
|
|
-0.027 |
0.043 |
|
|
0.023 |
3.991 |
Table 3.
Parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE), biases, and root mean squared errors (RMSE) for both Bayesian and MLE methods were evaluated at different censoring levels () of 10%, 20%, 30%, 40%, and 50% for unimodal hazard type.
Table 3.
Parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE), biases, and root mean squared errors (RMSE) for both Bayesian and MLE methods were evaluated at different censoring levels () of 10%, 20%, 30%, 40%, and 50% for unimodal hazard type.
|
|
Bayesian |
MLE |
Parameter |
|
Estimate |
SD |
Bias |
RMSE |
Estimate |
SE |
Bias |
RMSE |
|
|
3.956 |
0.019 |
-0.044 |
0.048 |
16.078 |
0.354 |
12.078 |
12.083 |
|
|
0.315 |
0.015 |
0.015 |
0.021 |
1.066 |
1.844 |
0.766 |
1.997 |
|
10% |
2.038 |
0.028 |
0.038 |
0.048 |
5.577 |
3.687 |
3.577 |
5.137 |
|
|
2.930 |
0.035 |
-0.070 |
0.078 |
0.226 |
0.350 |
-2.774 |
2.796 |
|
|
3.018 |
0.018 |
0.018 |
0.026 |
0.573 |
1.167 |
-2.427 |
2.693 |
|
|
|
|
-0.008 |
0.044 |
|
|
2.244 |
4.941 |
|
|
3.954 |
0.021 |
-0.047 |
0.051 |
2.479 |
2.450 |
-1.521 |
2.883 |
|
|
0.329 |
0.015 |
0.029 |
0.032 |
0.412 |
0.275 |
0.112 |
0.297 |
|
20% |
2.036 |
0.028 |
0.036 |
0.045 |
2.901 |
5.622 |
0.901 |
5.694 |
|
|
2.928 |
0.035 |
-0.072 |
0.080 |
2.869 |
4.027 |
-0.131 |
4.029 |
|
|
3.028 |
0.022 |
0.028 |
0.036 |
1.809 |
1.962 |
-1.191 |
2.296 |
|
|
|
|
-0.005 |
0.049 |
|
|
-0.366 |
3.040 |
|
|
3.952 |
0.022 |
-0.048 |
0.053 |
3.264 |
1.574 |
-0.736 |
1.738 |
|
|
0.332 |
0.015 |
0.032 |
0.036 |
0.304 |
0.339 |
0.004 |
0.339 |
|
30% |
2.038 |
0.029 |
0.038 |
0.048 |
1.640 |
0.972 |
-0.360 |
1.037 |
|
|
2.920 |
0.038 |
-0.080 |
0.088 |
1.944 |
0.787 |
-1.056 |
1.317 |
|
|
3.023 |
0.022 |
0.023 |
0.032 |
3.748 |
5.723 |
0.748 |
5.772 |
|
|
|
|
-0.007 |
0.051 |
|
|
-0.280 |
2.041 |
|
|
3.951 |
0.022 |
-0.049 |
0.054 |
2.462 |
0.854 |
-1.538 |
1.759 |
|
|
0.329 |
0.015 |
0.029 |
0.033 |
0.266 |
0.144 |
-0.035 |
0.148 |
|
40% |
2.038 |
0.030 |
0.038 |
0.049 |
1.737 |
1.341 |
-0.263 |
1.366 |
|
|
2.916 |
0.039 |
-0.085 |
0.093 |
2.204 |
0.752 |
-0.796 |
1.096 |
|
|
3.024 |
0.022 |
0.024 |
0.033 |
3.832 |
2.858 |
0.832 |
2.976 |
|
|
|
|
-0.008 |
0.052 |
|
|
-0.360 |
1.469 |
|
|
3.938 |
0.023 |
-0.062 |
0.066 |
3.264 |
0.583 |
-0.722 |
0.928 |
|
|
0.352 |
0.016 |
0.052 |
0.054 |
0.304 |
0.047 |
-0.210 |
0.215 |
|
50% |
2.042 |
0.032 |
0.042 |
0.053 |
1.640 |
4.560 |
4.567 |
6.454 |
|
|
2.910 |
0.044 |
-0.090 |
0.100 |
1.944 |
0.289 |
-0.731 |
0.786 |
|
|
3.027 |
0.024 |
0.027 |
0.036 |
3.748 |
3.781 |
3.643 |
5.250 |
|
|
|
|
-0.006 |
0.062 |
|
|
1.309 |
2.727 |
Table 4.
Parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE), biases, and root mean squared errors (RMSE) for both Bayesian and MLE methods were evaluated at different censoring levels () of 10%, 20%, 30%, 40%, and 50% for bathtub hazard type.
Table 4.
Parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE), biases, and root mean squared errors (RMSE) for both Bayesian and MLE methods were evaluated at different censoring levels () of 10%, 20%, 30%, 40%, and 50% for bathtub hazard type.
|
|
Bayesian |
MLE |
Parameter |
|
Estimate |
SD |
Bias |
RMSE |
Estimate |
SE |
Bias |
RMSE |
|
|
0.724 |
0.016 |
0.024 |
0.029 |
0.852 |
0.713 |
0.152 |
0.729 |
|
|
0.892 |
0.036 |
-0.108 |
0.114 |
2.094 |
1.288 |
1.094 |
1.690 |
|
10% |
0.556 |
0.018 |
-0.044 |
0.048 |
1.370 |
0.522 |
0.770 |
0.930 |
|
|
1.010 |
0.011 |
0.010 |
0.015 |
0.760 |
0.941 |
-0.241 |
0.972 |
|
|
0.941 |
0.015 |
-0.060 |
0.061 |
0.184 |
0.103 |
-0.816 |
0.823 |
|
|
|
|
-0.035 |
0.053 |
|
|
0.192 |
1.029 |
|
|
0.710 |
0.016 |
0.010 |
0.019 |
0.831 |
0.574 |
0.131 |
0.589 |
|
|
0.897 |
0.037 |
-0.103 |
0.110 |
0.873 |
0.647 |
-0.127 |
0.659 |
|
20% |
0.576 |
0.020 |
-0.024 |
0.031 |
0.818 |
0.598 |
0.218 |
0.637 |
|
|
1.011 |
0.011 |
0.011 |
0.016 |
1.166 |
1.339 |
0.166 |
1.349 |
|
|
0.969 |
0.012 |
-0.032 |
0.034 |
0.696 |
0.330 |
-0.304 |
0.448 |
|
|
|
|
-0.027 |
0.042 |
|
|
0.017 |
0.736 |
|
|
0.685 |
0.015 |
-0.015 |
0.021 |
0.344 |
0.684 |
-0.356 |
0.771 |
|
|
0.889 |
0.043 |
-0.111 |
0.119 |
1.381 |
7.424 |
0.381 |
7.434 |
|
30% |
0.597 |
0.022 |
-0.003 |
0.022 |
0.821 |
3.444 |
0.221 |
3.451 |
|
|
0.997 |
0.012 |
-0.003 |
0.012 |
1.878 |
16.820 |
0.878 |
16.843 |
|
|
0.968 |
0.012 |
-0.032 |
0.034 |
0.403 |
0.427 |
-0.598 |
0.734 |
|
|
|
|
-0.033 |
0.042 |
|
|
0.105 |
5.847 |
|
|
0.680 |
0.017 |
-0.020 |
0.026 |
0.457 |
0.515 |
-0.243 |
0.570 |
|
|
0.885 |
0.044 |
-0.115 |
0.123 |
1.148 |
2.762 |
0.148 |
2.766 |
|
40% |
0.606 |
0.023 |
0.006 |
0.024 |
0.662 |
1.520 |
0.062 |
1.521 |
|
|
0.986 |
0.014 |
-0.014 |
0.020 |
1.085 |
3.978 |
0.085 |
3.979 |
|
|
0.974 |
0.014 |
-0.026 |
0.030 |
0.598 |
0.423 |
-0.402 |
0.584 |
|
|
|
|
-0.034 |
0.045 |
|
|
-0.070 |
1.884 |
|
|
0.653 |
0.021 |
-0.047 |
0.052 |
0.344 |
1.623 |
1.187 |
2.011 |
|
|
0.901 |
0.038 |
-0.099 |
0.106 |
1.381 |
0.110 |
-0.384 |
0.400 |
|
50% |
0.651 |
0.031 |
0.051 |
0.060 |
0.821 |
0.978 |
0.219 |
1.002 |
|
|
0.969 |
0.017 |
-0.031 |
0.035 |
1.878 |
0.355 |
-0.631 |
0.724 |
|
|
0.997 |
0.018 |
-0.003 |
0.018 |
0.403 |
1.272 |
0.413 |
1.337 |
|
|
|
|
-0.026 |
0.054 |
|
|
0.161 |
1.095 |
Table 5.
Parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE) for the student dropout dataset.
Table 5.
Parameter estimates, standard deviations (SD) for Bayesian estimates, standard errors (SE) for Maximum Likelihood Estimates (MLE) for the student dropout dataset.
|
Bayesian |
MLE Estimate |
Parameter |
Estimate |
SD |
Estimate |
SE |
|
0.233 |
0.008 |
0.536 |
0.137 |
|
2.149 |
0.002 |
0.704 |
0.149 |
|
1.149 |
0.002 |
0.720 |
0.154 |
|
0.635 |
0.005 |
0.759 |
0.129 |
|
0.358 |
0.003 |
1.190 |
0.196 |
Table 6.
Goodness of fit for for the five-parameter exponentiated-Weibull Weibull (EWW) alongside the Exponential Weibull (EW), Weibull Weibull (WW), and Weibull Exponential (WE) distributions for the student dropout dataset.
Table 6.
Goodness of fit for for the five-parameter exponentiated-Weibull Weibull (EWW) alongside the Exponential Weibull (EW), Weibull Weibull (WW), and Weibull Exponential (WE) distributions for the student dropout dataset.
Model |
LogLikelihood |
AIC |
BIC |
EWW |
-750.75 |
1511.5 |
1532.57 |
WW |
-1799.22 |
3606.44 |
3623.3 |
WE |
-1859.54 |
3725.09 |
3737.73 |
EW |
-1029.02 |
2064.04 |
2076.68 |
Table 7.
Median, 25-percentile, 75-percentile and Interquartile range (IQR) of student dropout times (in months) for the five-parameter exponentiated-Weibull Weibull (EWW) alongside the Exponential Weibull (EW), Weibull Weibull (WW), and Weibull Exponential (WE) distributions.
Table 7.
Median, 25-percentile, 75-percentile and Interquartile range (IQR) of student dropout times (in months) for the five-parameter exponentiated-Weibull Weibull (EWW) alongside the Exponential Weibull (EW), Weibull Weibull (WW), and Weibull Exponential (WE) distributions.
|
EWW |
WW |
WE |
EW |
25% |
0.3 |
1.6 |
1.5 |
1.2 |
Median
|
0.8 |
2.0 |
1.9 |
2.1 |
75% |
1.6 |
2.5 |
2.5 |
3.1 |
IQR |
1.3 |
0.9 |
1.0 |
1.9 |