Preprint
Article

Eigenvalue Distributions in Random Confusion Matrices: Applications to Machine Learning Evaluation

Altmetrics

Downloads

127

Views

54

Comments

0

A peer-reviewed article of this preprint also exists.

Submitted:

13 April 2024

Posted:

16 April 2024

You are already at the latest version

Alerts
Abstract
In this paper, we derive the distribution of eigenvalues for a 2×2 random confusion matrix stemming from a machine learning (ML) evaluation problem. Additionally, we present the distributions of both the matrix’s trace and the difference between the two traces of random confusion matrices. We also illustrate the application of these derived distributions in calculating the superiority probability of four baseline ML models.
Keywords: 
Subject: Computer Science and Mathematics  -   Probability and Statistics

MSC:  60E05; 62H30

1. Introduction

The distribution of eigenvalues of a confusion matrix is an interesting and important concept in machine learning (ML), particularly in the evaluation of classification models [1]. Confusion matrices are widely used to assess the performance of a classification algorithm by providing a detailed breakdown of the predicted and actual class labels [2]. The eigenvalues of a confusion matrix offer insights into the underlying structure and characteristics of the classification results [3,4,5]. Eigenvalues are a mathematical concept used to analyze linear transformations, and in the context of confusion matrices, they can reveal information about the matrix’s behaviour [6,7]. The distribution of eigenvalues provides a quantitative measure of the spread and concentration of information in the matrix. Understanding the distribution of eigenvalues of a confusion matrix can be valuable for various purposes, including model assessment, variable selection, high-dimensional analysis, dimension reduction, model comparison, anomaly detection, and generalization or overfitting issues [1,7,8,9,10,11].
For example, in Ref. [1], the significance of eigenvalue analysis for selecting important features in big data was explored. The authors emphasize the importance of understanding the patterns of eigenvalues in covariance matrices for various analytical purposes, such as model comparison and anomaly detection. They highlight how eigenvalues provide insights into the underlying structure of classification results, contributing to an overall understanding of model performance. In a similar study by [7], the authors utilized eigenvalue analysis in conjunction with principal component analysis (PCA) methods to reduce the dimensionality of big data before exploring the performances of several classification methods. The results of their analysis revealed that the outcomes from eigenvalue and PCA are much superior to those from the linear discriminant analysis (LDA) procedure.
In another study by [11], various eigenvalue-based dimension reduction techniques were compared for high-dimensional analysis. Specifically, the authors investigated the performances of PCA, LDA, and singular value decomposition (SVD). The findings from the study validate the utility of eigenvalue-based dimension reduction techniques in handling high-dimensional data. By comparing the effectiveness of PCA, LDA, and SVD, the research underscores the importance of eigenvalue analysis in addressing the challenges posed by high-dimensional datasets. Moreover, [12] utilized eigenvalue analysis to tackle the generalization error problem in two-layered neural networks for high-dimensional analysis. By leveraging eigenvalue properties, the study aimed to enhance the understanding of how neural networks generalize from training data to unseen data. Eigenvalue analysis in this context provides valuable insights into the behavior and performance of neural networks, particularly in high-dimensional spaces. Ref. [12] approach highlights the significance of incorporating eigenvalue-based techniques in optimizing and refining machine learning models for complex data analysis tasks.
In a different context within high-dimensional analysis, Ref. [10] employed eigenvalue and eigenvector analyses to improve the performance of high-dimensional LDA classifier in spiked covariance model. The author introduced a modified regularized R-LDA that is based on eigenvalue and eigenvector analyses. Numerical simulations, using both real and simulated data, revealed that the proposed classifier yields better classification performance than the classical R-LDA while requiring lower computational complexity. In a similar context, Ref. [9] increased the performance of support vector machine (SVM) by employing eigenvalue analysis of the features covariance matrices and subsequently performing PCA to reduce the dimension of the features. This approach helps to increase the prediction accuracy of hepatitis disease.
In a Bayesian analysis of confusion matrices, Ref. [13] delve into Bayesian methods for analyzing confusion matrices in machine learning. While Bayesian approaches are widely used in various aspects of ML, their application to confusion matrices provides a unique perspective on model evaluation. Ref. [13] provided Bayesian interpretations of various evaluation metrics derived from confusion matrices of machine learning models. The authors presented posterior distributions for these metrics from the confusion matrix and used them to compare the performances of several ML models.
The findings of various studies reviewed indicate a significant body of work on eigenvalue analysis within the context of dimension reduction, particularly in Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA). However, there is a notable gap in research concerning eigenvalue analysis of confusion matrices arising from machine learning (ML) models. In high-dimensional analysis and variable selection, dimension reduction serves as a filtering mechanism wherein techniques like eigenvalue analysis are employed to select important variables before training a classification model. Many authors, including [2,14,15,16,17,18,19,20,21], among others, have criticized this approach. They argue that it eliminates the possibility of interaction effects present in variables. Therefore, embedded and wrapper variable selection methods, which combine selection techniques and ML models, are preferred. In this regard, comparing ML models based on confusion matrices from trained model will be beneficial than covariance matrix of a pre-trained model.
Moreover, by leveraging eigenvalue analysis, researchers can objectively compare different machine learning models, discerning their relative strengths and weaknesses based on the underlying structure of their confusion matrices [22,23,24,25]. Hence, this paper presents the distribution of eigenvalues for a 2 × 2 random confusion matrix arising from a machine learning evaluation scenario. Furthermore, we provide distributions for both the matrix’s trace and the difference between the traces of two random confusion matrices. We also demonstrate how these distributions can be utilized to compute the superiority probability of ML models.

2. Distribution of Eigenvalues of Random Confusion Matrix

Suppose we have a learning problem given by data D = { X i , Y i } , where i 1 , 2 , , n , X i is the matrix of the features, and Y i is the response vector which we assume to be categorical with k classes. For simplicity, we consider the binary case with k = 2 as the derivation in this paper can be easily generalized to the multiclass k classes. In any binary classification problem, the goal is to predict the Y i based on new information x X using a classifier y ^ : f ( x ) . Consider a testing dataset denoted as T = { ( X i , Y i ) } i = 1 n T , comprising n T independent samples drawn from an unknown distribution F ( X , Y ) . To assess the accuracy of predictions made by y ^ on the samples in T, we introduce a loss function L : y × y ^ { a , b , c , d } . Let y belong to { θ 0 , θ 1 } as the true class, and y ^ belong to { θ 0 , θ 1 } as the predicted class. Following convention in [13], we define the mapping of the L function as follows:
L = a , if y = θ 1 and y ^ = θ 1 b , if y = θ 1 and y ^ = θ 0 c , if y = θ 0 and y ^ = θ 1 d , if y = θ 0 and y ^ = θ 0
where a denotes true positive, b denotes false negative, c denotes false positive and d denotes true negative. The elements of vector L can be presented in a 2 × 2 matrix often referred to as a confusion matrix. Let A represents the 2 × 2 confusion matrix obtained from a classification learning problem defined above; A can be defined as:
A = a b c d
The obvious properties of A are that (a.) it is not symmetric (b.) it is square and it is also random. Now, if we assume A is diagonalizable such that there exist a scalar λ and vector V that we can use to decompose A using
A V = λ V
then λ = { λ 1 , λ 2 } and V = v 11 v 12 v 13 v 13 are the eigenvalues and eigenvectors of A respectively. One interesting property of eigenvalues of this type of diagonalizable square matrix is that the sum of the eigenvalues equals the trace of the matrix. That is;
t r ( A ) = j = 1 2 λ j .
The t r ( A ) is very useful in evaluating the accuracy of a classifier in a machine learning problem most especially if the categories of the response variable is balance that is p k = 1 / k . In a balance binary classification problem with n T test cases, the accuracy ( ϕ ^ ) of a classifier can be computed using:
ϕ ^ = n T 1 t r ( A ) = n T 1 j = 1 2 λ j .
Note that, since elements of A are resultants of random outcomes of randomly generated test instances used to validate classifier y ^ , A can be regarded as a random matrix. Also, since only n T is the only known parameter, the elements of A can be assumed to be multinomially distributed with parameters n T , π a , π b , π c and π d . Thus, the joint density function of the elements in the random matrix A can be given as:
P ( a , b , c , d | n T , π a , π b , π c , π d ) = n T ! a ! · b ! · c ! · d ! · a π a · b π b · c π c · d π d · I ( a + b + c + d = n T ) .
The last part of the RHS of (6) implies that it is required that the total sum of the four cells be equal to the number of test instances for it to be a proper pdf.
Theorem 1.
The joint probability density function (pdf) of the eigenvalues ( λ 1 , λ 2 ) of a 2 × 2 confusion matrix is given by:
f ( λ 1 , λ 2 ) = 1 4 s 2 π e 1 2 s 2 ( λ 1 2 + λ 2 2 2 A ¯ ( λ 1 + λ 2 ) + 2 A ¯ 2 ) | λ 1 λ 2 | λ 1 , λ 2 .
Proof. 
We begin this proof by standardizing the element of the confusion matrix A as follows:
z = s 1 ( A A ¯ ) ,
where A ¯ is the mean of all elements in A and s is the standard deviation of each element from their mean. If the confusion matrix is balanced such that p k = 1 / 4 for all four elements, the mean A ¯ and standard deviation s are n 4 and 3 n 16 respectively. If otherwise, the mean A ¯ and standard deviation s are computed as follows
A ¯ = a + b + c + d 4 , s = ( a A ¯ ) 2 + ( b A ¯ ) 2 + ( c A ¯ ) 2 + ( d A ¯ ) 2 3 .
The next step involves the symmetrization of z to achieve symmetry as expected for a Gaussian Orthogonal Ensemble ( G O E ) [26,27].
z s = ( z + z T ) / 2 .
where z s is the standardized symmetrized confusion matrix, z is the standardized confusion matrix and z T is it transpose. The elements of z s is explicitly defined as:
z s = a b b d .
Now, that we have established that z s is a G O E with joint p d f of ( a , b , d ) given by
f ( a , b , d ) = 1 2 π e t r ( z s 2 ) 2 ; a , b , d ,
we can proceed to derive the distribution of eigenvalues of z s and subsequently the distribution of eigenvalues of A . Note that by using the change of variable rule, the distribution of eigenvalues ( η 1 , η 2 ) of z s is given by
f ( η 1 , η 2 ) = f ( a , b , d ) | d e t ( J ) | ,
where J is the change of variable Jacobian matrix. Thus, since z s matrix is invariant under orthogonal transformation such that
z s = P T z s η P
where P = cos ( θ ) sin ( θ ) sin ( θ ) cos ( θ ) is an orthogonal matrix and z s η = η 1 0 0 η 2 is a diagonal matrix of the eigenvalues of matrix z s , we have
a b b d = cos ( θ ) sin ( θ ) sin ( θ ) cos ( θ ) η 1 0 0 η 2 cos ( θ ) sin ( θ ) sin ( θ ) cos ( θ ) = η 1 cos 2 ( θ ) + η 2 sin 2 ( θ ) ( η 1 η 2 ) sin ( θ ) ( η 1 η 2 ) sin ( θ ) η 1 sin 2 ( θ ) + η 2 cos 2 ( θ ) .
As we’re moving from z s to z s η , it is required to normalized the resultant p d f of ( η 1 , η 2 ) using the Jacobian determinant det ( J ) . The change of variable Jacobian J is given as
J = a η 1 a η 2 a θ d η 1 d η 2 d θ b η 1 b η 2 b θ = cos 2 ( θ ) sin 2 ( θ ) ( η 2 η 1 ) sin ( 2 θ ) sin 2 ( θ ) cos 2 ( θ ) ( η 1 η 2 ) sin ( 2 θ ) 1 2 sin ( 2 θ ) 1 2 sin ( 2 θ ) ( η 1 η 2 ) cos ( 2 θ ) .
Subsequently, the determinant of the Jacobian is given by
det ( J ) = det cos 2 ( θ ) sin 2 ( θ ) ( η 2 η 1 ) sin ( 2 θ ) sin 2 ( θ ) cos 2 ( θ ) ( η 1 η 2 ) sin ( 2 θ ) 1 2 sin ( 2 θ ) 1 2 sin ( 2 θ ) ( η 1 η 2 ) cos ( 2 θ ) = cos 2 ( θ ) cos 2 ( θ ) ( η 1 η 2 ) sin ( 2 θ ) 1 2 sin ( 2 θ ) ( η 1 η 2 ) cos ( 2 θ ) sin 2 ( θ ) sin 2 ( θ ) ( η 1 η 2 ) sin ( 2 θ ) 1 2 sin ( 2 θ ) ( η 1 η 2 ) cos ( 2 θ ) + ( η 2 η 1 ) sin ( 2 θ ) sin 2 ( θ ) cos 2 ( θ ) 1 2 sin ( 2 θ ) 1 2 sin ( 2 θ ) = ( η 1 η 2 ) ( cos 2 ( 2 θ ) + sin 2 ( 2 θ ) ) = η 1 η 2 .
Therefore, the corresponding joint p d f of ( η 1 , η 2 ) for matrix z s is given by
f ( η 1 , η 2 ) = 1 4 π e 1 2 ( η 1 2 + η 2 2 ) | η 1 η 2 | ; η 1 , η 2 .
Now that we have the distribution of the eigenvalues for the transformed matrix z s , we can obtain the distribution of eigenvalues for the required confusion matrix A as follows
A = s z s + A ¯ .
From (19), it can be seen that there is a one-one correspondence between matrices A and z s , thus, we can define the eigenvalues of A as a function of eigenvalues of z s . This implies
λ = s η + A ¯ .
where λ = ( λ 1 , λ 2 ) and η = ( η 1 , η 2 ) . Therefore, the joint p d f of eigenvalues of A is given by
f ( λ 1 , λ 2 ) = f η 1 , η 2 ( λ 1 , λ 2 ) | d η d λ | = 1 4 π e 1 2 s 2 ( λ 1 2 + λ 2 2 2 A ¯ ( λ 1 + λ 2 ) + 2 A ¯ 2 ) | λ 1 λ 2 s | | 1 s | f ( λ 1 , λ 2 ) = 1 4 s 2 π e 1 2 s 2 ( λ 1 2 + λ 2 2 2 A ¯ ( λ 1 + λ 2 ) + 2 A ¯ 2 ) | λ 1 λ 2 | ; λ 1 , λ 2 ,
where η 1 2 + η 2 2 = λ 1 A ¯ s 2 + λ 2 A ¯ s 2 , η 1 η 2 = λ 1 A ¯ s λ 2 A ¯ s and d η d λ = 1 s .
Remark 1.
Equation (21) implies f ( λ 1 , λ 2 ) is a shifted G O E with mean and variance A ¯ and s 2 respectively.

2.1. Distribution of trace of a random confusion matrix

Theorem 2.
The probability density function (pdf) of the trace t = t r ( A ) of a 2 × 2 random confusion matrix A is given by:
f ( t ) = 1 4 π s 2 e 1 4 s 2 ( t 2 A ¯ ) 2 ; t ,
Proof. 
Again, considering the standardized symmetrized confusion matrix z s defined in (11). The eigenvalues ( η 1 , η 2 ) of z s can be estimated from the characteristics equation
η 2 ( a + d ) η + ( a d b 2 ) = 0 .
Solving (23), gives
η 1 = ( a + d ) + ( a + d ) 2 4 b 2 2 , η 2 = ( a + d ) ( a + d ) 2 4 b 2 2 .
Recall that, the trace ( w ) for matrix z s is given by
t r ( z s ) = η 1 + η 2 w = η 1 + η 2 = ( a + d ) + ( a + d ) 2 4 b 2 2 + ( a + d ) ( a + d ) 2 4 b 2 2 w = a + d
Lemma 1.
Suppose matrix z s is a G O E , thus its elements ( a , d ) are independent and identically distributed as normal, N ( 0 , 1 ) , and b is distributed normally as N ( 0 , 1 / 2 ) .
Remark 2.
Lemma (1) implies that the distribution of the trace of the standardized symmetrized matrix z s is the sum of two normal distributions denoted by N ( 0 , 2 ) . Thus,
f ( w ) = 1 2 π e w 2 / 4 ; w .
Again, by change of variable, we can derive the distribution of the trace of matrix A as follows
f ( t ) = f w ( t ) | d w d t | = 1 2 π e ( t 2 A ) 2 / 4 s 2 ¯ | d w d t | = 1 2 π e ( t 2 A ¯ ) 2 / 4 s 2 | 1 s | f ( t ) = 1 4 π s 2 e ( t 2 A ¯ ) 2 / 4 s 2 ; t
Remark 3.
Equation (27) implies f ( t ) is normally distributed with mean and variance 2 A ¯ and 2 s 2 respectively and it is denoted by N ( 2 A ¯ , 2 s 2 ) .
Lemma 2.
The cumulative distribution function F ( t ) for the trace of matrix A is given by
F ( t ) = t f ( t ) d t = t 1 4 π s 2 e ( t 2 A ¯ ) 2 / 4 s 2 d t F ( t ) = Φ t 2 A ¯ 2 s 2 ,
where Φ is the c d f of standardized normal distribution with mean 0 and variance 1.
Figure 1. Graphs of the p d f of the trace of a random 2 × 2 confusion matrix for different diagonal probabilities π 1 and π 4 .
Figure 1. Graphs of the p d f of the trace of a random 2 × 2 confusion matrix for different diagonal probabilities π 1 and π 4 .
Preprints 103848 g001
Figure 2. Graphs of the c d f of the trace of a random 2 × 2 confusion matrix for different diagonal probabilities π 1 and π 4 .
Figure 2. Graphs of the c d f of the trace of a random 2 × 2 confusion matrix for different diagonal probabilities π 1 and π 4 .
Preprints 103848 g002

2.2. Distribution of Difference of Two Traces of Random Confusion Matrices

In machine learning, it is often valuable to compare the confusion matrices of two algorithms, such as decision trees and random forests [2,14,15]. Understanding the distribution of differences is crucial because it quantifies the degree of superiority one algorithm holds over the other. Therefore, in this section, we have developed the distribution of differences between two sets of 2 × 2 random confusion matrices.
Theorem 3.
The probability density function ( p d f ) of the difference of two traces of 2 × 2 random confusion matrices A and B denoted by m = t r ( A ) t r ( B ) is given by:
f ( m ) = 1 4 π S A + B 2 e 1 4 S A + B 2 ( m 2 A ¯ + 2 B ¯ ) 2 ; m ,
where S A + B 2 = S A 2 + S B 2 .
Proof. 
This proof follows from the earlier distribution of t which follows N ( 2 A ¯ , 2 s 2 ) .
Lemma 3.
Suppose the traces of matrices A and B are independently distributed normal N ( 2 A ¯ , 2 s 2 ) , then the distribution of their difference is also normal with mean 2 A ¯ 2 B ¯ and variance S A + B 2 = S A 2 + S B 2 .
Remark 4.
Lemma (3) implies that the p d f of the difference of two traces of 2 × 2 random confusion matrices A and B is N ( 2 A ¯ 2 B ¯ , S A + B 2 ) .
Thus, the p d f of the difference of two traces of 2 × 2 random confusion matrices A and B denoted by m = t r ( A ) t r ( B ) is given by:
f ( m ) = 1 4 π S A + B 2 e 1 4 S A + B 2 ( m 2 A ¯ + 2 B ¯ ) 2 ; m ,
Lemma 4.
The cumulative distribution function F ( m ) for the difference of two traces of 2 × 2 random confusion matrices A and B denoted by m = t r ( A ) t r ( B ) is given by
F ( m ) = m f ( m ) d m = m 1 4 π S A + B 2 e 1 4 S A + B 2 ( m 2 A ¯ + 2 B ¯ ) 2 d m F ( m ) = Φ m 2 A ¯ + 2 B ¯ S A + B 2 ,
where Φ is the c d f of standardized normal distribution with mean 0 and variance 1.
Figure 3. Graphs of the p d f of for the difference of two traces of 2 × 2 random confusion matrices for different effect size δ = t r ( A ) t r ( B ) .
Figure 3. Graphs of the p d f of for the difference of two traces of 2 × 2 random confusion matrices for different effect size δ = t r ( A ) t r ( B ) .
Preprints 103848 g003
Figure 4. Graphs of the c d f of for the difference of two traces of 2 × 2 random confusion matrices for different effect size δ = t r ( A ) t r ( B )
Figure 4. Graphs of the c d f of for the difference of two traces of 2 × 2 random confusion matrices for different effect size δ = t r ( A ) t r ( B )
Preprints 103848 g004

3. Example

To demonstrate our approach, let’s examine an example featuring two classifiers, A and B, generating the following confusion matrices on the identical testing dataset T, where the size of T is n T = 200 :
A = 62 36 51 51 B = 50 53 50 47 .
The eigenvalues of matrices A and B are denoted by ( λ 1 A , λ 2 A ) and ( λ 1 B , λ 2 B ) respectively. Correspondingly, the traces of matrices A and B can be computed as follows:
t r ( A ) = λ 1 A + λ 2 A t r ( B ) = λ 1 B + λ 2 B .
The estimates for the eigenvalues and traces of matrices A and B are as follows: ( λ 1 A = 99.7 , λ 2 A = 13.3 , tr ( A ) = 113 ) and ( λ 1 B = 100 , λ 2 B = 3 , tr ( B ) = 97 ) respectively. With these trace values, we can compute the accuracies of the two classifiers: ( ϕ A = 0.57 , ϕ B = 0.49 ) . According to this criterion, it seems that classifier A outperforms B. However, without sufficient information, we cannot conclusively determine whether this superiority is genuine or merely a result of chance. By analyzing the distribution of the difference between the two traces, as shown in (30) and (31), we can quantify the extent to which classifier A is superior to classifier B. Therefore, the probability that classifier A genuinely outperforms B is given by:
P [ ( ϕ A = 0.57 ϕ B = 0.49 ) > 0 ] = 1 P [ ( ϕ A = 0.57 ϕ B = 0.49 ) < 0 ] = 1 P ( m 0 ) = 1 F ( m ) = 1 Φ m 2 A ¯ + 2 B ¯ S A + B 2 = 1 Φ m 0 S A + B 2 = 1 Φ 0.08 0.0775 = 0.8492 .
This estimated probability value suggests a strong likelihood that model A significantly surpasses model B in terms of accuracy performance.

4. Applications

We utilize the following datasets to demonstrate the practical application of analyzing the distribution of differences between two traces of random confusion matrices in machine learning, particularly within the field of medicine and health:
  • Heart disease [28]: This dataset comprises information from 303 patients with heart disease at Cleveland Hospital, including 14 features. The objective is to determine the presence or absence of heart disease.
  • Breast cancer [29]: Originating from the University Medical Centre, Institute of Oncology, Ljubljana, Yugoslavia, this dataset contains data from 286 patients with breast cancer, encompassing 9 features. The goal is to predict the presence or absence of breast cancer recurrence.
  • Liver disease [30]: This dataset consists of 584 patient records from the NorthEast region of Andhra Pradesh, India, across 10 features. The objective is to predict whether a patient has liver disease using various biochemical markers.
The aim of this section is to implement and compare four baseline machine learning algorithms applied to these datasets: logistic regression (LR), decision trees (DT), random forest classification (RF), and XGboost classification (XG) [31]. The evaluation criterion utilized to compare the ML algorithms is accuracy. In addition, the supremacy of each of the algorithms is computed by computing the probability distribution in (31). Note that this probability can empirically be computed by bootstrapping the dataset L times and then obtaining the empirical distribution of the difference between two accuracies or traces. Thus, the approximate bootstrap estimate of (31) is
F ^ ( m ) = L 1 l = 1 L ( m l < m ) ,
where L = 5000 is set as the bootstrap sample size. The significance of the approach presented in this study lies in its provision of a closed-form solution for this distribution. This solution offers a faster and more accurate method for calculating the distribution of differences between two accuracies.
Table 1 presents bootstrap accuracy estimates, denoted as ϕ ^ L , along with their standard errors, S E ( ϕ ^ L ) , and accuracy estimates for eigenvalue distribution, denoted as ϕ ^ λ , along with their standard errors, S E ( ϕ ^ λ ) , for the three datasets using the four baseline ML methods. The results indicate that the accuracy estimates and associated standard errors using both the bootstrap and eigenvalue distribution approaches are similar across the machine learning (ML) methods and datasets. This finding empirically validates the eigenvalue distribution approach for estimating the accuracy of an ML method based on the eigenvalue of a confusion matrix.
Table 2 presents pairwise comparison results of the accuracies of the four ML methods using both bootstrap and eigenvalue distribution approaches. Again, the estimates of the pairwise differences are similar in most cases in terms of direction (positive or negative). However, significant differences exist in the estimates of the superiority probability between the bootstrap and eigenvalue distribution approaches. On average, the results are approximately similar for positive differences but exhibit distinct differences for negative differences. The bootstrap tends to be conservative on average when the difference between the accuracies of two ML methods is negative but restrictive when the difference is positive. It is worth noting that bootstrap estimates are approximations to the distribution of the difference of ML accuracy, while the eigenvalue distribution provides the actual distribution of the difference based on Theorem (3). Thus, the results of the superiority probability obtained using the eigenvalue distribution are more reliable than bootstrap estimates, which have been reported in previous studies to have potentially biased estimates [32,33].
In terms of ML performance based on superiority probability, XG is on average better than LR and DT, while RF is on average better than XG. Thus, RF emerges as the best among the four ML methods across the three datasets in terms of prediction accuracy and superiority of accuracy on several replications of the experiment.

5. Conclusion

In conclusion, our study presents bootstrap accuracy estimates and accuracy estimates for eigenvalue distribution across various machine learning (ML) methods and datasets. The findings suggest that both approaches yield similar accuracy estimates and standard errors, validating the effectiveness of the eigenvalue distribution method for ML accuracy estimation based on confusion matrix eigenvalues. Pairwise comparison results reveal consistent estimates of differences between ML methods, yet significant variations exist in superiority probability estimates between bootstrap and eigenvalue distribution approaches. Notably, the bootstrap method tends to be conservative for negative differences and restrictive for positive ones. This underscores the importance of considering the actual distribution provided by the eigenvalue approach for more reliable superiority probability assessments.

Author Contributions

Conceptualization, O.R.O., A.R.R.A., M.R.A; methodology, O.R.O., A.R.R.A.; software, O.R.O.; validation, M.R.A, O.R.O. and A.R.R.A.; formal analysis, O.R.O.; investigation, M.R.A, O.R.O. and A.R.R.A.; resources, M.R.A. and A.R.R.A.; data curation, O.R.O.; writing—original draft preparation, O.R.O.; writing—review and editing, M.R.A, O.R.O. and A.R.R.A.; visualization, O.R.O; supervision, O.R.O.; project administration, O.R.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The authors confirm that the data supporting the findings of this study are available within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, R.C.; Dewi, C.; Huang, S.W.; Caraka, R.E. Selecting critical features for data classification based on machine learning methods. Journal of Big Data 2020, 7, 52. [Google Scholar] [CrossRef]
  2. Olaniran, O.R.; Abdullah, M.A.A. Bayesian weighted random forest for classification of high-dimensional genomics data. Kuwait Journal of Science 2023, 50, 477–484. [Google Scholar] [CrossRef]
  3. Koço, S.; Capponi, C. On multi-class classification through the minimization of the confusion matrix norm. Asian Conference on Machine Learning. PMLR, 2013, pp. 277–292.
  4. García-Balboa, J.L.; Alba-Fernández, M.V.; Ariza-López, F.J.; Rodríguez-Avi, J. Analysis of thematic similarity using confusion matrices. ISPRS international journal of geo-information 2018, 7, 233. [Google Scholar] [CrossRef]
  5. Übeyli, E.D.; Güler, İ. Features extracted by eigenvector methods for detecting variability of EEG signals. Pattern Recognition Letters 2007, 28, 592–603. [Google Scholar] [CrossRef]
  6. Sayyad, S.; Shaikh, M.; Pandit, A.; Sonawane, D.; Anpat, S. Confusion matrix-based supervised classification using microwave SIR-C SAR satellite dataset. Recent Trends in Image Processing and Pattern Recognition: Third International Conference, RTIP2R 2020, Aurangabad, India, January 3–4, 2020, Revised Selected Papers, Part II 3. Springer, 2021, pp. 176–187.
  7. Reddy, G.T.; Reddy, M.P.K.; Lakshmanna, K.; Kaluri, R.; Rajput, D.S.; Srivastava, G.; Baker, T. Analysis of dimensionality reduction techniques on big data. Ieee Access 2020, 8, 54776–54788. [Google Scholar] [CrossRef]
  8. Golub, G.H.; Van Loan, C.F. Matrix computations; JHU press, 2013. [Google Scholar]
  9. Alamsyah, A.; Fadila, T. Increased accuracy of prediction hepatitis disease using the application of principal component analysis on a support vector machine. Journal of Physics: Conference Series 2021, 1968, 012016. [Google Scholar] [CrossRef]
  10. Sifaou, H.; Kammoun, A.; Alouini, M.S. High-dimensional linear discriminant analysis classifier for spiked covariance model. Journal of Machine Learning Research 2020, 21, 1–24. [Google Scholar]
  11. Hasan, S.N.S.; Jamil, N.W. A Comparative Study of Hybrid Dimension Reduction Techniques to Enhance the Classification of High-Dimensional Microarray Data. 2023 IEEE 11th Conference on Systems, Process & Control (ICSPC). IEEE, 2023, pp. 240–245.
  12. Lu, J.; Lu, Y. A priori generalization error analysis of two-layer neural networks for solving high dimensional Schrödinger eigenvalue problems. Communications of the American Mathematical Society 2022, 2, 1–21. [Google Scholar] [CrossRef]
  13. Caelen, O. A Bayesian interpretation of the confusion matrix. Annals of Mathematics and Artificial Intelligence 2017, 81, 429–450. [Google Scholar] [CrossRef]
  14. Olaniran, O.R.; Alzahrani, A.R.R. On the Oracle Properties of Bayesian Random Forest for Sparse High-Dimensional Gaussian Regression. Mathematics 2023, 11, 4957. [Google Scholar] [CrossRef]
  15. Olaniran, O.; Abdullah, M. Subset selection in high-dimensional genomic data using hybrid variational Bayes and bootstrap priors. Journal of Physics: Conference Series. IOP Publishing, 2020, Vol. 1489, p. 012030.
  16. Pudjihartono, N.; Fadason, T.; Kempa-Liehr, A.W.; O’Sullivan, J.M. A review of feature selection methods for machine learning-based disease risk prediction. Frontiers in Bioinformatics 2022, 2, 927312. [Google Scholar] [CrossRef] [PubMed]
  17. Mehmood, T.; Sæbø, S.; Liland, K.H. Comparison of variable selection methods in partial least squares regression. Journal of Chemometrics 2020, 34, e3226. [Google Scholar] [CrossRef]
  18. Chen, C.W.; Tsai, Y.H.; Chang, F.R.; Lin, W.C. Ensemble feature selection in medical datasets: Combining filter, wrapper, and embedded feature selection results. Expert Systems 2020, 37, e12553. [Google Scholar] [CrossRef]
  19. Wang, G.; Sarkar, A.; Carbonetto, P.; Stephens, M. A simple new approach to variable selection in regression, with application to genetic fine mapping. Journal of the Royal Statistical Society Series B: Statistical Methodology 2020, 82, 1273–1300. [Google Scholar] [CrossRef]
  20. Sauerbrei, W.; Perperoglou, A.; Schmid, M.; Abrahamowicz, M.; Becher, H.; Binder, H.; Dunkler, D.; Harrell, F.E.; Royston, P.; Heinze, G.; others. State of the art in selection of variables and functional forms in multivariable analysis—outstanding issues. Diagnostic and prognostic research 2020, 4, 1–18. [Google Scholar] [CrossRef] [PubMed]
  21. Chowdhury, M.Z.I.; Turin, T.C. Variable selection strategies and its importance in clinical prediction modelling. Family medicine and community health 2020, 8. [Google Scholar] [CrossRef] [PubMed]
  22. Peyrache, A.; Rose, C.; Sicilia, G. Variable selection in data envelopment analysis. European Journal of Operational Research 2020, 282, 644–659. [Google Scholar] [CrossRef]
  23. Montoya, A.K.; Edwards, M.C. The poor fit of model fit for selecting number of factors in exploratory factor analysis for scale evaluation. Educational and psychological measurement 2021, 81, 413–440. [Google Scholar] [CrossRef]
  24. Greenacre, M.; Groenen, P.J.; Hastie, T.; d’Enza, A.I.; Markos, A.; Tuzhilina, E. Principal component analysis. Nature Reviews Methods Primers 2022, 2, 100. [Google Scholar] [CrossRef]
  25. Popoola, J.; Yahya, W.B.; Popoola, O.; Olaniran, O.R. Generalized self-similar first order autoregressive generator (gsfo-arg) for internet traffic. Statistics, Optimization & Information Computing 2020, 8, 810–821. [Google Scholar]
  26. Sarkar, A.; Kothiyal, M.; Kumar, S. Distribution of the ratio of two consecutive level spacings in orthogonal to unitary crossover ensembles. Physical Review E 2020, 101, 012216. [Google Scholar] [CrossRef] [PubMed]
  27. Grimm, U.; Römer, R.A. Gaussian orthogonal ensemble for quasiperiodic tilings without unfolding: r-value statistics. Physical Review B 2021, 104, L060201. [Google Scholar] [CrossRef]
  28. Janosi, Andras, S.W.P.M.; Detrano, R. Heart Disease. UCI Machine Learning Repository, 1988. [CrossRef]
  29. Zwitter, M.; Soklic, M. Breast Cancer. UCI Machine Learning Repository, 1988. [CrossRef]
  30. Ramana, B.; Venkateswarlu, N. ILPD (Indian Liver Patient Dataset). UCI Machine Learning Repository, 2012. [CrossRef]
  31. Ding, N.; Sadeghi, P. A submodularity-based agglomerative clustering algorithm for the privacy funnel. arXiv, arXiv:1901.06629 2019.
  32. Navarro, C.L.A.; Damen, J.A.; Takada, T.; Nijman, S.W.; Dhiman, P.; Ma, J.; Collins, G.S.; Bajpai, R.; Riley, R.D.; Moons, K.G.; others. Risk of bias in studies on prediction models developed using supervised machine learning techniques: systematic review. bmj 2021, 375. [Google Scholar]
  33. Tantithamthavorn, C.; McIntosh, S.; Hassan, A.E.; Matsumoto, K. An empirical comparison of model validation techniques for defect prediction models. IEEE Transactions on Software Engineering 2016, 43, 1–18. [Google Scholar] [CrossRef]
Table 1. The bootstrap accuracy estimate ϕ ^ L along with its standard error S E ( ϕ ^ L ) , and the accuracy estimate for eigenvalue distribution ϕ ^ λ along with its standard error S E ( ϕ ^ λ ) , for the three datasets using the four baseline methods.
Table 1. The bootstrap accuracy estimate ϕ ^ L along with its standard error S E ( ϕ ^ L ) , and the accuracy estimate for eigenvalue distribution ϕ ^ λ along with its standard error S E ( ϕ ^ λ ) , for the three datasets using the four baseline methods.
Heart disease Breast cancer Liver disease
ϕ ^ L ϕ ^ λ ϕ ^ L ϕ ^ λ ϕ ^ L ϕ ^ λ
Method ( S E ( ϕ ^ L ) ) ( S E ( ϕ ^ λ ) ) ( S E ( ϕ ^ L ) ) ( S E ( ϕ ^ λ ) ) ( S E ( ϕ ^ L ) ) ( S E ( ϕ ^ λ ) )
LR 0.83 0.88 0.70 0.72 0.71 0.72
(0.031) (0.032) (0.043) (0.031) (0.029) (0.029)
DT 0.77 0.76 0.71 0.73 0.72 0.67
(0.042) (0.025) (0.031) (0.024) (0.031) (0.021)
RF 0.82 0.84 0.82 0.79 0.82 0.80
(0.031) (0.029) (0.025) (0.030) (0.025) (0.030)
XG 0.77 0.82 0.74 0.70 0.75 0.72
(0.036) (0.029) (0.030) (0.031) (0.030) (0.030)
Table 2. Estimates of the difference between pairwise accuracies ( m ^ = ϕ A ϕ B ) and their respective superiority probabilities ( 1 F ( m ^ ) ) using both bootstrap and eigenvalue distribution approaches across the three datasets.
Table 2. Estimates of the difference between pairwise accuracies ( m ^ = ϕ A ϕ B ) and their respective superiority probabilities ( 1 F ( m ^ ) ) using both bootstrap and eigenvalue distribution approaches across the three datasets.
Heart disease Breast cancer Liver disease
m ^ L m ^ λ m ^ L m ^ λ m ^ L m ^ λ
Pair ( 1 F ^ ( m ^ L ) ) ( 1 F ( m ^ λ ) ) ( 1 F ^ ( m ^ L ) ) ( 1 F ( m ^ λ ) ) ( 1 F ^ ( m ^ L ) ) ( 1 F ( m ^ λ ) )
XG - LR -0.05 -0.06 0.04 -0.02 0.03 0.01
(0.058) (0.408) (0.855) (0.473) (0.835) (0.509)
XG - RF -0.05 -0.02 -0.08 -0.08 -0.08 -0.08
(0.055) (0.468) (0.001) (0.367) (0.002) (0.375)
XG - DT 0.00 0.06 0.02 -0.03 0.02 0.05
(0.481) (0.598) (0.728) (0.453) (0.719) (0.588)
LR - RF 0.01 0.04 -0.12 -0.07 -0.11 -0.08
(0.536) (0.561) (0.000) (0.393) (0.000) (0.365)
LR - DT 0.06 0.11 -0.02 -0.01 -0.01 0.04
(0.895) (0.685) (0.300) (0.481) (0.316) (0.579)
RF - DT 0.05 0.08 0.10 0.06 0.10 0.13
(0.888) (0.629) (0.999) (0.595) (0.999) (0.715)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated