Preprint
Article

g.ridge: An R Package for the Generalized Ridge Regression for Sparse and High-Dimensional Linear Models

Altmetrics

Downloads

144

Views

62

Comments

0

A peer-reviewed article of this preprint also exists.

Submitted:

15 January 2024

Posted:

15 January 2024

You are already at the latest version

Alerts
Abstract
Ridge regression is one of the most popular shrinkage estimation methods for linear models. Ridge regression effectively estimates regression coefficients in the presence of high-dimensional regressors. Recently, a generalized ridge estimator was suggested by generalizing the uniform shrinkage of ridge regression to the non-uniform shrinkage, which was shown to perform well under sparse and high-dimensional linear models. In this paper, we introduce our newly developed R package “g.ridge” (the first version published on 2023-12-07 at https://cran.r-project.org/package=g.ridge) that implements both the ridge estimator and generalized ridge estimators. The package equips with the generalized cross-validation for automatic estimation of shrinkage parameters. The package also includes a convenient tool for generating a design matrix. By simulations, we test the performance of the R package under sparse and high-dimensional settings with the normal and skew-normal error distributions. From the simulation results, we conclude that the generalized ridge estimator is superior to the benchmark ridge estimator based on “glmnet”, and hence, it can be the most recommended estimator under sparse and high-dimensional models. We demonstrate the package using the intracerebral hemorrhage data.
Keywords: 
Subject: Computer Science and Mathematics  -   Probability and Statistics

MSC:  62J05; 62J07; 62P10; 62-04

1. Introduction

In linear regression, the least squares estimator (LSE) may not be suitable to estimate regression coefficients if the number of regressors p is higher than the sample size n (i.e., p > n ). Ridge regression (Hoerl and Kennard 1970; Montgomery et al. 2021) is an effective alternative for the high-dimensional ( p > n ) data, and is widely employed in such data encountered in genetics (Arashi et al. 2021; Vishwakarma et al. 2021), medical studies (Friedrich et al. 2023), electromagnetic signals (Gao et al. 2023), grain yields (Hernandez et al., 2015), and others.
Hoerl and Kennard (1970) originally proposed the ridge estimator to reduce the problem in multicollinearity. Later, the ridge estimator is naturally adopted to the high-dimensional ( p > n ) problem (Golub et al., 1979; Hastie et al., 2009) as a way to avoid overfitting phenomenon.
Ridge regression is a shrinkage estimator that shrinks all regression coefficients uniformly toward zero (Hoerl and Kennard 1970; Hastie et al., 2009). This approach is particularly suitable for modeling microarrays or single nucleotide polymorphism (SNP) data, where many coefficients are close to zero (sparse models). For instance, Cule et al. (2011) applied the ridge estimator to the high-dimensional SNP data and performed significance testing for selecting an informative subset of SNPs. Similar applications of ridge regression to high-dimensional data include Whittaker et al. (2000), Cule and De Lorio (2013), and Yang and Emura (2017).
Unlike the ordinary ridge regression that shrinks all regression coefficients uniformly, the generalized ridge regression allows different degrees of shrinkage under multiple shrinkage parameters. Allen (1974) and Loesgen (1990) demonstrated that the multiple shrinkage parameters in the generalized ridge estimator arise naturally by utilizing prior information about regression coefficients; see also an extensive review paper of Yüzbaşı et al. (2020) for generalized ridge regression methods.
However, multiple shrinkage parameters are difficult to be estimated for high-dimensional cases. To overcome this difficulty, Yang and Emura (2017) suggested reducing the number of shrinkage parameters for the case of p > n in their formulation of a generalized ridge estimator. The idea behind their approach is to assign two different weights (1 or 0.5) for shrinkage parameters by preliminary tests (Saleh and Kibria 1993; 2019; Norouzirad and Arashi 2019; Shih et al., 2021;2023; Taketomi et al. 2023). While this approach is shown to be promising due to its sound statistical performance under sparse and high-dimensional models, none of the software packages were implemented for the generalized ridge estimator.
This paper introduces our R package “g.ridge” (https://cran.r-project.org/package=g.ridge) that implements both the ridge estimator (Hoerl and Kennard 1970) and generalized ridge estimator (Yang and Emura 2017). The package equips with the generalized cross-validation for automatic estimation of shrinkage parameters. The package also includes a convenient tool for generating a design matrix. By simulations, we test the performance of the R package under the sparse and high-dimensional models, and the normal and skew-normal distributions for error terms. We conclude that the generalized ridge estimator is superior to the benchmark ridge estimator based on “glmnet”, and hence, it can be the most recommended estimator under sparse and high-dimensional models. We illustrate the package via the intracerebral hemorrhage data.
The remainder of the paper is structured as follows. Section 2 reviews the ridge and generalized ridge estimators. Section 3 introduces the proposed R package. Section 4 tests the package via simulations. Section 5 gives a real data example. Section 6 concludes the paper.

2. Ridge regression and generalized ridge regression

This section introduces the ridge regression method proposed by Hoerl and Kennard (1970) and the generalized ridge regression proposed by Yang and Emura (2017).

2.1. Linear regression

Consider the linear regression model y = X β + ε , where
y = y 1 y n ,     X = x 1 T x n T = x 11 x 1 p x n 1 x n p ,     β = β 1 β p ,     ε = ε 1 ε n ;
X is a design matrix, x i T = ( x i 1 , , x i p ) is the transpose of the vector x i , β is a vector of regression coefficients, and ε is a vector of errors with E [ ε ] =   0 . In some case, we assume C o v [ ε ] = σ 2 I n , where I n is the identify matrix of dimension n . The assumption of the covariance is necessary to obtain the standard error (SE) of regression coefficients and testing their significance. Assume that X is standardized such that i = 1 n x i j = 0 and i = 1 n x i j 2 = c for j = 1 , . . . , p , where c is n or n 1 . Also, we assume that X does not include the intercept term (see Section 3.3 for details). These settings for X are usually imposed in ridge regression (Hastie et al. 2009).
If X T X is invertible (non-singular), the LSE is defined as
β ^ L S E = ( X T X ) 1 X T y .
Clearly, the LSE is not available when X T X is singular, especially when p > n .

2.2. Ridge regression

Ridge estimator is an alternative to the LSE, which can be computed even when p > n . Hoerl and Kennard (1970) originally defined the ridge regression estimator as
β ^ ( λ ) = ( X T X + λ I n ) 1 X T y ,
where λ > 0 is called shrinkage parameter. A diagonal matrix λ I n added to X T X makes all the components of β ^ ( λ ) shrunk toward zero. Theorem 4.3 of Hoerl and Kennard (1970) shows that there exists λ > 0 such that the ridge estimator yields strictly smaller mean squared error (MSE) than that of the LSE.
Golub et al. (1979) suggested choosing λ on the basis of the predicted residual error sum of squares, or Allen’s PRESS (Allen, 1974) statistics. Golub et al. (1979) proposed the rotation-invariant version of the PRESS statistics, and called it as the generalized cross-validation (GCV) criterion (Golub et al.,1979). The GCV function defined by Golub et al. (1979) is
V λ = 1 n | | { I n A ( λ ) } y | | 2 / 1 n Tr I n A λ 2 .
where A ( λ ) = X ( X T X + λ I n ) X T is the “hat matrix”. The GCV estimator of λ is defined as
λ ^ = a r g m i n λ 0 V λ .
Therefore, the resultant ridge estimator is
β ^ ( λ ^ ) = ( X T X + λ ^ I n ) 1 X T y .
The GCV theorem (Golub et al. 1979) guarantees the use of the above ridge estimator under the p > n setup. Noted that the GCV criterion is different from the cross-validation criterion that is available in the widely used R function “cv.glmnet(.)” in the R package “glmnet”. There are many other available criteria for choosing λ (Cule et al. 2013; Wong and Chiu 2015; Kibria and Banik 2016; Assaf et al. 2019; Michimae and Emura 2020), most of which are not applicable for the p > n setup. Therefore, we will adopt the GCV criterion for the following discussions, which works well for both p < n and p > n setups.

2.3. Generalized ridge regression

Yang and Emura (2017) suggests relaxing the uniform shrinkage to the non-uniform shrinkage to yield the generalized ridge regression. For this purpose, Yang and Emura (2017) replaced the identity matrix I n with the diagonal matrix W ^ ( Δ ) (defined later), and proposed
β ^ ( λ , Δ ) = { X T X + λ W ^ Δ } 1 X T y ,
where λ > 0 is a shrinkage parameter and Δ 0 is called the “threshold” parameter. The diagonal components of W ^ ( Δ ) were suggested to be larger values (stronger shrinkage) for the components β closer to zero, yielding W ^ ( Δ ) = d i a g { w ^ 1 ( Δ ) ,   . . . , w ^ p ( Δ ) } , where
w ^ j ( Δ ) = 1 / 2   if z j Δ , 1 if z j < Δ
where z j = β ^ j 0 / S D β ^ 0 , and S D β ^ 0 = j = 1 p ( β ^ j 0 j = 1 p β ^ j 0 / p   ) 2 / p 1 1 / 2 for j = 1 ,   . . . ,   p , and β ^ 0 = (   β ^ 1 0 ,   . . . ,   β ^ p 0   ) T , defined as β ^ j 0 = X j T y / X j T X j , where X j is the j -th row of X . Note that β ^ 0 is called “compound covariate estimator” (Chen and Emura 2017).
The optimal value of ( λ , Δ ) is estimated by the modified GCV function defined as
V λ , Δ = 1 n | | { I n A ( λ , Δ ) } y | | 2 / 1 n Tr I n A λ , Δ 2 ,
where A ( λ , Δ ) = X { X T X + λ W ^ ( Δ ) } 1 X T . Then the estimators ( λ ^ , Δ ^ ) are defined as
λ ^ , Δ ^ = a r g m i n λ 0 ,   Δ 0 V λ , Δ .
Computation of ( λ ^ , Δ ^ ) is not difficult. Given Δ , the function V ( λ , Δ ) is continuous in λ , and hence it is easily minimized using any optimization scheme, such as the R function “optim(.)” to get λ ^ ( Δ ) . Under the sparse model ( β 0 ), the histogram of β ^ j 0 / S D ( β ^ 0 ) , j = 1 ,   . . . ,   p , is well-approximated by N ( 0,1 ) . This implies that | β ^ j 0 | / S D ( β ^ 0 ) falls in the range [ 0,3 ] with nearly 99.73%, and hence, a search range Δ [ 0,3 ] suffices. Since V ( λ ^ ( Δ ) , Δ ) is discontinuous in Δ , a grid search was suggested on the grid D = { 0 ,   3 / 100 , . . . ,   300 / 100 } .
Finally, the generalized ridge estimator is defined as
β ^ ( λ ^ , Δ ^ ) = { X T X + λ ^ W ^ Δ ^ } 1 X T y .
Also, the error variance can be estimated by
σ ^ 2 = y X β ^ λ ^ , Δ ^ 2   / ν ,
where ν = Tr { I n A ( λ ^ , Δ ^ ) } 2 and A ( λ ^ , Δ ^ ) = X { X T X + λ ^ W ^ ( Δ ^ ) } 1 X T .

2.4. Significance test

Ridge and generalized ridge estimators provide methods to test the null hypothesis
H 0 j :   β j = 0   vs .   H 1 j :   β j 0 ,
for j = 1 ,   . . . ,   p . One can perform significance testing, allowing one to access P-values of all the p regressors (Cule et al. 2011; Cule and De Lorio 2013; Yang and Emura 2017). Since the significance tests are similar between the ridge and generalized ridge estimators, we will explain the significance tests based on the generalized ridge estimator below.
Let β ^ j ( λ ^ , Δ ^ ) be the j -th component of β ^ ( λ ^ , Δ ^ ) . As in Cule et al. (2011) and Yang and Emura (2017), we define a Z-value Z j = β ^ j ( λ ^ , Δ ^ ) / S E { ( λ ^ , Δ ^ ) } , where S E { β ^ j ( λ ^ , Δ ^ ) } is the square root of the j -th diagonal of
C o v { β ^ ( λ ^ , Δ ^ ) } = σ ^ 2 { X T X + λ ^ W ^ ( Δ ^ )   } 1 X T X { X T X + λ ^ W ^ ( Δ ^ ) } 1 ,
We define the P-value as P j = 2 { 1 Φ Z j } , where Φ (.) is the cumulative distribution function of N 0 ,   1 . One can then select a subset of regressors by specifying a significance level.

3. R package: g.ridge

We implemented the methods of Section 2 in the R package “g.ridge”. This section explains the main R function “g.ridge(.)“, and the other function “X.mat(.)”. Appendix A explains another function “GCV(.)” that may not be directly used, but is useful for internal computing.

3.1. Generating data

The design matrix X and response vector y are necessary to perform linear regression (Section 2.1). Therefore, following Section 5 of Yang and Emura (2017, p.6093), we implemented a convenient R function that generates X having three independent blocks:
Preprints 96378 i001
where the first block consists of q 0 correlated regressors (correlation = 0.5), and the second block consists of r 0 correlated regressors (correlation = 0.5), and the third block consists of p q r > 0 independent regressors. That is,
C o r r ( x i j ,   x i k ) =   0.5 if j ,   k { 1 , . . . , q } ,   0.5 if j ,   k { q + 1 , . . . , q + r } ,   0 otherwise .
The design matrix mimics the “gene pathway” for two blocks of correlated gene expressions often used in simulation studies (Binder et al. 2009; Emura et al. 2012; 2016; 2023). The values q = 0 and r = 0 give a design matrix with all independent columns.
The marginal distributions of p regressors follow N ( 0,1 ) , which is achieved by
x i T = 1 2 z i 1 + u i ,   . . .   , 1 2 z i q + u i   1 2 z i , q + 1 + v i ,   . . . , 1 2 z i , q + r + v i     z i , q + r + 1 ,   . . . , z i p ,
where z i 1 ,   . . . ,   z i p and u i ,   v i all independently follow N ( 0,1 ) for i = 1 , . . . , n .
Figure 1 shows an example for generating design matrices using “X.mat(.)”. One can simply input n , p , q , and r into “X.mat(.)” by noting the constraint q + r < p .

3.2. Performing regression

The ridge and generalized ridge estimators can be computed by the R function “g.ridge(.)” whose input requires a design matrix X and a response vector y . Specifically, the command “g.ridge(X, Y, method = "HK", kmax = 500)” can calculate the ridge estimator β ^ ( λ ^ ) , where “HK” stands for “Hoerl and Kennard”, and “kmax=500” means the range 0 λ 500 for estimating λ . The output of the command includes λ ^ [ 0 ,   500 ] and β ^ ( λ ^ ) . Similarly, the command “g.ridge(X, Y, method = "YE", kmax = 500)” can calculate β ^ ( λ ^ , Δ ^ ) , where “YE” stands for “Yang and Emura”. The output also displays the plot of V ( λ ) against λ [ 0 ,   500 ] , and its minimizer λ ^ [ 0 ,   500 ] (the plot of V ( λ , Δ ^ ) for the generalized ridge).
The R function “g.ridge(.)” can also compute the SE of β ^ , Z-value and P-value for significance tests (Section 2.4), and the estimate of the error variance σ ^ 2 (Section 2.3). As in the typical practice of ridge regression, we used the centered responses “Y-mean(Y)” rather than “Y” (will be explained in Section 3.3). Also, if “X” were not generated by “X.mat”, the scaled design matrix “scale(X)” would be recommended rather than “X” itself.
Figure 2 shows the code and output for the ridge estimator. The output shows λ ^ = 31.66314 , β ^ λ ^ = 0.581485036 , 0.083260387 , SE, Z-value, P-value, and σ ^ 2 = 1.778618 . The output also displays the GCV function V ( λ ) against λ [ 0 ,   200 ] , showing its minimum at λ ^ = 31.66314 . In the code, we changed “kmax” from the default value (500) to 200 for a better visualization of the plot. In many cases, users will need to try different values of “kmax” to reach a desirable plot for the GCV function.
The output of the generalized ridge estimator is similar to that of the ridge estimator. Only the difference is an additional parameter estimate Δ ^ [ 0 ,   3 ] shown at “delta$”.

3.3. Technical remarks on centering and standardization

We assume that X is standardized and does not include the intercept term (Section 2.1). If X is generated by “X.mat”, it is already standardized, and hence there is no need to do the standardization. However, in other cases, X has to be standardized, e.g. by the R command “scale(X)”. By assumption, the design matrix X does not include the intercept term (a column of ones) since one can always use the reduced model y ( i = 1 n y i / n ) 1 = X β + ε for the intercept model y = β 0 1 + X β + ε ( β 0 is estimated by i = 1 n y i / n ), where 1 is a vector of ones. This means that the usual unbiased estimator is applied to the intercept.

4. Simulations

We conducted simulation studies to test our R package “g.ridge”. The goals of the simulations are to show: (a) the generalized ridge estimators in our package exhibits superior performance over the ridge estimator in the package “glmnet(.)”, and (b) the sound performance under the normally and non-normally distributed errors (the skew-normal distribution will be considered). Supplementary Materials include the R code to reproduce the results of the simulation studies.

4.1. Simulation settings

We generated X by “X.mat” with n = 100 , p { 50 ,   100 ,   150 ,   200 } , and q = r = 10 based on Equation (3). Given X , we set y = X β + ε , with β defined by the sparse model:
β = (   b / q ,   . . . , b / q q ,   d / r ,   . . . , d / r r ,   0 , . . . , 0 p q r ) T ,
for four cases: (I) b = d = 5 ; (II) b = d = 10 ; (III) b = 5 , d = 5 ; (IV) b = 10 , d = 10 . Errors ε were generated independently from the normal distribution, or the skew-normal distribution (Azzalini 2014); both distributions have mean zero and standard deviation one, and the skew-normal distribution has the slant parameter ten (alpha=10 in the R function “rsn(.)”). Figure 3 shows the remarkable difference of the two distributions. The skew-normal distribution was not previously examined in the simulation setting of Yang and Emura (2017).
For a given dataset ( X , y ) , we computed β ^ that can be (i) the ridge estimator by “g.ridge(.)”, (ii) the generalized ridge estimator by “g.ridge(.)”, or (iii) the ridge estimator by “glmnet(.)”. Based on 500 replications (on random errors ε ), the performances of the three proposed estimators were compared by the total mean squared error (TMSE) criterion defined as
T M S E β ^ = E β ^ β 2 = E j = 1 p β ^ j β j 2 ,
where E [ . ] was performed by the Monte Carlo average over 500 replications.

4.2. Simulation results

Table 1 compares the MSE among the three estimators: (i) the ridge by “g.ridge(.)”, (ii) the generalized ridge by “g.ridge(.)”, and (iii) the ridge by “glmnet(.)”. We see that the generalized ridge estimator is superior to the ridge estimator since the former has smaller TMSE values for all cases. Also, the generalized ridge estimator is superior to the “glmnet(.)” estimator for cases of p 100 ,   150 ,   200 , while it is not the case for p = 50 . This conclusion holds consistently across the four parameter settings (I)-(IV) and two error distributions (normal and skew-normal). In conclusion, the generalized ridge estimator in the R proposed package is the most recommended estimator for data with sparse and high-dimensional settings.

5. Data analysis

This section analyzes a real dataset to illustrate the generalized ridge estimator in the proposed package. We retrospectively analyzed a dataset from patients with intracerebral hemorrhage who were hospitalized at Saiseikai Kumamoto Hospital, Kumamoto city, Japan. The study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Ethics Committee of Saiseikai Kumamoto Hospital on March 29, 2023. Saiseikai Kumamoto Hospital is a regional tertiary hospital that serves patients with stroke in southern Japan and provides acute care in a comprehensive stroke care unit.
The outcome variables ( y ) are the changes of the blood volume (in mL) from the initial value and to the follow-up value, which were measured by the CT scan. Excluding patients with less than 10 mL blood volume at the initial CT scan and other inclusion/exclusion criteria, we arrive at n = 172 patients. Regressor variables ( X ) consist of p = 35 continuous measurements, including histological variables (e.g., age), vital signs (e.g., blood pressure, mmHG; respiratory rate, times/minute), and blood measurements (e.g. albumin, g/dL; Gamma-glutamyl transpeptidase (Gamma-GT), IU/L; lactate dehydrogenase, IU/L; Prothrombin time, second; Blood platelet count, 10-3/μL; C-reactive protein, mg/dL). The responses and regressors are centered and standardized before fitting a linear model as explained in Section 2.1 and Section 3.3.
Figure 3 displays scatter plots for the centered responses y ( i = 1 n y i / n ) 1 against the predictors X β ^ based on the ridge estimator and generalized ridge estimator. We observe that the predictors reasonably captured the variability of the changes in the blood volume (response variables). However, the figure also shows that the changes in the blood volumes were not fully explained by the predictors since some residual errors remain. Figure 4 depicts the residuals constructed by the generalized ridge estimator. We observe that the residuals are asymmetric around zero. The asymmetry in the errors does not yield any problem as the proposed ridge estimators was verified for the asymmetric errors (Section 4).
Table 2 shows the fitted results for estimated regression coefficients sorted by P-values (only with P < 0.05). The fitted results are quite similar between the ridge and generalized ridge estimators. The most significant regressor is the lactate dehydrogenase. This variable was previously reported as an important predictor of hematoma expansion (Chu et al. 2019). An interesting difference is found in the number of significant regressors: 6 regressors for the generalized ridge estimator, and 5 regressors for the ridge estimator. That is, the C-reactive protein as a regressor selected solely by the generalized ridge estimator. A study conducted in a large population in China reported that a high C-reactive protein level was an independent risk factor for the severe intracerebral hemorrhage (Wang et al. 2021). The selection of C-reactive protein as a variable associated with increased hematoma volume may reflect the tendency of patients with severe intracerebral hemorrhage to have easier hematoma expansion.
Finally, we compared the performance of the ridge and generalized ridge estimators by means of the residual mean squared error (RMSE) defined as
R M S E = y i = 1 n y i n 1 X β ^ 2 .
The RMSE were 2.288 (the ridge estimator) and 2.284 (the generalize ridge estimator). Therefore, the predictor constructed from the generalized ridge estimator had slightly better performance over the one from the ridge estimator.
Therefore, we have demonstrated that the ridge and generalize ridge estimators in the proposed R package provide statistically and biologically sound conclusions on the real data analysis.

6. Conclusion

This paper introduced the new R package “g.ridge” that can perform the ridge and generalized ridge regression methods. We showed that the generalized ridge estimator in the proposed package performs better than the widely used ridge estimator in the “glmnet” package. Furthermore, the ridge and generalized ridge estimators in the proposed package can deal with asymmetric error distributions. With the abundance of sparse and high-dimensional data (Kim et al. 2007; Zhang et al. 2021; Vishwakarma et al. 2021; Bhattacharjee 2022; Bhatnagar et al. 2023; Emura et al. 2023) and asymmetrically distributed data (Abe et al. 2012; Huynh et al. 2021; Yoshiba et al. 2023; Jimichi et al. 2023), the proposed package may provide a reliable statistical tool for statisticians and data scientists.
The generalized ridge estimator considered in this article may be extended to logistic regression, Poisson regression, and Cox regression. Such extensions have not been explored yet. While the extensions might not be technically difficult, well-designed simulation studies and implementations in some software packages will be necessary to fully justify the advantage and usefulness over the usual ridge estimator that is widespread via the “glmnet” package.

Data Availability Statement:

Supplementary Materials include the R code to reproduce the results of the simulation studies. The dataset used in Section 5 is available from the corresponding author on reasonable request.

Institutional Review Board Statement

The study (of Section 5) was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Ethics Committee of Saiseikai Kumamoto Hospital (Approval No. 1156 and March 29, 2023).

Informed Consent Statement

Given the retrospective nature of the study and the use of anonymized patient data in Section 5, written informed consent was not required. Patients preferred not to be included in this study were given the opportunity to opt out.

Acknowledgements

Emura T is supported financially by the research grant from the JSPS KAKENHI (grant no. 22K11948, and grant no. 20H04147).

Conflicts of Interest

The authors declare that they have no competing interests.

Appendix A. GCV function

The GCV functions are defined as V λ and V ( λ , Δ ) in Equations (1) and (2), respectively. The ridge estimator uses V λ while the generalized ridge estimator uses V ( λ , Δ ) for estimating shrinkage parameters. Therefore, we made an R function “CGV(X,Y,k,W)” to help computing V λ and V ( λ , Δ ) , where “X” is a design matrix, “Y” is a response vector, “k” is actually λ (because “lambda” is a long name). Note that “W” can be any square matrix of dimension p to allow for general use. Thus, what we actually compute in “CGV(X,Y,k,W)” is
V k , W = 1 n | | { I n A ( k , W ) } y | | 2 / 1 n Tr I n A k , W 2 ,
where A ( k , W ) = X { X T X + k W } 1 X T for any symmetric matrix W . However, we normally use W = I n for the ridge or W = W ^ ( Δ ) for the generalized ridge as defined in Section 2.3.
Figure A1 shows the R code for using “CGV(X,Y,k,W)”. The default for “W” is W = I n and therefore “GCV(X,Y,k=1)” can compute
V ( 1 , I n ) = 1 n | | { I n A ( 1 , I n ) } y | | 2 / 1 n Tr { I n A ( 1 , I n ) } 2
for A ( 1 , I n ) = X { X T X + I n } 1 X T , or equivalently,
V 1 = 1 n | | { I n A ( 1 ) } y | | 2 / 1 n Tr I n A 1 2
for A ( 1 ) = X { X T X + I n } 1 X T .
Figure A1. An example for using “CGV(X,Y,k,W)”.
Figure A1. An example for using “CGV(X,Y,k,W)”.
Preprints 96378 g0a1

References

  1. Abe, T.; Shimizu, K.; Kuuluvainen, T.; Aakala, T. Sine-skewed axial distributions with an application for fallen tree data. Environ. Ecol. Stat. 2012, 19, 295–307. [Google Scholar] [CrossRef]
  2. Allen, D.M. The relationship between variable selection and data augmentation and a method for prediction. Technometrics 1974, 16, 125–127. [Google Scholar] [CrossRef]
  3. Arashi, M.; Roozbeh, M.; Hamzah, N.A.; Gasparini, M. Ridge regression and its applications in genetic studies. PLOS ONE 2021, 16, e0245376. [Google Scholar] [CrossRef]
  4. Assaf, A.G.; Tsionas, M.; Tasiopoulos, A. Diagnosing and correcting the effects of multicollinearity: Bayesian implications of ridge regression. Tour. Manag. 2019, 71, 1–8. [Google Scholar] [CrossRef]
  5. Azzalini, A.; Capitanio, A. The Skew-Normal and Related Families; IMS Monographs series; Cambridge University Press (CUP): Cambridge, UK, 2014. [Google Scholar]
  6. Binder, H.; Allignol, A.; Schumacher, M.; Beyersmann, J. Boosting for high-dimensional time-to-event data with competing risks. Bioinformatics 2009, 25, 890–896. [Google Scholar] [CrossRef]
  7. Bhattacharjee, A. Big Data Analytics in Oncology with R; CRC Press: Boca Raton, FL, USA, 2022. [Google Scholar]
  8. Bhatnagar, S.R.; Lu, T.; Lovato, A.; Olds, D.L.; Kobor, M.S.; Meaney, M.J.; O'Donnell, K.; Yang, A.Y.; Greenwood, C.M. A sparse additive model for high-dimensional interactions with an exposure variable. Comput. Stat. Data Anal. 2023, 179. [Google Scholar] [CrossRef]
  9. Chen, A.-C.; Emura, T. A modified Liu-type estimator with an intercept term under mixture experiments. Commun. Stat. - Theory Methods 2017, 46, 6645–6667. [Google Scholar] [CrossRef]
  10. Chu, H.; Huang, C.; Dong, J.; Yang, X.; Xiang, J.; Dong, Q.; Tang, Y. Lactate Dehydrogenase Predicts Early Hematoma Expansion and Poor Outcomes in Intracerebral Hemorrhage Patients. Transl. Stroke Res. 2019, 10, 620–629. [Google Scholar] [CrossRef]
  11. Cule, E.; De Iorio, M. Ridge regression in prediction problems: automatic choice of the ridge parameter. Genetic Epidemiology 2013, 37, 704–714. [Google Scholar] [CrossRef]
  12. Cule, E.; Vineis, P.; De Iorio, M. Significance testing in ridge regression for genetic data. BMC Bioinform. 2011, 12, 372–372. [Google Scholar] [CrossRef]
  13. Emura, T.; Chen, Y.-H.; Chen, H.-Y. Survival Prediction Based on Compound Covariate under Cox Proportional Hazard Models. PLOS ONE 2012, 7, e47627. [Google Scholar] [CrossRef]
  14. Emura, T.; Chen, Y.-H. Gene selection for survival data under dependent censoring: A copula-based approach. Stat. Methods Med Res. 2016, 25, 2840–2857. [Google Scholar] [CrossRef]
  15. Emura, T.; Hsu, W.-C.; Chou, W.-C. A survival tree based on stabilized score tests for high-dimensional covariates. J. Appl. Stat. 2023, 50, 264–290. [Google Scholar] [CrossRef]
  16. Friedrich, S.; Groll, A.; Ickstadt, K.; Kneib, T.; Pauly, M.; Rahnenführer, J.; Friede, T. Regularization approaches in clinical biostatistics: A review of methods and their applications. Stat. Methods Med Res. 2023, 32, 425–440. [Google Scholar] [CrossRef]
  17. Gao, S.; Zhu, G.; Bialkowski, A.; Zhou, X. Stroke Localization Using Multiple Ridge Regression Predictors Based on Electromagnetic Signals. Mathematics 2023, 11, 464. [Google Scholar] [CrossRef]
  18. Golub, G.H. , Heath, M.; Wahba, G. Generalized cross-validation as a method for choosing a good ridge parameter. Technometrics 1979, 21, 215–223. [Google Scholar] [CrossRef]
  19. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning; Springer: New York, NY, USA, 2009. [Google Scholar]
  20. Hernandez, J.; Lobos, G.A.; Matus, I.; Del Pozo, A.; Silva, P.; Galleguillos, M. Using Ridge Regression Models to Estimate Grain Yield from Field Spectral Data in Bread Wheat (Triticum aestivum L.) Grown under Three Water Regimes. Remote Sens. 2015, 7, 2109–2126. [Google Scholar] [CrossRef]
  21. Hoerl, A.E.; Kennard, R.W. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
  22. Huynh, U.; Pal, N.; Nguyen, M. Regression model under skew-normal error with applications in predicting groundwater arsenic level in the Mekong Delta Region. Environ. Ecol. Stat. 2021, 28, 323–353. [Google Scholar] [CrossRef]
  23. Jimichi, M.; Kawasaki, Y.; Miyamoto, D.; Saka, C.; Nagata, S. Statistical Modeling of Financial Data with Skew-Symmetric Error Distributions. Symmetry 2023, 15, 1772. [Google Scholar] [CrossRef]
  24. Kibria, B.M.G.; Banik, S. Some Ridge Regression Estimators and Their Performances. J. Mod. Appl. Stat. Methods 2016, 15, 206–238. [Google Scholar] [CrossRef]
  25. Kim, S.Y.; Lee, J.W. Ensemble clustering method based on the resampling similarity measure for gene expression data. Stat. Methods Med Res. 2007, 16, 539–564. [Google Scholar] [CrossRef] [PubMed]
  26. Loesgen, K.H. A generalization and Bayesian interpretation of ridge-type estimators with good prior means. Stat. Pap. 1990, 31, 147–154. [Google Scholar] [CrossRef]
  27. Montgomery, D.C.; Peck, E.A.; Vining, G.G. Introduction to Linear Regression Analysis; John Wiley & Sons: Hoboken, NJ, USA, 2021. [Google Scholar]
  28. Michimae, H.; Emura, T. Bayesian ridge estimators based on copula-based joint prior distributions for regression coefficients. Comput. Stat. 2022, 37, 2741–2769. [Google Scholar] [CrossRef]
  29. Norouzirad, M.; Arashi, M. Preliminary test and Stein-type shrinkage ridge estimators in robust regression. Stat. Pap. 2019, 60, 1849–1882. [Google Scholar] [CrossRef]
  30. Saleh, E.A.M.; Kibria, G.B.M. Performance of some new preliminary test ridge regression estimators and their properties. Communications in Statistics-Theory and Methods 1993, 22, 2747–2764. [Google Scholar] [CrossRef]
  31. Saleh, A.M.E.; Arashi, M.; Kibria, B.G. Theory of Ridge Regression Estimation with Applications; John Wiley & Sons: Hoboken, NJ, USA, 2019. [Google Scholar]
  32. Shih, J.-H.; Lin, T.-Y.; Jimichi, M.; Emura, T. Robust ridge M-estimators with pretest and Stein-rule shrinkage for an intercept term. Jpn. J. Stat. Data Sci. 2021, 4, 107–150. [Google Scholar] [CrossRef]
  33. Shih, J.H. , Konno, Y., Chang, Y.T.; Emura, T. A class of general pretest estimators for the univariate normal mean. Communications in Statistics-Theory and Methods 2023, 52, 2538–2561. [Google Scholar] [CrossRef]
  34. Taketomi, N.; Chang, Y.-T.; Konno, Y.; Mori, M.; Emura, T. Confidence interval for normal means in meta-analysis based on a pretest estimator. Jpn. J. Stat. Data Sci. 2023, 1–32. [Google Scholar] [CrossRef]
  35. Vishwakarma, G.K.; Thomas, A.; Bhattacharjee, A. A weight function method for selection of proteins to predict an outcome using protein expression data. J. Comput. Appl. Math. 2021, 391, 113465. [Google Scholar] [CrossRef]
  36. Wang, D.; Wang, J.; Li, Z.; Gu, H.; Yang, K.; Zhao, X.; Wang, Y. C-Reaction Protein and the Severity of Intracerebral Hemorrhage: A Study from Chinese Stroke Center Alliance. Neurol. Res. 2021, 44, 285–290. [Google Scholar] [CrossRef] [PubMed]
  37. Whittaker, J.C.; Thompson, R.; Denham, M.C. Marker-assisted selection using ridge regression. Genet. Res. 2000, 75, 249–252. [Google Scholar] [CrossRef] [PubMed]
  38. Wong, K.Y.; Chiu, S.N. An iterative approach to minimize the mean squared error in ridge regression. Comput. Stat. 2015, 30, 625–639. [Google Scholar] [CrossRef]
  39. Yang, S.-P.; Emura, T. A Bayesian approach with generalized ridge estimation for high-dimensional regression and testing. Commun. Stat. - Simul. Comput. 2017, 46, 6083–6105. [Google Scholar] [CrossRef]
  40. Yoshiba, T.; Koike, T.; Kato, S. On a Measure of Tail Asymmetry for the Bivariate Skew-Normal Copula. Symmetry 2023, 15, 1410. [Google Scholar] [CrossRef]
  41. Yüzbaşı, B.; Arashi, M.; Ahmed, S.E. Shrinkage Estimation Strategies in Generalised Ridge Regression Models: Low/High-Dimension Regime. Int. Stat. Rev. 2020, 88, 229–251. [Google Scholar] [CrossRef]
  42. Zhang, Q.; Ma, S.; Huang, Y. Promote sign consistency in the joint estimation of precision matrices. Comput. Stat. Data Anal. 2021, 159, 107210. [Google Scholar] [CrossRef]
Figure 1. Examples for generating design matrices using “X.mat(.)”.
Figure 1. Examples for generating design matrices using “X.mat(.)”.
Preprints 96378 g001
Figure 2. The R code and output for calculating the ridge estimator using “g.ridge(.)”.
Figure 2. The R code and output for calculating the ridge estimator using “g.ridge(.)”.
Preprints 96378 g002aPreprints 96378 g002b
Figure 3. The histogram of the normal and skew-normal distributions (the slant parameter ten; alpha=10 in the R function “rsn(.)”). Both distributions have mean 0 and standard deviation 1.
Figure 3. The histogram of the normal and skew-normal distributions (the slant parameter ten; alpha=10 in the R function “rsn(.)”). Both distributions have mean 0 and standard deviation 1.
Preprints 96378 g003
Figure 3. The responses y ( i = 1 n y i / n ) 1 against the predictors X β ^ based on the ridge estimator and generalized ridge estimator applied to the intracerebral hemorrhage data.
Figure 3. The responses y ( i = 1 n y i / n ) 1 against the predictors X β ^ based on the ridge estimator and generalized ridge estimator applied to the intracerebral hemorrhage data.
Preprints 96378 g003b
Figure 4. The fitted residuals for the generalized ridge estimator applied to a dataset on patients with intracerebral hemorrhage.
Figure 4. The fitted residuals for the generalized ridge estimator applied to a dataset on patients with intracerebral hemorrhage.
Preprints 96378 g004
Table 1. The total mean squared error (TMSE) of the three estimators: (i) the ridge by “g.ridge(.)”, (ii) the generalized (g-) ridge by “g.ridge(.)”, and (iii) the ridge by “glmnet(.)”. The TMSE is computed by the Monte Carlo average over 500 replications.
Table 1. The total mean squared error (TMSE) of the three estimators: (i) the ridge by “g.ridge(.)”, (ii) the generalized (g-) ridge by “g.ridge(.)”, and (iii) the ridge by “glmnet(.)”. The TMSE is computed by the Monte Carlo average over 500 replications.
Error distribution Regression coefficients p (i) ridge (ii) g-ridge (iii) glmnet
Normal (I) b = d = 5 50 0.463 0.385 0.306
100 0.950 0.682 2.182
150 1.146 0.658 1.996
200 1.520 0.920 2.199
(II) b = d = 10 50 0.855 0.681 0.545
100 2.151 1.562 8.688
150 3.008 1.482 7.904
200 4.929 2.687 8.691
(III) b = 5 and d = 5 50 0.602 0.539 0.388
100 0.990 0.628 2.025
150 1.219 0.703 2.132
200 1.589 0.953 2.226
(IV) b = 10 and d = 10 50 1.541 1.290 0.737
100 2.398 1.580 8.046
150 3.231 1.614 8.434
200 4.651 2.770 8.804
Skew-normal (I) b = d = 5 50 0.440 0.361 0.294
100 0.957 0.670 2.182
150 1.162 0.678 2.000
200 1.500 0.910 2.197
(II) b = d = 10 50 0.821 0.655 0.527
100 2.285 1.705 8.691
150 3.021 1.509 7.905
200 4.883 2.673 8.686
(III) b = 5 and d = 5 50 0.576 0.519 0.376
100 0.974 0.622 2.029
150 1.233 0.721 2.137
200 1.582 0.949 2.243
(IV) b = 10 and d = 10 50 1.504 1.273 0.720
100 2.449 1.508 8.054
150 3.224 1.616 8.453
200 4.618 2.731 8.860
NOTE: We set the sample size n = 100 all cases.
Table 2. The fitted results for estimated regression coefficients (only with P-value < 0.05) sorted by P-values applied to a dataset on patients with intracerebral hemorrhage.
Table 2. The fitted results for estimated regression coefficients (only with P-value < 0.05) sorted by P-values applied to a dataset on patients with intracerebral hemorrhage.
Ridge Generalized ridge
β ^ j SE P-value β ^ j SE P-value
Lactate dehydrogenase 0.122 0.047 0.008 0.145 0.055 0.008
Gamma-GT 0.116 0.048 0.016 0.143 0.056 0.010
Respiratory rate -0.120 0.052 0.020 -0.140 0.059 0.018
Prothrombin time 0.077 0.036 0.031 0.083 0.040 0.038
Blood platelet count -0.100 0.049 0.040 -0.114 0.056 0.044
C-reactive protein None None > 0.05 0.112 0.057 0.049
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated