1. Introduction
The concept of extropy and its use have been increased rapidly in the recent years. It measures the uncertainty contained in the probability distributions. For a non negative random variable (
)
X, the entropy introduced by [
29] is given by
where
is the probability density function (
) of a
X. This measure is shift independent, that is, it is same for both
X and
and it cannot be applied in some fields such as in neurology. Thus [
4] introduced the notion of weighted entropy measure as
They pointed out that occurrence of an event has an impact on uncertainty in two ways. Being quantitative and qualitative, it first shows the probability of occurrence of an event and later it shows utility in achieving qualitative characteristics of a goal. The variable
x in the integral emphasize the weight related to the occurrence of the event
. Here it assigns more significance to large values of
X. It is important to note that the information obtained when a device fails to operate or a neuron fails to release spikes in a specific time interval, differs significantly from the information obtained when such events occur in other equally wide intervals. This is why there is a need, in some cases, to employ a shift-dependent information measure that assigns varying measures to these distributions.The importance of the presence of weighted measures of uncertainty has been exhibited by [
9].
The concept of extropy for a continuous
X has been presented and discussed across numerous works in the literature. The differential extropy defined by [
15] is:
One can refer [
23] for the extropy properties of order statistics and record values. The applications of extropy in automatic speech recognition can be found in [
3]. Various literature sources have presented a range of extropy measures and their extensions. Analogous to weighted entropy, [
1] introduced the concept of weighted extropy (
) in the literature. It is given as
which again can be alternatively expressed as
They also illustrated that there exist some distributions with same extropy but different
and vice versa. In the literature extropy, its different versions and their applications, have been studied by several authors (see, for instance [
6],[
1],[
14], [
7]). In particular, a unified version of extropy in classical theory and in Dempster-Shafer theory has been studied by [
5].
Let us define a
X with unknown
. We assume that
X is defined on
R and
is continuously differentiable. Suppose
be a sequence of independent and identically distributed (
)
. The most commonly used estimator of
is the kernel density estimator (
), given by [
22] and [
27] as
where
is the kernel function which satisfies the following conditions.
dx=1
dx=0
dx=1
dx <
Here, is a sequence of positive bandwidths such that → 0 and → as n→.
When probability density functions are estimated in a non parametric way, standard kernel density estimators are frequently used. However, when we deal with data that fits distributions with heavy tails, multiple modes, or skewness, particularly those with positive values these estimators may lose their effectiveness. In all of these scenarios, applying a transformation, we can yield more consistent results. Such a transformation involves a logarithmic transformation to create a non parametric kernel estimator. An important aspect of the logarithmic transformation is its ability to compress the right tail of the distribution. The obtained
are called logarithmic
(denoted as
), (check, [
8]). Let us define
Y=
,
=
;
and let
be the
of
Y. The
is defined as,
where
is the log-kernel function with bandwidth
0, at location parameter
z. For any
,
satisfies the conditions:
for all
and
.
For any
,
where
.
There are several papers available in the literature that delve into the estimation of extropy and its various versions. [
24] proposed estimators of extropy and also worked on its application by testing uniformity. [
26] approached the concept of length biased sampling in estimating extropy. Research on non parametric estimation using dependent data are also well explored in the literature. These covered the estimation of residual and past extropy both under
-mixing dependence condition and can be found in works such as [
18] and [
11]. Additionally, a work by [
19] explained about the recursive and non recursive kernel estimation of negative cumulative extropy under
-mixing dependence condition. Recently, [
20] discussed about the kernel estimation of extropy function using
-mixing dependent data. Moreover, [
12] introduced the log kernel estimation of extropy.
Even if there are several works available on the literature related to the estimation of extropy, just a little has been done on
and its estimation until now. There are situations in which we are forced to use
instead of extropy. Unlike extropy, the qualitative characteristics of information are also represented here. [
28] demonstrated the significance of employing
as opposed to regular extropy in certain scenarios. There are instances where certain distributions possess identical extropy values but exhibit distinct
values. In such situations, it becomes necessary to opt for weighted extropy. The estimators of
can also be used in the selection of models in the reliability analysis also. Here we tried to find some estimators for
and validated it using simulation study and data analysis.
The paper is organized as follows: In
Section 2, we have introduced the log kernel estimation of
. In
Section 3 the estimation of
using empirical kernel smoothed estimator is given. A simulation study is conducted to evaluate the estimators and we also made a comparison between log kernel and kernel estimation of
in
Section 4.
Section 5 devotes to the real data analysis to examine the proposed estimators. Finally, we conclude the study in
Section 6
3. Empirical Estimation Of Weighted Extropy
Non parametric estimation is a widely employed technique in various research papers for estimating extropy and its associated measures. One common approach within non parametric estimation is the use of kernel density estimation, which is a popular method in the literature in order to obtain smoothed estimates.
In this section, we introduce the empirical method for estimating
to assess
. This estimation is achieved through the utilization of non parametric kernel estimator. The empirical kernel smoothed estimator for
is,
where
is the kernel estimator given by [
22] and
is the
order statistic of the random sample.
Example 1.
Let the samples be from the distribution with = . Then follows standard uniform distribution. Moreover, is a beta distribution with mean and variance, respectively, and . Then, the mean and variance of is given by
and,
where is defined in equation (6).
Table 1.
Mean and Variance of for the distribution with =.
Table 1.
Mean and Variance of for the distribution with =.
|
n |
Mean |
Variance |
|
10 |
-0.58602 |
0.03039 |
|
20 |
-0.57820 |
0.02182 |
|
30 |
-0.55622 |
0.01888 |
|
40 |
-0.56181 |
0.01083 |
|
50 |
-0.52369 |
0.00777 |
|
100 |
-0.52239 |
0.00597 |
Table 1 shows the values of mean and variance of the samples of Example 1. Hence it is clear that the mean is approaching towards the theoretical mean and the variance is tending to zero when the sample size increases.
Example 2.
Suppose X is a Rayleigh distribution with parameter 1. Then follows exponential distribution and is distributed as exponential distribution with mean=, for . The mean and variance of are
Table 2.
Mean and Variance of for Rayleigh distribution with parameter=1.
Table 2.
Mean and Variance of for Rayleigh distribution with parameter=1.
|
n |
Mean |
Variance |
|
10 |
-0.46220 |
0.02453 |
|
20 |
-0.33956 |
0.01626 |
|
30 |
-0.29213 |
0.00173 |
|
40 |
-0.28919 |
0.00121 |
|
50 |
-0.26677 |
0.00094 |
|
100 |
-0.23975 |
0.00024 |
From the table 2 it is clear that the variance is decreasing to zero and the mean is advanced towards the theoretical value -0.25 in the case of Rayleigh distribution with parameter one.
Remark 1. Based on the examples 1 and 2, it is clear that the proposed estimator is consistent (see, table 1 and 2), since the mean of the estimator is approaching towards the theoretical value and the variance tends to zero when the sample size increases.
4. Simulation Study
We manage a simulation study to evaluate the performance of the presented estimators. Random samples are generated corresponding to different sample sizes from some standard distributions, then both bias and are calculated for 10000 samples.
It is obvious that the standard kernel estimation method performs better in many situations, but sometimes log kernel estimation methods outperforms it. So here to enable a comparison between log kernel and kernel estimators of weighted extropy we again proposed a kernel estimator for
using equation (
6). The estimator is given by
where
is the kernel estimator given by [
22]. Using the consistency property of kernel estimator it is clear that the proposed kernel estimator for
is also consistent. To lay the ground work for comparison, we had generated samples from exponential distribution, log normal distribution-a heavy tailed distribution and uniform distribution. The Gaussian log transformed kernel and Gaussian kernel are the kernel functions used for simulation.
Table 3.
Estimated value, and MSE of and from standard exponential distribution with =-0.125
Table 3.
Estimated value, and MSE of and from standard exponential distribution with =-0.125
|
|
n |
H |
|
MSE |
H |
|
MSE |
50 |
-0.11909 |
0.00591 |
0.00031 |
-0.13364 |
0.00864 |
0.00030 |
100 |
-0.11830 |
0.00670 |
0.00017 |
-0.12979 |
0.00479 |
0.00013 |
150 |
-0.11886 |
0.00614 |
0.00011 |
-0.12872 |
0.00372 |
0.00009 |
200 |
-0.11904 |
0.00596 |
0.00009 |
-0.12806 |
0.00306 |
0.00006 |
250 |
-0.11978 |
0.00522 |
0.00007 |
-0.12741 |
0.00241 |
0.00005 |
300 |
-0.11952 |
0.00548 |
0.00007 |
-0.12720 |
0.00220 |
0.00004 |
350 |
-0.11975 |
0.00525 |
0.00006 |
-0.12728 |
0.00228 |
0.00004 |
400 |
-0.12025 |
0.00475 |
0.00005 |
-0.12679 |
0.00179 |
0.00003 |
450 |
-0.12028 |
0.00472 |
0.00005 |
-0.12646 |
0.00146 |
0.00003 |
500 |
-0.12061 |
0.00439 |
0.00004 |
-0.12648 |
0.00148 |
0.00002 |
Table 4.
Estimated value, and MSE of and from lognormal distribution with =-0.14105.
Table 4.
Estimated value, and MSE of and from lognormal distribution with =-0.14105.
|
|
n |
H |
|
MSE |
H |
|
MSE |
50 |
-0.14370 |
0.00265 |
0.00025 |
-0.14621 |
0.00517 |
0.00025 |
100 |
-0.14243 |
0.00139 |
0.00012 |
-0.14375 |
0.00271 |
0.00013 |
150 |
-0.14189 |
0.00084 |
0.00008 |
-0.14241 |
0.00136 |
0.00008 |
200 |
-0.14199 |
0.00095 |
0.00006 |
-0.14207 |
0.00103 |
0.00006 |
250 |
-0.14175 |
0.00070 |
0.00005 |
-0.14204 |
0.00099 |
0.00006 |
300 |
-0.14171 |
0.00067 |
0.00004 |
-0.14180 |
0.00075 |
0.00005 |
350 |
-0.14155 |
0.00051 |
0.00003 |
-0.14169 |
0.00064 |
0.00004 |
400 |
-0.14140 |
0.00036 |
0.00003 |
-0.14158 |
0.00053 |
0.00004 |
450 |
-0.14127 |
0.00022 |
0.00003 |
-0.14153 |
0.00048 |
0.00004 |
500 |
-0.14121 |
0.00016 |
0.00001 |
-0.14137 |
0.00032 |
0.00004 |
Table 5.
Estimated value, and MSE of and from standard uniform distribution with =-0.25.
Table 5.
Estimated value, and MSE of and from standard uniform distribution with =-0.25.
|
|
n |
H |
|
MSE |
H |
|
MSE |
50 |
-0.2097 |
0.0403 |
0.00286 |
-0.22576 |
0.02424 |
0.0017 |
100 |
-0.21562 |
0.03438 |
0.00184 |
-0.22786 |
0.02214 |
0.00100 |
150 |
-0.21826 |
0.03174 |
0.00143 |
-0.22829 |
0.02171 |
0.00080 |
200 |
-0.22059 |
0.02941 |
0.00120 |
-0.23056 |
0.01955 |
0.00067 |
250 |
-0.22285 |
0.02715 |
0.00099 |
-0.23045 |
0.01944 |
0.00066 |
300 |
-0.22277 |
0.02723 |
0.00097 |
-0.23201 |
0.01799 |
0.00053 |
350 |
-0.22426 |
0.02574 |
0.00088 |
-0.23276 |
0.01724 |
0.00047 |
400 |
-0.22511 |
0.02489 |
0.00079 |
-0.23325 |
0.01675 |
0.00044 |
450 |
-0.22575 |
0.02425 |
0.00076 |
-0.23399 |
0.01621 |
0.00039 |
500 |
-0.22668 |
0.02332 |
0.00068 |
-0.23379 |
0.01601 |
0.00039 |
Table 6.
Estimated value, and MSE of from standard exponential distribution with =-0.125
Table 6.
Estimated value, and MSE of from standard exponential distribution with =-0.125
|
n |
H |
|
MSE |
50 |
-0.16379 |
0.03879 |
0.00256 |
100 |
-0.14418 |
0.01918 |
0.00052 |
150 |
-0.13857 |
0.01357 |
0.00027 |
200 |
-0.13496 |
0.00996 |
0.00015 |
250 |
-0.13285 |
0.00785 |
0.00010 |
300 |
-0.13234 |
0.00734 |
0.00009 |
350 |
-0.13111 |
0.00611 |
0.00007 |
400 |
-0.13043 |
0.00543 |
0.00006 |
450 |
-0.13005 |
0.00505 |
0.00005 |
500 |
-0.12976 |
0.00476 |
0.00005 |
Table 7.
Estimated value, and of from lognormal distribution with =-0.14105.
Table 7.
Estimated value, and of from lognormal distribution with =-0.14105.
|
n |
H |
|
|
50 |
-0.22300 |
0.08195 |
0.06302 |
100 |
-0.17574 |
0.03469 |
0.00255 |
150 |
-0.16491 |
0.02386 |
0.00108 |
200 |
-0.15942 |
0.01837 |
0.00075 |
250 |
-0.15744 |
0.01639 |
0.00070 |
300 |
-0.15401 |
0.01296 |
0.00040 |
350 |
-0.15218 |
0.01113 |
0.00025 |
400 |
-0.15072 |
0.00967 |
0.00020 |
450 |
-0.15015 |
0.00911 |
0.00021 |
500 |
-0.14888 |
0.00783 |
0.00014 |
Table 8.
Estimated value, and of from standard uniform distribution with =-0.25.
Table 8.
Estimated value, and of from standard uniform distribution with =-0.25.
|
n |
H |
|
|
50 |
-0.22669 |
0.02331 |
0.00172 |
100 |
-0.22713 |
0.02287 |
0.00107 |
150 |
-0.22892 |
0.02108 |
0.00085 |
200 |
-0.23084 |
0.01916 |
0.00066 |
250 |
-0.23194 |
0.01806 |
0.00056 |
300 |
-0.23168 |
0.01832 |
0.00055 |
350 |
-0.23195 |
0.01805 |
0.00049 |
400 |
-0.23317 |
0.01683 |
0.00045 |
450 |
-0.23335 |
0.01665 |
0.00041 |
500 |
-0.23361 |
0.01639 |
0.00039 |
From the above tables 3, 4, 5, 7 and 8, it is clear that the and bias of both the estimators are decreasing with sample size. The decreasing indicates that the estimator’s predictions are getting closer to the true values with larger sample sizes, demonstrating enhanced accuracy and efficiency in estimation. The decreasing bias also shows the accuracy of the estimators.
The comparison of bias and between kernel and logkernel estimators in simulation for reveals that the logkernel estimator outperforms the kernel estimator in certain scenarios, particularly when dealing with heavy-tailed distributions.