Preprint
Article

Cumulative Entropy of Past Lifetime for Coherent Systems at the System Level

Altmetrics

Downloads

97

Views

13

Comments

0

A peer-reviewed article of this preprint also exists.

This version is not peer-reviewed

Submitted:

21 August 2023

Posted:

23 August 2023

You are already at the latest version

Alerts
Abstract
In this paper, we investigate the cumulative historical entropy of the lifetime of an n-component coherent system under the assumption that all system components have failed at time t. Using the system signature, we can calculate the cumulative historical entropy of the system’s lifespan, which is an important aspect in determining a system’s predictability. This research produces a number of significant results, including expressions, bounds, and ordering aspects for this measure. In addition, we offer a method for finding a preferred system based on cumulative Kullback-Leibler discriminating information that is closely related to the parallel system. These findings provide important insights into the predictability and dependability of coherent systems, emphasizing the importance of this field.
Keywords: 
Subject: Computer Science and Mathematics  -   Probability and Statistics

MSC:  60E05; 62B10; 62N05; 94A17

1. Introduction

The Shannon differential entropy has gained widespread adoption as a measure across various fields of research. It was introduced by Shannon in his article [1], and since then, it has become a cornerstone of probability theory. Suppose we have a non-negative random variable X, which is absolutely continuous with a probability density function f ( x ) . If the expected value of the logarithm exists, then the expression H ( X ) = E [ log f ( X ) ] is referred to as the Shannon differential entropy. This is the situation being described in this scenario. This definition allows us to quantify the uncertainty associated with the random variable X by measuring the amount of information required to describe it. Owing to its adaptability and usefulness, this metric has gained widespread adoption in a multitude of research domains, making it an essential tool for any researcher looking to explore probability distributions.
While differential entropy has many advantages, a fascinating substitute for conventional entropy was suggested by Rao and colleagues in their paper [2], which they called cumulative residual entropy (CRE). Unlike differential entropy, which uses f ( x ) , CRE uses S ( x ) = P ( X > x ) to obtain a measure of entropy. The definition of CRE is given by
CE ( X ) = 0 S ( x ) log S ( x ) d x .
With its ability to capture the residual uncertainty of a distribution, CRE is particularly useful in situations where the tails of the distribution are of interest. The CRE has emerged as the preferred measure for characterizing information dispersion in problems related to reliability theory, which is robust metric has been extensively employed in numerous studies, such as those carried out by [3,4,5,6,7,8], to name a few. A representation of the cumulative entropy (CE), which is an information measure similar to equation (1), is provided by Di Crescenzo and Longobardi in their work [9] given by
CE ( X ) = 0 F ( x ) log F ( x ) d x ,
where F ( · ) is the cumulative distribution function (cdf) of a random variable X . Like CRE, CE is nonnegative and CE ( X ) = 0 if and only if X is a constant.
Several measures of dynamic information have been proposed to characterize the uncertainty in suitable functionals of the random lifetime X of a system. We review two such measures, namely the dynamic cumulative residual entropy (introduced by Asadi and Zohrevand [10]) and the cumulative past entropy (proposed by Di Crescenzo and Longobardi in [11]), which are defined as the dynamic cumulative entropy of X t = [ X | X > t ] and X t = [ X | X t ] , respectively. In this case, the dynamic cumulative residual entropy is defined as
E ( X t ) = t F ¯ t ( x ) log F ¯ t ( x ) d x
= t F ¯ ( x ) F ¯ ( t ) log F ¯ ( x ) F ¯ ( t ) d x , t > 0 ,
where F ¯ t ( x ) = F ¯ ( x ) / F ¯ ( t ) , x > t , is the survival function of X t = [ X | X > t ] . Di Crescenzo and Longobardi [11] pointed out that in many realistic scenarios, uncertainty is not necessarily related to the future. For example, if a system that starts operating at time 0 is only observed at predetermined inspection times and is found to be "down" at time t, then the uncertainty is dependent on the specific moment within ( 0 , t ) at which it failed. When faced with such scenarios, the cumulative past entropy proves to be a valuable metric. It is defined as follows:
CE ( X t ) = 0 t F t ( x ) log F t ( x ) d x = 0 t F ( x ) F ( t ) log F ( x ) F ( t ) d x ,
where F t ( x ) = F ( x ) / F ( t ) , 0 < x t , is the cdf of X t = [ X | X t ] .
Numerous researchers have demonstrated a strong inclination toward investigating the information characteristics of coherent systems which can be seen in [8,12,13,14,15,16], and the references therein. Recently, Kayid [17] investigated the Tsallis entropy of coherent systems when all components are alive at time t . Mesfioui et al. [18] have also delved into the Tsallis entropy such systems, adding to the current body of captivating research in this area. More recently, Kayid and Shrahili [19,20] investigated the Shannon differential and Renyi entropy of coherent systems when all components have failed at time t . The aim of this study is to examine the uncertainty properties of the lifetimes of coherent systems, with a particular focus on the cumulative past entropy. In fact, when all components have failed at time t , we focus on coherent systems consisting of n components by utilizing the system signature.
Therefore, the result of this paper is organized as follows: In Section 2, we provide an expression for the CE of a coherent system’s lifetime when component lifetimes are independent and identically distributed, given that all components of the system have failed at time t by implementing the concept of system signature. In Section 3, some useful bounds are presented. A new criterion is represented to choose a preferable among coherent systems in Section 4. Some concluding remarks are also given in Section 5.

2. CE of the past lifetime

In this section, we present by applying the system signature concept to define the past lifetime CE of a coherent system, which can have an arbitrary structure. To this aim, we assume that at a specific time t , all components of the system have failed and employ the concept of system signature which is described by an n-dimensional vector p = ( p 1 , , p n ) , where the i-th element is defined as p i = P ( T = X i : n ) , i = 1 , 2 , , n . (see [21]).
Consider a coherent system whose component lifetimes X 1 , , X n are independent and identically distributed (i.i.d.), and whose signature vector p = ( p 1 , , p n ) is known. Assuming that the coherent system have failed at time t, we can represent the past lifetime of the system as T t = [ t T | X n : n t ] . Khaledi and Shaked [22] have established results that allow us to express the cumulative distribution function of T t in terms of their findings as
F T t ( x ) = i = 1 n p i F T t i ( x ) ,
where
F T t i ( x ) = P ( t X i : n x | X n : n t ) = k = 0 i 1 n k F ( t x ) F ( t ) k 1 F ( t x ) F ( t ) n k , 0 < x < t ,
denotes the cdf of T t i = [ t X i : n | X n : n t ] , i = 1 , 2 , , n . It is important to note that T t i , i = 1 , 2 , , n , represents the elapsed time since the failure of the component with a lifetime of X i : n in the system, given that the system has failed at or before time t . Furthermore, according to (6), T t i corresponds to the ith order statistic of n i.i.d. components with a cumulative distribution function of F ( t x ) F ( t ) , 0 < x < t . Hereafter, we provide a formula for computing the cumulative entropy of T t . For this purpose, we define F t ( x ) = F ( x ) F ( t ) , 0 < x < t , and use V = F t ( T t ) , which is essential to our approach. Using this transformation, we can express the cumulative entropy of T t in terms of V, as shown in the upcoming theorem.
Theorem 1.
We can express the CE of T t as follows
CE ( T t ) = 0 1 ψ ( S V ( v ) ) f t ( F t 1 ( v ) ) d v ,
where ψ ( v ) = v log v , 0 < v < 1 , for all t > 0 and
S V ( v ) = i = 1 n p i S i ( v ) , 0 u 1 ,
represents the survival function of V = F t ( T t ) such that
S i ( v ) = k = 0 i 1 n k v k ( 1 v ) n k , 0 < v < 1 .
is the survival function of i-th order statistics where lifetimes are uniformly distributed. Moreover, F t 1 ( u ) = inf { x ; F t ( x ) u } is the quantile function of F t ( x ) .
Proof. 
Substituting z = t x and using equations (5) and (6), we obtain:
CE ( T t ) = 0 t ψ ( F T t ( x ) ) d x = 0 t ψ i = 1 n p i F T t i ( x ) d x = 0 t ψ i = 1 n p i k = 0 i 1 n k F ( t x ) F ( t ) k 1 F ( t x ) F ( t ) n k d u , = 0 t ψ i = 1 n p i k = 0 i 1 n k F ( x ) F ( t ) k 1 F ( x ) F ( t ) n k d u , = 0 1 ψ i = 1 n p i k = 0 i 1 n k u k 1 u n k f t ( F t 1 ( u ) ) d u , = 0 1 ψ i = 1 n p i S i ( v ) f t ( F t 1 ( u ) ) d u ,
where the last equality is obtained by using the change of variables u = F t ( x ) . Furthermore, S i ( v ) represents the survival function of U i : n = F t ( T t i ) as given in (9). By utilizing (8), we can derive the relationship (7), which serves to conclude the proof.    □
Suppose we examine an i-out-of-n system with a system signature of p = ( 0 , , 0 , 1 i , 0 , , 0 ) , where i = 1 , 2 , , n . Them, we obtain a special case of Equation (7), which reduces to
CE ( T t i ) = 0 1 ψ ( S i ( u ) ) f t ( F t 1 ( u ) ) d u , t > 0 .
The following theorem is a direct consequence of Theorem 1 and characterizes the aging properties of the system’s components. It is noteworthy to mention that a random variable X is said to have a decreasing reversed hazard rate (DRHR) if its reversed hazard rate function τ ( x ) = f ( x ) / F ( x ) declines for x > 0 .
Theorem 2.
If X is DRHR, then CE ( T t ) is increasing in t.
Proof.By noting that f t ( F t 1 ( x ) ) = x τ t ( F t 1 ( x ) ) , Equation (7) can be rewritten as
CE ( T t ) = 0 1 ψ ( S V ( u ) ) u τ t ( F t 1 ( u ) ) d u ,
for all t > 0 . We can easily confirm that F t 1 ( u ) = F 1 ( u F ( t ) ) holds for all 0 < u < 1 . Therefore, we obtain:
τ t ( F t 1 ( u ) ) = τ ( F 1 ( u F ( t ) ) ) , 0 < u < 1 .
If t 1 t 2 , then F 1 ( u F ( t 1 ) ) F 1 ( u F ( t 2 ) ) . Therefore, if X has a DRHR property, then:
0 1 ψ ( S V ( u ) ) u τ t 1 ( F t 1 1 ( u ) ) d u = 0 1 ψ ( S V ( u ) ) u τ ( F 1 ( u F ( t 1 ) ) d u ( ) 0 1 ψ ( S V ( u ) ) u τ ( F 1 ( u F ( t 2 ) ) d u d u = 0 1 ψ ( S V ( u ) ) u τ t 2 ( F t 2 1 ( u ) ) d u .
By utilizing Equation (11), we can infer that CE ( T t 1 ) ( ) CE ( T t 2 ) for all t 1 t 2 , thus completing the proof.    □
We provide an example to demonstrate the applications of Theorems 1 and 2 in engineering systems. This example highlights how these theorems can be utilized for analyzing the CE of a failed coherent system and for investigating the aging characteristics of a system.
Example 1.
Suppose we have a coherent system with a system signature p = ( 1 / 4 , 1 / 4 , 1 / 2 , 0 ) . To compute the precise value of CE ( T t ) using equation (7), we require the lifetime distributions of the system components. For this purpose, let us assume the following lifetime distributions.
(i) Let X follow the uniform distribution in [ 0 , 1 ] . From (7), we immediately obtain
CE ( T t ) = 0.2245 t , t > 0 .
The analysis indicates that the cumulative entropy of T t increases as time t increases, which aligns with previous research on the behavior of cumulative entropy for specific categories of random variables. For instance, it is established that the uniform distribution has a DRHR property, indicating that the CE of T t should increase as time t increases, as per Theorem 2.
(b) Let’s examine a random variable X, whose cdf is defined as follows:
F ( x ) = e x k , x > 0 , k > 0 .
After performing some algebraic manipulations, we have
CE ( T t ) = 0 1 ψ ( S V ( v ) ) k u t k log u k + 1 k d u , t > 0 .
Calculating this relationship explicitly is challenging, thus we must rely on numerical methods to proceed. In Figure 2, we illustrate the cumulative entropy of T t for different values of k. It is evident that X exhibits a DRHR property for all k > 0 . Referring back to Theorem 2, we can see that CE ( T t ) rises with increasing t when k > 0 .
Figure 1. A coherent system with signature p = ( 1 / 4 , 1 / 4 , 1 / 2 , 0 ) .
Figure 1. A coherent system with signature p = ( 1 / 4 , 1 / 4 , 1 / 2 , 0 ) .
Preprints 82990 g001
Figure 2. Cumulative entropy of T t with respect to t for various values of k > 0 , using the cdf Part (b) from Example 1.
Figure 2. Cumulative entropy of T t with respect to t for various values of k > 0 , using the cdf Part (b) from Example 1.
Preprints 82990 g002
The example above provides insight into the complex interplay between the CE of a random variable and time, emphasizing the significance of accounting for the decreasing reversed hazard rate property when analyzing such systems. The results suggest that the DRHR property of X is a critical factor in determining the temporal dynamics of the CE of T t , which could have significant consequences in various areas of research including the study of intricate systems and the development of efficient data compression methods.
The notion of duality is a valuable tool in engineering reliability to reduce the computational workload of computing the signatures of all coherent systems of a given magnitude by roughly half (as demonstrated, for instance, in Kochar et al. [23]). Particularly, if T t stands for lifetime of a coherent system with signature p = ( p 1 , , p n ) , then its dual system with lifetime T t D , n has a signature p D = ( p n , , p 1 ) . By utilizing the principle of duality, we have the following theorem which facilitates the computational intricacy entailed in calculating the past cumulative entropy of T t .
Theorem 3.
If f t ( F t 1 ( u ) ) = f t ( F t 1 ( 1 u ) ) holds true for all 0 < u < 1 and t, then, for all p and n , we can conclude that CE ( T t ) = CE ( T t D ) .
Proof. 
It is crucial to emphasize that the equation S i ( 1 u ) = S n i + 1 ( u ) valids for all i = 1 , , n and 0 < u < 1 . Furthermore, since f t ( F t 1 ( u ) ) = f t ( F t 1 ( 1 u ) ) holds for all 0 < u < 1 , we can leverage (7) to derive the subsequent expression:
CE ( T t D ) = 0 1 ψ i = 1 n p n i + 1 S i ( u ) f t ( F t 1 ( u ) ) d u = 0 1 ψ r = 1 n p r S n r + 1 ( u ) f t ( F t 1 ( u ) ) d u = 0 1 ψ r = 1 n p r S r ( 1 u ) f t ( F t 1 ( 1 u ) ) d u = 0 1 ψ r = 1 n p r S r ( u ) f t ( F t 1 ( u ) ) d u = CE ( T t ) ,
and this completes the proof. □

3. Bounding the Cumulative Entropy of Past Lifetime

When working with extremely intricate systems that consist of numerous components, calculating CE ( T t ) can be a difficult undertaking. To handle this difficulty, investigators have newly proposed uncertainty bounds for the component lifetimes in coherent systems. This approach is explored in works such as [16,19,20] and related literature. In the next theorem, we obtain bounds for CE ( T t ) based on the CE of the parent distribution CE ( X t ) . These bounds can be useful in certain circumstances, particularly when calculating the precise cumulative past entropy is challenging.
Proposition 1.
If T t = [ t T | X n : n t ] denotes the past lifetime of the system, then
B 1 CE ( X t ) CE ( T t ) B 2 CE ( X t )
where B 1 = inf u ( 0 , 1 ) ψ ( S V ( u ) ) ψ ( u ) , B 2 = sup u ( 0 , 1 ) ψ ( S V ( u ) ) ψ ( u ) and ψ ( u ) = u log ( u ) .
Proof. 
The upper bound is given by:
CE ( T t ) = 0 1 ψ ( S V ( u ) ) f t ( F t 1 ( u ) ) d u = 0 1 ψ ( S V ( u ) ) ψ ( u ) ψ ( u ) f t ( F t 1 ( u ) ) d u sup u ( 0 , 1 ) ψ ( S V ( u ) ) ψ ( u ) 0 1 ψ ( u ) f t ( F t 1 ( u ) ) d u = B 2 CE ( X t ) .
The constant B 2 is dependent on the distribution of V. Using the same method, we can also derive a lower bound for CE ( T t ) . □
It is worth noting that B 1 in the aforementioned theorem is obtained by computing the infimum of ψ ( S V ( u ) ) / ψ ( u ) for u ( 0 , 1 ) and B 2 is obtained by computing the supremum of the same ratio. The given bounds offer valuable means of approximating CE ( T t ) based on CE ( X t ) .
Remark 1.
It is worth noting that the lower bound B 1 in Proposition 1 equals zero for all coherent systems that satisfy either p 1 = 0 or p n = 0 . This is especially true for coherent systems with more than one i.i.d. component, as outlined in [8].
Proposition 2.
The lower bound for the CE of T t is given by
CE ( T t ) CE L ( T t ) ,
where CE L ( T t ) = i = 1 n p i CE ( T t i ) .
Proof. 
Utilizing (7) and the concavity of the function ψ ( u ) = u log ( u ) , we can derive the following lower bound:
CE ( T t ) = 0 1 ψ ( S V ( u ) ) f t ( F t 1 ( u ) ) d u 0 1 i = 1 n p i ψ ( S i : n ( u ) ) f t ( F t 1 ( u ) ) d u = i = 1 n p i CE ( T t i ) ,
where CE ( T t i ) is the cumulative entropy of T t i . □
Notice that equality in (12) holds for i-out-of-n systems in the sense that we have p j = 0 , for j i , and p j = 1 , for j = i and then CE ( T t ) = CE ( T t i ) . When the lower bounds in both parts of Theorems 1 and 2 can be computed, one may use the maximum of the two lower bounds.
Example 2.
In this example, we analyze a coherent system with the signature p = ( 0 , 3 10 , 5 10 , 2 10 , 0 ) , consisting of n = 5 i.i.d. component lifetimes drawn from a uniform distribution in the interval [ 0 , 1 ] . If T t = [ t T | X 5 : 5 t ] denotes the past lifetime of this system, Remark 1 yields B 1 = 0 , while B 2 = 1.25 and CE ( X t ) = t 4 , 0 < t < 1 . As per Theorem 1, we deduce that CE ( T t ) is bounded accordingly:
0 CE ( T t ) 1.25 t , 0 < t < 1 .
Moreover, as i = 1 n p i CE ( U i : n ) = 0.164 , we can represent the lower bound provided in (12) as:
CE ( T t ) 0.164 t , 0 < t < 1 .
Recalling the lower bound (14) and the upper bound (13), we have the following inequality
0.164 t CE ( T t ) 1.25 t ,
for all 0 < t < 1 . This inequality supplies a precise and secure boundary for the conditional CE of T t over the entire interval ( 0 , 1 ) .

4. Preferable system

Hereafter, we consider two nonnegative random variables, X and Y, that represent the lifetimes of two items having the same supports ( 0 , ) . For any given time t > 0 , we define their respective past lifetimes as X t = [ X | X t ] and Y t = [ Y | Y t ] . To this end, we also introduce the distribution functions of X t and Y t by
F t ( x ) = F ( x ) F ( t ) , S t ( x ) = G ( x ) G ( t ) , 0 x t ,
respectively. In their groundbreaking work, Di Crescenzo and Longobardi [24] introduced a novel concept that utilizes the mean past lifetimes of nonnegative random variables X and Y with cdfs F and G, by μ X ( t ) = E [ t X | X t ] and μ Y ( t ) = E [ t Y | Y t ] , respectively. They proposed a past version of the cumulative Kullback-Leibler information measure that is defined as a function of the mean past lifetimes of X and Y as follows:
CE ( X t , Y t ) = 0 t F t ( x ) log F t ( x ) S t ( x ) d x + μ X ( t ) μ Y ( t ) ,
provided that F t ( x ) = 0 whenever S t ( x ) = 0 . In order to advance our findings, we define a novel measure of distance that is symmetric and applicable to two distributions. This measure is called Symmetric Past Cumulative (SPC) Kullback-Leibler divergence and is denoted by the shorthand SPC ( X t , Y t ) .
Definition 1.
Consider two non-negative past random variables, X t and Y t , with shared support and cumulative distribution functions F t and S t , respectively. In this case, we introduce the SPC Kullback-Leibler divergence as a measure of distance between the two variables as follows:
SPC ( X t , Y t ) = CE ( X t , Y t ) + CE ( X t , Y t ) = 0 t F t ( x ) S t ( x ) log F t ( x ) S t ( x ) d x ,
for all t > 0 .
The proposed measure, defined as (17), possesses several desirable properties. First and foremost, it is nonnegative and symmetric. Moreover, the value of SPC ( X t , Y t ) is equal to zero if and only if F t ( x ) and S t ( x ) are almost everywhere equal. In addition to these properties, we also observe the following.
Lemma 1.Suppose we have three random variables, X t , Y t , and Z t , each with cumulative distribution functions F t , S t , and H t , respectively. If the stochastic ordering X t s t Y t s t Z t holds, then the following statement is true
(i)
0 SPC ( X t , Y t ) SPC ( X t , Z t ) ,
(ii)
0 SPC ( Y t , Z t ) SPC ( X t , Z t ) ,
for all t > 0 .
Proof. Given that the function ( x 1 ) log ( x ) is decreasing in the interval ( 0 , 1 ) and increasing in ( 1 , ) , we can make an important inference from the condition F t ( x ) S t ( x ) H t ( x ) for 0 x t . Specifically, we can conclude that
(i)
S t ( x ) F t ( x ) log S t ( x ) F t ( x ) H t ( x ) F t ( x ) log H t ( x ) F t ( x ) , 0 x t ,
(ii)
0 S t ( x ) H t ( x ) log S t ( x ) H t ( x ) F t ( x ) H t ( x ) log F t ( x ) H t ( x ) , 0 x t .
By integrating both sides of the descriptions (i) and (ii), we can obtain the desired result. □
As T t 1 s t T t s t T t n holds for any coherent system, from Lemma 1, we have the following theorem.
Theorem 4.If T t is the lifetime of a coherent system based on X 1 , , X n , then:
(i)
0 SPC ( T t , T t 1 ) SPC ( T t 1 , T t n ) ,
(ii)
0 SPC ( T t , T t n ) SPC ( T t 1 , T t n ) .
Traditional stochastic ordering may not be sufficient for pairwise comparisons of system performance, particularly for certain system structures. In some cases, pairs of systems remain incomparable using stochastic orders. In this case, alternative metrics for comparing system performance are being explored. Hereafter, we will introduce a new method to choose a preferable system. It is worth to note that engineers typically favor those that offer extended operational time. As a result, it is crucial to ensure that the systems being compared possess comparable attributes. Furthermore, assuming that all other attributes are equal, we can choose parallel system lifetime since it has a longer past lifetime than alternative systems. In other words, from (6), we have
F T t 1 ( x ) F T t ( x ) F T t n ( x ) , x > 0 .
Rather than relying on comparisons between two systems at a time t , we can explore a system that has a structure or distribution that is more similar to that of the parallel system. Essentially, our goal is to determine which of these systems is more similar or closer in configuration to the parallel system, while also having a dissimilar configuration from that of the series system. To this aim, one can employ the idea of SPC given in Definition 1. Consequently, we put forth the following Symmetric Past Divergence (SPD) measure for T t as
SPD ( T t ) = SPC ( T t , T t 1 ) SPC ( T t , T t n ) SPC ( T t 1 , T t n ) .
Theorem 3 establishes that 1 SPD ( T t ) 1 . It is evident that SPD ( T t ) = 1 if and only if T t = s t T t n and SPD ( T t ) = 1 if and only if T t = s t T t 1 . Put simply, we can deduce that if SPD ( T t ) is closer to 1, the distribution of T t is more akin to the parallel system’s distribution. On the other hand, if SPD ( T t ) is closer to 1 , the distribution of T t is more similar to the series system’s distribution. Based on this, we suggest the subsequent definition.
Definition 2.
Consider two coherent systems, each with n component lifetimes that are independent and identically distributed, alongside signatures p 1 and p 2 . Let T 1 , t and T 2 , t represent their respective past lifetimes. At time t, we assert that T 2 , t is more desirable than T 1 , t with regard to the Symmetric Past Distance (SPD) measure, indicated by T 1 , t S P D T 2 , t , if and only if SPD ( T 1 , t ) SPD ( T 2 , t ) for all t > 0 .
It is important to note that the equivalence of SPD ( T 1 , t ) and SPD ( T 2 , t ) does not always mean that T 1 , t = s t T 2 , t . In accordance with Definition 2, we can define PD ( T ) = SPC ( T , T t 1 ) SPC ( T , T t n ) . From equation (17) and the aforementioned conversions, we obtain
SPC ( T t , T t i ) = 0 1 S V ( u ) S i : n ( u ) ] f t ( F t 1 ( u ) ) log S V ( u ) S i : n ( u ) d u ,
for i = 1 , n . Then, from (19), we get
PD ( T ) = 0 1 [ S V ( u ) S 1 : n ( u ) ] f t ( F t 1 ( u ) ) log S V ( u ) S 1 : n ( u ) d u 0 1 [ S V ( u ) S n : n ( u ) ] f t ( F t 1 ( u ) ) log S V ( u ) S n : n ( u ) d u ,
SPC ( T t 1 , T t n ) = 0 1 [ S n : n ( u ) S 1 : n ( u ) ] f t ( F t 1 ( u ) ) log S n : n ( u ) S 1 : n ( u ) d u .
If we assume that the components are i.i.d., we can derive the equations ψ ( S 1 : n ( u ) ) = u n and ψ ( S n : n ( u ) ) = 1 ( 1 u ) n . By referencing Theorem 4, we can arrive at an intriguing result.
Proposition 3.
If T t is the lifetime of a coherent system based on X 1 , , X n , then SPC ( T t , T t i ) SPC ( T t 1 , T t n ) , for i = 1 , n .
The next theorem is readily apparent.
Theorem 5.
Assuming the conditions outlined in Definition 2, if the X 1 is exponentially distributed, then the SPD measure is independent of time t . In other words, SPD ( T t ) = SPD ( T ) holds true for all t > 0 .
Proof. 
Using the memoryless property, we can deduce that f t ( F t 1 ( u ) ) = f ( F 1 ( u ) ) holds true for all t > 0 . Consequently, we can conclude that the result holds true. □
Example 3.
Consider two coherent systems with past lifetimes T 1 , t and T 2 , t , in which the component lifetimes are exponentially distributed with cdf F ( x ) = 1 e x , x > 0 . The signatures of these systems are given by p 1 = ( 1 / 4 , 1 / 4 , 1 / 2 , 0 ) and p 2 = ( 0 , 2 / 3 , 1 / 3 , 0 ) , respectively. Although these systems are not comparable using traditional stochastic orders, so, we can compare them by the SPD measure. By using numerical computation, we obtain SPD ( T 1 , t ) = 0.0914 and SPD ( T 2 , t ) = 0.1179 . This indicates that the system with lifetime T 1 , t is less similar to the parallel system than the system with lifetime T 2 , t .
Theorem 6.
Suppose that T 1 , t and T 2 , t denote the lifetimes of two coherent systems with signatures p 1 and p 2 , respectively, based on n i.i.d. components with the same cdf F. If p 1 s t p 2 , we can assert that T 1 , t S P D T 2 , t .
Proof. 
We can derive the desired outcome from Theorem 2.3 of Khaledi and Shaked [22]. Specifically, if we have two probability vectors denoted p 1 and p 2 , where p 1 s t p 2 , then we have T t 1 s t T 1 , t s t T 2 , t s t T t n . By applying Lemma 1, we can obtain SPC ( T 1 , t , T t 1 ) SPC ( T 2 , t , T t 1 ) and SPC ( T 1 , t , T t n ) SPC ( T 2 , t , T t n ) . These relations enable us to arrive at the desired outcome due to the relation defined in equation (18). □
An exciting discovery is that the comparison based on SPD can perform as a necessary condition for the usual stochastic order. Let us assume we have two coherent systems, denoted by T 1 , t and T 2 , t , each composed of several components with lifetimes X 1 , , X n . If we find that T 1 , t is less reliable than T 2 , t in the stochastic order, denoted as T 1 , t s t T 2 , t , we can conclude that the system T 1 , t is also less reliable than T 2 , t in the SPD order, i.e., T 1 , t S P D T 2 , t . This comparison of systems through the SPD order can be useful when comparing systems that are difficult to assess otherwise. It is worth noting that if we find that two systems T 1 , t and T 2 , t are equally reliable in the stochastic order, i.e., T 1 , t = s t T 2 , t , then they are also equivalently reliable in the SPD order, i.e., T 1 , t = S P D T 2 , t . This highlights the potential power of the SPD order in system analysis.

5. Conclusion

In the past few years, there has been a growing interest in quantifying the uncertainty related to the lifetime of engineering systems. Hence, this criterion can be used to predict the uncertainty of the lifetime of the system. The measurement of CE is an extension of Shannon’s entropy, making it a highly useful tool in these types of situations. This paper presented the CE presentation for the system’s lifetime if all components of the system have failed at time t. In addition, we have studied the properties of the proposed metrics including expressions and bounds in terms of CE uncertainties using the concept of system signature. We presented a series of examples to demonstrate our findings. In conclusion, we introduced criteria based on relative CE to identify a preferred system that is more akin to the parallel system.

Author Contributions

Conceptualization, M.S.; methodology, M.S.; software, M.K.; validation, M.S. and M.K.; formal analysis, M.S.; investigation, M.K.; resources, M.K.; writing—original draft preparation, M.S.; writing—review and editing, M.S. and M.K.; visualization, M.K.; supervision, M.K.; project administration, M.S.; funding acquisition, M.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Researchers Supporting Project number (RSP2023R464), King Saud University, Riyadh, Saudi Arabia.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

The authors acknowledge financial support from the Researchers Supporting Project number (RSP2023R464), King Saud University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shannon, C.E. A mathematical theory of communication. The Bell system technical journal 1948, 27, 379–423. [Google Scholar] [CrossRef]
  2. Rao, M.; Chen, Y.; Vemuri, B.C.; Wang, F. Cumulative residual entropy: a new measure of information. IEEE transactions on Information Theory 2004, 50, 1220–1228. [Google Scholar] [CrossRef]
  3. Rao, M. More on a new concept of entropy and information. Journal of Theoretical Probability 2005, 18, 967–981. [Google Scholar] [CrossRef]
  4. Asadi, M.; Zohrevand, Y. On the dynamic cumulative residual entropy. Journal of Statistical Planning and Inference 2007, 137, 1931–1941. [Google Scholar] [CrossRef]
  5. Baratpour, S. Characterizations based on cumulative residual entropy of first-order statistics. Communications in Statistics—Theory and Methods 2010, 39, 3645–3651. [Google Scholar] [CrossRef]
  6. Baratpour, S.; Rad, A.H. Testing goodness-of-fit for exponential distribution based on cumulative residual entropy. Communications in Statistics-Theory and Methods 2012, 41, 1387–1396. [Google Scholar] [CrossRef]
  7. Navarro, J.; del Aguila, Y.; Asadi, M. Some new results on the cumulative residual entropy. Journal of Statistical Planning and Inference 2010, 140, 310–322. [Google Scholar] [CrossRef]
  8. Toomaj, A.; Zarei, R. Some new results on information properties of mixture distributions. Filomat 2017, 31, 4225–4230. [Google Scholar] [CrossRef]
  9. Di Crescenzo, A.; Longobardi, M. On cumulative entropies. Journal of Statistical Planning and Inference 2009, 139, 4072–4087. [Google Scholar] [CrossRef]
  10. Asadi, M.; Ebrahimi, N. Residual entropy and its characterizations in terms of hazard function and mean residual life function. Statistics & probability letters 2000, 49, 263–269. [Google Scholar]
  11. Di Crescenzo, A.; Longobardi, M. Entropy-based measure of uncertainty in past lifetime distributions. Journal of Applied probability 2002, 39, 434–440. [Google Scholar] [CrossRef]
  12. Abdolsaeed, T.; Doostparast, M. A note on signature-based expressions for the entropy of mixed r-out-of-n systems. Naval Research Logistics (NRL) 2014, 61, 202–206. [Google Scholar]
  13. Toomaj, A.; Doostparast, M. On the Kullback Leibler information for mixed systems. International Journal of Systems Science 2016, 47, 2458–2465. [Google Scholar] [CrossRef]
  14. Asadi, M.; Ebrahimi, N.; Soofi, E.S.; Zohrevand, Y. Jensen–Shannon information of the coherent system lifetime. Reliability Engineering & System Safety 2016, 156, 244–255. [Google Scholar]
  15. Toomaj, A. Renyi entropy properties of mixed systems. Communications in Statistics-Theory and Methods 2017, 46, 906–916. [Google Scholar] [CrossRef]
  16. Toomaj, A.; Chahkandi, M.; Balakrishnan, N. On the information properties of working used systems using dynamic signature. Applied Stochastic Models in Business and Industry 2021, 37, 318–341. [Google Scholar] [CrossRef]
  17. Kayid, M.; Alshehri, M.A. Tsallis entropy of a used reliability system at the system level. Entropy 2023, 25, 550. [Google Scholar] [CrossRef]
  18. Mesfioui, M.; Kayid, M.; Shrahili, M. Renyi Entropy of the Residual Lifetime of a Reliability System at the System Level. Axioms 2023, 12, 320. [Google Scholar] [CrossRef]
  19. Kayid, M.; Shrahili, M. Rényi Entropy for Past Lifetime Distributions with Application in Inactive Coherent Systems. Symmetry 2023, 15, 1310. [Google Scholar] [CrossRef]
  20. Kayid, M.; Shrahili, M. On the Uncertainty Properties of the Conditional Distribution of the Past Life Time. Entropy 2023, 25, 895. [Google Scholar] [CrossRef]
  21. Samaniego, F.J. System signatures and their applications in engineering reliability; Vol. 110, Springer Science & Business Media, 2007.
  22. Khaledi, B.E.; Shaked, M. Ordering conditional lifetimes of coherent systems. Journal of Statistical Planning and Inference 2007, 137, 1173–1184. [Google Scholar] [CrossRef]
  23. Kochar, S.; Mukerjee, H.; Samaniego, F.J. The “signature” of a coherent system and its application to comparisons among systems. Naval Research Logistics (NRL) 1999, 46, 507–523. [Google Scholar] [CrossRef]
  24. Di Crescenzo, A.; Longobardi, M.; others. Some properties and applications of cumulative Kullback-Leibler information. Applied stochastic models in business and industry 2015, 31, 875–891. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated