Parameters Estimation of the Exponentiated Chen Distribution Based on Upper Record Values

Farhad Yousaf1, 2, Sajid Ali1,*, Ismail Shah3, 1 and Saba Riaz4

1Department of Statistics, Quaid-i-Azam University, Islamabad 45320, Pakistan
2Department of Social Sciences, University of Naples Federico II, Naples, Italy
3Department of Statistical Sciences, University of Padua, 35121, Padova, Italy
4Department of Statistics, Rawalpindi Women University, 6th Road, Satellite Town, Rawalpindi, Punjab, Pakistan
E-mail: farhadyousaf@stat.qau.edu.pk; farhad.yousaf@unina.it; sajidali.qau@hotmail.com; shah@unipd.it; ishah@qau.edu.pk; saba.riaz@f.rwu.edu.pk
*Corresponding Author

Abstract

This article discusses the Bayesian and frequentist inferences for the exponentiated Chen distribution assuming upper record values. Due to unavailability of the compact form of marginal posterior distributions, a Markov Chain Monte Carlo algorithm is designed to compute the posterior summaries. Prediction of future record values under Bayesian and frequentist methods is also discussed mathematically and numerically. Further, a sensitivity analysis to assess the effect of prior on the estimated parameters is also a part of this study. Besides the simulation studies, the importance of the present study is illustrated with the help of a real data example. It is noted that the Bayes estimates outperform the frequentist inference.

Keywords: Asymptotic intervals, Bayesian prediction, exponentiated Chen distribution, record values.

1 Introduction

The ordered observations which strictly exceed the previous values are called the record values, and thus the record statistics are closely related to theory of order statistics. The study of records gained popularity in the literature because of their practical usage, such as industrial stress testing, meteorology, seismology, hydrology, sporting events, economics, life testing, oil and mining surveys (Ahsanullah and Nevzorov, 2015). In statistics, Chandler (1952) developed the concept of record statistics and also discuss record values, record times and inter record times. After that numerous studies discussed record values and related statistical inference. For instance, Dziubdziela and Kopociński (1976), Nagaraja (1977), Srivastava (1979), Nevzorov (1988), Arshad and Jamal (2019), Yousaf et al. (2019), etc. For more detailed information on record statistics, and their applications, we refer to Arnold et al. (2011) and Ahsanullah and Nevzorov (2015). Soliman and Al-Aboud (2008) used upper record values from a Rayleigh distribution and discussed Bayesian and non-Bayesian approaches to obtain the estimators of the parameter. Seo and Kim (2017) discussed objective Bayesian analysis based on upper record values from two-parameter Rayleigh distribution with partial information. The authors provided a pivotal quantity and an algorithm based on the pivotal quantity to predict the behavior of future survival records. Al-Duais (2021) used the inverse Weibull distribution for the Bayesian analysis of upper record values using balanced loss function. Pak et al. (2022) considered the records as well as the corresponding inter-record times to develop inference procedures for R=P(X>Y) assuming X and Y come from Weibull distribution. Kumar and Gupta (2023) discussed Bayesian analysis of inverse Rayleigh distribution using noninformative prior for different loss functions. Alhamidah et al. (2023) considered the problem of E-Bayesian estimation and its expected posterior mean squared error (E-PMSE) in a Burr type XII model on the basis of record values.

The exponentiated Chen distribution (ECD) is introduced by Chaubey and Zhang (2015), which is in fact is an extension of the two-parameter Chen distribution (Chen, 2000). The ECD is a positively skewed and more flexible than the two-parameter Chen model. The ECD has a bathtub hazard function, which decrease initially, then remains constant, and finally increase. In practice, the most accurate failure rate is bathtub-shaped, and in the literature, many researchers proposed different models for bathtub-shaped hazard rate, see, for example, Hjorth (1980), Rajarshi and Rajarshi (1988), Xie and Lai (1996), Chen (2000), and references cited therein.

Suppose X1,X2,,Xn be the n upper record values from the ECD with probability density function

g(x;γ,μ,φ) =γμφxμ-1exp(xμ)exp(φ(1-exp(xμ)))
(1-exp(φ(1-exp(xμ))))γ-1
x0,γ,μ,φ>0 (1)

where φ>0 is the scale and γ,μ>0 are the shape parameters. By setting γ = 1 in Equation (1), the distribution reduces to

g(x)=μφxμ-1exp(xμ)exp(φ(1-exp(xμ)))  x,μ,φ>0 (2)

which is the two-parameter Chen distribution. Thus, the Chen distribution is a special case of the ECD. The distribution function of the ECD is given as

G(x;γ,μ,φ)=(1-exp(φ(1-exp(xμ))))γ  x0 (3)

The hazard, reliability, and quantile function of ECD are given, respectively, by

h(x) =γμφxμ-1exp(xμ)exp(φ(1-exp(xμ)))(1-exp(φ(1-exp(xμ))))γ-11-(1-exp(φ(1-exp(xμ))))γx0,μ,φ>0
R(x) =1-(1-exp(φ(1-exp(xμ))))γ  x0,μ,φ>0 (4)
Xw =[log(1-1φlog(1-w1/γ))]1/μ  w(0,1) (5)

The graphical depiction of the ECD and its failure is given in Figure 1. To be more specific, the plot on the left, i.e., Figure 1(a), is the probability density function plot of the ECD with parameters γ=35,μ=0.25,φ=1 plotted as the yellow solid line, γ=35,μ=0.25,φ=1.5 plotted as the red dashed line, γ=35,μ=0.25,φ=2 plotted as the green dotted line, γ=35,μ=0.25,φ=2.5 plotted as the blue dot-dashed line and finally, γ=35,μ=0.25,φ=3 plotted as the gray long-dashed line. Similarly, in Figure 1(b) the hazard function is plotted assuming the scale and the first shape parameter fixed at γ=35,μ=0.25, while assuming the second shape parameter φ = 1, 1.5, 2, 2.5 and 3, respectively.

images

Figure 1 Density and hazard plots for different parameter values.

The development of new probability models and studying their different properties has gained a lot of attention from research during the last decade. For example, Xie et al. (2002) introduced the three-parameter Weibull distribution by adding a new scale parameter to baseline distribution. Later, Chaubey and Zhang (2015) also extended the Chen distribution by adding the shape parameter and named the proposed distribution as the ECD. Dey et al. (2017) studied different properties and estimation methods for the ECD and concluded that the maximum product of spacing (MPS) method is more efficient than the other estimation methods. Khan et al. (2016) introduced the transmuted exponentiated distribution. Recently, Khan et al. (2018) introduced a new five parameter Kumaraswamy exponentiated Chen distribution by adding two new shape parameters to the base line ECD. Eliwa et al. (2021) proposed exponentiated odd Chen-G family of distribution and studied different properties. Later, Awodutire (2022) used the exponential distribution as a baseline distribution to study the properties of exponentiated odd Chen-G family. Recently, Azimi et al. (2023) proposed inverted exponentiated Chen distribution with application to cancer data. Méndez-González et al. (2023) introduced the additive Chen distribution and discussed its properties and applications.

The aim of this study is to present parameter estimation of the ECD based on upper record values under Bayesian and non-Bayesian methods. Further, a framework for prediction of the future record observations is also discussed. As we have shown in the next section, Bayesian computations cannot be done in the traditional manner, a Markov Chain Monte Carlo (MCMC) algorithm is devised to obtain different posterior summaries.

The rest of the study is organized as follows. Section 2 derives the estimators of the unknown parameters for the ECD by the maximum likelihood method. Section 3 discusses Bayes estimators under the squared error loss function (SELF) using informative and non-informative priors. Sensitivity analysis is discussed in Section 4. The prediction of future record values for the ECD under Bayesian and frequentist approaches is discussed in Section 5. The numerical results of simulated as well as for real data are presented in Sections 6 and 7, respectively. Finally, Section 8 presents conclusion.

2 Maximum Likelihood Estimation

For the estimation of γ,μ and φ assuming the upper record values using the maximum likelihood estimation method, suppose X1,X2,X3, be a sequence of independent and identically distributed (iid) random variables with PDF g(x) and CDF G(x) from the Exponentiated Chen distribution. Suppose Yn=max(min){X1,X2,X3,,Xn} for n0. The observation Xk is an upper(lower) record value of this sequence if Yk>(<)Yk-1,k>1, i.e., Xk is the upper(lower) value if its value is greater(smaller) than that of all the preceding observations, where Xk(n) and k(n) denote the upper record values and upper record times, respectively. Assuming a sample of size n record values X={Xk(1),Xk(2),,Xk(n)} from Equation (1), the likelihood can be written as (Ahsanullah, 1995).

L(γ,μ,φ|x)=g(xk(n);γ,μ,φ)i=1n-1g(xk(i);γ,μ,φ)1-G(xk(i),γ,μ,φ)
0xk(1)<xk(2)<<xk(n)< (6)

By replacing the PDF and CDF of the ECD in Equation (6), the likelihood function can be written as

L(γ,μ,φ|x)=γnμnφni=1nxk(i)μ-1exp(xk(i)μ)exp[φ{1-exp(xk(i)μ)}][1-exp{φ(1-exp(xk(i)μ))}]γ-1i=1n-1[1-{1-exp(φ(1-exp(xk(i)μ)))}γ] (7)

2.1 Likelihood Estimators

For deriving the maximum likelihood estimators (MLE), take the partial derivatives of the logarithm of the likelihood function l(γ,μ,φ|x)=logL(γ,μ,φ|x) with respect to unknown parameters. It is to be noted that the terms without γ,μ and φ can be ignored.

l(γ,μ,φ|x) =nlog(γ)+nlog(μ)+nlog(φ)+i=1nxk(i)μ
+(γ-1)i=1nlog(1-exp(φ(1-exp(xk(i)μ))))
+(μ-1)i=1nlog(xk(i))
-i=1n-1log(1-(1-exp(φ(1-exp(xk(i)μ))))γ)
+φi=1n(1-exp(xk(i)μ)) (8)

Next, partially differentiate Equation (7) with respect to the parameters and equating the resulting equations to zero. The resulting normal equations can be written as.

nγ-i=1n-1log(1-exp(φ(1-exp(xk(i)μ)))γlog(1-exp(φ(1-exp(xk(i)μ)))1-(1-exp(φ(1-exp(xk(i)μ))))γ
  +i=1nlog(1-exp(φ(1-exp(xk(i)μ)))=0 (9)
nμ+i=1nlog(xk(i))+i=1nxk(i)μlog(xk(i))+(γ-1)
  i=1nφxk(i)μlog(xk(i))exp(φ(1-exp(xk(i)μ))+xk(i)μ)1-exp(φ(1-exp(xk(i)μ)))
  -i=1n-1γφxk(i)μlog(xk(i))exp(φ(1-exp(xk(i)μ))+xk(i)μ)(1-exp(φ(1-exp(xk(i)μ))))γ-11-(1-exp(φ(1-exp(xk(i)μ))))γ
  +i=1nlog(xk(i))-φi=1nexp(xk(i)μ)xk(i)μlog(xk(i))=0 (10)
nφ-i=1n-1γ(1-exp(xk(i)μ))exp(φ(1-exp(xk(i)μ)))(1-exp(φ(1-exp(xk(i)μ))))γ-11-(1-exp(φ(1-exp(xk(i)μ))))γ
  +i=1n(1-exp(xk(i)μ))-(γ-1)
  i=1n(1-exp(xk(i)μ))exp(φ(1-exp(xk(i)μ)))1-exp(φ(1-exp(xk(i)μ)))=0 (11)

Solving Equations (9)–(11), one can get the MLE for the ECD parameters γ,μ and φ, denoted as γ^MLE, μ^MLE and φ^MLE, respectively. An iterative procedure like Newton Raphson is suggested to solve the above equations because these equations do not have analytical solution.

2.2 Asymptotic Confidence Intervals

Since the MLE of γ,μ and φ cannot be solved analytically, the exact distribution of these estimators is difficult to obtain. Therefore, the exact confidence intervals for the parameters γ,μ and φ cannot be obtained. Alternatively, asymptotic confidence intervals are obtained by using the large sample approximation. It is well known fact that the asymptotic distribution of the MLE, say, ψ is (ψ^-ψ)N(ψ,I-1(ψ)) (Lawless, 2011), where I-1(ψ) is the inverse of the observed Fisher information matrix of the unknown parameters ψ = (γ,μ,φ), defined as

I-1=(-2l(γ,μ,φ|x)γ2-2l(γ,μ,φ|x)γμ-2l(γ,μ,φ|x)γφ-2l(γ,μ,φ|x)μγ-2l(γ,μ,φ|x)μ2-2l(γ,μ,φ|x)μφ-2l(γ,μ,φ|x)φγ-2l(γ,μ,φ|x)φμ-2l(γ,μ,φ|x)φ2)|(γ,μ,φ)=(γ^,μ^,φ^)-1

or

I-1=(var(γ^)cov(γ^,μ^)cov(γ^,φ^)cov(μ^,γ^)var(μ^)cov(μ^,φ^)cov(φ^,γ^)cov(φ^,μ^)var(φ^))

where

2l(γ,μ,φ|x)γ2 =γ(l(γ,μ,φ|x)γ),
2l(γ,μ,φ|x)γμ =γ(l(γ,μ,φ|x)μ)
2l(γ,μ,φ|x)μ2 =μ(l(γ,μ,φ|x)μ),
2l(γ,μ,φ|x)μφ =μ(l(γ,μ,φ|x)φ)
2l(γ,μ,φ|x)φ2 =φ(l(γ,μ,φ|x)φ),
2l(φ,μ,γ|x)φγ =φ(l(γ,μ,φ|x)γ)

The asymptotic 100(1-ψ)% confidence intervals of the parameter γ,μ and φ are of the forms

γ^±zψ/2var(γ^)
μ^±zψ/2var(μ^)

and

φ^±zψ/2var(φ^)

where zψ/2 is the upper (ψ/2)th percentile of the standard normal distribution.

3 Bayesian Estimation

The aim of this section is to present Bayesian analysis for the unknown parameters of the ECD based on the upper record values. Selection of a suitable prior for the unknown parameters is an important task in Bayesian inference. There are many types of priors, like informative and non-informative priors which can be used in the Bayesian inference. A sufficient knowledge about a problem at hand is known to be the informative prior while vice versa for the noninformative prior case. The Bayes theorem is used to combine the currently observed information with the prior information to get an updated form of the information, which is known as the posterior distribution. For simplicity, in this study it is assumed that the parameters γ,μ and φ are independent and follow gamma distribution, i.e., γ gamma(a,b), μ gamma(c,d) and φ gamma(e,f) with PDFs

g(γ)γa-1exp(-bγ)    γ>0 (12)
g(μ)μc-1exp(-dμ)    μ>0 (13)

and

g(φ)φe-1exp(-fφ)    φ>0 (14)

where a, b, c, d, e and f are known and non-negative hyperparameters. It is worth mentioning that the gamma prior is considered for illustration and any other suitable prior can be used instead of this. Furthermore, dependent type priors can also be assumed. To this end, the joint prior distribution can be written as follows.

g(γ,μ,φ)γa-1exp(-bγ)μc-1exp(-dμ)φe-1exp(-fφ)
γ,μ,φ>0 (15)

The joint posterior distribution is obtained by combining Equation (7) and Equation (15) and the resulting expression is given in Equation (16).

π(γ,μ,φ|x)L(γ,μ,φ|x)×g(γ,μ,φ)
π(γ,μ,φ|x)γn+a-1μn+c-1φn+e-1exp(-bγ-dμ-fφ)T (16)

where

T=i=1nxk(i)μ-1exp(xk(i)μ)exp(φ(1-exp(xk(i)μ)))(1-exp(φ(1-exp(xk(i)μ)))γ-1i=1n-1(1-(1-exp(φ(1-exp(xk(i)μ))))γ) (17)

For deriving the marginal posterior distribution of the parameter of interest, integrate the joint posterior distribution with respect to nuisance parameters. The marginal posterior distribution of γ,μ and φ can be written as follows.

π1(γ|x) γn+a-1exp(-bγ)00μn+c-1φn+e-1
×exp(-dμ-fφ)Tdμdφ (18)
π2(μ|x) μn+c-1exp(-dμ)00μn+c-1φn+a-1
×exp(-bγ-fφ)Tdγdφ (19)

and

π3(φ|x) φn+e-1exp(-fφ)00γn+a-1μn+c-1
×exp(-bγ-dμ)Tdγdμ (20)

Next, the specification of a suitable loss function is required to estimate the unknown parameters. In Bayesian, one of the most commonly used loss function is the squared error loss function (SELF). Assuming the SELF, the Bayes estimators for the parameter γ, μ and φ can be obtained as γ^BGE=Eγ|x(γ), μ^BGE=Eμ|x(μ) and φ^BGE=Eφ|x(φ), respectively, provided that E(γ), E(μ) and E(φ) are finite and do exist. These estimators can be written as γ^BGE=Tγ, μ^BGE=Tμ and γ^BGE=Tφ where

Tγ =000γn+aμn+c-1φn+e-1exp(-bγ-dμ-fφ)Tdγdμdφ (21)
Tμ =000γn+a-1μn+cφn+e-1exp(-bγ-dμ-fφ)Tdγdμdφ (22)

and

Tφ=000γn+a-1μn+c-1φn+eexp(-bγ-dμ-fφ)Tdγdμdφ (23)

Since the Bayes estimators cannot be obtained analytically, a numerical procedure is required to solve them numerically. To this end, the use of Metropolis-Hastings (MH) algorithm is suggested to obtain the Bayes estimators. Within Markov Chain Monte Carlo (MCMC), Metropolis algorithm is a very flexible method to obtain Bayes estimates (Metropolis et al., 1953; Hastings, 1970). We use MH algorithm to generate the parameter γ because its distribution is not a well-known distribution. We express the marginal posterior distribution of each parameter as follows

g(γ|μ,φ,x) γn+a-1exp(-bγ)i=1n
×exp(xk(i)μ)i=1n(1-exp(φ(1-exp(xk(i)μ))))γ-1i=1n-1(1-(1-exp(φ(1-exp(xk(i)μ))))γ) (24)
g(μ|x) Gamma(n+c,d+ln1i=1nxk(i)) (25)

and

g(φ|x)Gamma(n+e,f-(1-exp(xk(i)μ)) (26)

To apply the MH algorithm, first, initialize the parameter γ as the initial value γτ and draw the next value γ with probability density g(γ|γτ) which is known as the transition kernel. It is to be noted that gamma distribution is taken as a transition kernel for illustration and any other suitable distribution can be considered. Then, calculate m=min(m1m2,1), where m1=g(γ)/g(γτ) defining the probability ratio between the present and the last sample γτ, and m2=g(γτ|γ)g(γ|γτ). If the result of m2=1, the density is symmetric. A new state is rejected if m=1, γτ+1=γ, i.e.,

γτ+1={γ  with probability mγτ  with probability 1-m

To diminish the effect of initial state, repeat the above process a large number of times. To obtain more reliable results, some initial state values can be discarded which is known as the burn-in period. The Markov Chain will converge when the variance of the parameter values decreased or becomes too small. We suggest the following steps to calculate the posterior estimates.

(1) Guess the initial value of γ as γ0.

(2) Generate μ and φ from g(μ|x) and g(φ|x).

(a) To calculate γ, estimate the acceptance probability by g(γ(i),γ)=min(1,g(γ|x)g(γ(i)|γ)g(γ(i)|x)g(γ|γ(i))), where g(γ|x) has been defined above.

(b) Generate random numbers U(0,1).

(c) If g(γ(i),γ)U,γ(i+1)=γ(i)

(3) Suppose γ, μ and φ take the values γi,μi and φi at the ith step. Now, one can generate g(γi+1|γi,μi,φi,x).

(4) Repeat the above steps N-times.

(5) Obtain the Bayes estimators of h(γ,μ,φ) by 1K-Li=L+1Kg(γτ,μτ, φτ), where L denote the number of burn-in sample.

A great advantage of the MH approach over Gibbs sampler is that the conditional distribution of parameters is notessential. In this study, we considered 50,000 as the MCMC iteration size and first 5,000 observations were discarded as the burn-in samples.

3.1 Bayes Intervals

This section discusses the Bayes intervals, which give a probabilistic statement with a specified probability whether a parameter of interest is included in the intervals or not. Thus, Bayes intervals do not depend on the repetition concept as the classical intervals and also known as the credible intervals. After deriving the marginal posterior distribution of the parameter γ, a symmetric 100(1-ϕ)% two-sided Bayes interval estimate of γ, denoted by [γL,γU], can be obtained by solving the following two equations (Martz and Waller, 1982).

0γLπ1(ω|x)dω=ϕ2 (27)

and

γUπ1(ω|x)dω=ϕ2 (28)

for the limits γL and γU, and subject to π1(γL|x)=π1(γU|x). Similarly, a 100(1-ϕ/2)% two-sided Bayes interval estimate of μ, denoted by [μL,μU], can be obtained by solving

0μLπ2(ω|x)dω=ϕ2 (29)

and

μUπ2(ω|x)dω=ϕ2 (30)

for the limits μL and μU subject to π2(μL|x)=π2(μU|x). Finally, a 100(1-ϕ/2)% two-sided Bayes interval estimate of φ, denoted by [φL,φU] can be obtained by solving

0φLπ3(ω|x)dω=ϕ2 (31)

and

φUπ3(ω|x)dω=ϕ2 (32)

for the limits φU and φU subject to π3(φL|x)=π3(φU|x). Since the equations of the Bayes intervals cannot be solved analytically, the aforementioned steps can be used to compute them numerically, i.e., order the Bayes estimates and then obtain the specified lower and upper quantile points.

4 Sensitivity Analysis

To assess the change in the output by changing the output(s) is known as the sensitivity analysis. As the model is kept unchanged, this type of analysis provides insights about the data that are used to estimate the unknown quantities. For instance, if one model does not change too much by changing its inputs is considered better than a model that changes a lot by changing its inputs. In other words, the first model is less sensitive to the changes in the input data.

Since the prior information plays a crucial role in the Bayesian analysis, it is recommended to assess the effect of prior on the posterior summaries. Hence, prior sensitivity can be assessed to select more appropriate hyperparameters. To this end, we recompute the Bayes estimates and other posterior summaries by assuming different values of parameters of the gamma prior. Then, we compare the results with the posterior summaries obtained assuming informative prior. This aim of this sort of comparison is to find how much the results deviate by changing the original prior. We discuss numerical results of sensitivity analysis in the simulation study section.

5 Estimation of Future Record values

One of the aim of statistical inference is to make a prediction of a future unknown value. Thus, this section discusses the prediction of future record values for the ECD under Bayesian and frequentist approaches. In the literature, Basak and Balakrishnan (2003) studied the problem of future record statistic estimation by using the maximum likelihood method. Similarly, Madi and Raqab (2004) and Ahmadi et al. (2009) discussed the problem of estimating future record statistic for geometric and Burr type XII distribution, respectively.

Let n upper record values Xk(1),Xk(2),,Xk(n) from a population with density function f(x;ϕ) are available and our interest is to predict the mth record value, say, y=Xk(m),m>n. To this end, Basak and Balakrishnan (2003) defined the following joint predictive likelihood function of y=Xk(m).

L(y,ϑ,X)=i=1nh(Xu(i),ϑ)[H(y;ϑ)-H(Xk(n);ϑ)]m-n-1Γ(m-n)g(y;ϑ)

where

g(y;γ,μ,φ)=γμφyμ-1exp(yμ)exp(φ(1-exp(yμ)))
  (1-exp(φ(1-exp(yμ))))γ-1
S(y;γ,μ,φ)=1-(1-exp(φ(1-exp(yμ))))γ
G(y;γ,μ.φ)=(1-exp(φ(1-exp(yμ))))γ
H(y;ϑ)=-ln(1-G(y;ϑ))

Here H(y;ϑ) denotes the cumulative hazard function and the hazard function can be derived as

h(Xk(i);ϑ)=g(Xk(i);γ,μ,φ)S(Xk(i);γ,μ,φ)

The predictive likelihood function for the ECD is simplified as follows.

L(y,γ,μ,φ) =i=1n[γμφxk(i)μ-1exp(xk(i)μ)exp(φ(1-exp(xk(i)μ)))(1-exp(φ(1-exp(xk(i)μ))))γ-11-(1-exp(φ(1-exp(xk(i)μ))))γ]
×[-ln(1-(1-exp(φ(1-exp(yμ))))γ)+ln(1-(1-exp(φ(1-exp(xk(i)μ))))γ)Γ(m-n)](m-n-1)
×γμφyμ-1exp(yμ)exp(φ(1-exp(yμ)))
×(1-exp(φ(1-exp(yμ))))γ-1 (33)

The logarithm predictive likelihood function is given as

log(L) =(n+1)log(γ+μ+φ)+(μ-1)
×(i=1nlog(xk(i))+log(y))+i=1nlog(xk(i)μ)+log(yμ)
+i=1nφ(1-exp(xk(i)μ))+(γ+1)
×(i=1n(1-exp(φ(1-exp(xk(i)μ))))
+(1-exp(φ(1-exp(yμ)))))
+φ(1-exp(yμ))+(m-n-1)
×log[log(1-(1-exp(φ(1-exp(xk(i)μ))))γ)(1-(1-exp(φ(1-exp(yμ))))γ)]
-logΓ(m-n) (34)

To obtain the maximum likelihood predictors of the ECD parameters γ,μ and φ, partially differentiate Equation (34) with respect to γ,μ and φ and equating them to zero. As the resulting equations do not have explicit form, an iterative procedure like Newton Raphson is required to obtain the numerical values of predictions.

5.1 Bayes Prediction

This section presents prediction of future record values using the SELF. The prediction of future record values using the Bayesian approach is discussed by many scholars, see, for example, Mousa et al. (2002), Al-Hussaini and Ahmad (2003), Madi and Raqab (2004), El-Din et al. (2015), and references cited therein. Suppose n upper records Xk(1),Xk(2),,Xk(n) are available from the ECD and our interest is to find the Bayes estimate and Bayesian prediction intervals for the future upper record Xk(m), for some m>n, with a specified confidence. The conditional probability density function of y=Xk(m) for a given z=Xk(n) can be written as (Ahsanullah, 1995)

g(y|z;γ,μ,φ)=[logS(z;γ,μ,φ)-logS(y;γ,μ,φ)]m-n-1Γ(m-n)g(y;γ,μ,φ)S(z;γ,μ,φ)
y>z (35)

where

S(z;γ,μ,φ)=1-(1-exp(φ(1-exp(zμ))))γ

and

S(y;γ,μ,φ) =1-(1-exp(φ(1-exp(yμ))))γ
g(y|z;γ,μ,φ) =[log1-(1-exp(φ(1-exp(zμ))))γ-1-(1-exp(φ(1-exp(yμ))))γ]m-n-1Γ(m-n)
×γμφyμ-1exp(yμ)exp(φ(1-exp(yμ)))(1-exp(φ(1-exp(yμ))))γ-11-(1-exp(φ(1-exp(zμ))))γ (36)

As the record values follow the Markovian property, the future upper record y=Xk(m) given that X=Xk(1),Xk(2),,Xk(n) depends only on the current upper record z=Xk(n). Therefore, the conditional probability density function of y given x and the conditional probability density function of y given z is the same. The predictive probability density function of y given x is

g*(y|x)=000g(y|z,γ,μ,φ)×π(γ,μ,φ|x)dγdμdφ (37)

where g(y|z,γ,μ,φ) and π(γ,μ,φ|x) are defined previously. To obtain the 100(1-ϕ)% two-sided Bayes predicted intervals, solve the following

zyLg*(y|x)dy=ϕ2

and

yUg*(y|x)dy=ϕ2

where the lower limit is yL and the upper limit is yU.

5.2 Median Prediction

For calculating the conditional median prediction of the future record value, suppose that Xk(1),Xk(2),,Xk(n) be the upper record values from the ECD and our interest is to predict z=Xk(m), where m>n. As discussed previously that the record values have a Markovian structure, the conditional distribution of y given z is the same as given in Equation (5.1). The CDF of y given z by assuming the parameters γ,μ and φ are known, can be written as

G(z|y,γ,μ,φ)=0yg(τ|z,γ,μ,φ)dτ (38)

Further, supposing the inverse of CDF exists, i.e., G-1(v), the z^ can be obtained by solving G-1(v)=1/2 for v. Since this cannot be calculated easily in our case, a numerical procedure is used to compute it.

6 Results

Here, the results obtained by a simulation study are presented and discussed. Furthermore, sensitivity analysis for evaluating the impact of hyperparameters on the Bayes estimates is also discussed.

6.1 Simulation Study

Here, we present the performance of the maximum likelihood and Bayes estimators for the ECD based on record values assuming different sample sizes. Before presenting a detailed analysis, it is worth mentioning the generation of upper record samples from the ECD. Abdi and Asgharzadeh (2018) discussed a method for generating upper record values from any absolutely continuous distribution. To discuss Abdi and Asgharzadeh (2018) method for generating K1,K2,,Kn upper record values from an absolutely continuous probability distribution, the following steps are suggested.

(1) Generate m independent Exponential(1) observations Z1,Z2,,Zn.

(2) Compute Mj=Z1+Z2++Znforj=1,2,,n.

(3) Calculate Yj=1-exp(-Mj)forj=1,2,,m. Then, Y1,Y2,,Yn are the desired upper record values from the uniform(0,1) distribution.

(4) Finally, set Wj=G-1(Yj)forj=1,2,,n. Then, W1,W2,,Wn is the upper record values from the desired distribution.

For the ECD, Wj=[log{1-1φlog(1-Yj1/γ)}]1/μ.

An alternative procedure for generating upper record values is introduced by Pakhteev and Stepanov (2016). To this end, suppose the sequence Y1,Y2, of iid random variables with a continuous distribution G(.). Furthermore, we let the sequence Y1,Y2, satisfies the Markov Chain property, that is

P(Yn+1yn+1|Y(m)=yn)=G(yn+1)-G(yn)1-G(yn)  (yn+1>yn) (39)

Then, by using Equation (6.1), the formula to generate the future record values from ECD can be written as.

Yi =[log[1-1φlog[1-(i=1nWi((1-exp(φ(1-exp(x(n)μ))))γ
+(1-exp(φ(1-exp(x(n)μ))))γ))1/γ]]]1/μ

where Wi(0,1) and Yi are the desired record values from the ECD. It is to be noted that the accuracy of both methods is assessed and noticed that upper records generated by these methods are almost the same.

Table 1 Frequentist estimates of ECD with confidence intervals

n γ^MLE μ^MLE φ^MLE CIγ CIμ CIφ
10 1.77461 0.08225 1.04103 (0.0136, 3.5355) (0.0815, 0.0829) (0.9819, 1.1001)
(0.89844) (0.00037) (0.03016) (3.5219) (0.0014) (0.1182)
20 1.81214 0.09771 1.06165 (1.4575, 2.1668 ) (0.0974, 0.09798) (1.0586, 1.0646)
(0.18099) (0.00014) (0.00056) (0.7093) (0.0005) (0.0060)
30 1.39382 0.13089 1.14443 (1.3155, 1.4720) (0.1302, 0.1315) (1.1413, 1.1474)
(0.03992) (0.00035) (0.00155) (0.1565) (0.0013) (0.0061)
40 1.9193 0.23141 1.14219 (1.2062, 2.6323) (0.2290, 0.2337) (1.1187, 1.1656)
(0.36381) (0.00073) (0.01195) (1.4261) (0.0047) (0.0469)
50 1.9489 0.22391 1.65084 (1.8675, 2.0302) (0.2235, 0.2242) (1.6482, 1.6534)
(0.04150) (0.00017) (0.00131) (0.1627) (0.0007) (0.0052)

To generate record values from the ECD, the parameters γ=1, μ=1 and φ=1 are fixed and different sample sizes, n=10,20,30,40 and 50 are generated to obtain the maximum likelihood and Bayes estimates. Before proceeding further, it is worth mentioning that 50 is considered as the maximum sample size in our analysis, although records are of small sample size in practice. As 50 is a large sample, we generated this to evaluate the accuracy of our estimators. To compute the MLEs and their standard errors along with 95% confidence limits, we use the Newton Raphson method. The computed the maximum likelihood estimates and the 95% confidence limits using different sample size are listed in Table 1, where the maximum likelihood estimates are reported in the first row while the corresponding standard error in the second row within parentheses. In addition, two-sided 95% confidence limits are also reported in the first row while the corresponding confidence width in the second row. In the table, the abbreviation MLE refers to the maximum likelihood estimates and the CI stands for confidence intervals. As noticed previously, the estimates of the parameters increase with the sample size increases and the standard errors also fluctuate. In particular, the estimates of the parameters μ and φ gradually increased by increasing sample size. Columns 4-6 represent the confidence limits of the parameters and the corresponding confidence width. It is noticed that the maximum likelihood estimates are not consistent as claimed in the literature as they increase by increasing sample size.

images

Figure 2 Trace, density and history plots for simulated data.

Modern Bayesian analysis is heavily dependent on the Markov Chain Monte Carlo (MCMC) methods, which make this approach very flexible. In the past, solving complex integration to have an analytical answer was a major problem in Bayesian studies. Since the Bayes estimators cannot be obtained in closed form, we use a Metropolis-Hasting approach to compute the posterior summaries. Metropolis-Hasting is a powerful MCMC approach that does not depends on the conditional distribution compared to the Gibbs sampling. In particular, informative prior with the hyperparameters a=b=c=2 and d=e=f=2, which result into mean=1 and variance=0.5, is used for the parameters γ,μ and φ. The Bayes estimates are calculated using the SELF and results are tabulated in Table 2. To asses the convergence of the process, we also use graphical diagnostics, like trace, density, and history plots of MCMC which are plotted in Figure 2. The trace plot are depicted in Figure 2 for each of the parameters of the ECD to show the behavior of different sampled values over time. The density plots reflect the fact that the MH algorithm uses the proposal distribution to generate a mixture distribution. History plot gives the detailed view of the trace plot for each parameter. In our study, one can see from a graphical assessment that the chains converge quite well.

Table 2 Posterior summaries using informative prior

n Node Mean SD MC Error 2.5% Median 97.5%
10 γ 1.005 0.7078 0.00745 0.1191 0.8414 2.739
μ 1.012 0.7124 0.00658 0.1262 0.8479 2.837
φ 1.002 0.7149 0.00671 0.1186 0.8506 2.746
20 γ 1.006 0.7093 0.00512 0.1222 0.8474 2.791
μ 1.008 0.7150 0.00491 0.1204 0.8425 2.827
φ 0.999 0.7111 0.00453 0.1211 0.8456 2.747
30 γ 1.003 0.7082 0.00412 0.1209 0.8470 2.768
μ 1.004 0.7134 0.00419 0.1213 0.8369 2.824
φ 1.003 0.7128 0.00401 0.1213 0.8431 2.768
40 γ 1.001 0.7067 0.00369 0.1218 0.8446 2.771
μ 1.005 0.7163 0.00361 0.1207 0.8374 2.838
φ 1.003 0.7132 0.00348 0.1211 0.8424 2.777
50 γ 0.997 0.7044 0.00329 0.1206 0.8421 2.767
μ 1.002 0.7134 0.00325 0.1201 0.8357 2.822
φ 1.004 0.7146 0.00302 0.1213 0.8416 2.784

Table 3 Posterior summaries using non-informative priors

n Node Mean SD MC Error 2.5% Median 97.5%
10 γ 1.344 1.337 0.01353 0.0325 0.9290 4.876
μ 1.332 1.304 0.01167 0.0347 0.9412 4.809
φ 1.332 1.332 0.01307 0.0332 0.9159 4.914
20 γ 1.335 1.333 0.01023 0.0342 0.9213 4.857
μ 1.330 1.319 0.00817 0.0337 0.9351 4.903
φ 1.334 1.332 0.00914 0.0342 0.9218 4.920
30 γ 1.324 1.321 0.00745 0.0347 0.9167 4.808
μ 1.335 1.326 0.00737 0.0330 0.9346 4.924
φ 1.341 1.337 0.00773 0.0337 0.9304 4.933
40 γ 1.323 1.321 0.00698 0.0339 0.9124 4.822
μ 1.335 1.329 0.00590 0.0329 0.9339 4.938
φ 1.344 1.338 0.00613 0.0338 0.9331 4.954
50 γ 1.320 1.317 0.00606 0.0344 0.9122 4.814
μ 1.335 1.327 0.00551 0.0330 0.9324 4.949
φ 1.343 1.335 0.00548 0.0338 0.9314 4.952

To assess the effect of noninformative priors, the Bayes estimates are also computed using a=c=1 and b=d=0.75, which result into mean=1.33 and variance=1.78. It is noticed that the results obtained by using the non-informative prior, Table 3, are different than the informative prior case, Table 2, in the sense of higher standard error and MC error. As discussed in the previous section that proper assessment of prior sensitivity is very important, we re-computed the Bayes estimates assuming different hyperparameters values. The results are listed in Table 4 where the first row represents the Bayes estimates using hyperparameters that results 0.5 variance. Similarly, the other hyperparameters used in this study are listed in the first cell of each row of Table 4. To discuss results more precisely, we restrict our discussion to two cases, that is, informative versus non-informative prior. It can be seen from rows 2–4 of the table that if the variance is between 0.5 and 1, the Bayes estimates are approximately the same whereas the standard errors are slightly high. However, rows 5–6 indicate that when the variance of the prior distribution exceeds 1, the Bayes estimates are noticed larger than the prefixed nominal values and the standard errors are also quite large. Thus, it is concluded that the first case is less sensitive because the change in the priors from row 2–4 does not make any significant difference as compared to the results of row 1. Contrary to this, the second case is more sensitive, as the priors used in the last two rows produce different results than the rest.

Table 4 Sensitivity analysis using different priors

Prior Mean SD MC Error 2.5% Median 97.5%
γ(2,2) 0.997 0.7044 0.00329 0.1206 0.8421 2.767
μ(2,2) 1.002 0.7134 0.00325 0.1201 0.8357 2.822
φ(2,2) 1.004 0.7146 0.00302 0.1213 0.8416 2.784
γ(1.5,1.5) 1.001 0.7594 0.00341 0.0939 0.817 2.950
μ(1.5,1.5) 1.004 0.7635 0.00337 0.0957 0.8212 2.961
φ(1.5,1.5) 1.002 0.7576 0.00324 0.0987 0.8198 2.938
γ(1.2,1.2) 1.002 0.9165 0.00377 0.0417 0.7386 3.446
μ(1.2,1.2) 1.004 0.9124 0.00417 0.0442 0.7475 3.400
φ(1.2,1.2) 1.004 0.9209 0.00369 0.0449 0.7438 3.439
γ(1,1) 0.989 0.9877 0.00454 0.0258 0.6841 3.611
μ(1,1) 1.001 0.9955 0.00413 0.0248 0.6993 3.712
φ(1,1) 1.007 1.0010 0.00411 0.0253 0.6985 3.714
γ(1,0.75) 1.320 1.3170 0.00606 0.0344 0.9122 4.814
μ(1,0.75) 1.335 1.3271 0.00551 0.0330 0.9324 4.949
φ(1,0.75) 1.343 1.3350 0.00548 0.0338 0.9314 4.952
γ(2,1) 1.995 1.4091 0.00658 0.2412 1.6840 5.534
μ(2,1) 2.004 1.4270 0.00650 0.2401 1.6711 5.645
φ(2,1) 2.008 1.4290 0.00605 0.2427 1.6830 5.567

To obtain the prediction of a future record value under the methods discussed in Section 5, 50 random observations of upper record values are generated from the ECD with γ=μ=φ=1. The generated observations are given below.

0.002274756, 0.006270403, 0.045944343, 0.051727181, 0.058437952,
0.063922531, 0.101346749, 0.108894021, 0.111617942, 0.146417615,
0.189417577, 0.192859033, 0.210283739, 0.231513991, 0.268381071,
0.270724046, 0.294144448, 0.332739834, 0.359738550, 0.365599712,
0.381377851, 0.394667327, 0.401582031, 0.409507793, 0.453030212,
0.458217009, 0.459556415, 0.470093353, 0.476450090, 0.494024277,
0.536271210, 0.569521517, 0.614420238, 0.620284956, 0.627418804,
0.636715619, 0.637374831, 0.655008180, 0.656443801, 0.667602313,
0.673574564, 0.675018108, 0.675135805, 0.680501538, 0.681368387,
0.695071138, 0.711500379, 0.716533342, 0.726979873, 0.728704883.

Since the goal is to predict 51st record value on the basis of generated 50 upper record observations, we use a=b=c=2, and d=e=f=2 as the hyperparameter values. The Bayesian future prediction is 1.796 while the conditional median prediction is 1.421. The 95% Bayesian predictive interval for the future record is (0.1048-5.5780).

7 Real Data Analysis

A real data set taken from Lawless (2011) to illustrate the proposed methodology. The data set comprises of 24 observations, 0.014, 0.034, 0.059, 0.061, 0.069, 0.08, 0.123, 0.142, 0.165, 0.21, 0.381, 0.464, 0.479, 0.556, 0.574, 0.839, 0.917, 0.969, 0.991, 1.064, 1.088, 1.091, 1.174, 1.27, which represents the quantity of 1000s of cycles to failure for electrical appliances in a life test.

To assess the goodness of fit of the ECD and select appropriate model, we compare ECD with two parameter Chen, Weibull, exponential and generalized exponential distributions using the real data set. For model selection and goodness of fit assessment, we used Akaike Information Criterion (AIC), Bayesian Information Criterion (BIC), Kolmogorov-Smirnov (KS) statistic, Cramer-von Mises (CVM) statistic and Anderson-Darling (AD) statistic. It is noticed that among the mentioned distributions, Exponentiated Chen is a decent contender to fit this data set as the least value of all information criteria is observed in the case of ECD.

Table 5 Model selection for the real data

Goodness-of-fit ExpChen Chen Weibull Exp G Exp
AIC 13.60546 19.26221 21.84817 19.87926 21.86240
BIC 17.13962 21.61831 24.20428 21.05731 24.21850
KS 0.126029 0.155553 0.167998 0.167247 0.166566
CVM 0.060910 0.135710 0.145850 0.138593 0.133706
AD 0.381488 0.871534 0.917504 0.861168 0.826501

images

Figure 3 Visual goodness of fit assessment.

Table 6 Estimates for γ,μ and φ based on real data

Parameter MLE BIP BNIP
γ 1.2643 0.99771 1.32001
(2.8127) (0.70441) (1.31701)
μ 02127 1.00210 1.33511
(0.0036) (0.71342) (1.32702)
φ 1.0740 1.00402 1.34300
(0.1480) (0.71460) (1.33513)

The frequentist and Bayes estimates are listed in Table 5. For the Bayesian case, SELF is used to compute the Bayes estimates assuming informative and non-informative priors. To be more specific, independent gamma priors with with hyperparameters a=b=c=d=e=f=2 and a=c=e=1,b=d=f=0.75 to have informative and non-informative priors are considered. In the table, the abbreviations MLE, BIP, and BNIP are used for the maximum likelihood estimates, Bayes estimates using informative prior and Bayes estimates using non-informative prior, respectively.

The credible intervals for γ, μ and φ are (0.1191,2.739), (0.1262,2.837) and (0.1186,2.746), respectively. Similarly, the predicted future record value is 1.954 while the conditional median prediction is 1.696. The 95% Bayesian predictive interval for the future record value is (0.09344,4.4331). Thus, both the future prediction and the future conditional median fall within the 95% Bayesian interval. In addition, the graphical diagnostics using the Trace, Density and History plots are depicted in Figure 4. It is worth mentioning that the trace plots of each parameter of the ECD tell about the values that the parameter took during the runtime of the chain.

images

Figure 4 Trace, density and history plots for real data.

8 Conclusion

This study discussed the statistical inference for the three-parameter exponentiated Chen distribution based on upper record values. The reason of considering this positively skewed distribution is its flexibility in terms of bathtub-shaped hazard rate. Bayesian and frequentist methods are used to obtain the point estimates as well as the two-sided confidence intervals. To be more specific, the Bayes estimates are computed by using informative and non-informative priors, where independent gamma priors are assumed for the scale and the shape parameters. As the Bayes and ML estimators cannot be solved analytically, Newton Raphson and MCMC procedure are used to calculate the MLE and Bayes estimators. To evaluate the impact of hyperparameters, a sensitivity analysis is also presented in this article. In addition, two different methods for generating record values are presented in this study. A simulation study as well as a real data set is used to assess the efficiency of the numerical methods. For the real data set, the model selection is done using AIC, BIC, KS, CVM and AD measures. The results suggest that the exponentiated Chen distribution performs better than the other competitive lifetime distributions. In the future, this work can be extended by using the lower record values.

References

Abdi and Asgharzadeh (2018) Abdi, M. and Asgharzadeh, A. (2018). Rayleigh confidence regions based on record data. Journal of Statistical Research of Iran, 14(2):171–188.

Ahmadi et al. (2009) Ahmadi, J., Jozani, M. J., Marchand, É., and Parsian, A. (2009). Prediction of k-records from a general class of distributions under balanced type loss functions. Metrika, 70(1):19–33.

Ahsanullah (1995) Ahsanullah, M. (1995). Record Statistics. Nova Science Publishers.

Ahsanullah and Nevzorov (2015) Ahsanullah, M. and Nevzorov, V. B. (2015). Records via Probability Theory. Springer.

Al-Duais (2021) Al-Duais, F. S. (2021). Bayesian analysis of record statistic from the inverse Weibull distribution under balanced loss function. Mathematical Problems in Engineering, 2021(15):1–9.

Al-Hussaini and Ahmad (2003) Al-Hussaini, E. K. and Ahmad, A. E.-B. A. (2003). On Bayesian interval prediction of future records. Test, 12(1):79–99.

Alhamidah et al. (2023) Alhamidah, A., Qmi, M. N., and Kiapour, A. (2023). Comparison of E-Bayesian estimators in Burr XII model using E-PMSE based on record values. Statistics, Optimization & Information Computing, 1:709–718.

Arnold et al. (2011) Arnold, B. C., Balakrishnan, N., and Nagaraja, H. N. (2011). Records, volume 768. John Wiley & Sons.

Arshad and Jamal (2019) Arshad, M. and Jamal, Q. A. (2019). Estimation of common scale parameter of several heterogeneous Pareto populations based on records. Iranian Journal of Science and Technology, Transactions A: Science, 43:2315–2323.

Awodutire (2022) Awodutire, P. O. (2022). Statistical properties and applications of the exponentiated Chen-G family of distributions: Exponential distribution as a baseline distribution. Austrian Journal of Statistics, 51(2):57–90.

Azimi et al. (2023) Azimi, R., Esmailian, M., and Gallardo, D. (2023). The inverted exponentiated Chen distribution with application to cancer data. Japanese Journal of Statistics and Data Science, 6:213–241.

Basak and Balakrishnan (2003) Basak, P. and Balakrishnan, N. (2003). Maximum likelihood prediction of future record statistic. In Mathematical and Statistical Methods in Reliability, pages 159–175. World Scientific.

Chandler (1952) Chandler, K. N. (1952). The distribution and frequency of record values. Journal of the Royal Statistical Society, Series B (Methodological), 14(2):220–228.

Chaubey and Zhang (2015) Chaubey, Y. P. and Zhang, R. (2015). An extension of Chen’s family of survival distributions with bathtub shape or increasing hazard rate function. Communications in Statistics-Theory and Methods, 44(19):4049–4064.

Chen (2000) Chen, Z. (2000). A new two-parameter lifetime distribution with bathtub shape or increasing failure rate function. Statistics & Probability Letters, 49(2):155–161.

Dey et al. (2017) Dey, S., Kumar, D., Ramos, P. L., and Louzada, F. (2017). Exponentiated Chen distribution: Properties and estimation. Communications in Statistics-Simulation and Computation, 46(10):8118–8139.

Dziubdziela and Kopociński (1976) Dziubdziela, W. and Kopociński, B. (1976). Limiting properties of the k-th record values. Applicationes Mathematicae, 2(15):187–190.

El-Din et al. (2015) El-Din, M. M., Abdel-Aty, Y., Shafay, A., and Nagy, M. (2015). Bayesian inference for the left truncated exponential distribution based on ordered pooled sample of records. Journal of Statistics Applications & Probability, 4(1):1–11.

Eliwa et al. (2021) Eliwa, M. S., El-Morshedy, M., and Ali, S. (2021). Exponentiated odd Chen-G family of distributions: Statistical properties, Bayesian and non-Bayesian estimation with applications. Journal of Applied Statistics, 48(11):1948–1974.

Hastings (1970) Hastings, W. K. (1970). Monte carlo sampling methods using markov chains and their applications. Biometrika, 57(1):97–197.

Hjorth (1980) Hjorth, U. (1980). A reliability distribution with increasing, decreasing, constant and bathtub-shaped failure rates. Technometrics, 22(1):99–107.

Khan et al. (2016) Khan, M. S., King, R., and Hudson, I. L. (2016). Transmuted exponentiated Chen distribution with application to survival data. ANZIAM Journal, 57:268–290.

Khan et al. (2018) Khan, M. S., King, R., and Hudson, I. L. (2018). Kumaraswamy exponentiated Chen distribution for modelling lifetime data. Applied Mathematics & Information Sciences, 12(3):617–623.

Kumar and Gupta (2023) Kumar, R. and Gupta, R. (2023). Bayesian analysis of inverse Rayleigh distribution under non-informative prior for different loss functions. Thailand Statistician, 21(1):76–92.

Lawless (2011) Lawless, J. F. (2011). Statistical Models and methods for lifetime data, volume 362. John Wiley & Sons.

Madi and Raqab (2004) Madi, M. T. and Raqab, M. Z. (2004). Bayesian prediction of temperature records using the pareto model. Environmetrics, 15(7):701–710.

Martz and Waller (1982) Martz, H. and Waller, R. (1982). Bayesian Reliability Analysis. John Wiley & Sons, Inc.

Metropolis et al. (1953) Metropolis, N., Rosenbluth, A. W., Rosenbluth, M. N., Teller, A. H., and Teller, E. (1953). Equation of state calculations by fast computing machines. The Journal of Chemical Physics, 21(6):1087–1092.

Mousa et al. (2002) Mousa, M. A., Jaheen, Z., and Ahmad, A. (2002). Bayesian estimation, prediction and characterization for the gumbel model based on records. Statistics: A Journal of Theoretical and Applied Statistics, 36(1):65–74.

Méndez-González et al. (2023) Méndez-González, L. C., Rodríguez-Picón, L. A., Pérez-Olguín, I. J. C., and Vidal Portilla, L. R. (2023). An additive Chen distribution with applications to lifetime data. Axioms, 12(2).

Nagaraja (1977) Nagaraja, H. (1977). On a characterization based on record values. Australian Journal of Statistics, 19(1):70–73.

Nevzorov (1988) Nevzorov, V. B. (1988). Records. Theory of Probability & Its Applications, 32(2):201–228.

Pak et al. (2022) Pak, A., Raqab, M., Mahmoudi, M. R., Band, S. S., and Mosavi, A. (2022). Estimation of stress-strength reliability r=p(x>y) based on Weibull record data in the presence of inter-record times. Alexandria Engineering Journal, 61(3):2130–2144.

Pakhteev and Stepanov (2016) Pakhteev, A. and Stepanov, A. (2016). Simulation of gamma records. Statistics & Probability Letters, 119:204–212.

Rajarshi and Rajarshi (1988) Rajarshi, S. and Rajarshi, M. (1988). Bathtub distributions: A review. Communications in Statistics-Theory and Methods, 17(8):2597–2621.

Seo and Kim (2017) Seo, J. I. and Kim, Y. (2017). Objective Bayesian analysis based on upper record values from two-parameter Rayleigh distribution with partial information. Journal of Applied Statistics, 44(12):2222–2237.

Soliman and Al-Aboud (2008) Soliman, A. A. and Al-Aboud, F. M. (2008). Bayesian inference using record values from Rayleigh model with application. European Journal of Operational Research, 185(2):659–672.

Srivastava (1979) Srivastava, R. (1979). Two characterizations of the geometric distribution by record values. Sankhyā: The Indian Journal of Statistics, Series B, 40(3):276–278.

Xie and Lai (1996) Xie, M. and Lai, C. D. (1996). Reliability analysis using an additive Weibull model with bathtub-shaped failure rate function. Reliability Engineering & System Safety, 52(1):87–93.

Xie et al. (2002) Xie, M., Tang, Y., and Goh, T. N. (2002). A modified Weibull extension with bathtub-shaped failure rate function. Reliability Engineering & System Safety, 76(3):279–285.

Yousaf et al. (2019) Yousaf, F., Ali, S., and Shah, I. (2019). Statistical inference for the Chen distribution based on upper record values. Annals of Data Science, 6: 831–853.

Biographies

Farhad Yousaf is currently pursuing his PhD statistics from University of Naples Federico II, Naples, Italy. Prior to this he completed his MPhil in Statistics from Department of Statistics, Quaid-i-Azam University, Islamabad, Pakistan. His research interest is focused on Bayesian inference, socia network analysis, and data science.

Sajid Ali is Associate Professor at the Department of Statistics, Quaid-i-Azam University (QAU), Islamabad, Pakistan. He graduated (PhD Statistics) from Bocconi University, Milan, Italy. His research interest is focused on Bayesian inference, construction of new flexible probability distributions, electricity market modeling, and process monitoring.

Ismail Shah is currently serving as a researcher at the university of Padua. He is also an Associate Professor with the Department of Statistics, Quaid-i-Azam University, Islamabad, Pakistan. He received PhD degree in Statistics from University of Padova, Italy. His research interest areas are: Functional Data Analysis, Time Series Modeling, Electricity Market Modeling and Applied Statistics. Currently, he is also the editor of Journal of Quantitative Methods https://ojs.umt.edu.pk/index.php/jqm.

Saba Riaz is currently an assistant professor at Department of statistics, Rawalpindi Women University. She completed her PhD in Statistics from Quaid-i-Azam University (QAU), Islamabad, Pakistan. Her research interests are: survey sampling, and applied statistics.

Abstract

1 Introduction

images

2 Maximum Likelihood Estimation

2.1 Likelihood Estimators

2.2 Asymptotic Confidence Intervals

3 Bayesian Estimation

3.1 Bayes Intervals

4 Sensitivity Analysis

5 Estimation of Future Record values

5.1 Bayes Prediction

5.2 Median Prediction

6 Results

6.1 Simulation Study

images

7 Real Data Analysis

images

images

8 Conclusion

References

Biographies