Exact Two Sample Bayesian Prediction Bounds For The Exponential Distribution With Random Sample Size
Exact Two Sample Bayesian Prediction Bounds For The Exponential Distribution With Random Sample Size
Key Words and Phrases: Two sample, Order statistics, random sample size, pre-
dictive distribution, exponential model, probability coverage, average width.
Abstract:
This paper is concerned with the Bayesian prediction limits problem of the sth
order statistics when the size of the future sample is a random variable. The one pa-
rameter exponential distribution with constant failure rate is considered. Using the
gamma prior, the paper derives the posterior distribution for that failure rate and
hence the predictive distribution of future observations. Available data are from type
II censored sampling. The analysis depends mainly on assuming that the future sam-
ple size m is a random variable having Poisson or binomial distribution. Simulation
study and numerical examples are used to illustrate the procedure.
1 Introduction
In this paper, we consider the Bayesian prediction for future observations from ex-
ponential distribution in both (1) fixed sample size FSS and (2) random sample size
RSS. In next section, we derive the exact form for Bayesian predictive function in FSS
see, for example, Lawless (1971), Likes (1974), Lingappaiah (1973), Abd Ellah (1999,
2003), Abd Ellah and Soliman (2005),Lee (1997), Lindsey (1996), Huber -Carol, Bal-
akrishnan, Nikulin and Mesbah (2002), Johnson, Kotz and Balakrishnan (1994). In
section 3, we derive the Bayesian predictive function when the sample size is dis-
tributed as a Poisson pdf. In section 4, we derive the Bayesian predictive function
1
when the sample size is distributed as a binomial pdf. In section 5, we shown the
classical approach in pervious work. In section 6, in order to show the usefulness of
our results presented in the paper, we discuss simulation study and numerical exam-
ples. Finally, we present applications to illustrate the methods of inference developed
here. In many biological and quality control problems, the sample size of the future
sample can not always be taken as fixed. Lingappaiah (1986) assumed that the sam-
ple size of the future sample is a random variable. Under this assumption, he derived
the Bayesian predictive distribution of the range when the parent distribution is a
one parameter exponential distribution see, for example, Buhrman (1973), Raghu-
nandanan and Patil (1972). Upadhyay and Pandey (1989) have determined similar
prediction intervals for one parameter exponential distribution with fixed future sam-
ple size. Problems for constructing prediction limits have also been considered by
Hahn and Meeker (1991), Giesser (1993)and Soliman and Abd-Ellah (1996, 2006), .
Several authors, including Aitchison and Dunsmore (1975), Nelson (1982), Lawless
(1982), Cohen and Whitten (1988), Balakrishnan and Cohen (1991), Cohen (1991),
Bain and Engelhardt (1991), Meeker and Escobar (1998), Gelman, Carlin, Proschan
(1963), Stern and Rubin (1995), Gertsbakh (1989), Nagaraja (1995), Nelson (1982),
Patel (1989), Balakrishnan and Lin (2002, 2003) .
Let x1:n ≤ x2:n ≤ · · · ≤ xr:n , the observed ordered lifetimes of the first r compo-
nents to fail in a sample of n (r ≤ n) components. Let y1:m ≤ y2:m ≤ · · · ≤ xm:m , be
a second independent random sample of size m of future observation from the same
distribution, we assume that n,m components whose lifetimes have the one parameter
exponential distribution
Suppose that the survival time of some components may be modelled by equation
(1.1) and suppose that n new components have been established at a given point in
time and that the survival times of the first r of these to fail are Let x1:n ≤ x2:n ≤
· · · ≤ xr:n , the problem under consideration is to make inferences about the survival
times of the future ys:m in the future sample of m lifetimes. Assuming that the survival
times are independent, the likelihood function for a type II censored sample is
n!
L(x|λ) = λr e−λSr (1.2)
(n − r)!
Pr
where Sr = i=1 xi:n + (n − r)xr:n . We assume a 2 parameter gamma prior for λ
1
g(λ|a, b) = λa−1 e−λ/b , a, b > 0 (1.3)
Γ(a)ba
Where a and b known.
2
Combining the prior distribution in (1.3) with the likelihood function in (1.2) via
Bayes theorem yields the posterior distribution of λ
1 R−1 R −λA
P (λ|x) = λ A e (1.4)
Γ(R)
where
R = r+a
A = Sr + 1/b (1.5)
The Bayes predictive density function of y given x denoted by
Z ∞
h(y|x) = f (x|λ)P (λ|x) dλ (1.6)
0
3
where A and R are given in (1.5). The percentage points of the predictive cdf given
in (2.2) can be easily obtained by solving the nonlinear equation
G1 (t) = 1 − τ (2.3)
Then the exact 100τ % two sided Bayesian interval for the future ys:m can be con-
stracted as
tτ /2 , t1−τ /2 (2.4)
P r(y ≤ t1 ) = τ /2
P r(y ≥ t2 ) = τ /2 (2.5)
where t1 = tτ /2 and t2 = t1−T au/2 are the lower and upper percentage points of
G1 (t) . More precisely, given a value of τ , we may determine values of t1 and t2 .
4
exponential with mean 1 as follows: .107, .107, .569, .929, 1.061, 1.240, 1.546, 1.626,
1.778, 1.951, 2.413. The first 6 from the above order statistics are used together with
the percentage points of Bayesian that calculated from (2.2), then the predictive con-
fidence interval (2.5) is constructed for r = 6 and s = 1, 2, 3, 4, 5 with hyperparameter
of prior (known) a=1.01 and b=2.01 in the table given below:
r s 1% 2.5% 5% 10% 90% 95% 97.5% 99 %
6 1 1.2468 1.2572 1.2749 1.3124 3.2993 4.1343 5.0849 6.5390
2 1.3291 1.3871 1.4591 1.5754 4.8862 6.1256 7.5171 3.6254
3 1.4837 1.5936 1.7187 1.9077 6.4412 8.0685 9.8854 12.6276
4 1.6798 1.8403 2.0168 2.2756 8.0145 10.0296 12.2706 15.6433
5 1.8765 1.9403 2.178 2.756 8.4215 10.9816 12.2706 15.6433
From the above table, we see that, both the upper and lower 1 − τ bounds of the
predictive intervals are greater than xs:11 . Also, the bounds increase as the significant
level increases as well as s increases.
5
Hence to find the prediction limits for ys:m = y the sth smallest of a set m future
observations with pdf in (3.3) we choose t7 and t8 such that
P r(t7 ≤ y ≤ t8 ) = 1 − τ (3.5)
P r(y ≤ t7 ) = τ /2
P r(y ≥ t8 ) = τ /2 (3.6)
From (3.6)
Z t7
g2 (y) dy = τ /2
Z 0∞
g2 (y) dy = τ /2 (3.7)
t8
where A and R are given in (1.5). The percentage points of the predictive cdf given
in (3.8) where t7 = tτ /2 and t8 = t1−T au/2 are the lower and upper percentage points
of G2 (t).
e−θ X
∞
A Rm
G2 (t) = 1 − (3.9)
Ps m=1 A + t
G2 (t9 ) = τ /2
G2 (t10 ) = 1 − τ /2 (3.10)
6
3.2 Prediction Limits of ym:m with m ∼ Poisson(θ)
Put s=m in (3.8), and from (3.3)
m−1
e ∞ m−1
−θ X X (−1)i i
A
Ri
G2 (t) = 1 − (3.11)
Ps m i=0 i A+t
and
G2 (t11 ) = τ /2
G2 (t12 ) = 1 − τ /2 (3.12)
From the above table, we see that, both the upper and lower 1 − τ bounds of the
predictive intervals are greater than xs:11 . Also, the bounds increase as the significant
level increases as well as s increases.
7
using (3.1), (3.3) and (4.1)
m M
M
sq RA R M
X (p/q)m s m
A
R s−1
A
R(m−s)
g3 (y|x) = R+1
1− (4.2)
Bs m=s (A + y) A+y A+y
where
M
!
X M m M −m
Bs = p q (4.3)
m=s m
Hence to find the prediction limits for ys:m = y the sth smallest of a set m future
observations with pdf in (4.2) we choose t13 and t14 such that
P r(t13 ≤ y ≤ t14 ) = 1 − τ (4.4)
If we consider equal tail limits, the probability in (4.4) gives
P r(y ≤ t13 ) = τ /2
P r(y ≥ t14 ) = τ /2 (4.5)
From (4.5)
Z t13
g3 (y) dy = τ /2
Z0 ∞
g3 (y) dy = τ /2 (4.6)
t14
G3 (t15 ) = τ /2
G3 (t16 ) = 1 − τ /2 (4.9)
Using (4.8) and (4.9), we can determin values of LB and UB.
8
4.2 Prediction Limits of ym:m with m∼ binomial(m;M,p)
Put s=m in (4.7), and from (4.2)
m−1
M
R X
M m−1 (−1)i
! Ri
A A
m
X i
G3 (t) = 1 − (p/q) (4.10)
Bs m m i=0 i A+t
and
G3 (t17 ) = τ /2
G3 (t18 ) = 1 − τ /2 (4.11)
Using (4.10) and (4.11), we can determin t17 and t18 of LB and UB of G3 (t)
Example 3
Follow the technique used in the above example, we use the truncated binomial
B(m; 15, 0.5) as a generator for the sample size, we get n = 12. Next we get 12
order statistics from exponential as:.024, .049, .085, .124, .154, .253, .279, .735, 1.107,
1.750, 2.258, 2.406. By making use the first 6 of above order statistics and the calcu-
lated percentage points of Bayesian given in (4.7), the predictive confidence intervals
for the future order statistics when r = 6 and s = 1, 2, 3, 4, 5 are estimated according
(4.6) with hyperparameter of prior (known) a=1.01 and b=2.01 as given below:
The entries of the above table tell that, both the upper and lower (1 − τ ) bounds of
the predictive confidence intervals are greater than xs:12 and increase as the significant
level increases as well as s increases.
my1:m Sr
∼ F (2, 2r), where x̄ = (5.12)
x̄ r
9
we choose l1 and l2 such that
my1:m
P r l1 < < l2 = 1 − τ (5.13)
x̄
For equal tail limits l1 and l2 are the lower and upper τ /2 cdf points of F (2, 2r).
These are the classical prediction limits for y1:m
P r C1 < y1:m < C2 = 1 − τ (5.14)
Thus knowing l1 and l2 we can obtain Cl and C2 the lower and upper limits for
y1:m .
The percentage points (factors) of the distributions of Classical and Bayesian when
the sample size n is distributed as P (m; λ) and b(m; M, p) were calculated. The
routine DZREAL from Mathematica were used for solving the nonlinear equations
when n distributed as:
(i) Possion P (m; θ) with θ = 1, 2, 3, (ii) Binomial B(m; M; p) with M = 15 and
p = 0.25, 0.50, 0.75. Tables for different choices of r, s and α are available with the
authors upon request. In order to examine the efficiency of our technique when the
sample size is random variable, the probability coverages and average width of the
predictive confidence intervals are calculated based on 10000 simulations in Tables
1, 2, 3 and 4 see, for example, Rubinstein (1981), Van Dorp and Mazzuchi (2004),
Witkovsky (2002), Sugita (2002), Stuart and Ord (1994). From Tables 1 and 2, we
see that the probability coverages are quite close to the corresponding confidence
levels including 90%, 95%, 97.5% and 99%. From Tables 3 and 4, we see that in each
case the average width decreases as both θ and p increase. Also, the average width
increases as r increases and s decreases. For seek of completeness and comparisons,
the probability coverages and the average width in the case of fixed sample size are
also calculated along in Tables 1,2 ,3 and 4. In conclusion, we can say that our
new technique presented in the papers in the case of random sample size is work
satisfactory as well as the technique when the sample size in fixed. By making use of
the percentage points, we simulate numerical examples of random sample sizes from
the exponential lifetime when the sample size distributed as P (m; 2) and B(m; 15, 0.5)
as follows:
10
Example 4
In this example, follows the same technique in Example 1 but when m distributed as
binomial. Then the truncated binomial generator B(m; 15, 0.25) from Mathematica
gives the sample size to be 11. Again, we generate 11 order statistics from exponential
as:.017, .173, .183, .223, .382, .391, .499, .581, .687, 2.165, 2.632. Once again, by
making use the first 6 of above order statistics together with the calculated percentage
points of Bayesian by solving the nonlinear equation (4.7). Then predictive confidence
interval for the future order statistics when r = 6 and s = 1, 2, 3, 4, 5 are estimated
according (4.6) as given below:
Again, from the above table, we see that, both the upper and lower (1 − τ ) bounds of
the predictive confidence intervals are greater than xs:11 . Also, these bounds increase
as the significant level increases as well as s increases.
Example 5
The truncated Poisson distribution P (m; 2) is used again to generate the sample size
to be 12. Next, 12 order statistics are generated from exponential as: .024, .049, .085,
.253, .279, .735, 1.107, 1.750, 1.926, 2.258, 2.406, 3.149. The first 6 from the above
sample of order statistics are used together with the percentage points of Bayesian
calculated from (3.8). Then the predictive confidence intervals given in (3.7) are
calculated for r = 6 and s = 1, 2, 3, 4, 5 in the table given below:
The entries of the above table tell that, both the upper and lower (1 − τ ) bounds of
the predictive confidence intervals are greater than xs:12 and increase as the significant
level increases as well as s increases.
11
7 Applications
In this section, we apply our predictive technique to some real data that follow expo-
nential distribution as listed below:
1. Consider the data given in Lawless (1971) in which the test is terminated after 4
failures, they are 30, 90, 120 and 170 hours. By using these four times and apply
our predictive confidence intervals given in Sections 3 ,4 and 5, we may able
to predict the 95% and 99% predictive confidence intervals for the fifth failure
when the sample size m is distributed as P (m; 1) and B(m; 15, 0.25). Then
predictive confidence intervals when the sample size is fixed is also calculated
as given below:
From the above table, we see that the lower bound of the predictive intervals
for the same level of significant are somewhat closed. Also, both of Classical
and Bayesian are almost give the same results for both fixed and random sam-
ple sizes. In addition, the interval width increases as the level of prediction
increases.
2. The following data given by Proschan (1963), are times between successive
failure of air conditioning equipment in a Boeing 720 airplane, arranged in
increase order of magnitude. The first 9 of the data are 12, 21, 26, 29, 29,
48, 57, 59 and 70. Upon using the percentage points of Classical and Bayesian
given in Sections 3,4 and 5, we predict 95% and 99% confidence intervals for
the 10th observation as given below:
12
From the above table, we see that the predictive confidence intervals are quite
closed for the same level of significant and their widths increase as the prediction
level increases. We also notice that, Classical gives intervals more wider than
the intervals based on Bayesian.
4. Let us consider the following data that is the failure times (in minutes) for a
specific type of electrical insulation in an experiment in which the insulation
was subjected to a continuously increasing voltage stress [Lawless (1982)]. The
first 9 observations of the data are 12.3, 21.8, 24.4, 28.6, 43.2, 46.9, 70.7, 75.3
and 98.1. Once again by using the percentage points of Classical and Bayesian
given in Sections 2 and 3, we predict 95% and 99% confidence intervals for the
13
10th observation as given below:
We see from the above table that the predictive confidence intervals are closed
for the same level of significant and their widths increase as the prediction level
increases. Also, we notice that Classical gives intervals more wider than the
intervals based on Bayesian.
5. Let us consider the data obtained from an experiment on insulating fluid break-
down [Nelson (1982)]. Of the 12 specimens tested at 45KV, 3 failed before 1
second and the time to breakdown (in second) of the remaining 9 specimens
were as: 2, 2, 3, 9, 13, 47, 50, 55 and 71. By using the percentage points
of Classical and Bayesian given in Sections 2 and 3, we predict 95% and 99%
confidence intervals for the 5th observation as given below:
From the above table, we see that the predictive confidence intervals are closed
for the same level of significant and their widths increase as the prediction level
increases. We also, notice that Classical gives intervals more wider than the
intervals based on Bayesian.
14
sample size is random variable, the probability coverages and average width of the
predictive confidence intervals are calculated based on 10000 simulations in Tables 1,
2, 3 and 4. From Tables 1 and 2, we see that the probability coverages are quite close
to the corresponding confidence levels including 90%, 95%, 97.5% and 99%. From
Tables 3 and 4, we see that in each case the average width decreases as both θ and p
increase. Also, the average width increases as r increases and s decreases. For seek
of completeness and comparisons, the probability coverages and the average width
in the case of fixed sample size are also calculated along in Tables 1,2 ,3 and 4. In
conclusion, we can say that our new technique presented in the papers in the case of
random sample size is work satisfactory as well as the technique when the sample size
in fixed. By making use of the percentage points, we simulate numerical examples of
random sample sizes from the exponential lifetime when the sample size distributed
as P (m; 2) and B(m; 15, 0.5) as follows:
Example 6
By using the Mathematica, the truncated Poisson P (m; 2) gives the sample size n =
11. Next, we generate 11 order statistics from exponential with mean 1 as follows:
.107, .107, .569, .929, 1.061, 1.240, 1.546, 1.626, 1.778, 1.951, 2.413. The first 6 from
the above order statistics are used together with the percentage points of Bayesian
that calculated from (3.8), then the predictive confidence interval (3.7) is constructed
for r = 6 and s = 1, 2, 3, 4, 5 in the table given below:
From the above table, we see that, both the upper and lower 1 − τ bounds of the
predictive intervals are greater than xs:11 . Also, the bounds increase as the significant
level increases as well as s increases. intervals.
Example 7
In this example, follows the same technique in Example 1 but when m distributed as
binomial. Then the truncated binomial generator B(m; 15, 0.25) from IMSL routines
gives the sample size to be 11. Again, we generate 11 order statistics from exponential
as:.017, .173, .183, .223, .382, .391, .499, .581, .687, 2.165, 2.632. Once again, by
making use the first 6 of above order statistics together with the calculated percentage
points of Bayesian by solving the nonlinear equation (4.7). Then predictive confidence
15
interval for the future order statistics when r = 6 and s = 1, 2, 3, 4, 5 are estimated
according (4.6) as given below:
Again, from the above table, we see that, both the upper and lower (1 − τ ) bounds of
the predictive confidence intervals are greater than xs:11 . Also, these bounds increase
as the significant level increases as well as s increases.
Example 8
The truncated Poisson distribution P (n; 2) is used again to generate the sample size
to be 12. Next, 12 order statistics are generated from exponential as:.024, .049, .085,
.253, .279, .735, 1.107, 1.750, 1.926, 2.258, 2.406, 3.149. The first 6 from the above
sample of order statistics are used together with the percentage points of Classical
calculated from (3.8). Then the predictive confidence intervals given in (3.7) are
calculated for r = 6 and s = 1, 2, 3, 4, 5 in the table given below:
The entries of the above table tell that, both the upper and lower (1 − τ ) bounds of
the predictive confidence intervals are greater than xs:12 and increase as the significant
level increases as well as s increases.
Example 9
Follow the technique used in the above example, we use the truncated binomial
B(m; 15, 0.5) as a generator for the sample size, we get n = 12. Next we get 12
order statistics from exponential as:
.024, .049, .085, .124, .154, .253, .279, .735, 1.107, 1.750, 2.258, 2.406.
By making use the first 6 of above order statistics and the calculated percentage
points of Classical given in (4.7), the predictive confidence intervals for the future
16
order statistics when r = 6 and s = 1, 2, 3, 4, 5 are estimated according (4.6) as given
below:
The entries of the above table tell that, both the upper and lower (1−τ ) bounds of
the predictive confidence intervals are greater than xr:12 and increase as the significant
level increases as well as s increases.
Comment
1. If r increases, the upper and lower prediction limits increases for all cases.
3. If m and θ increases, the upper and lower prediction limits using random sample
size is greater than that obtained by using fixed sample size.
REFERENCES
Abd Ellah, A. H. (1999) A prediction problem concerning sample from the exponen-
tial distribution with random sample size Bull.Fac.SCi., Assiut Univ.28(1-c),
pp1-8
Abd Ellah, A.H. (2003) Bayesian one sample prediction bounds for Lomax distribu-
tion, Indian J. Pure Appl. Math., 43(1), 101-109.
Abd Ellah, A.H. and Sultan, K. S (2005) Exact Bayesian prediction of exponential
lifetime based on fixed and random sample sizes, QTQM Journal , vol 2,no.
2,pp 161-174.
17
Aitchison, J. and Dunsmore , I.R. (1975).Statistical Prediction analysis, Cambridge
university press, Cambridge.
Balakrishnan, N. and Lin, C. T. (2002). Exact linear inference and prediction for
exponential distributions based on general progressively Type-II censored sam-
ples, The Journal of Statistics Computation and Simulation, 72,, 8, 677-686.
Buhrman, J.M. (1973). Order statistics when the sample size has a binomial distri-
bution, Statistica Neerlandica , 27, 125-126.
David, H.A. and Nagaraja, H.N. (2003). Order Statistics, John Wiley & Sons, New
York.
Dunsmore, I.R. (1974). The Bayesian predictive distribution in life testing model,
Technometrics, 16, 455-460.
Escobar, L.A. and Meeker, W.Q. (1999). Sattistical prediction based on censored
Life data, Technometrics, 41(2), pp. 113-124.
18
Hastings, K. (2001).Introduction to Probability with Mathematica, Chapman & Hall.
Huber-Carol, C., Balakrishnan, N., Nikulin, M.S. and Mesbah, M. (2002). Goodness
- of- Fit Tests and Model Validity, Mirkhäuser Boston. Basel . Berlin.
Lawless, J.F. (1971). A prediction problem concerning samples from the exponential
distribution with application in life testing, Technometrics, 13, 725-730.
Lawlees, J.F. (1982). Statistical Models & Methods For Lifetime Data, Wiley &
Sons, New York.
Likes, J. (1974). Prediction of sth ordered observation for the two parameter expo-
nential distribution, Technometrics, 16,241-244.
Gelman, A., Carlin, J.B., Stern, H.S. and Rubin, D. B.(1995).Bayesian Data Anal-
ysis, New York: Chapman & Hall
Meeker, W.Q. and Escobar, L.A (1998).Statistical Methods for Reliability Data,John
Wiley & Sons, New York.
Nelson, W. (1982). Applied Life Data Analysis,,John Wiley & Sons, New York.
19
Patel, J.K. (1989).Prediction intervals - A review, Commun. Statist. - Theor. Meth.
, 18, 2393-2465.
Raghunandanan, K. and Patil, S.A. (1972).On order statistics for random sample
size, Statistica Neelandica, 26, 121-126.
Rubinstein, R.Y.(1981).Simulation and the Monte Carlo Method ,John Wiley &
Sons, New York.
Soliman,A.A. and Abd Ellah, A.H. (1996).Additional factors for calculating pre-
diction intervals for sth ordered observation from one parameter exponential
distribution, Bull. Fac. Sci., Assiut Univ, 25(2-c), 1-18.
Stuart, A.and Ord, K.J. (1994).Kendall’s Advanced Theory of Statistics, New York:
Edward.
Van Dorp, J.R. and Mazzuchi, T.A. (2004) A general Bayes exponential inference
model for accelerated life testing, J. Stat. Pla. and Inf.
20
Table 1: Probability coverage
21
Table 2: Probability Coverage
22
Table 3: Average Width
23
Table 4: Average Width
24