Hong Chapter
Hong Chapter
net/publication/335420172
CITATIONS READS
0 789
4 authors:
Lu Lu Yili Hong
University of South Florida Virginia Tech (Virginia Polytechnic Institute and State University)
74 PUBLICATIONS 741 CITATIONS 131 PUBLICATIONS 2,111 CITATIONS
All content following this page was uploaded by Yili Hong on 30 August 2019.
Abstract
Traditional reliability analysis has been using time to event data, degradation data,
and recurrent event data, while the associated covariates tend to be simple and constant
over time. Over the past years, we have witnessed the rapid development of sensor
and wireless technology, which enables us to track how the product has been used and
under which environmental conditions it has been used. Nowadays, we are able to
collect richer information on covariates which provides opportunities for better reliability
predictions. In this chapter, we first review recent development on statistical methods
for reliability analysis. We then focus on introducing several specific methods that were
developed for different types of reliability data with covariate information. Illustrations
of those methods are also provided using examples from industry. Test planning is also
an important part of reliability analysis. In addition to data analysis, we also provide a
briefly review on recent developments of test planning and then focus on illustrating the
sequential Bayesian design with an example of fatigue testing for polymer composites.
The chapter is concluded with some discussions and remarks.
Key Words: Degradation data, Dynamic covariates, Lifetime data, Recurrent
events data, Reliability prediction, Sequential test planning.
1
1 Introduction
1.1 Background
Traditional reliability data analysis mainly use time to event data, degradation data, and
recurrent event data to make reliability predictions [1]. The covariate information involved in
the reliability analysis is usually time-invariant and the number of covariates is typically small.
For time to event data, parametric models such as the Weibull and lognormal distributions are
popular and accelerated failure time models are often used to incorporate covariate information
on accelerating factors. For degradation data, the general path models and stochastic models
are the common choices and the covariate information is often incorporated through regression
type of models. The recurrent event data are often modeled by the event intensity models or
mean cumulative functions with regression type of models that are often used to incorporate
covariates.
With technological advances, new information on covariates become available. Products
and systems can be equipped with sensors and smart chips to keep track of various information
on the field usage of product units, number of transfers, and environmental conditions such as
temperature and humidity. Such covariate information often change over time, so we refer to
them as dynamic covariate information. Because the dynamic covariates often come in large
volume and variety, it presents big data opportunities and challenges in the area of reliability
analysis (e.g., [2] and [3]). Dynamic covariate data can be used for modeling and prediction
of reliability because units under heavy usage often fail sooner than those lightly used. In
recent years, more statistical methods for dynamic covariates have been being developed to
make use of this new type of covariate data.
Another important area of reliability analysis is about test planning, which focuses on
how to efficiently collect various types of data to make better prediction of reliability. For
accelerated life tests (ALTs), it is especially challenging to timely collect sufficient failure data
because the data collection is a time-consuming process and often requires using expensive
equipment for testing units under elevated stress conditions. In some laboratories, there
are typically one or two machines available for testing certain material. In this case, it is
impractical to test multiple samples simultaneously and therefore limits the total obtainable
sample size. Another challenge with traditional test planning is that it typically relies on a
single set of best guess of the parameter values, which may lead to suboptimal designs when
the specified parameter values are not accurate. Due to these challenges, sequential designs
become popular where earlier test results can be utilized to determine the test conditions for
later runs. In addition, Bayesian methods can be used to leverage prior information from the
expert’s knowledge or related historical data to inform the test planning. The objective of
2
this chapter is to review current development and then introduce the statistical methods for
dynamic covariates and sequential Bayesian design (SBD) for ALT.
1.3 Overview
The rest of this chapter is organized as follows. Section 2 describes an application on time-
to-event data with dynamic covariates. Section 3 illustrates the modeling of degradation with
dynamic covariates. Section 4 describes the MTRP model for describing recurrent event data
with dynamic covariates. Section 5 introduces SBD strategies for ALTs. Section 6 contains
3
20
10.0
5.0
15
2.0
Use Rate
Unit No.
10
1.0
0.5
5
0.2
1
censored failed
0 10 20 30 40 50 60 70 0 10 20 30 40 50 60 70
Figure 1: (a) The event plot for a subset of the Product D2 failure-time data and (b) the
corresponding plot of the use-rate trajectories. Figure reproduced with permission.
4
2.2 Model for Time to Event and Parameter Estimation
Three sets of observable random variables: the failure time, censoring indicator and dynamic
covariate over time are are considered, which are denoted by { T, ∆, X(T ) }. The observed
data are described by { ti , δi , xi (ti ) }. Here n denotes the number of units in the dataset, ti is
the failure time or time in service, and δi is the observed censoring indicator (i.e., it equals
to 1 if unit i fails and 0 otherwise). The xi (ti ) = {xi (s) : 0 ≤ s ≤ ti } is the observed dynamic
covariate information of unit i from the time 0 to ti , where xi (s) is the observed covariate
value at time s for unit i. Particularly for Product D2, we use X(t) = log[R(t)/R0 (t)] as the
form of the covariate in the model, where R0 (t) is the baseline use-rate that is chosen to be a
typical constant use rate.
The cumulative exposure model in [45] is used to model the failure-time data with dynamic
covariate. The cumulative exposure u(t) is defined as,
Z t
u(t) = u[t; β, x(t)] = exp[βx(s)]ds,
0
where β represents the influence of the covariate on the exposure. When the cumulative
exposure of a unit reaches a random threshold U at time T , the unit fails. This establishes a
relationship between U and T , that is,
Z T
U = u(T ) = exp[βx(s)]ds. (1)
0
Under the above model and the covariate history x(∞), the cumulative distribution function
(cdf) of the failure time T is
and probability density function (pdf) is f (t; β, θ0 ) = exp[βx(t)]f0 {u[t; β, x(t)]; θ 0 } . Here θ 0
is the parameter in the baseline cdf of the cumulative exposure threshold U and f0 (u; θ0 ) is
the pdf of U. In the Product D2 application, the baseline cumulative exposure distribution
F0 (u; θ0 ) was modeled by the Weibull distribution, of which the cdf and pdf are
log(u) − µ0 1 log(u) − µ0
F0 (u; θ0 ) = Φsev and f0 (u; θ0 ) = φsev .
σ0 σ0 u σ0
In the above expression, θ 0 = (µ0 , σ0 )′ , where µ0 and σ0 are the location and scale parameters.
Also, Φsev (z) = 1 − exp[− exp(z)], and φsev (z) = exp[z − exp(z)]. Lognormal and other log-
location-scale distributions can also be used if they are considered appropriate for certain
applications.
5
2.3 Model for Covariates
To model the covariate process, we use the linear mixed effect model. In particular, X(t) is
modeled as
In model (2), η is the constant mean, and the term Zi (tij )w i is used to model variation at
individual level. Here Zi (tij ) = [1, log(tij )] and wi is the vector of random effects of the initial
covariate at time 0 and the changing rate for unit i. It is assumed that wi = (w0i , w1i )′ ∼
N(0, Σw ) with the covariance matrix
σ12 ρσ1 σ2
Σw = ,
ρσ1 σ2 σ22
The first component of (3) is the likelihood function of the failure-time data, which is
n
Y
L(θ T ) = {exp[βxi (ti )]f0 (u[ti ; β, xi (ti )]; θ 0 )}δi {1 − F0 (u[ti ; β, xi (ti )]; θ 0 )}1−δi . (4)
i=1
In the above equation, f1 ( · ) is the pdf of a univariate normal and f2 ( · ) is the pdf of a bivariate
normal distribution.
6
where θ denotes all parameters. Then ρi (s; θ) can be further expressed as
ρi (s; θ) = EX i (ti ,ti +s)|X i (ti )=xi (ti ) {Pr[ti < Ti ≤ ti + s|Ti > ti , X i (ti ), X i (ti , ti + s)]} (7)
EX i (ti ,ti +s)|X i (ti )=xi (ti ) {F0 (u[ti + s; β, X i (ti + s)]; θ0 )} − F0 (u[ti ; β, xi (ti )]; θ 0 )
=
1 − F0 (u[ti ; β, xi (ti )]; θ 0 )
where X i (t1 , t2 ) = {Xi (s) : t1 < s ≤ t2 }. Since model (2) is assumed for X i (ti , ti + s) and
X i (ti ) = xi (ti ), the multivariate normal distribution theory can be used to obtain the condi-
tional distribution.
b since an analytical expression for
The Monte Carlo simulation is used to evaluate ρi (s; θ)
b
ρi (s; θ) is unavailable. The following procedure is used to compute ρi (s; θ).
2. Let X ∗i (ti + s) = {xi (ti ), X ∗i (ti , ti + s)} be the simulated covariate process in the time
interval (ti , ti + s).
b = M −1 PM b
5. The estimate is computed by ρi (s; θ) m=1 ρ∗m
i (s; θ).
b ∗ = (θ
2. Let θ b∗′ , θ b∗ ) following the above algorithm.
b∗′ )′ and obtain ρ∗∗ (s; θ
T X i
Figure 2 shows the estimated DRL for two representative units. One unit has a higher
use rate which increases quickly over time (w
b0 = 0.4061, w
b1 = 0.4184) and the other has a
lower use rate which increases slowly over time (w
b0 = 0.1704, w
b1 = 0.0168). The trends in the
7
0.8
DRL for high use rate
0.4
0.2
0.0
0 20 40 60 80 100
Figure 2: The estimated DRLs and the 95% pointwise CIs for two representative at-risk units.
Figure reproduced with permission.
plot are consistent with our expectation that units with higher use rates tend to have higher
failure risk.
To assess the prediction variability, one may
h also want
i to calculate the prediction interval
e
(PI) of individual remaining life, denoted by S i , Si . A 100(1 − α)% PI of the remaining
e
lifetime can be obtained by using the method introduced by [46] as in
8
200
100
50
point prediction
90% pointwise PI
0 20 40 60 80 100
Figure 3: The point predictions and the pointwise 90% PIs for the cumulative number of
failures after the DFD out of the 1731 units at risk. Figure reproduced with permission.
9
length on the FTIR spectrum, the change in the height of the peak was used to measure the
decrease in the concentration of the compound. One of the compounds of interest for the
NIST data was C-O stretching of aryl ether, which was measured at the wavelength 1250 cm.
Figure 4(a) shows the degradation paths of nine representative specimens with varied starting
times in the study. We can observe very different trajectories with the degradation rate varies
over time and among different specimens as well. Figures 4(b)-(d) show the dynamic covariate
information on the daily UV dosage, RH, and temperature as well as the fitted smooth lines
for showing the mean process of one specimen over the study period. The vertical lines are
used to label time windows separated by every six months. We can observe both a seasonal
pattern and a random oscillation of the daily records for each individual covariate. There
are stronger seasonal patterns for the UV dosage and temperature than the RH. There also
appears to be a larger variation of the daily observations in the summer than in the winter,
which indicates a varied degree of variability of the covariates over different time periods.
where yi (tij ) for i = 1, · · · , n, j = 1, · · · , ni is the degradation measurement at time tij for unit
i, εi (tij ) ∼ N(0, σε2 ) denotes the measurement error, and xi (tij ) = [xi1 (tij ), . . . , xip (tij )]′ is the
vector containing the dynamic covariate information at the time tij . The actual degradation
level at tij is modeled by D[tij ; xi (tij )] + G(tij ; wi ) as the sum of a fixed component and
a random component. The fixed component is the population common degradation path,
modeled in a cumulative damage form given by
p Z
X tij
D[tij ; xi (tij )] = β0 + fl [xil (u); βl ]du. (10)
l=1 0
This model incorporates the dynamic covariates through the covariate-effect functions fl (·)
for l = 1, · · · , p. Here, β0 is the initial degradation, fl [xil (u); βl ] is the lth covariate-effect of
Rt
xil (u) on the degradation process at time u, and 0 ij fl [xil (u); βl ]du is the cumulative effect of
xil up to time tij . The random component includes the random effect terms for modeling the
unit-to-unit variation, which is specified in G(tij ; wi ) = w0i + w1i tij . Here, wi = (w0i , w1i )′ is
the vector of random effects for the initial degradation and the growth rate over time, and it
is assumed to follow a bivariate normal distribution N(0, Σw ) with the covariance matrix
σ02 ρσ0 σ1
Σw = .
ρσ0 σ1 σ12
10
0.0 70
data points fitted
60
−0.1
50
UV Dosage
−0.2
40
Damage
30
−0.3
20
−0.4
10
0
−0.5
0 50 100 150 200 5 10 15 20 25 30 35
Days since first measurement Months since 01JAN2002
100
50
40 80
Relative humidity (%)
Temperature (°C)
30
60
20
40
10
0 20
−10
5 10 15 20 25 30 35 5 10 15 20 25 30 35
Months since 01JAN2002 Months since 01JAN2002
Figure 4: Plots of (a) nine representative degradation paths and (b)-(d) dynamic covariate
information on the daily UV dosage, temperature and relative humidity for a single sample.
The black dots connected by green lines show the daily values. The vertical lines show the
time windows by every 6 months from January 2002. The red smooth curves are the estimated
mean process. Figure reproduced with permission.
11
Also we use σ w = (σ0 , σ1 , ρ)′ to denote all distinct parameters included in Σw .
The ML method is used for estimating the parameters. Since the degradation measure-
ments and the dynamic covariates are observed at discrete time points, the discrete version
of the degradation path model is used for computing the likelihood by replacing D[tij ; xi (tij )]
in (10) by
p
X X
D[tij ; xi (tij )] = β0 + fl [xil (uik ); βl ](uik − ui,k−1), (11)
l=1 uik ≤tij
where uik is the kth time point when the degradation and covariates are measured for unit i
and ui0 = 0. Let θ D = {β, σ w , σε } denote all the model parameters. Then the likelihood is
Yn Z Y 1 C[yi (tij ); xi (tij ), wi ]
L(θ D ) = φ gwi (wi ; σw ) dwi (12)
i=1 w i t ≤t
σε σε
ij ini
where C[yi (tij ); xi (tij ), wi ] = yi (tij ) − D[tij ; xi (tij )] − G(tij ; wi ), φ(·) and gwi (·) are the pdfs
of a standard normal distribution and a bivariate N(0, Σw ) distribution, respectively.
Considering there was not sufficient knowledge on what might be a sensible form for the
covariate-effect functions, the paper chose to estimate the fl (·) using a linear combination
of spline bases. To leverage the physical understanding of the relationships between the
degradation process and the covariates, the shape-restricted splines [48] were used to ensure
monotonic decreasing bases (I-splines) for the UV dosage and temperature and concave bases
(C-splines) for the RH. Let Blq [xil (uik )] for q = 1, · · · , al denote the spline bases for the
covariate xl , then the covariate-effect function is modeled as
al
X
fl [xil (uik ); βl ] = Blq [xil (uik )]βlq ,
q=1
P
where βlq ’s are the spline coefficients. Define Ulq (tij ) = uik ≤tij Blq [xil (uik )](uik − ui,k−1).
Then the model in (9) with D[tij ; xi (tij )] given in (11) can be written as a linear mixed effects
model in the form of y i = X i β + Z i wi + εi , where
1 U11 (ti1 ) · · · U1a1 (ti1 ) · · · Up1 (ti1 ) · · · Upap (ti1 ) 1 ti1
.. .. .. .. ..
Xi = . .
..
. .
..
. .
..
. . , Z i = ... ... ,
1 U11 (tini ) · · · U1a1 (tini ) · · · Up1 (tini ) · · · Upap (tini ) 1 tini
and the coefficient vector β = (β ′u , β′c )′ , where β u and β c denote the unconstrained and
constrained parameters, respectively.
bD that maxi-
The following algorithm was proposed [25] to obtain the ML estimate θ
mizes (12):
12
1. Initiallize σ w and σε by fitting a linear mixed-effects model with no constraints.
2. Compute Vi = Z i Σw Z ′i + σǫ2 I i .
3. The mixed primal-dual bases algorithm in [49] is used to estimate β. That is to minimize
Pn ′ −1
i=1 (y i − X i β) Vi (y i − X i β) subject to β c ≥ 0.
With the shape-restricted splines, the ML estimates of some parameters might locate
on the boundary of the parameter space. In this case, the bootstrap method is useful for
assessing the variability and making inference about the parameters. An adjusted bootstrap
procedure by [50] was applied to resample the residuals and the estimated random effects for
constructing bootstrap resamples of the original data to avoid underestimating variability and
producing too narrow CIs. Then the bias-corrected bootstrap CIs were constructed based on
obtaining the ML estimates of model parameters using the above mentioned algorithm for a
large number of bootstrap samples.
where Trj (t) models the long-term trend of the covariate process for the jth covariate, Sj (t)
captures the seasonal pattern, and ξj (t) depicts the random error which is assumed to be a
stationary process. For the NIST outdoor weathering data, there was no significant long-term
trend observed, and hence Trj (t) = µj for j = 1, 2, 3. However, the seasonal pattern was quite
prominent and there were seasonal effects observed for both the mean and variance of the
process. So two sine functions were included in both the seasonal and error terms (except for
RH which shows no seasonal effect assumed for the variation of the process from Figure 4) in
the following form
κ sin 2π (t − η ) 1
+ ν
1 + sin
2π
(t − ς )
ε1 (t)
S1 (t) 1
365 1
ξ1 (t) 1
365
2π 1
S2 (t) = κ2 sin 365 (t − η2 )
2π
, ξ2 (t) =
1 + ν2 1 + sin 365 (t − ς2 ) ε2 (t)
.
2π
S3 (t) κ3 sin 365 (t − η3 ) ξ3 (t) ε3 (t)
(13)
13
To capture the autocorrelation within and among the covariate processes, a lag-2 VAR model
[i.e. Var(2)] was used, where the error term was modeled by
ε1 (t) ε1 (t − 1) ε1 (t − 2) e1 (t)
ε2 (t) = Q1 ε2 (t − 1) + Q2 ε2 (t − 2) + e2 (t) (14)
ε3 (t) ε3 (t − 1) ε3 (t − 2) e3 (t)
In the above equation, Q1 and Q2 are regression coefficients matrices, and [e1 (t), e2 (t), e3 (t)]′ ∼
N(0, Σe ) are multivariate normal random errors that do not change over time.
The parameters in models (13) and (14) are estimated in two steps. First, the ML estimates
of the seasonal effects in the process mean and variance structures are obtained by ignoring the
autocorrelation in the error terms. Then the VAR model is fitted to the residuals calculated
from the first step using the multivariate least squares approach [51]. The bootstrap method
is used for obtaining the CIs of the parameters in the dynamic covariate process.
However, for a random unit, the covariate process X(∞) and w are random. Hence, the cdf
of the failure time, T = T [Df , X(∞), w], can be defined as
with θ = {θ D , θX } denoting all the unknown parameters. There is usually no closed form
b is estimated through Monte Carlo
expression of F (t; θ). Hence, the cdf at any estimated θ
simulation outlined in the following steps [25].
bX .
1. One need to simulate the covariate process based on the estimated parameter θ
2. Then one can simulate the random effects w from N(0, Σw ) with the estimated parameter
bD .
θ
3. Compute D[t; X(∞)] + G(t; w) based on the simulated covariate process and random
effects.
4. For the degradation path in step 3, determine the failure-time tD by Eqn. (15).
14
1.0
point estimates
95% pointwise CIs
0.8
0.6
Probability
0.4
0.2
0.0
0 50 100 150
Figure 5: The estimated cdf and corresponding 95% pointwise CIs for a population of units
with random starting time between 161 and 190 days. Figure reproduced with permission.
By using the bootstrap approach, the point estimates and the CIs of F (t; θ) can be calculated
b based
using the sample mean and quantiles of the bootstrap version of the estimates of F (t; θ)
b By using Df = −0.4, M = 200 Monte Carlo
on a large number of bootstrap estimates θ.
simulations, and 10000 bootstrap samples, Figure 5 shows the predicted F (t; θ) and its 95%
pointwise CIs for the NIST coating degradation data. We can see that for a population of
units with random starting time between 161 and 190 days, a majority of the population will
fail between 50 to 150 days in service.
15
ID No. current age
182 3.0
171
2.5
150
134
2.0
119
Response
107
1.5
41
40
1.0
31
16 0.5
Component event Subsystem event
0.0
0 20 40 60 80 100 0 20 40 60 80 100
Months Months
(a) Recurrent event processes for (b) Cumulative usage processes for
a subset of system units a subset of system units
Figure 6: Plots of (a) the event processes and (b) the cumulative usage processes for ten
randomly selected units in the Vehicle B fleet. Figure reproduced with permission.
failures) and/or event at component level (e.g., oil pump failures). In the field data, we have
n = 203 units from a 110-month observation period. There are 219 component events and 44
subsystem events observed during the study period. Figure 6(a) shows the event plot for ten
randomly selected units. We also have the cumulative usage information available for each
unit, which is dynamic covariate. The cumulative usage information is shown in Figure 6(b).
The goal is to make a prediction for the cumulative number of component event occurrences
at a future time.
We need some notation to introduce the MTRP model. Suppose there are n units under
observation from time 0 to τi . Let Xi (t) be the time-dependent covariate at time t for system
i. Let Nis (t) be the number of subsystem events and and Nic (t) be the number of component
events up to time t. The total number of replacement events is Ni (t) = Nis (t) + Nic (t). The
subsystem event time is sorted as 0 < tsi1 < · · · < tsi,Nis (τi ) < τi . The component event time is
sorted as 0 < tci1 < · · · < tci,Nic (τi ) < τi . Let 0 < ti1 < · · · < ti,Ni (τi ) < τi be the replacement
event times, regardless of the types.
16
4.2 The MTRP Model and Parameter Estimation
For a two-level repairable system, Xu et al. [36] propose the following MTRP model to describe
events occurred at component level. In particular, the intensity function is
c
c c s s s s
λi (t|Fi,t− ; θ ) = h Λi (t|Fi,t− ) − Λi ti,Ni (t− ) | Fi,t− ; θ λsi (t|Fi,ts − ; θ c ).
c
(17)
i,Ni (t− )
Here Fi,ts − denotes the historical information. In this multi-level model framework, the effect
of subsystem events on the component event process is modeled by λsi (t|Fi,ts − ; θ c ), which takes
the form
Here, θ c denotes the unknown parameters. The cumulative event intensity functions can be
Rt Rt
obtained as Λi (t) = 0 λi (u; θ c ) du, and Λsi (t|Fi,ts − ) = 0 λsi (u|Fi,u
s c
− ; θ ) du. The baseline
function λi (t; θ c ) models the intensity of the component process when there is no event ad-
justment, and the function hs (·) is used to model the adjustment for effect of events from
the subsystem. The renewal distribution function F c (·) is used to describe the distribution
of gap times under the transformed scale. The model in (18) can be extended to incorporate
dynamic covariates and random effects.
To model the dynamic covariates, the intensity function can be extended as
where λb (t) denotes intensity trend function under the baseline and γ is the regression coef-
ficient. In the Vehicle B application, we use g[Xi (t)] = log[Xi (t)]. To incorporate random
effects, the intensity function can be further extended as
Here wi ’s are independent and identically distributed with N(0, σr2 ). The MTRP with random
effects is referred to as HMTRP(F c , F s , λi ), in which the HMTRP stands for heterogenous
MTRP.
To estimate the model parameters, one need to construct the likelihood function. The
component events data can be denoted as {tij , δijc } with tij be the event time and δijc be the
component-event indicator. The event history is denoted as F = {Nic (u), Nis (u), Xi (u) : 0 <
u ≤ τi , i = 1, · · · , n}. The likelihood function is
(τi )+1 n
n Ni Y oδijc
Y
c c
L(θ ) == s
f c [Λsi (tij |Fi,t s s
− ) − Λi (ti,j−1 |F −
i,t
s
)]λsi (tij |Fi,t−; θ )
ij i,j−1 ij
i=1 j=1
n o1−δij c
c
× S [Λsi (tij |Fi,t
s
−) − Λsi (ti,j−1 |Fi,t
s
− )] . (21)
ij i,j−1
17
Xu et al. [36] use Bayesian methods with diffuse priors to estimate the model parameters.
The Metropolis-within-Gibbs algorithm is used to obtained the posterior distributions and
then the inference can be carried out using the Markov Chain Monte Carlo (MCMC) samples
from the posterior distributions.
Here Nic (t∗ ; θ) denotes the prediction for unit i. Because there is no closed form expression
for (22), the Monte Carlo simulation is used.
By fitting the MTRP model to the Vehicle B data using Bayesian estimation, one needs
to specify the prior distributions for the unknown parameters. The Weibull distribution was
used as renewal functions for F c and F s . To check the performance of prediction, first the last
15 months of the Vehicle B data were held back and only the first 95 months data were used to
estimate the MTRP model and then generate predictions for the last 15 months. Figure 7(a)
shows the prediction of the cumulative component events for the last 15 months based on the
earlier data. One can see that the actual observed cumulative numbers of component events
are closely located around the predicted values and also well bounded within the pointwise
prediction intervals. Figure 7(b) shows the predicted future events given all the observed
data for the next 30 months, which indicates that the total number of component events are
expected to range between 62 and 90 with a 95% confidence level.
18
50
80
40
Cumulative Number of Events
40
20
20
10
Actual number
Predicted number Predicted mean
0 95% PI 0 95% PI
0 5 10 15 0 5 10 15 20 25 30
(a) Back test based on an early subset of the data (b) Prediction of future events
Figure 7: Plots of the predicted cumulative number of component events for Vehicle b for (a)
the last 15 months based on the earlier 95 months data and (b) the future 30 months based
on all observed data. Figure reproduced with permission.
14 observations of E-glass are made including 11 failed and 3 right-censored units. Historical
data of the observations are show in Figure 8. Several other important factors in the test
are set as follow. Let R = σm /σM denote the stress ratio, where σm is the minimum stress
and σM is the maximum stress. The range of R can reveal different test type and it is set at
R = 0.1 for a tension-tension loading test in this application. The ultimate stress σult , where
the material breaks at the first cycle is set to be 1339.67 MPa. The frequency of the cyclic
stress testing (f ) is set at 2 Hz, and the angle (α) between the testing direction and material
is set at 0.
where Φ(·) and φ(·) are the standard cdf and pdf, respectively. The lognormal and Weibull
distributions are the common choices. In the ALT modeling, we assume a constant scale
parameter ν and the location parameter is µ = µβ (x), where x is the stress level and β is the
unknown parameter. The following nonlinear model for composite materials proposed in [52]
19
1200
14
1000
12
800
log(Cycles)
Stress
600
10
400
8
200
Exact observations
Exact observations Censored observations
Censored observations Fitted S−N curve by θ0
0
6
0 500000 1000000 1500000 2000000 0 200 400 600 800 1000 1200
Cycles Stress
where ξp,u is the pth quantile at the use condition u and zp is the pth quantile of the standard
distribution. Our goal is to propose test planning under multiple use conditions to approximate
the real scenarios. The use stress profile consists of a set of use levels, {u1 , · · · , uK }, with
P
weights {w1 , · · · , wK } and K
k=1 wk = 1.
Let (xi , ti , δi ) denote the data for the ith testing unit, where xi is the stress level of the
accelerating factor and ti is the observed cycles to failure (or censoring). Let δi be a censoring
indicator where δi = 1 if the observation is censored and δi = 0 if the observation fails. Then,
the log-likelihood function is given by
n
X
l (θ|xn , tn , δ n ) = (1 − δi ) [log φ (zi ) − log(ti ) − log(ν)] + δi log [1 − Φ (zi )] , (25)
i=1
20
b be the ML estimates of θ and log(ξbp,u ) be the ML
where zi = [log(ti ) − µβ (xi )] /ν. Let θ
estimate of the logarithm of the pth quantile at the use level u, obtained by substituting β
b and νb in (24). Given the use level u, the asymptotic variance of log(ξbp,u) is
and ν by β
h i
Avar log ξbp,u = c′ Σθ (xn ) c,
where c = [∂µβ (u)/∂A, ∂µβ (u)/∂B, zp ]′ , Σθ (xn ) = In−1 (θ), and In (θ) is the Fisher informa-
tion matrix based on n observed data. The details for calculating In (θ) can be found in [43].
A weighted version of asymptotic variance can be expressed as
K
X h i
wk Avar log ξbp,uk . (26)
k=1
where Σθ (q new ) = [In (θ, q n ) + I1 (θ, qnew )]−1 , q new = (q ′n , qnew )′ , q n = (q1 , . . . , qn )′ , and
π (θ|q n , tn , δ n ) is the posterior distribution of θ. Specifically,
21
where f (tn |θ, xn , δ n ) is the joint pdf of the historical data and π(θ) is the prior distribution
of θ. Then, the optimum (n + 1)th design point is determined by
∗
qn+1 = arg min ϕ (qnew ) . (28)
qnew ∈[qL ,qU ]
2. Evaluate the asymptotic variance. Use the technique of MCMC to approximate (27).
The details of the related algorithms can be found in [43].
∗
3. Determine the optimum next testing point qn+1 . Given a candidate set of design points,
their corresponding values of the objective function in (27) can be evaluated in Step 2.
Then, determine the optimum next design point, which has the smallest value of the
asymptotic variance.
∗ ∗
4. Obtain the failure data at the level qn+1 . Under the stress level qn+1 , conduct the exper-
iment and obtain the failure information (tn+1 , δn+1 ).
5. Repeat Steps 2 to 4 till the desired number of testing units are obtained. Add the new
∗
lifetime data, (qn+1 , tn+1 , δn+1 ), to the historical dataset and repeat Steps 2 to 4 till the
desired number of new design points are obtained.
1. Prior information: Let A and B be from the normal distributions, where A ∼ N(0.08,
0.0008) and B ∼ N(1, 0.0833). The prior distribution for ν 2 is Inverse Gamma(4.5, 3).
2. Historical data: In practical implementation, the sample size at the beginning of testing
is limited. Choose the three failed observations at stress levels x3 = (621, 690, 965) from
Figure 8 as the historical dataset.
3. Total size of design points: Let the sample size of the new design points be 12.
22
Simulation 1 Simulation 2
0.9
0.9
historical stress level historical stress level
10
sequential design sequential design
Stress level
Stress level
0.7
0.7
0.5
0.5
8
Simulation 1
0.3
0.3
Simulation 2
1 5 7 9 11 14 1 5 7 9 11 14 Simulation 3
6
Number of units in current data Number of units in current data Simulation 4
Avar
Simulation 3 Simulation 4
4
0.9
0.9
historical stress level historical stress level
sequential design sequential design
Stress level
Stress level
0.7
0.7
2
0.5
0.5
0.3
0.3
0
1 5 7 9 11 14 1 5 7 9 11 14 3 4 5 6 7 8 9 11 13 15
Number of units in current data Number of units in current data Number of units in the current data
4. Design candidate: The standardized levels of historical data are 0.46, 0.52, and 0.72,
and the candidate points are from qL = 0.35 to qU = 0.75 with a 5% increase.
For the illustrative purpose, assume that the true values of parameters are θ 0 . When an
optimum design point is determined, the new observation is generated from the lognormal
distribution with parameter θ 0 and the censoring time at 2 × 106 cycles. Repeat Steps 2 to
4 in Section 5.3 till 12 testing locations are obtained. Then, the results of 4 simulation trials
are shown in Figure 9. It consistently shows that only two stress levels at 0.35 and 0.75 are
selected, and 8 and 4 units are allocated to the levels 0.35 and 0.75, respectively. And the
resulting asymptotic variances decrease as the size of sequential runs increases.
Using the same historical data, the developed SBD is also compared with the local c-
optimality design. For the locally c-optimal design, the estimated values of parameters from
historical data are usually used as the planning values of the parameters. With only 3 obser-
vations available from the historical data, the ML estimates are θ̂ 1 = (0.0005, 0.7429, 0.1658)′ .
Hence, the local c-optimality design chooses 11 and 1 unit at the testing levels at 0.65 and
0.75, respectively. Now, we compare the performance on the value of the asymptotic variance
based on the ML estimates of the final dataset including the 12 new testing observations and
3 historical observations. With 100 simulations, the averages of asymptotic variances for the
SBD and the local c-optimality designs are 0.6048 and 4.0337, respectively. It shows that the
SBD is more efficient than the traditional local c-optimality design when there is too little
historical data available to provide accurate estimates of the model parameters. The proposed
23
SBD can be also applied when there is no historical data but only prior information based on
subject matter expertise.
6 Concluding Remarks
In this chapter, we review recent developments on statistical reliability analysis utilizing dy-
namic covariates and sequential test planning. For time to event data, we introduce a cumu-
lative damage model to account for the effect of dynamic covariates and illustrate the method
with the Product D2 application. For degradation data, we present the general path model for
incorporating dynamic covariates and illustrate the method with the NIST coating degrada-
tion data. We also introduce the MTRP model for recurrent events using dynamic covariates
and illustrate it with the Vehicle B data. Regarding to test planning for ALT, we focus on
the SBD and illustrate it with the ALT design for polymer composites fatigue testing.
Looking forward, more versatile data become available due to the rapid advance of modern
technology, and new statistical methods need to be developed to make use of those new data
for improving reliability modeling and prediction. As described in [3], many data types such
as spatial data, functional data, image data, and text data, all have great potential to be
used for reliability modeling and analysis. New methods that are available in statistics and
machine learning can also be transformed and integrated with reliability domain knowledge
for reliability analysis, which provides tremendous opportunity in reliability research.
References
[1] W. Q. Meeker, L. A. Escobar: Statistical Methods for Reliability Data (John Wiley &
Sons, Inc., New York 1998)
[2] W. Q. Meeker, Y. Hong: Reliability Meets Big Data: Opportunities and Challenges, with
Discussion, Quality Engineering 26, 102–116 (2014)
[3] Y. Hong, M. Zhang, W. Q. Meeker: Big data and reliability applications: The complexity
dimension, Journal of Quality Technology 50(2), 135–149 (2018)
[5] H. Guo, A. Monteforte, A. Mettas, D. Ogden: Warranty prediction for products with
random stresses and usages. In: IEEE Proceedings Annual Reliability and Maintainability
Symposium (IEEE, Fort Worth, TX 2009) pp. 72–77
24
[6] L. Lu, C. M. Anderson-Cook: Using Age and Usage for Prediction of Reliability of an
Arbitrary System from a Finite Population, Quality and Reliability Engineering Interna-
tional 27, 179–190 (2011)
[7] Y. Hong, W. Q. Meeker: Field-failure and Warranty Prediction Based on Auxiliary Use-
rate Information, Technometrics 52, 148–159 (2010)
[8] W. Nelson: Prediction of field reliability of units, each under differing dynamic stresses,
from accelerated test data. In: Handbook of Statistics 20: Advances in Reliability, ed. by
N. Balakrishnan, C. R. Rao (North-Holland, Amsterdam 2001) Chap. IX
[12] K. A. Doksum, A. Hóyland: Models for variable-stress accelerated life testing experiments
based on Wiener processes and the inverse Gaussian distribution, Technometrics 34, 74–
82 (1992)
[13] X. Wang: Wiener processes with random effects for degradation data, Journal of Multi-
variate Analysis 101, 340–351 (2010)
[14] J. F. Lawless, M. Crowder: Covariates and random effects in a gamma process model
with application to degradation and failure, Lifetime Data Analysis 10, 213–227 (2004)
[15] X. Wang, D. Xu: An Inverse Gaussian Process Model for Degradation Data, Technomet-
rics 52, 188–197 (2010)
[16] Z.-S. Ye, N. Chen: The Inverse Gaussian Process as a Degradation Model, Technometrics
56, 302–311 (2014)
25
[19] V. Bagdonavičius, M. S. Nikulin: Estimation in degradation models with explanatory
variables, Lifetime Data Analysis 7, 85–103 (2001)
[20] S. J. Bae, W. Kuo, P. H. Kvam: Degradation models and implied lifetime distributions,
Reliability Engineering and System Safety 92, 601–608 (2007)
[23] Y. Xie, C. B. King, Y. Hong, Q. Yang: Semi-parametric models for accelerated destructive
degradation test data analysis, Technometrics 60, 222–234 (2018)
[25] Y. Hong, Y. Duan, W. Q. Meeker, D. L. Stanley, X. Gu: Statistical Methods for Degrada-
tion Data with Dynamic Covariates Information and an Application to Outdoor Weath-
ering Data, Technometrics 57, 180–193 (2015)
[26] Z. Xu, Y. Hong, R. Jin: Nonlinear General Path Models for Degradation Data with
Dynamic Covariates, Applied Stochastic Models in Business and Industry 32, 153–167
(2016)
[27] R. Zhao, B. Liu: Renewal process with fuzzy interarrival times and rewards, International
Journal of Uncertainty, Fuzziness and Knowledge-Based Systems 11, 573–586 (2003)
[29] M. Kijima: Some results for repairable systems with general repair, Journal of Applied
Probability 26, 89–102 (1989)
[30] H. Wang, H. Pham: A quasi renewal process and its applications in imperfect mainte-
nance, International Journal of Systems Science 27, 1055–1062 (1996)
26
[31] L. Doyen, O. Gaudoin: Classes of imperfect repair models based on reduction of failure
intensity or virtual age, Reliability Engineering and System Safety 84, 45–56 (2004)
[32] B. Lindqvist, G. Elvebakk, K. Heggland: The trend-renewal process for statistical analysis
of repairable systems, Technometrics 45, 31–44 (2003)
[33] Q. Yang, Y. Hong, Y. Chen, J. Shi: Failure Profile Analysis of Complex Repairable
Systems with Multiple Failure Modes, IEEE Transactions on Reliability 61, 180–191
(2012)
[34] D. Pietzner, A. Wienke: The trend-renewal process: a useful model for medical recurrence
data, Statistics in Medicine 32, 142–152 (2013)
[35] Q. Yang, Y. Hong, N. Zhang, J. Li: A copula-based trend-renewal process model for
analysis of repairable systems with multitype failures, IEEE Transactions on Reliability
66(3), 590–602 (2017)
[37] W. Q. Meeker: A comparison of accelerated life test plans for Weibull and lognormal
distributions and type I censoring, Technometrics 26(2), 157–171 (1984)
[38] W. Nelson: Accelerated Testing: Statistical Models, Test Plans, and Data Analyses, (Re-
published in a paperback in Wiley Series in Probability and Statistics, 2004) (John Wiley
& Sons, New York 1990)
[39] Y. Zhang, W. Q. Meeker: Bayesian life test planning for the Weibull distribution with
given shape parameter, Metrika 61(3), 237–249 (2005)
[40] Y. Zhang, W. Q. Meeker: Bayesian methods for planning accelerated life tests, Techno-
metrics 48(1), 49–60 (2006)
[41] Y. Hong, C. B. King, Y. Zhang, W. Q. Meeker: Bayesian life test planning for log-
location-scale family of distributions, Journal of Quality Technology 47, 336–350 (2015)
[42] C. King, Y. Hong, S. P. Dehart, P. A. Defeo, R. Pan: Planning Fatigue Tests for Polymer
Composites, Journal of Quality Technology 48, 227–245 (2016)
[43] I.-C. Lee, Y. Hong, S.-T. Tseng, T. Dasgupta: Sequential Bayesian Design for Accelerated
Life Tests, Technometrics 60(4), 472–483 (2018)
27
[44] L. Lu, I. Lee, Y. Hong: Bayesian Sequential Design Based on Dual Objectives for Accel-
erated Life Tests, arXiv:1812.00055 (2019)
[45] V. Bagdonavičius, M. S. Nikulin: Accelerated Life Models: Modeling and Statistical Anal-
ysis (Chapman & Hall/CRC, Boca Raton, FL 2001)
[47] Y. Hong: On Computing the Distribution Function for the Poisson Bionomial Distribu-
tion, Computational Statistics and Data Analysis 59, 41–51 (2013)
[48] M. C. Meyer: Inference using shape-restricted regression splines, The Annals of Applied
Statistics 2, 1013–1033 (2008)
[50] J. R. Carpenter, H. Goldstein, J. Rasbash: A novel bootstrap procedure for assessing the
relationship between class size and achievement, Applied Statistics 52, 431–443 (2003)
[51] H. Lütkepohl: New Introduction to Multiple Time Series Analysis, Second edn. (Springer-
Verlag, Berlin 2005)
28