11Ch - Fote 1
11Ch - Fote 1
14
14.1 Introduction
This chapter gives an introduction to reliability data analysis, also known as
survival analysis and lifetime analysis. The dataset to be analyzed consists of
lifetimes that are measured from a starting time to an endpoint of interest. The
starting time is usually the time when the item is put into operation for the first
time, but may also be the time when we start observing the item. The endpoint of
interest is usually a failure event, sometimes restricted to a specific failure mode.
For many datasets, the data collection is stopped before all the items fail. This
means that some items are still in a functioning state when the data collection
stops. The recorded times for such items are therefore not times-to-failure, but
the time measured from the starting time until the data collection stopped. These
times are said to be censored times, and a dataset with one or more censored times
is called a censored dataset.
Some datasets include one or more explanatory variables such as pressure, tem-
perature, flow-rate, and vibration. These variables are called covariates and help
explain why there are differences between the times-to-failure of the same type of
items.
Reliability data analysis is a loosely defined term that encompasses a variety of
statistical methods for analyzing positive-valued datasets. The methods presented
in this chapter are also extensively used in biostatistics and medical research under
the heading survival analysis.
A high number of books have been published on this topic, but to recommend
one of these for further study is difficult and will depend on your particular appli-
cation.
To analyze reliability data, we need to use a suitable computer program. Many
programs are available, and it is difficult to claim that one is better than all the
others. In this book, we have chosen the program R because it covers most of the
System Reliability Theory: Models, Statistical Methods, and Applications, Third Edition.
Marvin Rausand, Anne Barros, and Arnljot Høyland.
© 2021 John Wiley & Sons, Inc. Published 2021 by John Wiley & Sons, Inc.
Companion Website: www.wiley.com/go/SystemReliabilityTheory3e
656 14 Reliability Data Analysis
techniques dealt with in this chapter, because it is used by many universities, and
because it is free software that can be run on all major computer platforms.
By searching the Internet for “survival analysis,” “survival models,” and simi-
lar terms, you find an almost endless number of presentations, lecture notes, and
slides. Most of these are made for medical applications, but they are in most cases
still relevant for the content of this chapter.
● All the brakes of the same type used in the railway rolling stock within a
country.
Model. To study an aspect of a population, we define a random variable X that may
give us information about this aspect. To be able to use statistical methods in our
study, we establish a probabilistic model, M, related to the random variable X.
The model may be parametric, nonparametric, or semiparametric. As a starting
point, we assume that the model M is parametric with some parameter 𝜃. The
parameter is fixed but unknown, applies for the population, and is sometimes
called a population parameter.
If X is a discrete variable, the model is formulated by a conditional probability
mass function Pr(X = x ∣ 𝜃), where 𝜃 is fixed, but unknown. If X is a continuous
variable, the model is formulated by a probability density function f (x ∣ 𝜃).
14.2 Some Basic Concepts 657
for a continuous variable. The expression for a discrete variable is left to the
reader.
Inference. Inference is a procedure to use information gathered from a sample to
make statements about the population from which the sample was taken. The
main concepts involved in statistical inference are shown in Figure 14.1.
14.2.1 Datasets
The starting point in this chapter is a dataset containing a random sample from
a population of independent items. Throughout this chapter, we assume that the
time-to-failure of an item is a nonnegative random variable T. In most applica-
tions, we assume that we observe n identical items with random times-to-failure
Population Sample
Sampling
Tests/measurements
Inference
Conclusions on the population Results from sample
True time-to-failure
Item no.
1 C
2 F
3 C
4 F
. .
. .
. .
n F
0 Time
(a)
Item no.
1 C
2 F
3 C
4. . F
. .
. .
n F
0 Time
(b)
Figure 14.3 An observed dataset (a), and the same dataset shifted to time 0 (b). F
denotes a failure and C denotes censoring.
We denote the vector of survival times survtime and the status vector status
and enter the following in the R script
For a dataset with many survival times, it may be wise to enter the data into a
spreadsheet program and save the file either as a CSV or as an Excel® file.1
> print(my.surv)
[1] 17.88 28.92+ 33.00 41.52 42.12 45.60+
Observe that + is added to the censored survival times. The + indicates that the
time-to-failure would have been somewhat longer if the survival time were not
censored.
Censoring of Type I
A life test of n numbered and identical items is carried to gain information about
the probability distribution of the time-to-failure T of the items. A specific time
interval [0, 𝜏] has been allocated for the test. After the test, only the times-to-failure
of those items that failed before 𝜏 are known.
1 Commands to import data files into R may be found by searching the Internet for “import data
into R.”
14.2 Some Basic Concepts 661
This type of censoring is called censoring of type I, and the information in the
dataset consists of s (≤ n) observed, ordered survival times
In addition, we know that (n − s) items have survived the time 𝜏, and this infor-
mation should also be used.
Because the number of items that fail before time 𝜏 obviously is random, there
is a chance that none or relatively few of the items will fail before 𝜏. This may be
a weakness of the test design.
Censoring of Type II
Consider the same life test as for censoring of type I, but assume that it has been
decided to continue the test until exactly r (< n) failures have occurred. The test
is therefore stopped when the rth failure occurs. This censoring is called censoring
of type II, and the dataset obtained from the test consists of
together with the fact that (n − r) items have survived the time t(r) .
In this case, the number r of recorded failures is not random. The price for
obtaining this is that the time t(r) to complete the test, is random. A weakness of
this design is therefore that we cannot know beforehand how long time the test
will last.
Censoring of Type IV
Consider a life test of n numbered identical items. Each item may either run to fail-
ure or be censored at a random time C. The time-to-failure T is, as before, assumed
to have distribution function FT (t) and probability density function fT (t), whereas
the censoring time C has distribution function FC (c) and probability density func-
tion fC (c). The two random variables T and C are assumed to be independent. The
survival time we observe is therefore the minimum of T and C.
This censoring is called censoring of type IV and is sometimes also called random
censoring. Many of the datasets that are relevant for reliability studies have random
censoring, especially when the datasets originate from systems in operation.
662 14 Reliability Data Analysis
Right Censoring
Right censoring means that the item is removed from the study before a failure
occurs, or that the study of the item ends before the item fails. For all the examples
in this chapter, the censoring is right censoring.
Informative Censoring
All the examples discussed in this chapter assume that the censoring is nonin-
formative. This means that the time-to-failure T is independent of the censoring
mechanism. The censoring may also be informative, for example when an item
is taken out of service because its level of performance is less than adequate, but
without failing.
Item no.
1 C
2 C
3 F
4 F
5 F
6 F
7
8 C
9 C
10 F
14.2.5 At-Risk-Set
The at-risk-set at time t is the set of items that have not failed or been censored
before time t, that is, the set of items that are at risk of failing at time t. When a
(single) failure or censoring occurs, one item is removed from the at-risk-set and
when a new item enters the study, the at-risk-set is increased by one item. The
number of items in the at-risk-set at time t is an important variable in several of
the survival analyses methods presented in this chapter.
An exploratory data analysis (EDA) is an essential first step in any data analysis.
The EDA gives a “first look at the data” before any modeling effort is done. An
EDA has two main parts: (i) calculation of a selection of sample statistics such as
the mean, median, and standard deviation, and (ii) data visualization in the form
of histograms, empirical distribution functions, Q–Q plots, and so on.
EDA helps the analyst to understand the underlying structure of the data, to
identify anomalies and outliers in the dataset, to assess the assumptions about the
data, and several more. The examination of the data helps seeing what the data
can tell us. EDA got increased importance following the publication of John W.
Tukey’s seminal book Exploratory Data Analysis (Tukey 1977).
664 14 Reliability Data Analysis
Instead of a textfile, you may alternatively create an Excel file or a CSV file (but
this requires another command to activate the data). ◻
Ties
Two or more continuously distributed survival times may sometimes be recorded
with the same value. This is called a tie and may be caused by common-cause
failures or by rounding-off. The dataset in Table 14.1 has a tie for 68.64, because
two survival times are recorded with the same value. The number of failures that
occur at time t(i) is called the multiplicity of the tie and is denoted by di . The dataset
may therefore be recorded in two different ways:
(1) The ordered dataset may be recorded as t(1) ≤ t(2) ≤ · · · ≤ t(n) with a survival
time for each item, thus realizing that some of the survival times may be equal.
(2) The ordered dataset may be recorded as nt ≤ n distinct survival times t(1) <
t(2) < · · · < t(nt ) associated with a vector giving the multiplicities of failures
d1 , d2 , … , dnt .
Most of the following sections use option 1.
Mean
The mean of a dataset is a measure of the central location of the data values and
is calculated as the sum of its data values divided by the number n of data values.
1∑
n
t= t. (14.2)
n i=1 i
The R command to obtain the mean of the dataset is mean(survtime) and for
the data in Table 14.1, we obtain t = 68.08.
Median
The median tm of a dataset is the value at the middle of the ordered data. For
odd number of values (i.e. n = 2k + 1), the median is the (k + 1)th smallest in the
ordered dataset, that is t(k) . For an even number of values (i.e. n = 2k), the median
is the average of the two values in the middle of the ordered dataset, that is the
average of t(k) and t(k+1) . If, for example the sorted dataset has the six values 2, 4, 5,
7, 8, 10, the median is (5 + 7)∕2 = 6. A more formal definition is given in (14.3).
⎧t for n = 2k + 1
⎪ (k+1)
tm = ⎨ t(k) + t(k+1) . (14.3)
⎪ for n = 2k
⎩ 2
666 14 Reliability Data Analysis
The ordered dataset in Table 14.1 has n = 22 values and the median is therefore
the average of t(11) and t(12) ,
t(11) + t(12)
Median = = 61.68.
2
The same result is obtained by the R function median(survtime). Observe that
the mean value is larger than the median for this dataset.
A simple summary, including the mean and the median, of the dataset surv-
time is obtained by the command summary(survtime). If you have created
the textfile dataset.txt as recommended in Remark 14.1, you may use the
script
survtime <-read.table("dataset.txt",header=F,dec=".")
summary(survtime)
and obtain
Min. : 17.88
1st Qu.: 46.30
Median : 61.68
Mean : 68.08
3rd Qu.: 90.87
Max. :138.04
Quartiles (Qu.) are introduced below.
The R commands to obtain the variance and the standard deviation of the dataset
are var(survtime) and sd(survtime), respectively. Observe that the stan-
dard deviation is measured with the same unit as the data values, whereas the
variance is measured with “squared units.” For the dataset in Table 14.1, the (sam-
ple) standard deviation obtained by sd(survtime) is 32.01.
14.3 Exploratory Data Analysis 667
Quantiles
For p ∈ (0, 1), the quantile of order p of the distribution FT (t) is the value tp such
that
FT (tp ) = p, which means that Pr(T ≤ tp ) = p.
For realistic life distributions, tp is unique.
Now, consider an ordered dataset t(1) ≤ t(2) ≤ · · · ≤ t(n) . The (sample) quantile of
order p may approximately be calculated as t([np]+1) , where [np] is the largest inte-
ger < np. The dataset in Table 14.1 has n = 22 values. To determine the (sample)
quantile of, say order p = 0.15, we first calculate np = 22 ⋅ 0.15 = 3.3. The largest
integer less than np is 3 and the quantile of order 0.15 is therefore t(4) = 41.52.
The (sample) quantile of order p is available in R by the function quan-
tile(survtime,p), which gives 41.61. This is not exactly the same result we
got by hand calculation because R applies a more elaborate and “correct” formula
based on interpolation of the ordered survival times. Interested readers may check
the help file in R, help(quantile).
Quartiles
The quantiles of order 0.25 and 0.75 are called the lower and upper quartiles,
respectively. The lower quartile (or 1st quartile), t0.25 is the value that cuts off the
first 25% of the ordered dataset, and the upper quartile (or 3rd quartile) t0.75 is the
value that cuts off the first 75% of the ordered dataset. Both are provided by the
command summary(survtime).
Interquartile Range
The distance between the upper and the lower quartile, t0.75 − t0.25 , called the
interquartile range, is a common measure for the dispersion of the dataset around
its mean or median. The interquartile range for the dataset in Table 14.1 is deter-
mined by quantile(survtime,0.75)-quantile(survtime,0.25) and
the result is 44.57.
library(moments)
survtime <-read.table("dataset.txt",header=F,dec=".")
k <-3 # Choose the order of the noncentral moment
moment(survtime,order = k,central=F)
moment(survtime,order=k,central=T)
2 5612.752 978.3606
3 534022.7 18720.53
Skewness
Skewness is a measure of the asymmetry of the dataset. The skewness value can be
positive or negative. When the distribution of the values of the dataset is symmet-
ric, the skewness is zero. When the values are predominantly large (but with some
small values), the skewness is negative and when the values are predominantly
small (with some large values), the skewness is positive.
The skewness 𝛾1 is defined by
m3,c
𝛾1 = 3∕2
, (14.8)
m2,c
where mk,c is the kth central sample moment of the dataset. The skewness 𝛾1 is
available in the R package moments by the command skewness(survtime).
The result is 0.6117442, which indicates that the dataset is slightly skewed to
the left.
Kurtosis
The kurtosis describes the shape of the tails of the distribution of the values in the
dataset. The normal distribution has zero kurtosis. Negative kurtosis indicate a
thin tail of the distribution and positive kurtosis indicate a thicker tail.
The kurtosis 𝛾2 is defined as
m4,c
𝛾2 = − 3, (14.9)
m23,c
14.3 Exploratory Data Analysis 669
where 𝜇k is the kth central moment of the dataset. The kurtosis is available in the
R package moments by the command kurtosis(survtime) and the result for
the dataset in Table 14.1 is 2.555 003.
14.3.3 Histogram
A histogram consists of parallel bars that graphically show the frequency distribu-
tion of a variable. As a default, all the bars have the same width. We may choose
the number of bars to display. We may also choose whether to (i) show the number
of values in the dataset that fall into the interval corresponding to the width of the
bar, or (ii) to show the relative number (or percentage) of the values that fall into
the interval. With option (ii), the histogram is said to show the relative frequency
distribution or the distribution density of the values in the dataset.
The information obtained from the histogram depends on the resolution, that
is how many intervals we choose. Figure 14.5 shows three different histograms of
the data in Table 14.1, with different numbers of columns.
0.004 0.008
0.010
Density
Density
0.000
0.000
20 60 100 140
Time to failure
(c)
Figure 14.5 Histogram of the dataset in Table 14.1 with different numbers of columns:
(a) 3 columns, (b) 7 columns, and (c) 26 columns.
670 14 Reliability Data Analysis
It is not always true that a higher resolution makes it easier to understand the
distribution of the data.
Histograms are established by the R script
The resulting plot is shown in Figure 14.6. The plot is made by an averaging
technique and is based on a set of input parameters. The current plot is made with
the default parameters of the density command. Other parameters and other
0.015
Sample density
0.010
0.005
0.000
0 20 40 60 80 100 120 140
Time t
averaging techniques may be chosen. Interested readers may consult the help file,
by the command help(density) in the R terminal (console).
̂(t) = N(t) .
R (14.11)
n
The random variable N(t) has a binomial distribution with probability mass func-
tion
( )
n
Pr(N(t) = m) = R(t)m [1 − R(t)]n−m for m = 0, 1, … , n,
m
with mean and variance
E[N(t)] = nR(t).
var[N(t)] = nR(t)[1 − R(t)].
The mean of the estimator is E[R ̂(t)] = nR(t)∕n = R(t) and the estimator is there-
fore unbiased. The variance of the estimator is
̂(t)] = var[N(t)] = R(t)[1 − R(t)] −−−−→ 0.
var[R
n2 n n→∞
The plot of Rn (t), as shown in Figure 14.7, for the dataset in Table 14.1 is also
called a survival curve and can be made with R using several different packages.
1.0
Survival probability
0.8
0.6
0.4
0.2
0.0
0 20 40 60 80 100 120 140
Time t
Figure 14.7 Empirical survivor function (survival curve) for the dataset in Table 14.1.
The authors prefer the package survival and the survival curve is obtained by
the script:
library(survival)
survtime <-read.table("dataset.txt",header=F,dec=".")
# Prepare the data and calculate required values
data<- Surv(survtime)
survfunct<- survfit(Surv(survtime)∼1,conf.type="none")
plot(survfunct, xlab="Time t", ylab="Survival
probability")
1.0
Survival probability
0.8
0.6
0.4
0.2
0.0
0 20 40 60 80 100 120 140
Time t
Figure 14.8 Empirical survivor function (survival curve) for the dataset in Table 14.1
with 95% confidence intervals.
14.3 Exploratory Data Analysis 673
survtime<-read.table("dataset.txt",header=F,dec=".")
x<-survtime$V1
qqnorm(x)
qqline(x)
The resulting plot for the dataset in Table 14.1 is shown in Figure 14.9. The Q–Q
plot shows a fairly good fit to the normal distribution for this particular dataset.
If we consider the fit to the normal distribution to be acceptable, the parameters
of the normal distribution can be estimated from the slope and intercept of the
straight line. This is not pursued any further here.5
Q–Q plots for general distributions may be obtained in R by using the function
qqplot. To use this function, we need to compare our dataset with a simulated
dataset from the distribution we want to compare our data with. Assume that we
want to compare the data in Table 14.1 with the exponential distribution with rate
𝜆 = 1. A random sample of size, say, 300 from the exponential distribution is gen-
erated in R by the function rexp(300,rate=1). The Q–Q plot comparing the
data in Table 14.1 with the exponential distribution is obtained by the script
survtime <-read.table("dataset.txt",header=F,dec=".")
y<- survtime$V1
qqplot(rexp(300,rate=1),y)
100
80
60
40
20
−2 −1 0 1 2
Theoretical quantiles
100
80
y
60
40
20
0 1 2 3 4 5 6
Simulated exponential distribution
Probability distributions usually have one or more quantities that we call parame-
ters. Examples of parameters include 𝜆 in the exponential distribution exp(𝜆) and
the mean 𝜇 and the standard deviation 𝜎 in the normal distribution (𝜇, 𝜎 2 ).
14.4 Parameter Estimation 675
Unbiased
An estimator 𝜃̂ is said to be an unbiased (point) estimator for 𝜃 if its expected value
̂ = 𝜃. An unbiased estimator will not sys-
is equal to the parameter, that is, if E(𝜃)
tematically overestimate or underestimate the “true” parameter.
An estimator that is not unbiased is said to be biased. The bias is calculated as
̂ = E(𝜃)
bn (𝜃) ̂ − 𝜃.
An estimator 𝜃̂ is said to be asymptotically unbiased if limn→∞ bn (𝜃)̂ = 0.
676 14 Reliability Data Analysis
Small Variance
The estimator 𝜃̂ should preferably have a small spread or variability, that is, a small
variance and standard deviation.
Consistency
An estimator 𝜃̂ is said to be a consistent (point) estimator for 𝜃 if 𝜃̂ → 𝜃 when the
sample size n increases. More formally, we say that the estimator 𝜃̂ is consistent if
we for all 𝜀 > 0 have that
̂ − 𝜽| > 𝜀) → 0
Pr(|𝜽 when n → ∞. (14.13)
This means that the distribution of 𝜃̂ becomes more and more concentrated around
the “true” value of 𝜃 as the sample size increases.
Chebyshev’s Inequality
Chebyshev6 showed that for all 𝜀 > 0
̂ − 𝜽)2
E(𝜽 MSE(𝜽)̂
̂ − 𝜽| ≥ 𝜀) ≤
Pr(|𝜽 = . (14.14)
𝜀2 𝜀2
̂ tends to 0 when n → ∞, then 𝜽
If we can prove that the MSE of 𝜽 ̂ is consistent.
Estimator properties are illustrated in the following example.
It may be natural to estimate p as the relative frequency of the outcomes that result
in A, and a natural estimator is therefore
X
̂
p= . (14.15)
n
This estimator is seen to be unbiased, because
E(X)
E(̂
p) = = p.
n
The estimator ̂
p is consistent because it is unbiased and
var(X) np(1 − p)
var(̂
p) = = → 0 when n → ∞.
n2 n2
If we, for example, carry out n = 50 independent Bernoulli trials and get x = 3 out-
comes A, we may put this dataset into the estimator and obtain the point estimate
̂
p = 3∕50 = 0.06. Again, observe that the estimator ̂
p is a random variable, whereas
the estimate is a numerical value. ◻
(2) Find the k first noncentral sample moments m1,nc , m2,nc , … , mk,nc .
(3) From the system of equations 𝜇i,nc = mi,nc , for i = 1, 2, … , k, solve for the
̂ = (𝜃̂1 , 𝜃̂2 , … , 𝜃̂k ).
parameters 𝜽 = (𝜃1 , 𝜃2 , … , 𝜃k ). The solution is the MME 𝜽
Recall that we – from the law of large numbers – know that the first sample
moment converges to the first population moment (i.e. the population mean).
1∑
n
m1,nc → 𝜇1,nc that is T → E(T) for n → ∞, (14.16)
n i=1 i
but we do not know much about the higher moments (i.e. for k ≥ 2).
We illustrate the MME procedure by two examples.
1∑
n
1
= T.
𝜆 n i=1 i
Solving for 𝜆, we obtain the MME
n
𝜆̂ = ∑n .
i=1 Ti
Assume that we have a complete dataset with n = 8 items that have run to failure
∑8
and with a total time in operation i=1 ti = 25 800 hours. The MME (estimate) of
the failure rate 𝜆 with this dataset is then
8
𝜆̂ = h−1 ≈ 3.10 × 10−4 h−1 .
25 800 ◻
Setting the first two population moments equal to the first two sample moments
yields
1∑
n
𝛼
= T.
𝜆 n i=1 i
1∑ 2
n
𝛼(𝛼 + 1)
= T .
𝜆2 n i=1 i
We may now solve the two equations to obtain
1 ∑n
i=1 Ti
̂
𝜆= n
( ∑ )2
1 ∑n 2 1 n
n i=1 Ti
− n i=1 Ti
and
( ∑ )2
1 n
Ti
1∑
n
n i=1
̂ = 𝜆̂
𝛼 T = ( ∑ )2 .
n i=1 i 1 ∑n 2 1 n
n i=1 Ti − n i=1 Ti
By using the dataset in Table 14.1, we may use the following R script to find the
estimates of 𝛼 and 𝜆.
survtime <-read.table("dataset.txt",header=F,dec=".")
a<-mean(survtime)
b<-mean(survtimeˆ2)
lambda<- a/(b-aˆ2)
print(lambda)
alpha<- lambda*a
print(alpha)
(5) The minimum number of moments we need equals the number of unknown
parameters.
(6) Sometimes, the MMEs may be meaningless.
Likelihood Function
We start with the likelihood function for a discrete, binomial model. This model
is based on a random variable X with probability mass function
( )
n x
Pr(X = x ∣ p) = p (1 − p)n−x for x = 0, 1, 2, … , n. (14.17)
x
In the classical setup, the parameter p has a deterministic but an unknown value.
In (14.17), the unknown parameter p is made visible in the probability mass func-
tion Pr(X = x ∣ p) to highlight that the probability is also a function of p. The num-
ber n of trials is considered to be a known number and therefore not a parameter.
Assume that the experiment has been carried out and the data has been
recorded. The data may, for example be n = 10 and x = 3. We may now wonder
which value of p that produced this particular result. To shed light on this prob-
lem, we calculate the probability of obtaining X = 3 for different values of p. The
probabilities may be calculated by using the function dbinom(3, size=10,
prob=p) in R.
p 0.1 0.2 0.3 0.4 0.5 0.6 0.7
Pr(X = 3 ∣ p) 0.0574 0.201 0.267 0.215 0.117 0.0425 0.009
The probabilities for p = 0.8 and p = 0.9 are very small and not included in this
table. Observe that with these p-values, the probability Pr(X = x ∣ p) is largest for
p = 0.3, which means that p = 0.3 is the most likely probability to have produced
X = 3.
Consider the probability Pr(X = 3 ∣ p) as function of p.
( )
10 3
L(p ∣ 3) = p (1 − p)7 for 0 ≤ p ≤ 1. (14.18)
3
We use the symbol L(p ∣ 3), because it seems natural to call this function the like-
lihood function of p for the observed data. It tells how likely it is that a particular
value of p has produced the observed result.
14.4 Parameter Estimation 681
0.25
0.20
Likelihood
0.15
0.10
0.05
0.00
0.0 0.2 0.4 0.6 0.8 1.0
p
Figure 14.11 Likelihood function for the binomial distribution (n = 10 and x = 3).
where the maximum is taken over all possible values of the parameter 𝜃. ◻
which means that 𝜃̂ is the value (the argument) 𝜃 that maximizes L(𝜃 ∣ data). The
MLE hence is the answer to the question: What value of the parameter 𝜃 makes
the data most likely to occur?
In many applications, the natural logarithm of the likelihood function, is more
convenient to work with. Because the logarithm log(⋅) is a monotonically increas-
ing function, the logarithm of L(𝜃 ∣ data) attains its maximum value at the same
point as the L(𝜃 ∣ data) and therefore the log-likelihood function can be used instead
of the likelihood function to obtain the MLE.
The log-likelihood function is written as
When plotting the log-likelihood function 𝓁(𝜃 ∣ data) it is most common to plot
the negative log-likelihood function, −𝓁(𝜃 ∣ data), such that the MLE 𝜃̂ is deter-
mined from the minimum value of this function. The negative log-likelihood func-
tion for the binomial distribution (14.18) is shown in Figure 14.12.
We now illustrate the maximum likelihood principle by some simple examples.
−4.4
−log Likelihood
−4.5
−4.6
−4.7
Figure 14.12 The negative log-likelihood function for the binomial distribution.
14.4 Parameter Estimation 683
• Assume that we have found the MLE 𝜃̂ of 𝜃 and that g(𝜃) is a one-to-one function.
̂
The MLE of g(𝜃) is then g(𝜃).
• An MLE is asymptotically unbiased. E(𝜃̂n ) → 𝜃 when the sample size n
increases.
• Under relatively mild conditions, the MLE is consistent.
• Under certain regularity conditions, the ML estimator has an asymptotically
normal distribution.
Interested readers may consult almost any good book on estimation theory to find
proofs and further properties.
MLE with R
In most cases, ML estimation results in explicit formulas and R may therefore
not be needed to compute the MLEs. If a computer support is deemed to be
required, MLE is available by using the R packages: stats4 , bbmle, or maxLik.
If you want to use one of these packages, please read carefully the package
manuals that are found on the Internet (e.g. by searching for “CRAN package
bbmle”).
To illustrate the analysis, a brief R script using the package bbmle to calculate
the MLE for p in the binomial distribution is shown. To calculate the maximum
likelihood estimate, bbmle uses the function mle2, which again is based on the
negative log-likelihood function.
ML estimation in the binomial model was illustrated in Example 14.5. With
R and the dataset size = 40 and mydata = 5, we can use the following
R script.
686 14 Reliability Data Analysis
The MLE is found by setting the derivative of the log-likelihood function equal to
zero.
n ∑
n
d
𝓁(𝜆 ∣ t) = − t = 0.
d𝜆 𝜆 i=1 i
Solving for 𝜆 gives the ML estimate
n
𝜆̂ = ∑n .
i=1 ti
The corresponding ML estimator is
n
𝜆̂ = ∑n . (14.23)
i=1 Ti
1 ∑n
The ML estimate can hence be expressed by the sample average t = n i=1 ti , as
1
𝜆̂ = .
t
When a complete dataset D = {t1 , t2 , … , tn } is available, the ML estimate can be
calculated in R as 1/mean(D), and a special R package is not required.
2.0e−43
1.5e−43
1.0e−43
5.0e−44
0.0e+00
0e+00 1e−04 2e−04 3e−04 4e−04 5e−04 6e−04
λ
Figure 14.13 Likelihood function for the exponential distribution in Example 14.8.
We now study the properties of the ML estimator and first find out whether or
not it is unbiased.
Because Ti ∼ exp(𝜆), 2𝜆Ti is 𝜒 2 distributed with two degrees of freedom for i =
∑n
1, 2, … , n (e.g. see Ross 2014). Because the Ti s are independent, 2𝜆 i=1 Ti is 𝜒 2
distributed with 2n degrees of freedom.
The ML estimator can be written as
n 2n𝜆
𝜆̂ = ∑n = ∑n ,
T
i=1 i 2𝜆 i=1 Ti
and has the same distribution as 2n𝜆∕Z, where Z is 𝜒 2 distributed with 2n degrees
of freedom. Accordingly,
( )
̂ = 2n𝜆 E 1 .
E(𝜆)
Z
Here,
( ) ∞
1 1 1 1 n−1 −z∕2
E = z e dz
Z ∫0 z 2n Γ(n)
∞
1 1
= zn−2 e−z∕2 dz
2(n − 1) ∫0 2n−1 Γ(n − 1)
1
= .
2(n − 1)
Therefore,
̂ = 2n𝜆 1 n
E(𝜆) = 𝜆.
2(n − 1) n − 1
The estimator 𝜆̂ is accordingly not unbiased, but the estimator 𝜆∗ , given by
n−1 ̂ n−1
𝜆∗ = 𝜆 = ∑n
n i=1 Ti
14.4 Parameter Estimation 689
and
( )
z1−𝜀∕2,2n z𝜀∕2,2n
Pr ∑n ≤ 𝜆 ≤ ∑n = 1 − 𝜀.
2 i=1 Ti 2 j=1 Tj
Thus, a 1 − 𝜀 confidence interval for 𝜆 is
( )
z1−𝜀∕2,2n z𝜀∕2,2n
∑n , ∑n . (14.26)
2 i=1 Ti 2 j=1 Tj
Total-Time-on-Test
Let T(1) ≤ T(2) ≤ · · · ≤ T(n) be the order statistics for the variables T1 , T2 , … , Tn ,
and similarly, let t(1) ≤ t(2) ≤ · · · ≤ t(n) be the ordered dataset that is obtained from
the experiment. Assume that all the n items are put into operation at the same
time t = 0.
690 14 Reliability Data Analysis
We introduce the symbol (t) for the accumulated time in operation in the inter-
val (0, t), and call (t) the total-time-on-test (TTT) at time t. At time t(1) , the n items
have accumulated a time in operation (t(1) ) = nt(1) . Just after time t(1) , there are
n − 1 items left in operation. The accumulated time in operation at time t(2) is
therefore (t(2) ) = nt(1) + (n − 1)(t(2) − t(1) ).
Let di = t(i) − t(i−1) be the time interval between the termination of the operation
of the (i − 1)th entry and the termination of the ith entry, such that
t(1) = d1
t(2) = d1 + d2
⋮ ⋮
t(r) = d1 + d2 + · · · + dr .
The TTT at time t(r) has two parts
(1) The time on test of the items that have failed in the interval (0, t(r) ], which is
∑r
i=1 t(i) = rd1 + (r − 1)d2 + · · · + dr .
(2) The time on test of the n − r items that are still in operation at time t(r) , which
∑r
is (n − r)t(r) = (n − r) i=1 di .
The TTT at time t(r) is therefore,
∑
r
(t(r) ) = t(i) + (n − r)t(r)
i=1
∑
r
= rd1 + (r − 1)d2 + · · · + dr + (n − r) di .
i=1
Tidying up this expression yields
∑
r
(t(r) ) = [n − (i − 1)]di . (14.27)
i=1
By introducing the corresponding random variables, we obtain
∑
r
(T(r) ) = [n − (i − 1)]Di . (14.28)
i=1
Censoring of Type II
For censoring of type II, the life test is terminated as soon as r failures have been
observed. The ordered dataset may be written as t(1) < t(2) < · · · < t(r) < t(r+1) <
· · · < t(n) , for r < n. The dataset contains r times-to-failure and n − r censored
times. This means that tr is the longest time-to-failure. The likelihood function
for this situation is (see Remark 14.4)
( [ r ])
∑
L(𝜆 ∣ t(1) , … , t(r) ) ∝ 𝜆 exp −𝜆
r
t(j) + (n − r)t(r)
j=1
Introducing
D∗j = [n − (j − 1)]Dj for j = 1, 2, … , r.
we know that 2𝜆D∗1 , 2𝜆D∗2 , … , 2𝜆D∗r are independent and 𝜒 2 distributed, each with
2 degrees of freedom. Hence, 2𝜆 (T(r) ) is 𝜒 2 distributed with 2r degrees of free-
dom, and we can utilize this to find E(𝜆∗II ).
( ) ( ) ( )
r 1 1
∗
E(𝜆II ) = E = 2𝜆rE = 2𝜆rE ,
(T(r) ) 2𝜆 (T(r) ) Z
where Z is 𝜒 2 distributed with 2r degrees of freedom. This implies that
( )
1 1
E = .
Z 2(r − 1)
692 14 Reliability Data Analysis
Hence,
1 r
E(𝜆∗II ) = 2𝜆r =𝜆 .
2r − 1 (r − 1)
The estimator 𝜆∗II is accordingly not unbiased, but
(r − 1)
𝜆∗II = (14.31)
(T(r) )
is seen to be unbiased. By the method used for a complete dataset, we find that
𝜆2
var(𝜆̂II ) = .
(r − 2)
Confidence intervals, as well as tests for standard hypotheses about 𝜆, may now be
derived from the fact that 2𝜆 (T(r) ) is 𝜒 2 distributed with 2r degrees of freedom.
Censoring of Type I
The fact that the number (S) of items failing before time t0 is random, makes this
situation more difficult to deal with from a probabilistic point of view. We therefore
confine ourselves to suggesting an intuitive estimator for 𝜆.
First, observe that the estimators for 𝜆, derived in the case of complete
datasets and of type II censored data, both could be written as a fraction with
numerator equal to “number of recorded failures −1” and denominator equal to
“total-time-on-test at the termination of the test.” It seems intuitively reasonable
to use the same fraction when we have type I censoring.
In this case, the number of failures is S and the TTT is
∑
S
(t0 ) = T(j) + (n − S)t0 . (14.32)
j=1
Hence,
S−1
𝜆̂I =
(t0 )
seems to be a reasonable estimator for 𝜆.
It can be shown that this estimator is biased for small samples, but asymptoti-
cally, it has the same properties as 𝜆̂II (see Mann et al. 1974, p. 173).
Complete Sample
Let T1 , T2 , … , Tn be a complete sample of lifetimes that are independent and iden-
tical to Weibull distribution with probability density
( ) [ ( )𝛼 ]
𝛼 t 𝛼−1 t
fT (t) = exp − for t > 0, 𝛼 > 0, 𝜃 > 0.
𝜃 𝜃 𝜃
The likelihood function is
( )𝛼−1 [ ( )𝛼 ]
∏
n
𝛼 tj tj
L(𝛼, 𝜃 ∣ t1 , t2 , … , tn ) = exp − , (14.33)
j=1
𝜃 𝜃 𝜃
n ∑ n
1 ∑ 𝛼
n
= − n log 𝜃 + log tj + 𝛼 t (log tj − log 𝜃).
𝛼 j=1
𝜃 j=1 j
library(WeibullR)
failtime<-c(31.7,39.2,57.5,65.8,70.0,101.7,109.2,130.0)
censored<-c(65.0,75.0,75.2,87.5,88.3,94.2,105.8,110.0)
# Prepare the data for analysis
data<-wblr.conf(wblr.fit(wblr(failtime,censored)),lwd=1)
plot(data)
The function wblr is used to prepare the dataset for usage in WeibullR. The
resulting plot is shown in Figure 14.14.
Observe that the plot in Figure 14.14 is obtained by a simplified procedure in
WeibullR using only default settings. The default names of parameters are “beta”
for 𝛼 and “eta” for 𝜃. The estimates obtained are 𝛼 = 2.35 and 𝜃 = 115.2. Con-
fidence bounds are supplied. To choose a more advanced estimation procedure
and to adjust the settings, the reader should read the WeibullR documentation
carefully.
Censoring of Type II
With censoring of type II, the dataset contains r times-to-failure, and n − r cen-
sored times, and the censoring takes place at time t(r) . Analogous with (14.33) the
likelihood function is proportional with
( )𝛼−1 ( ) ( ( )𝛼 )n−r
∏r
𝛼 t(j) t(j) 𝛼 t(r)
L(𝛼, 𝜃 ∣ t) ∝ exp − exp −
j=1
𝜃 𝜃 𝜃 𝜃
( ( )𝛼 )
∏r
t(r)
𝛼−1
= 𝛼 r 𝜃 −𝛼r t(j) exp −(n − r) ,
j=1
𝜃
14.4 Parameter Estimation 695
99
90
90
50
50
Unreliability (%)
ranks = median
n (fail | cens.) = 16 (8 | 8)
5 10 20
20
weibull (rr−xony)
beta = 2.35
eta = 115.2
10
r2 = 0.9715
prr = 88.15 by corr.
5
CI bounds, type = “pivotal−rr”
CI = 90 [%], S = 10000
B−life ssCL = 95 [%]
2
2
B10 = 22.15 | 44.21 | 65.46
B5 = 12.72 | 32.55 | 53.53
B1 = 3.607 | 16.27 | 34.67
1
1
10 50 100 500 1000
Time to failure
where t is the ordered dataset, that is, the r times-to-failure and the n − r censoring
times that are all equal to t(r) . The log-likelihood is
∑
r
𝓁(𝛼, 𝜆 ∣ t) = r log 𝛼 − r𝛼 log 𝜃 + (𝛼 − 1) log t(j)
j=1
( )𝛼 ( )𝛼
∑
r
t(j) t(r)
− − (n − r) .
j=1
𝜃 𝜃
Analogous with the complete data situation, we can determine the MLE estimates
𝛼 ∗ and 𝜆∗ from
( )1∕𝛼∗
r
𝜆 = ∑r 𝛼∗
∗
𝛼∗
(14.35)
j=1 t(j) + (n − r)t(r)
and
∑r 𝛼∗ 𝛼 ∗
For further details on ML estimation in the Weibull distribution, e.g. see Meeker
and Escobar (1998) and McCool (2012).
696 14 Reliability Data Analysis
A nonparametric estimate for the survivor function R(t) = Pr(T > t) was intro-
duced by Kaplan and Meier (1958) and is called the Kaplan–Meier estimate.7 A
valued feature of the Kaplan–Meier estimate is that it provides an intuitive graph-
ical representation. We first introduce the estimate for a complete dataset.
R(t(1) ) = Pr(T > t(1) ) = Pr(T > t(1) ∣ T > t(0) ) = R(t(1) ∣ t(0) ).
The probability of surviving the next interval (when it is known that it has survived
the first interval) is
and so on. This means that the survivor function at time t(i) can be expressed by
using the multiplication rule for conditional probabilities as
∏
i
R(t(i) ) = R(t(j) ∣ t(j−1) ), (14.37)
j=1
7 Named after the authors: Edward Lynn Kaplan (1920–2006) and Paul Meier (1924–2011).
14.5 The Kaplan–Meier Estimate 697
Based on the binomial model, we may then estimate the factors of (14.37) as
n1 − d1 d1 1
̂(t(1) ∣ t(0) ) =
R =1− =1−
n1 n1 n
n − d2 d2 1
̂(t(2)
R ∣ t(1) ) = 2 =1− =1− ,
n2 n2 n−1
and so on.
If we use this result in (14.37), we obtain a reformulated estimate for the empir-
ical survivor function
( )
∏ ∏ dj
̂(t) =
R ̂ (t(j) ∣ t(j−1) ) =
R 1−
j;t(j) <t j;t(j) <t
nj
( )
∏ 1
= 1− . (14.38)
j;t <t
n − j+1
(j)
The only difference from (14.38) is the values for dj and nj . If t(j) is a censoring
time, dj = 0, the factor (1 − dj ∕nj ) = 1 and does not directly influence the estimate
̂(t), but the censoring influences the at-risk-set before the next event (failure or
R
censoring).
We may rewrite the definition of the Kaplan–Meier estimate to include the infor-
mation of whether a survival time is a failure or a censoring time by including the
status 𝛿j in the formula
( )
∏ dj
̂(t) =
R 1− , (14.39)
j;t <t,𝛿 =1
nj
(j) j
where the product includes all items j that have a failure time (i.e. 𝛿j = 1) such that
t(j) < t. This formula clearly shows that the factors are only included for survival
times that represent failure. Survival times that represent censoring give a factor
equal to one and will hence not influence the estimate directly.
698 14 Reliability Data Analysis
With ̂
pj = 1 − dj ∕nj = (nj − dj )∕nj we may write (14.54) as
∏
̂(t) =
R ̂
pj . (14.40)
j;t(j) <t,𝛿j =1
The estimate R ̂ (t) (14.39) and (14.40) is known as the Kaplan–Meier estimate
and is also called the product limit (PL) estimate. The procedure to calculate the
Kaplan–Meier estimate is illustrated in Example 14.9.
library(survival)
survtime <- c(31.7,39.2,57.5,65.0,65.8,70,0,75.0,75.2,
87.5,88.3,94.2,101.7,105.8,109.2,110.0,130.0)
status <- c(1,1,1,0,1,1,0,0,0,0,0,0,1,0,1,0,1)
data<- Surv(survtime,status==1)
km <- survfit(Surv(survtime, status==1)∼1,conf
.type="none")
plot(km,xlab="Time t",ylab="Survival probability")
Table 14.2 Computation of the Kaplan–Meier Estimate (censored times are marked with
0 in column “Status”).
0 — — — 1 1.000
4 13 65.0 0 1 0.813
7 10 75.0 0 1 0.677
8 9 75.2 0 1 0.677
9 8 87.5 0 1 0.677
10 7 88.3 0 1 0.677
11 6 94.2 0 1 0.677
13 4 105.8 0 1 0.542
15 2 110.0 0 1 0.361
16 1 130.0 1 0 0.000
700 14 Reliability Data Analysis
t ̂
R(t)
We see from (14.39) that R ̂ (t) is a step function, continuous from the right, that
̂
equals 1 at t = 0. R(t) drops by a factor of (nj − 1)∕nj at each failure time t(j) . The
estimate R̂ (t) does not change at the censored times. The censored times influence
̂ (t).
the values of nj (i.e. the at-risk-set) and hence, the size of the steps in R
̂
A slightly problematic point is that R(t) never reduces to zero when the longest
survival time t(n) recorded is a censored time. For this reason, R ̂(t) is usually taken
to be undefined for t > t(n) . This issue is further discussed by Kalbfleisch and Pren-
tice (1980).
1.0
Survival probability
0.8
0.6
0.4
0.2
0.0
0 20 40 60 80 100 120
Time t
Figure 14.15 Kaplan–Meier plot for the data in Example 14.9. Made with R.
1.0
Survival probability
0.8
0.6
0.4
0.2
0.0
0 20 40 60 80 100 120
Time t
Figure 14.16 Kaplan–Meier plot of the dataset in Example 14.9 with 90% confidence
limits, made with R.
Let R(t) be the survivor function for a certain type of items, and assume that the
distribution is continuous with probability density f (t) = R′ (t), where f (t) > 0 for
702 14 Reliability Data Analysis
t > 0. No further assumptions are made about the distribution (i.e. a nonparamet-
ric model).
The failure rate function was defined in Section 5.3.2 as
f (t) d
z(t) = = − log R(t).
R(t) dt
The cumulative failure rate function is
t
Z(t) = z(u) du = − log R(t), (14.42)
∫0
and the survivor function may therefore be written as
R(t) = e−Z(t) .
Plotting Z(t) as a function of t gives a cumulative failure rate plot. If the plot is
convex when plotted on a linear scale, the failure rate function is increasing, and
if the plot is concave, the failure rate function is decreasing.
Plotted as a function of t on a linear scale, the plot of Z(t) is a straight line with
slope 𝜆. If we are able to determine an estimate Z ̂ (t), the plotted values should
follow a reasonably straight line. ◻
If Z(t) is plotted versus t on a log–log scale, the plot is a straight line with slope
𝛼. If we are able to determine an estimate Z ̂ (t), the plotted values should follow a
reasonably straight line on a log–log scale. ◻
The rest of this section is concerned with a particular type of cumulative failure
rate plots: the Nelson–Aalen plot.
14.6 Cumulative Failure Rate Plots 703
be the recorded ordered survival times until either failure or censoring, and let 𝛿j
be the status of survival time t(j) , for j = 1, 2, … , n.
The Nelson–Aalen estimate of the cumulative failure rate is then
∑ dj
̂ (t) =
Z , (14.44)
n
j;t <t,𝛿 =1 j
(j) j
where dj , as before, is the number of items that fail at time t(j) and nj is the number
of items at risk just before t(j) . The Nelson–Aalen estimator of the survivor function
at time t is
̂ (t)].
R∗ (t) = exp[−Z (14.45)
Before we give a justification for these estimators, we illustrate how they are
calculated in Example 14.12.
Table 14.4 Nelson–Aalen estimate for the censored dataset in Example 14.12,
compared with the Kaplan–Meier estimate.
Kaplan–
Survival Status Nelson–Aalen Nelson–Aalen Meier
j time 𝜹j ̂ )
estimate Z(t R∗ (t(j) ) ̂ )
R(t
j (j)
8 75.2 0
9 87.5 0
10 88.3 0
11 94.2 0
1 1 1
12 101.7 1 +···+ + = 0.5748 0.563 0.542
16 11 5
13 105.8 0
1 1 1
14 109.2 1 +···+ + = 0.9082 0.403 0.361
16 5 3
15 110.0 0
1 1 1
16 130.0 1 +···+ + = 1.9082 0.148 0.000
16 3 1
library(survival)
# Data to be analyzed
survtime<-c(31.7,39.2,57.5,65.0,65.8,70.0,75.0 75.2,
87.5,88.3,94.2,101.7,105.8,109.2,110.0,130.0)
status<-c(1,1,1,0,1,1,0,0,0,0,0,1,0,1,0,1)
# Prepare hazard data
revrank<-order(survtime,decreasing=T)
haz<- status/revrank
cumhaz<- cumsum(haz)
# Select only failures for plotting.
df<- data.frame(survtime status,cumhaz)
z<- subset(df,status==1)
# Generate cumulative failure rate plot for exp. distr.
plot(z$survtime, z$cumhaz,type="o",pch=19,xlab="Time",
ylab="Cumulative failure rate")
The plot obtained from this script is made with a linear scale on both axes.
This means that if the data come from an exponential distribution, the plot
should be approximately a straight line (see Example 14.10). The plot is shown
in Figure 14.17. The plot is rather far from linear, and we may conclude that the
underlying distribution is probably not exponential. To inspect the data used, you
may use the commands print(df) and print(z).
2.0
Cumulative failure rate
1.5
1.0
0.5
0.0
0 20 40 60 80 100 120 140
Time
2.0
1.5
1.0
Cumulative failure rate
0.5
30 50 70 90 110 130
Time
We may also make the Nelson–Aalen plot with log 10 scale on both axes. As
shown in Example 14.11, an approximately straight line would indicate that the
underlying distribution may be a Weibull distribution. The plot is obtained by
adding the option log="xy" to the plot( ) command in the R script above.
The resulting Nelson–Aalen plot is shown in Figure 14.18. The plot is not too far
from a straight line, so the Weibull distribution might be an adequate model.
∏
i
R(t(i) ) = R(t(j) ∣ t(j−1) ),
j=1
and assume that the failure rate function in the interval (t(j−1) , t(j) ) may be approx-
imated by a constant failure rate 𝜆j , for j = 1, 2, ….
For a time t, such that t(m) < t < t(m+1) , we get
R(t) = Pr(T > t(1) ∣ T > t(0) ) · · · Pr(T > t ∣ T > t(m) ). (14.46)
14.6 Cumulative Failure Rate Plots 707
As for the Kaplan–Meier estimate, the idea is to estimate each single factor on the
right-hand side of (14.46) and use the product of these estimates as an estimate of
R(t). What is now a reasonable estimate of pj = Pr(T > t(j+1) ∣ T > t(j) )? With the
same approach, we used to justify the Kaplan–Meier estimate, the only survival
times for which it is natural to estimate pj with something other than 1, are the
survival times t(j) where a failure occurs. Because we only consider the times where
something happens (failure or censoring), nj items are at risk in the whole interval
(t(j−1) , t(j) ).
The total functioning time in the (t(j−1) , t(j) ) is nj (t(j) − t(j−1) ). Because we assume
a constant failure rate 𝜆j in (t(j−1) , t(j) ) a natural estimate of 𝜆j is
No. of failures dj
𝜆̂j = = . (14.47)
Total functioning time nj (t(j+1) − t(j) )
A natural estimate of pj , when dj failures occur at t(j) is therefore
( )
dj
̂ ̂
pj = exp[−𝜆j (t(j) − t(j−1) )] = exp − . (14.48)
nj
Inserting these estimates in (14.46) gives
( )
∏ dj ⎡ ∑ dj ⎤
̂
R(t) = exp − = exp ⎢− ⎥. (14.49)
nj ⎢ t <t,𝛿 =1 nj ⎥
t(j) <t,𝛿j =1 ⎣ (j) j ⎦
Because R(t) = exp[−Z(t)], a natural estimate for the cumulative failure rate func-
tion is
∑ dj
̂ (t) =
Z , (14.50)
n
t <t,𝛿 =1 j
(j) j
It may be shown that both the Nelson–Aalen estimator and the variance estima-
tor are close to unbiased. For large samples, it may further be shown that the
Nelson–Aalen estimator at time t is approximately normally distributed. We may
̂ (t) as
therefore find a (1 − 𝜖) confidence interval for Z
̂ (t) ± u1−𝜖∕2 𝜎
Z ̂(t), (14.52)
where u1−𝜖∕2 is the 1 − 𝜖∕2 fractile of the standard normal distribution. More prop-
erties of the estimator may be found in Aalen et al. (2008).
708 14 Reliability Data Analysis
∑
i
(t) = t(j) + (n − i)t for i = 0, 1, … , n and t(i) ≤ t < t(i+1) ,
j=1
(14.53)
and t(0) is defined to be equal to 0 and t(n+1) = +∞.
(t) is the total observed lifetime of the n items at time t. We assume that all the
n items are put into operation at time t = 0 and that the observation is terminated
at time t. In the time interval (0, t], a number, i, of the items have failed. The total
∑i
functioning time of these i items is j=0 t(j) . The remaining n − i items survive the
time interval (0, t]. The total functioning time of these n − i items is thus (n − i)t.
The TTT at the ith failure is
∑
i
(t(i) ) = t(j) + (n − i)t(i) for i = 1, 2, … , n. (14.54)
j=1
In particular,
∑
n
∑
n
(t(n) ) = t(j) = tj .
j=1 j=1
The TTT at the ith failure, (t(i) ), may be scaled by dividing by (t(n) ). The scaled
TTT at time t is defined as (t)∕ (t(n) ).
If we plot the points
( )
i (t(i) )
, for i = 1, 2, … , n, (14.55)
n (t(n) )
we obtain the TTT plot of the dataset.
Example 14.13 Suppose that we have activated 10 identical items and observed
their lifetimes (in hours):
14.7 Total-Time-on-Test Plotting 709
To construct the TTT plot for this (complete) dataset, calculate the necessary
quantities and put them in a table as done in Table 14.5. The TTT plot for this
(complete) dataset is shown in Figure 14.19. ◻
To be able to interpret the shape of the TTT plot, we need the following results,
which we state without proofs.
have the same joint distribution as the (n − 1) ordered variables U(1) , U(2) , … ,
U(n−1) . For a proof, see Barlow and Campo (1975).
(2) If the underlying life distribution F(t) is exponential, then
(a) var[ (Ti )∕ (Tn )] is finite
(b) E[ (Ti )∕ (Tn )] = 1∕n for i = 1, 2, … , n
∑i ∑i i (t(i) )
i t(i) j=1 t(j) j=1 t(j) + (n − i)t(i) = (t(i) )
n (t(n) )
1.0
0.8
0.6
ϕF(v)
0.4
0.2
0.0
0.0 0.2 0.4 0.6 0.8 1.0
i/n (v)
[ i ( ) ]
∑ t(j)
j−1
=n 1− du
j=1
∫t(j−1) n
∑ i
= (n − j + 1)(t(j) − t(j−1) )
j=1
We now come to the heuristic part of the argument. First, let n equal 2m + 1, where
m is an integer. Then t(m+1) is the median of the dataset. What happens to the
integral
t(m+1)
[1 − Fn (u)] du when m → ∞.
∫0
When m → ∞, we can expect that
Fn (u) → F(u),
and that
t(m+1) → {median of F} = F −1 (1∕2),
and therefore that
F −1 (1∕2)
1
(t(m+1) ) → [1 − F(u)] du. (14.58)
n ∫0
Next, let n = 4m + 3. In this case, t(2m+2) is the median of the data, and t(m+1) and
t(3m+3) are the lower and upper quartiles, respectively.
When m → ∞, by arguing as we did above, we can expect the following:
F −1 (1∕4)
1
(t(m+1) ) → [1 − F(u)] du
n ∫0
F −1 (1∕2)
1
(t(2m+2) ) → [1 − F(u)] du (14.59)
n ∫0
F −1 (3∕4)
1
(t(3m+3) ) → [1 − F(u)] du.
n ∫0
In addition, we have that
∞ F −1 (1)
E(T) = 𝜇 = [1 − F(u)] du = [1 − F(u)] du. (14.60)
∫0 ∫0
When n → ∞, we can therefore expect that
1∑
n F −1 (1)
1
t = (t(n) ) → [1 − F(u)] du. (14.61)
n i=1 i n ∫0
712 14 Reliability Data Analysis
1 F(t)
1–v
HF –1(v)
v R(t) = 1 – F(t)
0
0 F –1(v) Time t
The integrals that we obtain as limits by this approach seem to be of interest and
we will look at them more closely. They are all of the type
F −1 (𝑣)
[1 − F(u)] du for 0 ≤ 𝑣 ≤ 1.
∫0
and hence
1
F −1 (𝑣) = − log(1 − 𝑣) for 0 ≤ 𝑣 ≤ 1.
𝜆
Thus, the TTT transform of the exponential distribution is
[− log(1−𝑣)]∕𝜆
1 −𝜆u |− 𝜆1 log(1−𝑣)
HF−1 (𝑣) = e−𝜆u du = − e |
∫0 𝜆 |0
|
1 1 𝜆 log(1−𝑣)∕𝜆
= − e
𝜆 𝜆
1 1 𝑣
= − (1 − 𝑣) = for 0 ≤ 𝑣 ≤ 1.
𝜆 𝜆 𝜆
Further
1
HF−1 (1) = .
𝜆
The scaled TTT transform for the exponential distribution is therefore
𝑣∕𝜆
= 𝑣 for 0 ≤ 𝑣 ≤ 1. (14.65)
1∕𝜆
The scaled TTT transform of the exponential distribution is thus a straight line
from (0, 0) to (1, 1), as shown in Figure 14.21. ◻
1.0
0.8
Scaled TTT transform
0.6
0.4
0.2
0.0
0.0 0.2 0.4 0.6 0.8 1.0
v
Figure 14.21 Scaled TTT transform of the exponential distribution (Example 14.16).
714 14 Reliability Data Analysis
0.8 α=5
α=3
α=2
0.6
ϕF(v)
0.4
α = 0.8
α = 0.5
0.2
α = 0.3
0
0 0.2 0.4 0.6 0.8 1
v
Figure 14.22 Scaled TTT transforms of the Weibull distribution for some selected
values of 𝛼.
Proof :
Because
F −1 (𝑣)
d −1 d
H (𝑣) = [1 − F(u)] du
d𝑣 F d𝑣 ∫0
d −1 1
= (1 − F[F −1 (𝑣)]) F (𝑣) = (1 − 𝑣) ,
d𝑣 f [F −1 (𝑣)]
then
d −1 1 1
H (𝑣)|𝑣=F(t) = [1 − F(t)] = .
d𝑣 F f (t) z(t)
From (14.67) we obtain
(2) If F(t) is a continuous life distribution, strictly increasing for F −1 (0) = 0 < t <
F −1 (1), then
(a) F ∼ IFR ⇐⇒ HF−1 (𝑣) concave; 0 ≤ 𝑣 ≤ 1
(b) F ∼ DFR ⇐⇒ HF−1 (𝑣) convex; 0 ≤ 𝑣 ≤ 1
716 14 Reliability Data Analysis
To estimate the scaled TTT transform of F(t) for different 𝑣 values on the basis
of the observed lifetimes, it is natural to use the estimator
Fn−1 (𝑣)
∫0 [1 − Fn (u)] du i
F −1 (1)
for 𝑣= , i = 1, 2, … , n. (14.68)
∫0 n [1 − Fn (u)] du n
Fn−1 (𝑣)
i
Hn−1 (𝑣) = [1 − Fn (u)] du for 𝑣= , i = 1, 2, … , n,
∫0 n
(14.69)
By comparing (14.70) with (14.64), it seems natural to call Hn−1 (𝑣)∕Hn−1 (1) the
empirical, scaled TTT transform of the distribution F(t).
The following result is useful when we wish to exploit the TTT plot to provide
information about the life distribution F(t):
(3) If F(t) is a continuous life distribution function, strictly increasing for F −1 (0) =
0 < t < F −1 (1), then
( )
Hn−1 ni (t(i) )
= for i = 1, 2, … , n, (14.71)
−1
Hn (1) (t(n) )
Proof :
According to (14.69), for i = 1, 2, … , n,
( ) Fn−1 ( ni )
i
Hn−1 = [1 − Fn (u)] du
n ∫0
T(i)
1
= [1 − Fn (u)] du = (T(i) ),
∫0 n
where as
Fn−1 (1)
Hn−1 (1) = [1 − Fn (u)] du
∫0
1∑
∞ n
1
= [1 − Fn (u)] du = (t(n) ) = t.
∫0 n n i=1 i
By introducing these results in (14.70), we get (14.71).
Therefore, the scaled TTT at time t(i) seems to be a natural estimate of the scaled
TTT transform of F(t) for 𝑣 = i∕n, for i = 1, 2, … , n. One way of obtaining an esti-
mate for the scaled TTT transform for (i − 1)∕n < 𝑣 < i∕n, is by applying linear
interpolation between the estimate for 𝑣 = (i − 1)∕n and 𝑣 = i∕n. In the following
we use this procedure.
Now suppose that we have access to a survival dataset. We first determine
(t(i) )∕ (t(n) ) for i = 1, 2, … , n as we did in Example 14.13, plot the points
[i∕n, (t(i) )∕ (t(n) )] and join pairs of neighboring points with straight lines. The
curve obtained is an estimate for HF−1 (𝑣)∕HF−1 (1) = 𝜇1 HF−1 (𝑣), for 0 ≤ 𝑣 ≤ 1.
We may now assess the shape of the curve (the estimate for HF−1 (𝑣)) in the light
of the result in (14.67) and its proof, and in this way obtain information about the
underlying distribution F(t).
A plot, such as the one shown in Figure 14.23a, shows that HF−1 (𝑣) is concave.
The plot therefore indicates that the corresponding life distribution F(t) is IFR.
Using the same type of argument, the plot in Figure 14.23b shows that HF−1 (𝑣) is
convex, so that the corresponding life distribution F(t) is DFR. Similarly, the plot
in Figure 14.23c indicates that HF−1 (𝑣) “is first convex” and “thereafter concave.”
In other words, the failure rate of the corresponding lifetime distribution has a
bathtub shape.
The TTT plot obtained in Example 14.13, therefore indicates that these data orig-
inate from a life distribution with bathtub shaped failure rate.
(a)
(c)
ϕF(v)
(b)
0 1
v
Figure 14.23 TTT plots indicating (a) increasing failure rate (IFR), (b) decreasing failure
rate (DFR), and (c) bathtub-shaped failure rate.
The TTT plot of the ball bearing data is presented in Figure 14.24. The TTT plot
indicates an IFR. We may try to fit a Weibull distribution to the data. The Weibull
parameters 𝛼 and 𝜆 are estimated to be 𝛼 ̂ = 2.10 and 𝜆̂ = 1.22 × 10−2 . The TTT
transform of the Weibull distribution with these parameters is plotted as an overlay
curve to the TTT plot in Figure 14.24. ◻
1.0
0.8
0.6
T(i/n)
0.4
0.2
0.0
0.0 0.2 0.4 0.6 0.8 1.0
i/n
Figure 14.24 TTT plot of the ball bearing data in Example 11.11 together with an
overlay curve of the TTT transform of the Weibull distribution with shape parameter
𝛼 = 2.10.
library(AdequacyModel)
# Enter the dataset
data <- c(17.88,28.92,33.00,41.52,42.12,45.60,48.40,
51.84,51.96,54.12,55.56,67.80,68.64,68.64,
68.88,84.12,93.12,98.64,105.12,105.85,127.92,
128.04,173.40)
# Make the TTT plot
TTT(data,lwd=1.5,grid=F,lty=3)
library(AdequacyModel)
# Generate a random sample from a Weibull distribution
data <- rweibull(8000,3,scale=1)
# Make the TTT transform
TTT(data, lwd=1.5,grid=F,lty=3)
ϕF(v)
–1 –c/k 0 v0 1
Figure 14.25 Determination of the optimal replacement age from the scaled TTT
transform.
The optimal value of 𝑣0 can now be read as the abscissa of the point where
the tangent touches the TTT transform. If 𝑣0 = 1, then t0 = ∞, and no preventive
replacements should be performed. The procedure is shown in Figure 14.25.
When a set of times-to-failure of the actual type of item has been recorded, we
may use this dataset to obtain the empirical, scaled TTT transform of the underly-
ing distribution function F(t), and draw a TTT plot. The optimal replacement age t0
may now be determined by the same procedure as described above. This is shown
in Figure 14.26. The procedure is further discussed, for example, by Bergman and
Klefsjö (1982,1984). ◻
–1 –c/k 0 i/n 1
Figure 14.26 Determination of the optimal replacement age from a TTT plot.
Technically, the plot is obtained as follows: Let t(1) , t(2) , … , t(k) denote the k
ordered failure times among t1 , t2 , … , tn and let
̂(t(i) )
𝑣(i) = 1 − R for i = 1, 2, … , k.
Define
t(i) ∑
i−1
̂ −1 (𝑣(i) ) =
H ̂ (u) du =
R ̂(t(j) ),
(t(j+1) − t(j) )R
∫0 j=1
where t(0) = 0.
The TTT plot is now obtained by plotting the points
( )
𝑣(i) Ĥ −1 (𝑣(i) )
, for 1 = 1, 2, … , k.
𝑣(k) Ĥ −1 (𝑣(k) )
Observe that when k = n, that is, when the dataset is complete, then
i
𝑣(i) = ,
n
̂ −1 (𝑣(i) ) = (t(i) ),
H
and we get the same TTT plot as we got for complete datasets.
(1) A time-dependent factor z0 (t), which is called the baseline failure rate and does
not depend on s. The baseline failure rate is usually not specified in the PH
model.
(2) A proportionality factor g(s), which is a function of the covariate vector s, and
not of time t.
Hazard Ratio
We may compare the effects of two covariate vectors s1 and s0 by the ratio:
z(t ∣ s1 ) g(s1 )
HR(s1 , s0 ) = = . (14.75)
z(t ∣ s0 ) g(s0 )
This expression is called the hazard ratio (HR) for the covariate vectors s1 and s0 .
The covariate vector s0 often refers to a basic and known application of the item,
called the baseline application, whereas the covariate vector s1 refers to the use of a
similar item in a new environment. The hazard ratio shows that the two failure rate
functions are proportional for any value of t. This proportionality is the reason for
calling the model a PH model. The factor of interest is how large g(s1 ) is compared
to g(s0 ) and not the value of each of them. Therefore, we often set g(s0 ) = 1, such
that g(s1 ) = HR(s1 , s0 ).
In cases where g(s0 ) = 1, and we study a single alternative covariate vector, this
vector is usually denoted s (i.e. without an index) and the hazard ratio is HR(s) =
g(s).
The effect of the covariate vector s is, therefore, determined by g(s), which
scales the baseline failure rate function z0 (t). Figure 14.27 shows a baseline failure
rate function z0 (t) for a Weibull distribution with shape parameter 𝛼 = 1.65 (fully
drawn line) together with the failure rate function (dotted line) for an item with
covariate vector s and hazard ratio g(s) = 2 based on a PH model. The failure rate
function for s is obtained by multiplying z0 (t) with HR = 2 for each point of time t.
0.6
0.4
z(t)
0.2
0.0
0 2 4 6 8 10
Time t
Figure 14.27 Failure rate function for the PH model. The baseline failure rate function
(fully drawn line) and for another condition with hazard ratio (HR)= 2 (dotted line). The
baseline is a Weibull distribution with shape parameter 𝛼 = 1.65.
14.8 Survival Analysis with Covariates 725
Survivor Function
Let R0 (t) be the survivor function for the baseline application. The survivor func-
tion for a new application with covariate vector s is from (14.74)
This result implies that if we know the survivor function in the baseline applica-
tion and if we are able to determine the hazard ratio, g(s), it is easy to find the sur-
vivor function – and all the related reliability measures – in the new environment s.
where ki is a constant that determines the effect of si on the failure rate, for i =
1, 2, … , m. To comply with our knowledge about the effect of the various influ-
ences, the covariates used may be transformed values of the physical variables.
The square of the voltage may, for example, be used as a covariate. ◻
𝜆 = 𝜆0 ⋅ 𝜋 S ⋅ 𝜋 T ⋅ 𝜋 E ⋅ 𝜋 Q ⋅ 𝜋 A , (14.78)
where
For the Cox model, the log-failure rate function is a linear function
log z(t ∣ s) = log z0 (t) + 𝛽1 s1 + 𝛽2 s2 + · · · + log 𝛽k sk . (14.80)
This indicates that (14.80) may be a suitable basis for some sort of regression anal-
ysis.
The Cox model is said to be a semiparametric model. It is not a parametric model
because the baseline failure rate z0 (t) is unspecified, and it is not nonparametric
because it is assumed how the failure rate function varies with the value of the
covariates. If we make special assumptions about the baseline failure rate func-
tion z0 (t), the Cox model becomes a parametric model. The baseline distribution
may, for example, be assumed to be an exponential or a Weibull distribution. The
advantage of the Cox model is that such assumptions can be avoided. Even though
z0 (t) is unspecified, our objective is to estimate the parameters 𝜷. One of the biggest
14.8 Survival Analysis with Covariates 727
advantages of the Cox model is that we can estimate the parameters 𝜷 that reflect
the effects of the covariates without having to make any assumptions about the
form of z0 (t).
where z(⋅) and R(⋅) are seen as functions of 𝜷 and “data” includes all the data avail-
able in the dataset, including ti , 𝛿i , and si for all the items. For the Cox model, the
likelihood function may be written as
∏
n
L(𝜷 ∣ data) = z0 (ti )[exp(𝜷si )]𝛿i [R0 (ti )]exp(𝜷si ) .
i=1
It is not possible to find the 𝜷 that maximizes this log-likelihood function unless
we specify the baseline failure rate function z0 (t). A detailed discussion of this
problem is given in Cox and Oakes (1984, chapter 7). Instead, Cox (1972) intro-
duced a partial likelihood function that does not depend on z0 (t). This function
uses the at-risk-set RS(t) at time t, that is, the set of all items that are functioning
and exposed to failure just before time t. Items that have failed or have been cen-
sored before time t are not members of RS(t). In this simplified introduction, we
assume that there are no ties in the dataset.
728 14 Reliability Data Analysis
and is the contribution to the partial likelihood [Lp (⋅)] from survival time ti . The
arguments used to arrive at the above result may be summarized as follows:
z(ti ∣ 𝜷si )
= ∑ .
j∈RS(ti ) z(ti ∣ 𝜷sj )
𝜓i = exp(𝜷si ) for i = 1, 2, … , n,
which is the factor we must multiply the baseline failure rate z0 (t) with to obtain
the failure rate for the covariate vector si , that is, z(ti ∣ 𝜷, si ) = 𝜓i z0 (t). The con-
tribution to the total partial likelihood function from failure time ti can hence be
9 Any item in RS(ti ) would do. We assume that item i corresponds to the ordered survival time
ti , so it is obviously a member of RS(ti ). We focus on item i to simplify the notation.
14.8 Survival Analysis with Covariates 729
written as
𝜓i
Lp (𝜷 ∣ ti , si ) = ∑ . (14.81)
j∈RS(ti ) 𝜓j
where censored times are excluded by the indicator 𝛿i = 0 (remember x0 = 1). The
partial likelihood function is obtained by multiplying the contributions (14.80)
from the actual failure times, but the censoring times are still important because
they enter into the at-risk-sets RS(t).
There are at least two reasons why Lp (𝜷 ∣ data) is called a partial likelihood:
• It is not a complete likelihood function for all parameters of the density function
(because the baseline failure rate function is not covered).
• All the data in the dataset is not used because the actual survival times play no
part in (14.81), but only their ranking, i.e. when they enter into the at-risk set.
More thorough treatments may be found in Cox and Oakes (1984), Lawless
(1982), and Kalbfleisch and Prentice (1980).
When there are many ties in the dataset, computation of maximum partial-
likelihood estimates is still possible but may become time-consuming. Of this
reason, the partial likelihood function is often approximated. Two commonly
employed approximations are due to Norman E. Breslow and to Bradley Efron.
Both approximations are available in the R survival package.
The procedures to find estimates for the various parameters are rather techni-
cal and are not presented in the current book. Readers who plan to use the Cox
model on a practical dataset are advised to consult a more specialized book and to
carefully read the documentation of the relevant R packages.
14.9 Problems
14.1 Assume that you have determined the lifetimes for a total of 12 identical
items and obtained the following results (given in hours): 10.2, 89.6, 54.0,
96.0, 23.3, 30.4, 41.2, 0.8, 73.2, 3.6, 28.0, 31.6
The dataset can be downloaded from the book companion site.
(a) Find the sample mean and the sample standard deviation for the
dataset. Can you draw any conclusions about the underlying distri-
bution F(t) by comparing the sample mean and the sample standard
deviation?
(b) Construct the empirical survivor function for the dataset.
(c) Plot the data on a Weibull paper.10 What conclusions can you draw
from the plot?
(d) Construct the TTT plot for the dataset. What conclusion can you draw
from the TTT plot about the corresponding life distribution?
14.2 Failure time data from a compressor were discussed in Example 10.2. All
compressor failures at a certain process plant in the time period from
1968 until 1989 have been recorded. In this period, a total of 90 critical
failures occurred. In this context, a critical failure is defined to be a fail-
ure causing compressor downtime. The compressor is very important for
the operation of the process plant, and every effort is taken to restart a
failed compressor as soon as possible. The 90 repair times (in hours) are
presented chronologically in Table 14.6. The repair time associated with
the first failure was 1.25 hours, the second repair time was 135.00 hours,
and so on. The dataset can be downloaded from the book companion
site.
(a) Plot the repair times in chronological order to check whether or not
there is a trend in the repair times. Is there any reason to claim that
the repair times increase with the age of the compressor?
(b) Assume now that the repair times are independent and identically
distributed. Construct the empirical distribution function for the
repair times
(c) Plot the repair times on a lognormal plotting paper. Is it reason to
believe that the repair times are lognormally distributed?
14.3 Consider the set of material strength data presented by Crowder et al.
(1991, p. 46) and given in Table 14.7. An experiment has been carried
14.4 Establish a graph paper such that the Nelson–Aalen plot of Weibull
distributed life data is close to a straight line. Describe how the Weibull
parameters 𝛼 and 𝜆 can be estimated from the plot.
14.8 Consider a homogeneous Poisson process (HPP) with rate 𝜆. Let N(t)
be the number of failures (events) in a time interval of length t. N(t) is
hence Poisson distributed with parameter 𝜆t. Assume that the process is
observed in a time interval of length t = 2 years. In this time period, a
total of seven failures have been observed.
(a) Find an estimate for 𝜆
(b) Determine a 90% confidence interval for 𝜆
14.10 Denote the distribution function of the Poisson distribution with parame-
ter 𝜆 by o (x; 𝜆), and the distribution function of the 𝜒 2 distribution with
𝜈 degrees of freedom by Γ𝜈 (z).
(a) Show that o (x ∣ 𝜆) = 1 − Γ2(x+1) (2𝜆). (Hint: First show that
∞ x
1 − Γ2(x+1) (2𝜆) = ∫2𝜆 ux! e−u du, and next apply repeated partial
integrations to the integral).
14.9 Problems 733
14.12 Reconsider the situation in Example 14.16, but assume that the
times-to-failure are those that are not starred. They are given in
Table 14.9. The dataset can be downloaded from the book companion
site.
(a) Determine the Kaplan–Meier estimate R̂(t) and display it graphically.
(b) Determine the Nelson–Aalen estimate R∗ (t) for the survivor function
and display it graphically.
14.13 Table 14.10 shows the intervals in operating hours between successive
failures of air-conditioning equipment in a Boeing 720 aircraft. The first
734 14 Reliability Data Analysis
interval is 413, the second is 14, and so on. The data are from Proschan
(1963). The dataset can be downloaded from the book companion
site.
(a) Establish the Nelson–Aalen plot (N(t) plot) of the dataset. Describe
(with words) the shape of the rate of occurrence of failures (ROCOF).
14.14 Suppose that the dataset in Problem 14.11 was obtained by simultane-
ously activating 20 identical items, but that the test was terminated at the
12th failure.
(a) What type of censoring is this?
(b) Estimate 𝜆 in this situation.
(c) Calculate a 95% confidence interval for 𝜆.
(d) Compare the results with those derived in Problem 14.11.
14.15 Establish a graph paper such that the Nelson–Aalen plot of normally dis-
tributed ( (𝜇, 𝜎 2 )) life data is close to a straight line. Describe how the
parameters 𝜇 and 𝜎 may be estimated from the plot.
14.16 Table 14.11 shows the intervals in days between successive failures of a
piece of software developed as part of a large data system. The first interval
is 9, the second is 12, and so on. The data are from Jelinski and Moranda
(1972).The dataset can be downloaded on the book companion site.
(a) Establish the Nelson–Aalen plot (N(t) plot) of the dataset. Is the
ROCOF increasing or decreasing?
(b) Assume that the ROCOF follows a log-linear model, and find the max-
imum likelihood estimates (MLE) for the parameters of this model.
14.9 Problems 735
9 12 11 4 7 2 5 8 5 7
1 6 1 9 4 1 3 3 6 1
11 33 7 91 2 1 87 47 12 9
135 258 16 35
(c) Draw the estimated cumulative ROCOF in the same diagram as the
Nelson–Aalen plot. Is the fit acceptable?
(d) Use the Laplace test to test whether the ROCOF is decreasing or not
(use a 5% level of significance).
14.18 A record of the times-to-failure (given in hours) of a sensor give the fol-
lowing historical dataset in Table 14.13. The dataset can be downloaded
from the book companion site.
1.2 × 104 9.3 × 104 0.5 × 104 0.2 × 104 1.1 × 104
2.6 × 104 9.4 × 104 1.2 × 104 4.9 × 104 9.6 × 104
4 4 4 4
0.9 × 10 8.6 × 10 6.5 × 10 0.5 × 10 1.0 × 104
0.1 × 104 0.8 × 104 3.6 × 104 3.2 × 104
736 14 Reliability Data Analysis
References