The RELIABILITY Procedure

Probability Plotting

Probability plots are useful tools for the display and analysis of lifetime data. See Abernethy (2006) for examples that use probability plots in the analysis of reliability data. Probability plots use a special scale so that a cumulative distribution function (CDF) plots as a straight line. Thus, if lifetime data are a sample from a distribution, the CDF estimated from the data plots approximately as a straight line on a probability plot for the distribution.

You can use the RELIABILITY procedure to construct probability plots for data that are complete, right censored, or interval censored (in readout form) for each of the probability distributions in Table 16.57.

A random variable Y belongs to a location-scale family of distributions if its CDF F is of the form

\[  Pr\{ Y \leq y\}  = F(y) = G\left(\frac{y-\mu }{\sigma }\right)  \]

where $\mu $ is the location parameter, and $\sigma $ is the scale parameter. Here, G is a CDF that cannot depend on any unknown parameters, and G is the CDF of Y if $\mu =0$ and $\sigma =1$. For example, if Y is a normal random variable with mean $\mu $ and standard deviation $\sigma $,

\[  G(u) = \Phi (u) = \int _{-\infty }^{u}\frac{1}{\sqrt {2\pi }}\exp \left(-\frac{u^{2}}{2}\right) \,  du  \]

and

\[  F(y) = \Phi \left(\frac{y-\mu }{\sigma }\right)  \]

Of the distributions in Table 16.57, the normal, extreme value, and logistic distributions are location-scale models. As shown in Table 16.58, if T has a lognormal, Weibull, or log-logistic distribution, then $\log (T)$ has a distribution that is a location-scale model. Probability plots are constructed for lognormal, Weibull, and log-logistic distributions by using $\log (T)$ instead of T in the plots.

Let $y_{(1)} \le y_{(2)} \le \ldots \le y_{(n)}$ be ordered observations of a random sample with distribution function $F(y)$. A probability plot is a plot of the points $y_{(i)}$ against $m_{i}=G^{-1}(a_{i})$, where $a_{i}=\hat{F}(y_{i})$ is an estimate of the CDF $F(y_{(i)})=G\left(\frac{y_{(i)}-\mu }{\sigma }\right)$. The points $a_{i}$ are called plotting positions. The axis on which the points $m_{i}$s are plotted is usually labeled with a probability scale (the scale of $a_{i}$).

If F is one of the location-scale distributions, then y is the lifetime; otherwise, the log of the lifetime is used to transform the distribution to a location-scale model.

If the data actually have the stated distribution, then $\hat{F} \approx F$,

\[ m_{i}=G^{-1}(\hat{F}(y_{i}))\approx G^{-1}\left(G\left(\frac{y_{(i)}-\mu }{\sigma }\right)\right)= \frac{y_{(i)}-\mu }{\sigma } \]

and points $(y_{(i)}, m_{i})$ should fall approximately on a straight line.

There are several ways to compute plotting positions from failure data. These are discussed in the next two sections.

Complete and Right-Censored Data

The censoring times must be taken into account when you compute plotting positions for right-censored data. The RELIABILITY procedure provides several methods for computing plotting positions. These are specified with the PPOS= option in the ANALYZE, PROBPLOT, and RELATIONPLOT statements. All of the methods give similar results, as illustrated in the section Expected Ranks, Kaplan-Meier, and Modified Kaplan-Meier Methods, the section Nelson-Aalen, and the section Median Ranks.

Expected Ranks, Kaplan-Meier, and Modified Kaplan-Meier Methods

Let $y_{(1)} \le y_{(2)} \le \ldots \le y_{(n)}$ be ordered observations of a random sample including failure times and censor times. Order the data in increasing order. Label all the data with reverse ranks $r_{i}$, with $r_{1}=n, \ldots , r_{n}=1$. For the failure corresponding to reverse rank $r_{i}$, compute the reliability, or survivor function estimate

\[  R_{i} = \left[\frac{r_{i}}{r_{i}+1}\right]R_{i-1}  \]

with $R_{0}=1$. The expected rank plotting position is computed as $a_{i}=1-R_{i}$. The option PPOS=EXPRANK specifies the expected rank plotting position.

For the Kaplan-Meier method,

\[  R_{i} = \left[\frac{r_{i}-1}{r_{i}}\right]R_{i-1}  \]

The Kaplan-Meier plotting position is then computed as $a’_{i}=1-R_{i}$. The option PPOS=KM specifies the Kaplan-Meier plotting position.

For the modified Kaplan-Meier method, use

\[  R’_{i} = \frac{R_{i} + R_{i-1}}{2}  \]

where $R_{i}$ is computed from the Kaplan-Meier formula with $R_{0}=1$. The plotting position is then computed as $a”_{i}=1-R’_{i}$. The option PPOS=MKM specifies the modified Kaplan-Meier plotting position. If the PPOS option is not specified, the modified Kaplan-Meier plotting position is used as the default method.

For complete samples, $a_{i}=i/(n+1)$ for the expected rank method, $a’_{i}=i/n$ for the Kaplan-Meier method, and $a”_{i}=(i-.5)/n$ for the modified Kaplan-Meier method. If the largest observation is a failure for the Kaplan-Meier estimator, then $F_{n}=1$ and the point is not plotted. These three methods are shown for the field winding data in Table 16.60 and Table 16.61.

Table 16.60: Expected Rank Plotting Position Calculations

Ordered

Reverse

$r_{i}/(r_{i}+1)$

$\times R_{i-1}$

$=R_{i}$

$a_{i}=1-R_{i}$

Observation

Rank

       

31.7

16

16/17

1.0000

0.9411

0.0588

39.2

15

15/16

0.9411

0.8824

0.1176

57.5

14

14/15

0.8824

0.8235

0.1765

65.0+

13

       

65.8

12

12/13

0.8235

0.7602

0.2398

70.0

11

11/12

0.7602

0.6968

0.3032

75.0+

10

       

75.0+

9

       

87.5+

8

       

88.3+

7

       

94.2+

6

       

101.7+

5

       

105.8

4

4/5

0.6968

0.5575

0.4425

109.2+

3

       

110.0

2

2/3

0.5575

0.3716

0.6284

130.0+

1

       

+ Censored Times


Table 16.61: Kaplan-Meier and Modified Kaplan-Meier Plotting Position Calculations

Ordered

Reverse

$(r_{i}-1)/r_{i}$

$\times R_{i-1}$

$=R_{i}$

$a’_{i}=1-R_{i}$

$a”_{i}$

Observation

Rank

         

31.7

16

15/16

1.0000

0.9375

0.0625

0.0313

39.2

15

14/15

0.9375

0.8750

0.1250

0.0938

57.5

14

13/14

0.8750

0.8125

0.1875

0.1563

65.0+

13

         

65.8

12

11/12

0.8125

0.7448

0.2552

0.2214

70.0

11

10/11

0.7448

0.6771

0.3229

0.2891

75.0+

10

         

75.0+

9

         

87.5+

8

         

88.3+

7

         

94.2+

6

         

101.7+

5

         

105.8

4

3/4

0.6771

0.5078

0.4922

0.4076

109.2+

3

         

110.0

2

1/2

0.5078

0.2539

0.7461

0.6192

130.0+

1

         

+ Censored Times


Nelson-Aalen

Estimate the cumulative hazard function by

\[  H_{i} = \frac{1}{r_{i}} + H_{i-1}  \]

with $H_0=0$. The reliability is $R_ i = \exp ( -H_ i )$, and the plotting position, or CDF, is $a”’_{i} = 1 - R_ i$. You can show that $R_{KM} < R_{NA}$ for all ages. The Nelson-Aalen method is shown for the field winding data in Table 16.62.

Table 16.62: Nelson-Aalen Plotting Position Calculations

Ordered

Reverse

$1/r_{i}$

$+ H_{i-1}$

$=H_{i}$

$a”’_{i}=1-\exp (-H_{i})$

Observation

Rank

       

31.7

16

1/16

0.0000

0.0625

0.0606

39.2

15

1/15

0.0625

0.1292

0.1212

57.5

14

1/14

0.1292

0.2006

0.1818

65.0+

13

       

65.8

12

1/12

0.2006

0.2839

0.2472

70.0

11

1/11

0.2839

0.3748

0.3126

75.0+

10

       

75.0+

9

       

87.5+

8

       

88.3+

7

       

94.2+

6

       

101.7+

5

       

105.8

4

1/4

0.3748

0.6248

0.4647

109.2+

3

       

110.0

2

1/2

0.6248

1.1248

0.6753

130.0+

1

       

+ Censored Times


Median Ranks

See Abernethy (2006) for a discussion of the methods described in this section. Let $y_{(1)} \le y_{(2)} \le \ldots \le y_{(n)}$ be ordered observations of a random sample including failure times and censor times. A failure order number $j_{i}$ is assigned to the ith failure: $j_{i}=j_{i-1}+\Delta $, where $j_{0}=0$. The increment $\Delta $ is initially 1 and is modified when a censoring time is encountered in the ordered sample. The new increment is computed as

\[  \Delta = \frac{(n+1) - \mbox{ previous failure order number }}{1 + \mbox{ number of items beyond previous censored item }}  \]

The plotting position is computed for the ith failure time as

\[  a_{i} = \frac{j_{i} - .3}{n + .4}  \]

For complete samples, the failure order number $j_{i}$ is equal to i, the order of the failure in the sample. In this case, the preceding equation for $a_{i}$ is an approximation to the median plotting position computed as the median of the ith-order statistic from the uniform distribution on (0, 1). In the censored case, $j_{i}$ is not necessarily an integer, but the preceding equation still provides an approximation to the median plotting position. The PPOS=MEDRANK option specifies the median rank plotting position.

For complete data, an alternative method of computing the median rank plotting position for failure i is to compute the exact median of the distribution of the ith order statistic of a sample of size n from the uniform distribution on (0,1). If the data are right censored, the adjusted rank $j_{i}$, as defined in the preceding paragraph, is used in place of i in the computation of the median rank. The PPOS=MEDRANK1 option specifies this type of plotting position.

Nelson (1982, p. 148) provides the following example of multiply right-censored failure data for field windings in electrical generators. Table 16.63 shows the data, the intermediate calculations, and the plotting positions calculated by exact ($a’_{i}$) and approximate ($a_{i}$) median ranks.

Table 16.63: Median Rank Plotting Position Calculations

Ordered

Increment

Failure Order

   

Observation

$\Delta $

Number $j_{i}$

$a_{i}$

$a’_{i}$

31.7

1.0000

1.0000

0.04268

0.04240

39.2

1.0000

2.0000

0.1037

0.1027

57.5

1.0000

3.0000

0.1646

0.1637

65.0+

1.0769

     

65.8

1.0769

4.0769

0.2303

0.2294

70.0

1.0769

5.1538

0.2960

0.2953

75.0+

1.1846

     

75.0+

1.3162

     

87.5+

1.4808

     

88.3+

1.6923

     

94.2+

1.9744

     

101.7+

2.3692

     

105.8

2.3692

7.5231

0.4404

0.4402

109.2+

3.1590

     

110.0

3.1590

10.6821

0.6331

0.6335

130.0+

6.3179

     

+ Censored Times


Interval-Censored Data

Readout Data

The RELIABILITY procedure can create probability plots for interval-censored data when all units share common interval endpoints. This type of data is called readout data in the RELIABILITY procedure. Estimates of the cumulative distribution function are computed at times corresponding to the interval endpoints. Right censoring can also be accommodated if the censor times correspond to interval endpoints. See the section Weibull Analysis of Interval Data with Common Inspection Schedule for an example of a Weibull plot and analysis for interval data.

Table 16.64 illustrates the computational scheme used to compute the CDF estimates. The data are failure data for microprocessors (Nelson, 1990, p. 147). In Table 16.64, $t_{i}$ are the interval upper endpoints, in hours, $f_{i}$ is the number of units failing in interval i, and $n_{i}$ is the number of unfailed units at the beginning of interval i.

Note that there is right censoring as well as interval censoring in these data. For example, two units fail in the interval (24, 48) hours, and there are 1414 unfailed units at the beginning of the interval, 24 hours. At the beginning of the next interval, (48, 168) hours, there are 573 unfailed units. The number of unfailed units that are removed from the test at 48 hours is $1414 - 2 - 573 = 839$ units. These are right-censored units.

The reliability at the end of interval i is computed recursively as

\[  R_{i} = (1-(f_{i}/n_{i}))R_{i-1}  \]

with $R_{0}=1$. The plotting position is $a_{i}=1-R_{i}$.

Table 16.64: Interval-Censored Plotting Position Calculations

Interval

Interval

$f_{i}/n_{i}$

$R^{}_{i}=$

$R_{i}=$

$a_{i}=1-R_{i}$

i

Endpoint $t_{i}$

 

$1-(f_{i}/n{i})$

$R^{}_{i}R_{i-1}$

 

1

6

6/1423

0.99578

0.99578

.00421

2

12

2/1417

0.99859

0.99438

.00562

3

24

0/1415

1.00000

0.99438

.00562

4

48

2/1414

0.99859

0.99297

.00703

5

168

1/573

0.99825

0.99124

.00876

6

500

1/422

0.99763

0.98889

.01111

7

1000

2/272

0.99265

0.98162

.01838

8

2000

1/123

0.99187

0.97364

.02636


Arbitrarily Censored Data

The RELIABILITY procedure can create probability plots for data that consists of combinations of exact, left-censored, right-censored, and interval-censored lifetimes. Unlike the method in the previous section, failure intervals need not share common endpoints, although if the intervals share common endpoints, the two methods give the same results. The RELIABILITY procedure uses an iterative algorithm developed by Turnbull (1976) to compute a nonparametric maximum likelihood estimate of the cumulative distribution function for the data. Since the technique is maximum likelihood, standard errors of the cumulative probability estimates are computed from the inverse of the associated Fisher information matrix. A technique developed by Gentleman and Geyer (1994) is used to check for convergence to the maximum likelihood estimate. Also see Meeker and Escobar (1998, chap. 3) for more information.

Although this method applies to more general situations, where the intervals may be overlapping, the example of the previous section will be used to illustrate the method. Table 16.65 contains the microprocessor data of the previous section, arranged in intervals. A missing (.) lower endpoint indicates left censoring, and a missing upper endpoint indicates right censoring. These can be thought of as semi-infinite intervals with lower (upper) endpoint of negative (positive) infinity for left (right) censoring.

Table 16.65: Interval-Censored Data

Lower

Upper

Number

Endpoint

Endpoint

Failed

.

6

6

6

12

2

24

48

2

24

.

1

48

168

1

48

.

839

168

500

1

168

.

150

500

1000

2

500

.

149

1000

2000

1

1000

.

147

2000

.

122


The following SAS statements compute the Turnbull estimate and create a lognormal probability plot:

data micro;         
   input t1 t2 f ;  
   datalines;           
. 6 6            
6 12 2           
12 24 0          
24 48 2          
24 .  1          
48 168 1         
48 .   839       
168 500 1        
168 .   150      
500 1000 2       
500 .    149     
1000 2000 1      
1000 . 147       
2000 . 122       
;                
proc reliability data=micro;
   distribution lognormal;
   freq f;
   pplot ( t1 t2 ) / itprintem 
                     printprobs
                     maxitem = ( 1000, 25 )
                     nofit
                     npintervals = simul
                     ppout;
run;

The nonparametric maximum likelihood estimate of the CDF can only increase on certain intervals, and must remain constant between the intervals. The Turnbull algorithm first computes the intervals on which the nonparametric maximum likelihood estimate of the CDF can increase. The algorithm then iteratively estimates the probability associated with each interval. The ITPRINTEM option along with the PRINTPROBS option instructs the procedure to print the intervals on which probability increases can occur and the iterative history of the estimates of the interval probabilities. The PPOUT option requests tabular output of the estimated CDF, standard errors, and confidence limits for each cumulative probability.

Figure 16.54 shows every 25th iteration and the last iteration for the Turnbull estimate of the CDF for the microprocessor data. The initial estimate assigns equal probabilities to each interval. You can specify different initial values with the PROBLIST= option. The algorithm converges in 130 iterations for this data. Convergence is determined if the change in the loglikelihood between two successive iterations less than $\Delta $, where the default value is $\Delta =10^{-8}$. You can specify a different value for delta with the TOLLIKE= option. This algorithm is an example of an expectation-maximization (EM) algorithm. EM algorithms are known to converge slowly, but the computations within each iteration for the Turnbull algorithm are moderate. Iterations will be terminated if the algorithm does not converge after a fixed number of iterations. The default maximum number of iterations is 1000. Some data may require more iterations for convergence. You can specify the maximum allowed number of iterations with the MAXITEM= option in the PROBPLOT, ANALYZE, or RPLOT statement.

Figure 16.54: Iteration History for Turnbull Estimate

The RELIABILITY Procedure

Iteration History for the Turnbull Estimate of the CDF
Iteration Loglikelihood (., 6) (6, 12) (24, 48) (48, 168) (168, 500) (500, 1000) (1000, 2000) (2000, .)
0 -1133.4051 0.125 0.125 0.125 0.125 0.125 0.125 0.125 0.125
25 -104.16622 0.00421644 0.00140548 0.00140648 0.00173338 0.00237846 0.00846094 0.04565407 0.93474475
50 -101.15151 0.00421644 0.00140548 0.00140648 0.00173293 0.00234891 0.00727679 0.01174486 0.96986811
75 -101.06641 0.00421644 0.00140548 0.00140648 0.00173293 0.00234891 0.00727127 0.00835638 0.9732621
100 -101.06534 0.00421644 0.00140548 0.00140648 0.00173293 0.00234891 0.00727125 0.00801814 0.97360037
125 -101.06533 0.00421644 0.00140548 0.00140648 0.00173293 0.00234891 0.00727125 0.00798438 0.97363413
130 -101.06533 0.00421644 0.00140548 0.00140648 0.00173293 0.00234891 0.00727125 0.007983 0.97363551


If an interval probability is smaller than a tolerance ($10^{-6}$ by default) after convergence, the probability is set to zero, the interval probabilities are renormalized so that they add to one, and iterations are restarted. Usually the algorithm converges in just a few more iterations. You can change the default value of the tolerance with the TOLPROB= option. You can specify the NOPOLISH option to avoid setting small probabilities to zero and restarting the algorithm.

If you specify the ITPRINTEM option, the table in Figure 16.55 summarizing the Turnbull estimate of the interval probabilities is printed. The columns labeled ’Reduced Gradient’ and ’Lagrange Multiplier’ are used in checking final convergence to the maximum likelihood estimate. The Lagrange multipliers must all be greater than or equal to zero, or the solution is not maximum likelihood. See Gentleman and Geyer (1994) for more details of the convergence checking.

Figure 16.55: Final Probability Estimates for Turnbull Algorithm

Lower Lifetime Upper Lifetime Probability Reduced Gradient Lagrange Multiplier
. 6 0.0042 0 0
6 12 0.0014 0 0
24 48 0.0014 0 0
48 168 0.0017 0 0
168 500 0.0023 0 0
500 1000 0.0073 -7.219342E-9 0
1000 2000 0.0080 -0.037063236 0
2000 . 0.9736 0.0003038877 0


Figure 16.56 shows the final estimate of the CDF, along with standard errors and confidence limits. Figure 16.57 shows the CDF and simultaneous confidence limits plotted on a lognormal probability plot.

Figure 16.56: Final CDF Estimates for Turnbull Algorithm

Cumulative Probability Estimates
Lower Lifetime Upper Lifetime Cumulative
Probability
Pointwise 95% Confidence
Limits
Standard Error
Lower Upper
6 6 0.0042 0.0019 0.0094 0.0017
12 24 0.0056 0.0028 0.0112 0.0020
48 48 0.0070 0.0038 0.0130 0.0022
168 168 0.0088 0.0047 0.0164 0.0028
500 500 0.0111 0.0058 0.0211 0.0037
1000 1000 0.0184 0.0094 0.0357 0.0063
2000 2000 0.0264 0.0124 0.0553 0.0101


Figure 16.57: Lognormal Probability Plot for the Microprocessor Data

Lognormal Probability Plot for the Microprocessor Data