- Research
- Open access
- Published:

# Bayesian test for hazard ratio in survival analysis

*SpringerPlus*
**volumeÂ 5**, ArticleÂ number:Â 649 (2016)

## Abstract

Over the decades, testing for equivalence of hazard functions has received a wide attention in survival analysis. In this paper, we proposed a Bayesian test to address this testing equivalence problem, Most of all, proposed test is methodologically flexible so that a procedure determining weights is not required when the proportional assumption is violated. In comparison with popularly exploited methods, the proposed test is shown to be more powerful and robust in testing differences of hazard functions, in spite of the presence of crossing hazard functions. Extensive applications to simulation and real data were conducted, demonstrating that the proposed test presents outstanding performance and hold desirable properties in terms of numerical aspects.

## Background

Inference of the survival function \(\mathrm{P}( T >t )\) is a main focus of survival analysis, where *T* follows the distribution *F* on \([0, \infty ).\) Survival functions play a key role in testing the effects of clinical therapies or drugs, reliability analysis in engineering, and estimating the risk of bankrupts.

If we let the hazard function of *T* be

and there exists a probability density function of *T*, *f* dominated by Lebesgue measure, then

where \(F(t) = \int _0^t f(s) ds.\) The survival function is

In practice, we often encounter a censoring random variable *C*,Â and observe \(X = \mathrm{min} (T, C).\)

If we have separate groups and our main interest aims at testing differences between hazard functions, we need to address testing the equality of the hazard functions. For this end, Mantel (1966) proposed the log rank test, and many analogous methods motivated by the log rank test (e.g., the weighted log rank tests) were studied by Gehan (1965), Peto and Peto (1972), and Prentice (1978). The log rank test commonly suffers low power when the ratio of the hazard functions differs in the time line. For this reason, the weighted log rank tests were developed to overcome the limitation of the log rank test, and various theoretical properties of these tests were introduced in Gill (1980), Harrington etÂ al. (1982), Fleming and Harrington (2005), and Andersen etÂ al. (1993) relying on martingale theories. More importantly, it was shown that the tests hold consistency and the test proposed in Harrington etÂ al. (1982) proved to be the locally most powerful rank test in the specific class of survival functions. However, power of aforementioned tests may possibly vary depending on types of the hazard functions. Also Renyi test motivated by RÃ©nyi (1953) has been widely used in practice. This test requires weights similar to weighted log rank tests.

In this paper, we primarily focus on testing equivalence of hazard functions through the Coxâ€™s proportional hazards model (Cox 1972) such that

Here \(\beta \in \mathbb {R}\) and *z* is a covariate. If we perform a test procedure for \(M_0 : \beta =0 \ \mathrm{against} \ M_1 : \beta \ne 0\) where *z* is an indicator variable for each group (0 = control group, 1 = treatment group), it is equivalent to test equivalence of the hazard functions against \(\lambda (t)/\lambda _0(t) =c \ne 1\) for all *t*. Thus this test may decrease power when \(\lambda (t)/\lambda _0(t)\) is a time-varying function, especially in the case of \(\lambda (t)/\lambda _0(t) = (t-1/2)\) on [0,Â 1],Â i.e, crossing hazards. Thus if we consider the time-varying Coxâ€™s model such as

incorporating a time-varying coefficient, and have a test procedure for the testing

then we can construct the test working well in spite of the crossing hazards. Inspired by the frequentist approach, Hess (1994) and Verweij and van Houwelingen (1995) studied time-varying coefficient model in Coxâ€™ regression, and provided the estimation methodology, In particular, Verweij and van Houwelingen (1995) proposed a test procedure using the B-spline basis functions. Also Yang and Prentice (2005) proposed the advanced semi-parametric model including the proportional hazards model and proportional odds model, and proposed a test procedure for detecting the crossing hazards. Yang and Prentice test has no adaptive step such as selecting weights, and shows efficient performance. Recently, Chauvel and Oâ€™Quigley (2014) studied the test based on Coxâ€™s regression with time-varying coefficients. They used the stochastic integral and its limit distribution to test \(\beta (\cdot )\equiv 0\).

When it comes to testing equivalence of hazards including crossing hazards, few Bayesian studies have been scarcely utilized. Although Kalbfleisch (1978), Hjort (1990), and Kim (2006) turned to the Bayesian methodology for estimation of hazard or survival function and Kim etÂ al. (2011) proposed the Bayesian test for monotone hazards, to our best knowledge, there are only a few studies done for testing equivalence of hazards including crossing hazards.

In the context of Bayesian approach, the testing

is equivalent to model selection using posterior probabilities of \(M_0\) and \(M_1\) where \(\mathcal {F}\) is a function class such as Sobolev space. Also Bayesian asymptotic theories proved that if data are randomly sampled, Bayesian test is consistent when

in probability \(\text { as } n \rightarrow \infty\) under \(M_1\), and

in probability \(\text { as } n \rightarrow \infty\) under \(M_0.\) So Bayesian test canâ€™t give the typical *p* value, but the construction of the test procedure is easy and interpretation of this test is straightforward.

In addition, theoretical studies of Kim (2012) imply consistency of this Bayesian test using only the partial likelihood when we use a prior of \(\pi (\beta )\) under \(M_1\) having the support on the function class absolutely bounded and spanned by the B-spline basis functions (obviously the prior for \(\beta\) under \(M_0\) is a Dirac measure at 0). Under regularity conditions and prior masses of *q* and \(1-q \ (0< q <1)\) for the model \(M_0\) and \(M_1\), respectively, Kim (2012) shows that we can have \(\mathcal {F}\) as the function class such that all derivatives from 0 to \(p \ ( \in \mathbb {N} )\) are absolutely bounded at a compact set in the time line.

In this paper, we construct the Bayesian test based on the results of Kim (2012). Considered model, data and test are explained. Priors and posteriors for Bayesian test are shown. We performed various simulation studies and real data analysis. Concluding remarks and discussions are presented in the last section.

## Model and Bayesian test

Assume that we have \(D_{1:n} = \{ ( X_i, \delta _i, z_i ) \}_{i=1}^n\) where

for \(t \in [0, \infty ),\) and \(F_{z_i}, G\) and *I* are distribution functions and an indicator function, respectively. Here \((C_i, \delta _i)\) is a random vector of censoring variable, censoring indicator and \(z_i\) is a group indicator, respectively. We also assume that for some \(0< \tau <\infty ,\)
\(G(t-) = G(t)\) on \(t \in [0, \tau )\) and \(G(\tau ) = 1.\) Note that we have no ties in the uncensored failure time \(X_i\)s, and observed \(X_i\)s are bounded by \(\tau\).

Since we have the survival function of \(T_i\) given \(z_i\):

we can consider the testing

where

and \(\beta ^{(p)}\) is the *p*th \(( p \in \mathbb {N} )\) derivative of \(\beta\).

## Partial likelihood, priors and posteriors for the test

Before a description of the test procedure, we observe the likelihood \(L \left( \beta , \lambda _0 \right)\) and the partial likelihood \(L \left( \beta \right)\) as

where \(R(t) = \{ k : X_k \ge t \}.\) The use of (12) requires difficult computations because of \(\lambda _0\) and integration in the exponent term. Especially in the Bayesian method, priors for \(\lambda _0\) can greatly increase computational burdens. On the other hand, using (13) is relatively easy to implement. We can refer to Andersen and Gill (1982) for the large sample property of partial likelihood. So it is attractive to use only the partial likelihood in Bayesian analysis, and Kim (2006, 2012) have reported that we can use the partial likelihood and \(\pi (\beta )\), the priors for \(\beta\) to obtain the Bayes estimators for \(\beta\) since \(L \left( \beta ; D_{1:n} \right) \pi ( \beta )\) is proportional to the marginal posterior of \(\beta\) when beta processes are used as priors for \(\lambda _0\).

For the test, we consider the expansion by the B-spline basis functions such that

for \(\beta (X_i)\), where \(B_{d, a_n, l}\) is the B-spline basis functions of degree *d* with equally spaced knots and \(\eta \in \{ 0, 1\}\). See de Boor (2001) and Lyche and MÃ¸rken (2008) for details of the B-spline, and Fig.Â 1 shows the B-spline basis functions of degree 1 and 2. We can use \(d \ge p-1\) to approximate the function in \(\mathcal {F}_{p, M}.\) Priors are then put on \(\eta \in \{ 0, 1\}\) and \(\gamma _l\)s, also let \(a_n = \left[ ( n / \log n )^{1/(2p+1)} \right]\) for the consistent Bayesian test (Kim 2012). If we obtain a posterior probability of \(\eta =1\), we can test \(\beta (\cdot ) \equiv 0\).

For the priors on \(\eta\) and \(\gamma _l\)s, we consider the following:

where \(\phi ( \cdot ; 0, \sigma ^2)\) is the probability density function of the normal distribution with mean 0 and variance \(\sigma ^2 >0\), and

for a large \(L >0\). Here \(\sigma ^2\) and *L* are hyper parameters.

We obtain posteriors by using only the partial likelihood of (13) instead of the full likelihood of (12). If we let

and

then we have

from the partial likelihood, and the posteriors are

Here the posterior probability of \(\eta =1\) is equivalent to the posterior of \(M_1\) where \(\pi ( \{ \gamma \}_{i=1}^{a_n} )\) is a prior for \(M_1\) (shown in the Appendix). Posteriors can be obtained from the Metropolis-Hastings algorithm (Metropolis etÂ al. 1953; Hastings 1970) or rejection sampling. Note that since \(z \beta\) is in the exponent term, too large value of it can breaks down the MCMC (Markov chain Monte Carlo) algorithm. Thus we choose moderate \(\sigma ^2>0\) with sufficiently large \(L>0.\)

Although we can put priors on \(\sigma ^2\), we instead use hyper-parameters for the simplicity of computation. We also use the Bayesian bootstrap proposed by Kim and Lee (2003) for posterior sampling for \(\gamma _l\)s since it is speedy and gives more stable results. Details of the Bayesian bootstrap and applications can be found in Kim and Lee (2003) and Kim etÂ al. (2011). After obtaining the posterior samples, we calculated the Bayes estimates of

If the estimates are over 0.5, we reject equivalence of hazard functions, i.e., we choose a model with higher posterior probability. Kass and Adrian (1995) proposed the procedures for model selection, but we have only two models. Thus this approach is reasonable though it can be seen a little liberal.

## Simulation studies

In this section, we performed numerical studies for various values of \(\beta (t)\). Let \(\tau =6, \ L= 10, \ \sigma ^2 = 1.0\), and the censoring random variables were generated from truncated exponential distributions. Knot points selection is commonly critical especially when censoring rates are high, but here we considered rather simple cases such that inner knots are equally spaced on \(\big [ l_n+\epsilon , u_n-\epsilon \big ]\) for very small \(\epsilon >0\), where

Note that data out of the range of \(\big [ l_n, u_n \big ]\) have no effect for testing \(\beta\). Emmanuel etÂ al. (2010) and Eduard and Paulo (2014) introduced adaptive knots selection whereas our simulations instead adopt simplified scenarios to address properties of the proposed test.

### Simple setups and results

First, we take into account typical cases of hazard function equivalence, proportional ratio of hazard function, and a changing ratio of hazard function with censoring rate 0.3. From these simulations, we were able to verify numerical properties of the proposed method. The details of the three models are below

and further details are illustrated in Fig.Â 2 along with the results summarized in Table 1. First, we conducted the log rank, Yang Prentice, Fleming and Harrington, and Renyi tests using *p* values (reject null hypothesis-equivalence of hazard if *p* value is not greater than 0.05). Also proposed test were implemented by a five B-spline basis functions.

All numbers in TableÂ 1 represent the rejection ratio of equivalence of hazards functions from 100 replications. In Fleming and Harrington tests, 1 and 2 mean that we use \(\hat{S}(t)\) and \((1-\hat{S}(t))\) as weights, respectively, where \(\hat{S}(t)\) is the pooled estimator of survival function. Also Renyi tests, 1 and 2 means giving more weight to differences early on and later on, respectively. As shown in Table 1, the power of the log rank test is outstandingly high when proportional assumption is true. Fleming and Harrington test seems to similar to log rank test under the proportional assumption, while its performances are variable in the case of a changing ratio. In the Fleming and Harrington tests, performance is very sensitive to weight selection. Behaviors of Renyi tests are similar to Fleming and Harrington tests, and its powers are slightly lower than Fleming and Harrington tests.The Yang and Prentice test largely performs well in a range of scenarios because it theoretically covers wider models than the proportional hazards model. It is also interesting to note that the proposed test performs well in the various simulation conditions, particularly when ratios of hazards functions are quite far away from 1 even though the ratio of hazard functions is not continuous.

### Crossing and diverging hazards

The following simulation setups are motivated by crossing hazards. For example, it is reported by Schein (1982) (Gastrointestinal Tumor Study Group) that, a trial that compared chemotherapy with combined chemotherapy and radiation therapy in the treatment produced a ratio of survival functions (denominated by former group) that varied from under 1 to over 1, crossing along the time line. Importantly, this argument implied that enduring radiation is somewhat risky, but increases the life expectancy of patients. It is a conventional problem of crossing hazards, which has tendency to cause low power of the standard log rank test. Crossing hazards are interesting topics including identification of changing points in the ratio of hazard functions and estimation of hazard functions, which is studied by Muggeo and Miriam (2010). Also we consider diverging hazards that hazard ratio is a monotone function but not being 1 (if the ratio may have 1, it is the same as the crossing hazard problems).

Here we consider examples of both crossing hazard functions and diverging hazard functions such as

where \(\lambda _0(t) = 0.25.\) The \(M_3\) and \(M_4\) are the examples of crossing hazards and diverging hazards, respectively. FigureÂ 3 shows the survival function generated from each model. The survival functions appear no crossing, despite the crossing hazards. Interestingly, however, the difference between survival functions is shown to have the variation in curvatures.

To examine numerical properties and testing powers, we increase data size in combination with varied censoring rates (e.g., \(n=50, 100,\) and 200 with censoring rates of 0.30, 0.50, and 0.70). We contrived simulation schemes similar to the previous section, and summarized simulation results in Table 2. The numbers in the table represent the rejection ratio of hazard function equivalence from 100 replications.

Most of all Table 2 clearly shows that increasing data sizes and lower censoring rates improve performance. Note that the Fleming and Harrington testsâ€™ performance and Ranyi testsâ€™ performance depend on the weights yet and it performed best with some appropriate weights. In contrast wrong weight selection tends to results in fairly poor performances. Moreover, we found that the proposed test performs better than the log rank, Yang and Prentice test for all simulation scenarios, when the censoring rate is not high. However, high censoring rates generally bring about attenuate performance with respect to other tests when data size is relatively large.

In the left side of each plot in Fig.Â 4, the proposed test is shown to perform best when censoring rate is lower or moderate. The proposed test is inferior to Fleming and Harrington test 2 when censoring rates are high. When it comes to diverging hazard functions, the proposed test, Yang and Prentice test, and log rank test revealed the similar patterns when censoring rate is not high. With high censoring rate, the proposed test did not outperform any other test in \(M_4\), while, it performed well in \(M_3.\) We omitted Ranyi tests in this figure because the performance is similar to Fleming and Harrington tests, but we draw the plot of Fig.Â 6 in Appendix to compare these two tests.

Since our test is based on B-spline basis functions, which concerns non-parametric in theory, data size therefore strongly associated with the testing power. In high censoring environments, non-censoring data are rare, and so it could reduce the efficiency of the proposed test primarily due to non-parametric nature.

When performing data analysis, it is integral to carefully select the number of knots and knot points to circumvent the shortcoming of the proposed method. Nevertheless, it is certain that the proposed method accommodate many challenging testing equivalence problems, which existing method cannot effectively address in many ways.

###
*Remark*

Each MCMC chain in simulations had a size of 200 obtained by 1000 burn-in and thinned by 25. We observed the posterior of \(\eta\) in one replication in Fig.Â 7 of the Appendix. The cumulative means became stable as the posterior sample became larger, implying the estimates of \(\mathrm{P} ( \eta =1 | D_{1:n})\) are stable. In addition, Fig.Â 8 in Appendix displays the posterior mean of \(\eta\) for 100 replications when the data size 100 and censoring rate 0.50. This result proves the stability of the Bayes estimates.

## Real data analysis

We consider the data set available in R package YPmodel by Yang and Prentice (2005). Data sets include 90 patients, half of whom were treated with chemotherapy, the other half with the chemotherapy combined with the radiation therapy. There were two censoring in the former group and six censoring in the latter group. Yang and Prentice (2005) showed that the survival functions crossed near 1000 days by Kaplanâ€“Meier estimates (Kaplan and Meier 1958) crossed near 1000 days on the *x*-axis in Fig.Â 5. This implies the strong evidence for the crossing hazards. In addition, we report that Fleming and Harrington test (1/2) give *p* values of 0.04 and 0.16, respectively. Also Renyi test (1/2) give *p* values of 0.01 and 0.30, respectively. It supports that crossing hazards can exist.

The posterior probability of \(\eta =1\) is 0.62 from the proposed test, and the Yang and Prentice test gives a *p* value of 0.03. The log rank test gives a *p* value over 0.25. Taken together, these results showed that the proposed test performs well, and the proposed, Yang and Prentice tests identify non-equivalence of hazard functions. In contrast to the success of the proposed and Yang and Prentice tests, the log rank test cannot detect the non-equivalence of hazard functions.

## Conclusions

We showed that Bayesian test worked well to test hazard function equivalence, especially when crossing hazards appeared. It is commonplace that Bayesian test suffer from computation complexity or inconsistent phenomenon. Even so, we can construct a consistent Bayesian test via the B-spline basis functions. However, we also found that selection of *p* and the number of the B-spline basis functions still remains controversial. Using P-splines or putting priors for *p* in \(a_n\) can be further considered, possibly giving better performance for high censoring environments.

In addition, we can extend the proposed test for more than three groups by modeling of

where \(\{ z_i \}_{i=1}^{k-1}\) are indicators to distinguish from the baseline group. In this paper, we are only allowed for testing \(\beta (\cdot )\equiv 0\), however, estimation of \(\beta\) and detecting the time of crossing are interesting works in medical research. Since the proposed approach is based on the Bayesian methodology, extension to estimation of \(\beta\) is feasible and tractable for implementation. We left Bayesian estimation and testing for crossing hazards for interesting future work.

## References

Andersen PK, Borgan O, Gill RD, Keiding N (1993) Statistical models based on counting processes. Springer, New York

Andersen PK, Gill RD (1982) Coxâ€™s regression model for counting processes: a large sample study. Ann Stat 10(4):1100â€“1120

Chauvel C, Oâ€™Quigley J (2014) Tests for comparing estimated survival functions. Biometrika 101(3):535â€“552

Cox DR (1972) Regression models and life-tables. J R Stat Soc Ser B 34(2):187â€“220

de Boor C (2001) A practical guide to splines. Springer, New York

Eduard B, Paulo S (2014) Adaptive priors based on splines with random knots. Bayesian Anal 9(4):859â€“882

Emmanuel S, Robert LS, David R, Mark C, Lakshmi H (2010) Bayesian adaptive B-spline estimation in proportional hazards frailty models. Electron J Stat 4:606â€“642

Fleming TR, Harrington DP (2005) Counting processes and survival analysis. Wiley, Hoboken

Gehan EA (1965) A generalized two-sample Wilcoxon test for doubly censored data. Biometrika 52(3):650â€“653

Gill R (1980) Censoring and stochastic integrals. Stat Neerl 34(2):124

Harrington DP, Fleming TR, Gill R (1982) A class of rank test procedures for censored survival data. Biometrika 69(3):553â€“566

Hastings EK (1970) Monte Carlo sampling methods using Markov chains and their applications. Biometrika 57(1):97â€“109

Hess KR (1994) Assessing time-by-covariate interactions in proportional hazards regression models using cubic spline functions. Stat Med 13(10):1045â€“1062

Hjort NL (1990) Nonparametric Bayes estimators based on beta processes for life history data. Ann Stat 18(3):1259â€“1294

Kalbfleisch JD (1978) Non-parametric Bayesian analysis of survival time data. J R Stat Soc Ser B 40(2):214â€“221

Kaplan E, Meier P (1958) Nonparametric estimation from incomplete observations. J Am Stat Assoc 53(282):457â€“481

Kass RE, Adrian ER (1995) Bayes factors. J Am Stat Assoc 90(430):773â€“795

Kim Y (2006) The Bernsteinâ€“von Mises theorem for the proportional hazard model. Ann Stat 34(4):1678â€“1700

Kim G (2012) Posterior contraction rate of the proportional hazards model having a nonparametric link and its applications. PhD thesis, Seoul National University, Department of Statistics

Kim Y, Lee JY (2003) Bayesian bootstrap for proportional hazards models. Ann Stat 31(6):1905â€“1922

Kim Y, Park JK, Kim G (2011) Bayesian analysis for monotone hazard ratio. Lifetime Data Anal 17(2):302â€“320

Lyche T, MÃ¸rken K (2008) Spline Methods (draft). Department of Informatics, Center of Mathematics for Applications, University of Oslo, Oslo

Mantel N (1966) Evaluation of survival data and two new rank order statistics arising in its consideration. Cancer Chemother Rep Part 1 50(3):163â€“170

Metropolis N, Rosenbluth AW, Rosenbluth MN, Teller AH, Teller E (1953) Equations of state calculations by fast computing machines. J Chem Phys 21(6):1087â€“1092

Muggeo VMR, Miriam T (2010) A flexible approach to crossing hazards problem. Stat Med 29(18):1947â€“1957

Peto R, Peto J (1972) Asymptotically efficient rank invariant test procedures (with discussion). J R Stat Soc Ser A 135(2):185â€“206

Prentice RL (1978) Linear rank test with right censored data. Biometrika 65(1):291â€“298

RÃ©nyi A (1953) On the theory of order statistics. Acta Math Hung 4(3â€“4):191â€“231

Schein PD (1982) A comparison of combination chemotherapy and combined modality therapy for locally advanced gastric carcinoma. Cancer 49(9):1771â€“1777

Verweij JM, van Houwelingen HC (1995) Time-dependent effects of fixed covariates in Cox regression. Biometrics 51:1088â€“1108

Yang S, Prentice R (2005) Semiparametric analysis of short-term and long-term hazard ratio with two-sample survival data. Biometrika 92(1):1â€“17

## Authorsâ€™ contributions

We construct the Bayesian test to test equality of hazard functions. It can be applicable in various circumstances including crossing hazards in spite of the non-proportionality. Thus it is better than log rank test having limitation in the case of non-proportionality. Both authors read and approved the final manuscript.

### Acknowledgements

Leeâ€™s research was supported by a National Research Foundation of Korea (NRF) grant funded by the Korean government (No. 2012-005741).

### Competing interests

The authors declare that they have no competing interests.

## Author information

### Authors and Affiliations

### Corresponding author

## Appendix

### Appendix

### Implementation of the proposed test

The implementation of the proposed test was done by FORTRAN codes. In addition, we conducted other tests by R functions such as survMisc, survival, YPmodel, and PwrGSD. Base codes of the proposed test can be shown at http://github.com/s88012/B-hazards-test/, and all codes are available on request.

### Proof of \(\pi ( \eta = k | D_{1:n} ) = \pi (M_k | D_{1:n} )\) for \(k = 0, 1.\)

Using only the partial likelihood, we have

where \(\mathbf{0}\) means the function having only zeros, and

in MCMC algorithm. Thus two are equal when letting \(\pi ^B\) be the prior of (15) for \(M_1\) and \(\pi (M_0) =\pi (\eta =0) = 1/2.\) This completes the proof.\(\square\)

## Rights and permissions

**Open Access** This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

## About this article

### Cite this article

Kim, G., Lee, SW. Bayesian test for hazard ratio in survival analysis.
*SpringerPlus* **5**, 649 (2016). https://doi.org/10.1186/s40064-016-2210-9

Received:

Accepted:

Published:

DOI: https://doi.org/10.1186/s40064-016-2210-9