DOI QR코드

DOI QR Code

Smoothing Kaplan-Meier estimate using monotone support vector regression

단조 서포트벡터기계를 이용한 카플란-마이어 생존함수의 평활

  • 황창하 (단국대학교 정보통계학과) ;
  • 심주용 (인제대학교 데이터정보학과)
  • Received : 2012.08.30
  • Accepted : 2012.09.30
  • Published : 2012.11.30

Abstract

Support vector machine is known to be the very useful statistical method in classification and nonlinear function estimation. In this paper we propose a monotone support vector regression (SVR) for the estimation of monotonically decreasing function. The proposed monotone SVR is applied to smooth the Kaplan-Meier estimate of survival function. Experimental results are then presented which indicate the performance of the proposed monotone SVR using survival functions obtained by exponential distribution.

서포트벡터 기계는 분류 및 비선형 함수추정에서 유용하게 사용되고 있는 통계적 기법이다. 본 논문에서는 두 개의 입력변수와 회귀함수의 단조 관계를 이용하여 단조 서포트벡터기계를 제안하고, Kaplan-Meier의 방법에 의해서 생존함수의 추정값이 주어진 경우 제안된 방법을 이용하여 생존 함수를 평활하는 방법 또한 제안한다. 모의실험에서는 실제 생존함수를 이용하여 Kaplan-Meier의 방법에 의한 생존함수의 추정값과의 성능을 비교함으로써 제안된 방법의 우수성을 보이기로 한다.

Keywords

References

  1. Buckley, J. and James, I. (1979). Linear regression with censored data. Biometrika, 66, 429-436. https://doi.org/10.1093/biomet/66.3.429
  2. Chakravarti, I. M., Laha, R. G. and Roy, J. (1967). Handbook of methods of applied statistics, John Wiley and Sons, New York.
  3. Cox, D. R. (1972). Regression models and life tables. Journal of the Royal Statistical Society, 34, 187-202.
  4. Cristianini, N. and Shawe-Taylor, J. (2000). An introduction to support vector machines and other kernelbased learning methods, Cambridge University Press, Cambridge.
  5. Green, P. J. and Silverman, B. W. (1994). Nonparametric regression and generalized linear models, Chapman & Hall, London.
  6. Gunn, S. R. (1998). Support vector machines for classification and regression, Technical Report, Department of Electronics and Computer Science, Southampton University.
  7. Hwang, C. and Shim, J. (2010). Semiparametric support vector machine for accelerated failure time model. Journal of the Korean Data & Information Science Society, 21, 467-477.
  8. Hwang, C. and Shim, J. (2011). Cox proportional hazard model with L1 penalty. Journal of the Korean Data & Information Society, 22, 613-618.
  9. Jo, D. H., Shim, J. and Seok, K. H. (2010). Doubly penalized kernel method for heteroscedastic autoregressive data. Journal of the Korean Data & Information Science Society, 21, 155-162.
  10. Kalbfleisch, J. D. and Prentice, R. L. (1980). The statistical analysis of failure time data, John Wiley & Sons Inc., New York.
  11. Kaplan, E. L. and Meier, P. (1958). Nonparametric estimation from incomplete observations. Journal of American Statistical Association, 53, 457-481. https://doi.org/10.1080/01621459.1958.10501452
  12. Kuhn, H. W. and Tucker, A. W. (1951). Nonlinear programming. In Proceedings of 2nd Berkeley Symposium, University of California Press, Berkeley, 481-491.
  13. Kim, M., Park, H., Hwang, C. and Shim, J. (2008). Claims reserving via kernel machine. Journal of the Korean Data & Information Science Society, 19, 1419-1427.
  14. Mercer, J. (1909). Functions of positive and negative type and their connection with theory of integral equations. Philosophical Transactions of Royal Society A, 415-446.
  15. Miller, R. G. (1981). Survival analysis, Wiley, New York.
  16. Moulton, L. H. and Dibley, M. J. (1997). Multivariate time-to-event models for studies of recurrent childhood diseases. International Journal of Epidemiology, 26, 1334-1339. https://doi.org/10.1093/ije/26.6.1334
  17. Saunders, C., Gammerman, A. and Vovk, V. (1998). Ridge regression learning algorithm in dual variables. In Proceedings of 15th International Conference on Machine Learning, Madison, WI, 515-521.
  18. Scholkopf, B. and Smola, A. (2002). Learning with kernels-support vector machines, regularization, optimization and beyond, MIT Press, Cambridge, MA.
  19. Stablein, D. M., Carter, W. H. and Novak, J. W. (1981). Analysis of survival data with nonproportional hazard functions. Controlled Clinical Trials, 2, 149-159. https://doi.org/10.1016/0197-2456(81)90005-2
  20. Suykens, J. A. K. and Vandewalle, J. (1999). Least square support vector machine classifier. Neural Processing Letters, 9, 293-300. https://doi.org/10.1023/A:1018628609742
  21. Vapnik, V. N. (1995). The nature of statistical learning theory, Springer, New York.
  22. Vapnik, V. N. (1998). Statistical learning theory, John Wiley, New York.
  23. Wahba, G. (1990). Spline models for observational data, SIAM, Philadelphia.

Cited by

  1. Deal price model in Deal-or-No-Deal game vol.25, pp.4, 2014, https://doi.org/10.7465/jkdi.2014.25.4.697
  2. Partially linear support vector orthogonal quantile regression with measurement errors vol.26, pp.1, 2015, https://doi.org/10.7465/jkdi.2015.26.1.209
  3. Support vector quantile regression ensemble with bagging vol.25, pp.3, 2014, https://doi.org/10.7465/jkdi.2014.25.3.677
  4. Estimating small area proportions with kernel logistic regressions models vol.25, pp.4, 2014, https://doi.org/10.7465/jkdi.2014.25.4.941