A study on the properties of sensitivity analysis in principal component regression and latent root regression

주성분회귀와 고유값회귀에 대한 감도분석의 성질에 대한 연구

  • Published : 2009.03.31

Abstract

In regression analysis, the ordinary least squares estimates of regression coefficients become poor, when the correlations among predictor variables are high. This phenomenon, which is called multicollinearity, causes serious problems in actual data analysis. To overcome this multicollinearity, many methods have been proposed. Ridge regression, shrinkage estimators and methods based on principal component analysis (PCA) such as principal component regression (PCR) and latent root regression (LRR). In the last decade, many statisticians discussed sensitivity analysis (SA) in ordinary multiple regression and same topic in PCR, LRR and logistic principal component regression (LPCR). In those methods PCA plays important role. Many statisticians discussed SA in PCA and related multivariate methods. We introduce the method of PCR and LRR. We also introduce the methods of SA in PCR and LRR, and discuss the properties of SA in PCR and LRR.

회귀분석에서 설명변수들 사이에 상관이 높으면 최소제곱추정법에서 구한 회귀계수들의 정도가 떨어진다. 다중공선성이라 불리는 이 현상은 실제 자료분석에서 심각한 문제를 야기시킨다. 이 다중공선성의 문제를 극복하기 위한 여러 가지 방법이 제안되었다. 능형회귀, 축소추정량 그리고 주성분분석에 기초한 주성분회귀와 고유값회귀등이 있다. 지난 수십 년간 많은 통계학자들은 일반적인 중 회귀에서 감도분석에 관해 연구하였으며, 주성분회귀, 고유값회귀와 로지스틱 주성분회귀에 대해서도 같은 주제로 연구하였다. 이 모든 방법에서 주성분분석은 중요한 역할을 하였다. 또한, 많은 통계학자들이 주성분분석과 관련된 다변량 방법에서 감도분석에 대해 연구를 하였다. 본 연구논문에서는 주성분회귀와 고유값회귀를 소개하고, 또한 주성분회귀와 고유값회귀에서 감도분석의 방법을 소개하고, 마지막으로 이들두방법에 대한 감도분석의 성질에 대해 논의하였다.

Keywords

References

  1. Chatterjee, S. and Price, B. (1977). Regression analysis by example, John Wiley & Sons, New York.
  2. Critchley, F. (1985). In uence in principal component analysis. Biometrika, 72, 627-636. https://doi.org/10.1093/biomet/72.3.627
  3. Hampel, F. R. (1974). The in uence curve and its role in robust estimation. Journal of the American Statistical Associations, 69, 383-393. https://doi.org/10.2307/2285666
  4. Jolliffe, I. T. (1986). Principal components analysis, Springer-Verlag, New York.
  5. Pack, P., Jolliffe, I. T. and Morgan, B. J. T. (1987). In uential observations in principal component analysis: A case study. Journal of Applied Statistics, 15, 37-50
  6. Pregibon, D. (1981). Logistic regression diagnostics. The Annals of Statistics, 9, 705-724. https://doi.org/10.1214/aos/1176345513
  7. Radhakrishnan, R. and Kshirsagar, A. M. (1981). In uence functions for certain parameters in multivariate analysis. Communication in Statistics: Theory and Methods, 10, 515-529. https://doi.org/10.1080/03610928108828055
  8. Shin, J. K. and Chang, D. J. (2003). Sensitivity analysis in principal component regression with quadratic approximation. Journal of Korean Data & Information Science Society, 14, 623-630.
  9. Shin, J. K. and Moon, S. H. (1999). Numerical investigations in choosing the number of principal component in principal component regression-case a. Journal of Korean Data & Information Science Society, 10, 163-172.
  10. Tanaka, Y. (1988). Sensitivity analysis in principal component analysis : In uence on the subspace spanned by principal components. Communication in Statistics: Theory and Methods, 17, 3157-3175. (Corrections, A 18(1989), 4305). https://doi.org/10.1080/03610928808829796
  11. Tanaka, Y. (1989). In uence functions related to eigenvalue problems which appear in multivariate methods. Communication in Statistics: Theory and Methods, 18, 3991-4010. https://doi.org/10.1080/03610928908830137
  12. Walker, E. and Birch, J. B. (1988). In uence measure in ridge regression. Technometrics, 30, 221-227. https://doi.org/10.2307/1270168
  13. Webster, J. T., Gunst, R. F. and Mason, R. L. (1974). Latent root regression analysis. Technometrics, 16, 513-522. https://doi.org/10.2307/1267602
  14. Williams, D. A. (1987). Generalized linear model diagnostics using the deviance and single case deletions. Applied Statistics, 36, 181-191. https://doi.org/10.2307/2347550