DOI QR코드

DOI QR Code

Applications of response dimension reduction in large p-small n problems

  • Minjee Kim (Department of Statistics, Ewha Womans University) ;
  • Jae Keun Yoo (Department of Statistics, Ewha Womans University)
  • 투고 : 2024.01.04
  • 심사 : 2024.02.05
  • 발행 : 2024.03.31

초록

The goal of this paper is to show how multivariate regression analysis with high-dimensional responses is facilitated by the response dimension reduction. Multivariate regression, characterized by multi-dimensional response variables, is increasingly prevalent across diverse fields such as repeated measures, longitudinal studies, and functional data analysis. One of the key challenges in analyzing such data is managing the response dimensions, which can complicate the analysis due to an exponential increase in the number of parameters. Although response dimension reduction methods are developed, there is no practically useful illustration for various types of data such as so-called large p-small n data. This paper aims to fill this gap by showcasing how response dimension reduction can enhance the analysis of high-dimensional response data, thereby providing significant assistance to statistical practitioners and contributing to advancements in multiple scientific domains.

키워드

과제정보

For Jae Keun Yoo and Minjee Kim, this work was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Korean Ministry of Education (RS-2023-00240564 and RS-2023-00217022).

참고문헌

  1. Benjamini Y and Hochberg Y (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing, Journal of the Royal Statistical Society, Series B, 57, 289-300.
  2. Cook RD (2007). Fisher lecture: Dimension reduction in regression, Statistical Science, 22, 1-26.
  3. Cook RD, Li B, and Chiaromonte F (2010). Envelope models for parsimonious and efficient multivariate linear regression, Statistics Sinica, 20, 927-960.
  4. Cook RD and Setodji CM (2003). A model-free test for reduced rank in multivariate regression, Journal of the American Statistical Association, 98, 340-351.
  5. Cook RD and Weisberg S (1991). Sliced inverse regression for dimension reduction: Comment, Journal of the American Statistical Association, 86, 328-332.
  6. Hooper JW (1959). Simultaneous equations and canonical correlation theory, Econometrica: Journal of the Econometric Society, 27, 245-256.
  7. Hung H and Huang SY (2019). Sufficient dimension reduction via random-partitions for the large-p-small-n problem, Biometrics, 75, 245-255.
  8. Im Y, Gang H, and Yoo JK (2015). High-throughput data dimension reduction via seeded canonical correlation analysis, Journal of Chemometrics, 29, 193-199.
  9. Kim B, Im Y, and Yoo JK (2021). SEEDCCA: An integrated R-package for canonical correlation analysis and partial least squares, The R Journal, 13, 7-20.
  10. Ko S and Yoo JK (2022). Projective resampling estimation of informative predictor subspace for multivariate regression, Journal of the Korean Statistical Society, 51, 1117-1131.
  11. Lee K, Choi Y, Um HY, and Yoo JK (2019). On fused dimension reduction in multivariate regression, Chemometrics and Intelligent Laboratory Systems, 193, 103828.
  12. Lenth RV (2016). Least-squares means: The R package lsmeans, Journal of Statistical Software, 69, 1-33.
  13. Li B and Wang S (2007). On directional regression for dimension reduction, Journal of the American Statistical Association, 102, 997-1008.
  14. Li B, Zha H, and Chiaromonte F (2005). Contour regression: A general approach to dimension reduction, The Annals of Statistics, 33, 1580-1616.
  15. Li K-C (1991). Sliced inverse regression for dimension reduction, Journal of the American Statistical Association, 86, 316-327.
  16. Li X, Gill R, Cooper NG, Yoo JK, and Datta S (2011). Modeling microRNA-mRNA interactions using PLS regression in human colon cancer, BMC Medical Genomics, 4, 1-15.
  17. Liland KH, Mevik B-H, Wehrens R, and Hiemstr P (2021). PLS: Partial Least Squares and Principal Component Regression. R package version 2.8.0. Available from: https://cran.r-project.org/web/packages/pls/pls.pdf
  18. Schafer J, Opgen-Rhein R, Zuber V, Ahdesmaki M, Silva APD, and Strimmer K (2017). corpcor: Efficient Estimation of Covariance and (Partial) Correlation. R package version 1.6.9. Available from: https://cran.r-project.org/web/packages/corpcor/corpcor.pdf
  19. Yin X and Hilafu H (2015). Sequential sufficient dimension reduction for large p, small n problems, Journal of the Royal Statistical Society, Series B, 77, 879-892.
  20. Yoo JK (2023). A theoretical view of the envelope model for multivariate linear regression as response dimension reduction, Journal of the Korean Statistical Society, 42, 143-148.
  21. Yoo JK (2018). Response dimension reduction: Model-based approach, Statistics, 52, 409-425.
  22. Yoo JK (2019). Unstructured principal fitted response reduction in multivariate regression, Journal of the Korean Statistical Society, 48, 561-567.
  23. Yoo JK and Cook RD (2007). Optimal sufficient dimension reduction for the conditional mean in multivariate regression, Biometrika, 94, 231-242.
  24. Yoo JK and Cook RD (2008). Response dimension reduction for the conditional mean in multivariate regression, Computational Statistics and Data Analysis, 53, 334-343.
  25. Yoo JK and Im Y (2014). Multivariate seeded dimension reduction, Journal of the Korean Statistical Society, 43, 559-566.
  26. Yoo JK, Lee K, and Wu S (2010). On the extension of sliced average variance estimation to multivariate regression, Statistical Methods & Applications, 19, 529-540.