TY - GEN
T1 - Noise Reduced Common PCA for High-Dimensional, Low-Sample Size Multi-View Data
AU - Hasegawa, Hiroki
AU - Kawamura, Homura
AU - Shin, Ryota
AU - Yata, Kazuyoshi
AU - Okada, Yukihiko
AU - Kunimatsu, Jun
N1 - Publisher Copyright:
© 2024, Avestia Publishing. All rights reserved.
PY - 2024
Y1 - 2024
N2 - High-Dimensional Low-Sample Size (HDLSS) data pose significant challenges in fields like medicine and neuroscience. Traditional principal component analysis (PCA) often fails under these conditions, leading to unstable eigenvalue estimation. This study introduces Noise Reduced-Common Principal Component Analysis (NR-CPCA), a method that combines Common Principal Component Analysis (CPCA) with a noise reduction technique to enhance eigenvalue stability and reliability in HDLSS data. By comparing eigenvalue estimations from NR-CPCA and traditional CPCA across various dimensions (1000, 2000, 3000) and sample sizes (10 to 120), we demonstrate that NR-CPCA mitigates noise effects more effectively, ensuring stable principal component selection. Simulation results confirm that NR-CPCA reduces variability in eigenvalue estimation, making it a valuable tool for dimensionality reduction in multi-view data. Despite limitations in simulation-based validation, NR-CPCA shows promise for real-world applications in data-intensive fields. Future research should focus on refining this method and applying it to diverse datasets to fully realize its potential. NR-CPCA provides an important advancement for researchers dealing with HDLSS data, promoting more accurate analysis and contributing to progress in data science, biology, and neuroscience.
AB - High-Dimensional Low-Sample Size (HDLSS) data pose significant challenges in fields like medicine and neuroscience. Traditional principal component analysis (PCA) often fails under these conditions, leading to unstable eigenvalue estimation. This study introduces Noise Reduced-Common Principal Component Analysis (NR-CPCA), a method that combines Common Principal Component Analysis (CPCA) with a noise reduction technique to enhance eigenvalue stability and reliability in HDLSS data. By comparing eigenvalue estimations from NR-CPCA and traditional CPCA across various dimensions (1000, 2000, 3000) and sample sizes (10 to 120), we demonstrate that NR-CPCA mitigates noise effects more effectively, ensuring stable principal component selection. Simulation results confirm that NR-CPCA reduces variability in eigenvalue estimation, making it a valuable tool for dimensionality reduction in multi-view data. Despite limitations in simulation-based validation, NR-CPCA shows promise for real-world applications in data-intensive fields. Future research should focus on refining this method and applying it to diverse datasets to fully realize its potential. NR-CPCA provides an important advancement for researchers dealing with HDLSS data, promoting more accurate analysis and contributing to progress in data science, biology, and neuroscience.
KW - Big Data Analytics
KW - Common Principal Component Analysis
KW - Dimensionality reduction
KW - High-Dimensional Data Analysis
KW - Low-sample Size data
KW - Noise Reduction Techniques
UR - http://www.scopus.com/inward/record.url?scp=85205722832&partnerID=8YFLogxK
U2 - 10.11159/icsta24.173
DO - 10.11159/icsta24.173
M3 - Conference contribution
AN - SCOPUS:85205722832
SN - 9781990800429
T3 - Proceedings of the International Conference on Statistics
BT - Proceedings of the 6th International Conference on Statistics
A2 - Samia, Noelle
PB - Avestia Publishing
T2 - 6th International Conference on Statistics: Theory and Applications, ICSTA 2024
Y2 - 19 August 2024 through 21 August 2024
ER -