A Feature Selection Method for High-Dimensional Medical Data Based on Adaptive Redundancy Penalty
DOI:
https://doi.org/10.54097/gp12fc89Keywords:
High-dimensional Medical Data, Feature Selection, Joint Mutual Information, Adaptive Redundancy Penalty, Classification AccuracyAbstract
High-dimensional medical data are often characterized by extremely high feature dimensionality, limited sample sizes, substantial redundant information, and strong noise interference. To address these challenges, this paper proposes a filter-based feature selection method, termed ARMR (Adaptive Redundancy-based Minimum Redundancy Maximum Relevance Feature Selection). The proposed method first employs mutual information to select initial features from the original feature set. It then uses joint mutual information to characterize the joint discriminative capability of candidate features and selected features with respect to class labels. Furthermore, an adaptive penalty factor, constructed from conditional mutual information and three-way interaction information, is introduced to dynamically weight the redundancy term, thereby enabling a more flexible balance between feature complementarity and redundancy suppression. To evaluate the effectiveness of the proposed method, experiments were conducted on six benchmark datasets and four medical datasets, with comparisons against several representative and recently developed methods, including CFR, DCSF, JMI, MRMD, and mRMR. Two classifiers, namely SVM and NB, were employed for performance assessment. The results show that the proposed method achieves superior overall performance. In particular, it attains the highest average classification accuracy of 84.58% with SVM and 76.36% with NB. Further analysis based on classification accuracy curves under different feature subset sizes, together with boxplots of F1-score and AUC, demonstrates that the proposed method exhibits strong performance in terms of classification accuracy, stability, and classifier adaptability.
Downloads
References
[1] Y. Kang, D. Zheng, H. Wang, Y. Peng, and S. Zhou, ‘Dual-metric guided multi-strategy hybrid optimization for feature selection on high-dimensional medical data’, Swarm Evol. Comput., vol. 98, p. 102118, Oct. 2025, doi: 10.1016/j.swevo.2025.102118.
[2] M. Wang, A. A. Heidari, and H. Chen, ‘A multi-objective evolutionary algorithm with decomposition and the information feedback for high-dimensional medical data’, Appl. Soft Comput., vol. 136, p. 110102, 2023, doi: https://doi.org/10.1016/j.asoc.2023.110102.
[3] D. Dhinakaran, L. Srinivasan, S. E. Raja, K. Valarmathi, and M. G. Nayagam, ‘Synergistic feature selection and distributed classification framework for high-dimensional medical data analysis’, MethodsX, vol. 14, p. 103219, 2025, doi: https://doi.org/10.1016/j.mex.2025.103219.
[4] X. Zhang and J. Wang, ‘A noise-robust feature selection using KNN and weighted fuzzy rough sets for imbalanced multi-scale data’, Appl. Soft Comput., vol. 194, p. 114964, 2026, doi: https://doi.org/10.1016/j.asoc.2026.114964.
[5] H. Ju et al., ‘Distributed multi-label feature selection via feature-label information granulation’, Inf. Sci., vol. 742, p. 123334, 2026, doi: https://doi.org/10.1016/j.ins.2026.123334.
[6] F. Deng, L. Zhao, N. Yu, Y. Lin, and L. Zhang, ‘Union With Recursive Feature Elimination: A Feature Selection Framework to Improve the Classification Performance of Multicategory Causes of Death in Colorectal Cancer’, Lab. Invest., vol. 104, no. 3, p. 100320, 2024, doi: https://doi.org/10.1016/j.labinv.2023.100320.
[7] S.-W. Lin, Z.-J. Lee, S.-C. Chen, and T.-Y. Tseng, ‘Parameter determination of support vector machine and feature selection using simulated annealing approach’, Appl. Soft Comput., vol. 8, no. 4, pp. 1505–1512, 2008, doi: https://doi.org/10.1016/j.asoc.2007.10.012.
[8] L. Gao, L. Li, and X. Chen, ‘Sparse learning of interval type-2 fuzzy model with embedded feature and rule selection’, Neurocomputing, vol. 677, p. 133102, 2026, doi: https://doi.org/10.1016/j.neucom.2026.133102.
[9] T. Dupré la Tour, M. Eickenberg, A. O. Nunez-Elizalde, and J. L. Gallant, ‘Feature-space selection with banded ridge regression’, NeuroImage, vol. 264, p. 119728, Dec. 2022, doi: 10.1016/j.neuroimage.2022.119728.
[10] H. Lyu, M. Wan, J. Han, R. Liu, and C. Wang, ‘A filter feature selection method based on the Maximal Information Coefficient and Gram-Schmidt Orthogonalization for biomedical data mining’, Comput. Biol. Med., vol. 89, pp. 264–274, 2017, doi: https://doi.org/10.1016/j.compbiomed.2017.08.021.
[11] H. Gong, Y. Li, J. Zhang, B. Zhang, and X. Wang, ‘A new filter feature selection algorithm for classification task by ensembling pearson correlation coefficient and mutual information’, Eng. Appl. Artif. Intell., vol. 131, p. 107865, May 2024, doi: 10.1016/j.engappai.2024.107865.
[12] U. Agrawal, V. Rohatgi, and R. Katarya, ‘Normalized Mutual Information-based equilibrium optimizer with chaotic maps for wrapper-filter feature selection’, Expert Syst. Appl., vol. 207, p. 118107, 2022, doi: https://doi.org/10.1016/j.eswa.2022.118107.
[13] W. Gao, L. Hu, P. Zhang, and J. He, ‘Feature selection considering the composition of feature relevancy’, Pattern Recognit. Lett., vol. 112, pp. 70–74, Sep. 2018, doi: 10.1016/j.patrec.2018.06.005.
[14] W. Gao, L. Hu, and P. Zhang, ‘Class-specific mutual information variation for feature selection’, Pattern Recognit., vol. 79, pp. 328–339, Jul. 2018, doi: 10.1016/j.patcog.2018.02.020.
[15] ‘Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy’. Accessed: Dec. 30, 2024. [Online]. Available: https://ieeexplore.ieee.org/document/1453511
[16] ‘Data visualization and feature selection | Proceedings of the 12th International Conference on Neural Information Processing Systems’. Accessed: Dec. 30, 2024. [Online]. Available: https://dl.acm.org/doi/10.5555/3009657.3009755
[17] M. Bennasar, Y. Hicks, and R. Setchi, ‘Feature selection using Joint Mutual Information Maximisation’, Expert Syst. Appl., vol. 42, no. 22, pp. 8520–8532, Dec. 2015, doi: 10.1016/j.eswa.2015.07.007.
[18] Z. Zeng, H. Zhang, R. Zhang, and C. Yin, ‘A novel feature selection method considering feature interaction’, Pattern Recognit., vol. 48, no. 8, pp. 2656–2666, Aug. 2015, doi: 10.1016/j.patcog.2015.02.025.
[19] W. Gao, L. Hu, and P. Zhang, ‘Feature redundancy term variation for mutual information-based feature selection’, Appl. Intell., vol. 50, no. 4, pp. 1272–1288, Apr. 2020, doi: 10.1007/s10489-019-01597-z.
[20] ‘Feature Selection: A Data Perspective: ACM Computing Surveys: Vol 50, No 6’. Accessed: Dec. 30, 2024. [Online]. Available: https://dl.acm.org/doi/10.1145/3136625
[21] Dias Daniel, Peres, Sarajane and H. Bscaro, ‘Libras Movement’. 2009.
[22] A. Tsanas, ‘LSVT Voice Rehabilitation’. 2014.
Downloads
Published
Issue
Section
License
Copyright (c) 2026 Journal of Computer Science and Artificial Intelligence

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.








