Interaction-Based Feature Selection Technique Using Fuzzy Discretization and Class Association Rule Mining for Breast Cancer Classification
DOI:
https://doi.org/10.11113/mjfas.v21n2.3787Keywords:
Breast cancer classification, class association rule, feature selection, feature interaction, fuzzy discretization.Abstract
Breast cancer is a leading global cause of cancer-related deaths highlighting the need for an accurate diagnostic system. Up to now, computer-aided diagnosis (CAD) system plays an essential role in supporting pathologists with prompt and accurate classification. Feature selection within the CAD system is crucial as it helps identify the most relevant data for subsequent classification tasks. This paper proposed a novel method that focuses on fuzzy discretization in handling continuous features and selecting relevant and interactive features while eliminating redundancy using Class Association Rule Mining (CARM). The proposed method, FD-CARI was compared with other feature selection techniques including CFS, FCBF, Consistency, Relief-F, and mRMR using five different machine learning classifiers such as Decision Tree (DT), Random Forest (RF), Logistic Regression (LR), Naive Bayes (NB), and Support Vector Machine (SVM). Performance evaluation metrics such as Accuracy (ACC), Sensitivity (SEN), Specificity (SPE), Precision, F1-Score, and AUC were then utilized. Results: The experimental findings consistently showed that the proposed method achieved high performance with an ACC of 96.21%, SEN of 94.26%, and SPE of 97.38% on the SVM classifiers, and an ACC of 96.05%, SEN of 93.82%, and SPE of 97.38% on the LR classifiers. It demonstrated similar effectiveness to Relief-F for DT and RF classifiers. However, FCBF achieved the highest performance on NB with ACC, SEN, and SPE values of 96.21%, 92%, and 96.60%, respectively. The proposed method efficiently selects relevant and interactive features while enabling classifiers to achieve better classification accuracy.
References
Sung, H., et al. (2021). Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA: A Cancer Journal for Clinicians, 71(3), 209–249. https://doi.org/10.3322/caac.21660
Siegel, R. L., Giaquinto, A. N., & Ahmedin, J. (2024). Cancer statistics, 2024. https://doi.org/10.3322/caac.21820
Yadav, R. K., Singh, P., & Kashtriya, P. (2023). Diagnosis of breast cancer using machine learning techniques: A survey. Procedia Computer Science, 218, 1434–1443. https://doi.org/10.1016/j.procs.2023.01.122
Oskouei, R. J., Kor, N. M., & Maleki, S. A. (2017). Data mining and medical world: Breast cancers’ diagnosis, treatment, prognosis, and challenges. American Journal of Cancer Research, 7(3), 610–627.
Khamparia, A., et al. (2021). Diagnosis of breast cancer based on modern mammography using hybrid transfer learning. Multidimensional Systems and Signal Processing, 32(2), 747–765. https://doi.org/10.1007/s11045-020-00756-7
Kuhl, C. K., et al. (2005). Mammography, breast ultrasound, and magnetic resonance imaging for surveillance of women at high familial risk for breast cancer. Journal of Clinical Oncology, 23(33), 8469–8476. https://doi.org/10.1200/JCO.2004.00.4960
Pulumati, A., Pulumati, A., Dwarakanath, B. S., Verma, A., & Papineni, R. V. L. (2023). Technological advancements in cancer diagnostics: Improvements and limitations. Cancer Reports, 6(2), 1–17. https://doi.org/10.1002/cnr2.1764
Drukker, K., Sennett, C. A., & Giger, M. L. (2009). Automated method for improving system performance of computer-aided diagnosis in breast ultrasound. IEEE Transactions on Medical Imaging, 28(1), 122–128. https://doi.org/10.1109/TMI.2008.928178
Li, J., et al. (2017). Feature selection: A data perspective. ACM Computing Surveys, 50(6). https://doi.org/10.1145/3136625
Dash, M., & Liu, H. (1997). Feature selection for classification. Intelligent Data Analysis, 1(3), 131–156. https://doi.org/10.3233/IDA-1997-1302
Vergara, J. R., & Estévez, P. A. (2014). A review of feature selection methods based on mutual information. Neural Computing and Applications, 24(1), 175–186. https://doi.org/10.1007/s00521-013-1368-0
Jakulin, A., & Bratko, I. (2004). Testing the significance of attribute interactions. In Proceedings of the Twenty-First International Conference on Machine Learning (pp. 409–416). https://doi.org/10.1145/1015330.1015377
Deisy, C., Baskar, S., Ramraj, N., Koori, J. S., & Jeevanandam, P. (2010). A novel information theoretic-interact algorithm (IT-IN) for feature selection using three machine learning algorithms. Expert Systems with Applications, 37(12), 7589–7597. https://doi.org/10.1016/j.eswa.2010.04.084
Gu, X., Guo, J., Wei, H., & He, Y. (2020). Spatial-domain steganalytic feature selection based on three-way interaction information and KS test. Soft Computing, 24(1), 333–340. https://doi.org/10.1007/s00500-019-03910-x
Zhao, Z., & Liu, H. (2007). Searching for interacting features. In Proceedings of the International Joint Conference on Artificial Intelligence (pp. 1156–1161).
Zeng, Z., Zhang, H., Zhang, R., & Yin, C. (2015). A novel feature selection method considering feature interaction. Pattern Recognition, 48(8), 2656–2666. https://doi.org/10.1016/j.patcog.2015.02.025
Wang, L., Jiang, S., & Jiang, S. (2021). A feature selection method via analysis of relevance, redundancy, and interaction. Expert Systems with Applications, 183, 115365. https://doi.org/10.1016/j.eswa.2021.115365
Wang, G., & Song, Q. (2012). Selecting feature subset via constraint association rules. In Lecture Notes in Computer Science: Vol. 7301. Advances in Knowledge Discovery and Data Mining (pp. 304–321). https://doi.org/10.1007/978-3-642-30220-6_26
Šikonja, M.-R., & Kononenko, I. (2003). Theoretical and empirical analysis of ReliefF and RReliefF. Machine Learning, 53, 23–69.
Peng, H., Long, F., & Ding, C. (2005). Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on Pattern Analysis and Machine Intelligence, 27(8), 1226–1238. https://doi.org/10.1109/TPAMI.2005.159
Wang, G., & Song, Q. (2013). A novel feature subset selection algorithm based on association rule mining. Intelligent Data Analysis, 17(5), 803–835. https://doi.org/10.3233/IDA-130608
Kira, K., & Rendell, L. A. (1992). The feature selection problem: Traditional methods and a new algorithm. In Proceedings of the National Conference on Artificial Intelligence (pp. 129–134). AAAI Press. https://doi.org/10.1201/9781003402848-6
Hall, M. A., & Smith, L. A. (1995). Feature selection for machine learning: Comparing a correlation-based filter approach to the wrapper. In Proceedings of the Florida Artificial Intelligence Research Society Conference (pp. 235–239).
Yu, L., & Liu, H. (2003). Feature selection for high-dimensional data: A fast correlation-based filter solution. In Proceedings of the Twentieth International Conference on Machine Learning (pp. 856–863).
Dash, M., & Liu, H. (2003). Consistency-based search in feature selection. Artificial Intelligence, 151(1–2), 155–176. https://doi.org/10.1016/S0004-3702(03)00079-1
Das, S., & Nath, B. (2008). Dimensionality reduction using association rule mining. In Proceedings of the IEEE Region 10 Colloquium and 3rd International Conference on Industrial and Information Systems (pp. 1–6). https://doi.org/10.1109/ICIINFS.2008.4798351
Xie, J., Wu, J., & Qian, Q. (2009). Feature selection algorithm based on association rules mining method. In Proceedings of the 8th IEEE/ACIS International Conference on Computer and Information Science (pp. 357–362). https://doi.org/10.1109/ICIS.2009.103
Karabatak, M., & Ince, M. C. (2009). A new feature selection method based on association rules for diagnosis of erythemato-squamous diseases. Expert Systems with Applications, 36(10), 12500–12505. https://doi.org/10.1016/j.eswa.2009.04.073
Sheikhan, M., Rad, M. S., & Shirazi, H. M. (2011). Application of fuzzy association rules-based feature selection and fuzzy ARTMAP to intrusion detection. Majlesi Journal of Electrical Engineering, 5(4). https://doi.org/10.1109/ICSMC.2006.385078
Waseem, S., Salman, A., & Muhammad, A. K. (2013). Feature subset selection using association rule mining and JRip classifier. International Journal of Physical Sciences, 8(18), 885–896. https://doi.org/10.5897/ijps2013.3842
Wang, L., & Guo, H. (2014). Feature selection based on fuzzy clustering analysis and association rule mining for soft-sensor. In Proceedings of the 33rd Chinese Control Conference (pp. 5162–5166). https://doi.org/10.1109/ChiCC.2014.6895819
Harikumar, S., Dilipkumar, D. U., & Kaimal, M. R. (2017). Efficient attribute selection strategies for association rule mining in high dimensional data. International Journal of Computational Science and Engineering, 15(3–4), 201–213. https://doi.org/10.1504/IJCSE.2017.087416
Li, Y., et al. (2019). Association rule-based feature mining for automated fault diagnosis of rolling bearing. Shock and Vibration, 2019. https://doi.org/10.1155/2019/1518246
Lin, Q., & Gao, C. (2023). Discovering categorical main and interaction effects based on association rule mining. IEEE Transactions on Knowledge and Data Engineering, 35(2), 1379–1390. https://doi.org/10.1109/TKDE.2021.3087343
Liewlom, P. (2021). Class-association-rules pruning by the profitability-of-interestingness measure: Case study of an imbalanced class ratio in a breast cancer dataset. Journal of Advances in Information Technology, 12(3), 246–252. https://doi.org/10.12720/jait.12.3.246-252
Farghaly, H. M., & Abd El-Hafeez, T. (2023). A high-quality feature selection method based on frequent and correlated items for text classification. Soft Computing, 27(16), 11259–11274. https://doi.org/10.1007/s00500-023-08587-x
Zhang, L. (2021). A feature selection algorithm integrating maximum classification information and minimum interaction feature dependency information. Computational Intelligence and Neuroscience, 2021. https://doi.org/10.1155/2021/3569632
Lv, Y., Lin, Y., Chen, X., Wang, C., & Li, S. (2021). Feature interaction based online streaming feature selection via buffer mechanism. Concurrency and Computation: Practice and Experience, 33(21), e6435. https://doi.org/10.1002/cpe.6435
Liu, J., Yang, S., Zhang, H., Sun, Z., & Du, J. (2023). Online multi-label streaming feature selection based on label group correlation and feature interaction. Entropy, 25(7), Article 1071. https://doi.org/10.3390/e25071071
Kianmehr, K., Alshalalfa, M., & Alhajj, R. (2010). Fuzzy clustering-based discretization for gene expression classification. Knowledge and Information Systems, 24(3), 441—465. https://doi.org/10.1007/s10115-009-0214-2
Zadeh, L. A. (1965). Fuzzy sets. Information and Control, 8, 338–353.
Agrawal, R., Imieliński, T., & Swami, A. (1993). Mining association rules between sets of items in large databases. ACM SIGMOD Record, 22(2), 207–216. https://doi.org/10.1145/170036.170072
Akbas, K. E., et al. (2022). Assessment of association rule mining using interest measures on the gene data. Medical Records, 4(3), 286–292. https://doi.org/10.37990/medr.1088631
Anusha, P. V., Anuradha, C., Chandra Murty, P. S. R., & Kiran, C. S. (2019). Detecting outliers in high dimensional data sets using Z-score methodology. International Journal of Innovative Technology and Exploring Engineering, 9(1), 48–53. https://doi.org/10.35940/ijitee.A3910.119119
Prasetiyowati, M. I., Maulidevi, N. U., & Surendro, K. (2021). Determining threshold value on information gain feature selection to increase speed and prediction accuracy of random forest. Journal of Big Data, 8(1). https://doi.org/10.1186/s40537-021-00472-4
Damtew, Y. G., Chen, H., & Yuan, Z. (2023). Heterogeneous ensemble feature selection for network intrusion detection system. International Journal of Computational Intelligence Systems, 16(1). https://doi.org/10.1007/s44196-022-00174-6
Ishibuchi, H., Yamamoto, T., & Nakashima, T. (2001). Fuzzy data mining: Effect of fuzzy discretization. In Proceedings of the IEEE International Conference on Data Mining (pp. 241–248). https://doi.org/10.1109/icdm.2001.989525
Kianmehr, K., Alshalalfa, M., & Alhajj, R. (2008). Effectiveness of fuzzy discretization for class association rule-based classification. In Lecture Notes in Computer Science: Vol. 4994. Advances in Artificial Intelligence (pp. 298–308). https://doi.org/10.1007/978-3-540-68123-6_33
Shanmugapriya, M., Nehemiah, H. K., Bhuvaneswaran, R. S., Arputharaj, K., & Sweetlin, J. D. (2017). Fuzzy discretization based classification of medical data. Research Journal of Applied Sciences, Engineering and Technology, 14(8), 291–298. https://doi.org/10.19026/rjaset.14.4953
Quinlan, J. R. (1986). Induction of decision trees. Machine Learning, 1(1), 81–106. https://doi.org/10.1023/A:1022643204877
Breiman, L. (2001). Random forests. Machine Learning, 45, 5–32.
Han, S., & Kim, H. (2019). On the optimal size of candidate feature set in random forest. Applied Sciences, 9(5). https://doi.org/10.3390/app9050898
Shobha, G., & Rangaswamy, S. (2018). Machine learning. In Handbook of Statistics: Vol. 38. Advances in Machine Learning and Data Science (1st ed.). Elsevier B.V. https://doi.org/10.1016/bs.host.2018.07.004
Berkson, J. (1944). Application of the logistic function to bio-assay. Journal of the American Statistical Association, 39(227), 357–365.
Cortes, C., & Vapnik, V. (1995). Support-vector networks. Machine Learning, 297(20), 273–297.
Tharwat, A. (2018). Classification assessment methods. Applied Computing and Informatics, 17(1), 168–192. https://doi.org/10.1016/j.aci.2018.08.003
Wolberg, W., Mangasarian, O., Street, N., & Street, W. (1995). Breast Cancer Wisconsin (Diagnostic). UCI Machine Learning Repository. https://doi.org/10.24432/C5DW2B
Garner, S. R. (1995). WEKA: The Waikato Environment for Knowledge Analysis. In Proceedings of the New Zealand Computer Science Research Students Conference (pp. 57–64).
The MathWorks. (2023). MATLAB (Version 9.14.0 R2023a) [Software]. Natick, MA: The MathWorks Inc.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 Shahiratul, A. Karim, Ummul, H. Mohammad, Puteri, N. E. Nohuddin

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.