439Accesses
1Citation
1Altmetric
Abstract
Dimension reduction methods is effective for tackling the complexity of models learning from high-dimensional data. Usually, they are presented as a black box, where the reduction process is unknown to the practitioners. Yet, this process potentially transmits a reliable framework for understanding the regularities behind the data. Furthermore, in some applications contexts, the available datasets are presented with a huge lack of records. Therefore, the classical and the deep dimension reduction methods often fall in the over-fitting trap. We propose to tackle these challenges under the Bayesian network paradigm associated with the latent variables learning. We propose an interpretable framework for learning a reduced dimension while ensuring the effectiveness against the curse of dimensionality. Our exhaustive experimental results, over benchmark datasets, prove that our dimension reduction algorithm yields a user-friendly model that not only minimizes the information loss due to the reduction process, but also escapes data overfitting due to the lack of records.
This is a preview of subscription content,log in via an institution to check access.
Access this article
Subscribe and save
- Get 10 units per month
- Download Article/Chapter or eBook
- 1 Unit = 1 Article or 1 Chapter
- Cancel anytime
Buy Now
Price includes VAT (Japan)
Instant access to the full article PDF.







Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Notes
These datasets are downloadable at:https://archive.ics.uci.edu/ml/datasets.php.
IBNA code is available online via this link:https://github.com/HasnaNjah/IBNA.
References
Oseledets IV, Tyrtyshnikov EE (2009) Breaking the curse of dimensionality, or how to use SVD in many dimensions. SIAM J Sci Comput 31(5):3744–3759
Scott DW (2008) The curse of dimensionality and dimension reduction. Multivar Density Estim Theory Pract Visual 1:195–217
Friedman N, Geiger D, Goldszmidt M (1997) Bayesian network classifiers. Mach Learn 29(2):131–163
Geiger D, Verma T, Pearl J (1990) D-separation: from theorems to algorithms. Mach Intell Pattern Recogn 10:139–148
Hausman DM, Woodward J (1999) Independence, invariance and the causal Markov condition. Br J Philos Sci 50(4):521–583
Fodor IK (2002) A survey of dimension reduction techniques.:Technical Report UCRL-ID-148494, Lawrence Livermore National Laboratory
Jolliffe IT (2002) Principal component analysis for special types of data. Springer, New York, pp 338–372
Dumais ST (2004) Latent semantic analysis. Ann Rev Inf Sci Technol 38(1):188–230
Spearman C (1904) The proof and measurement of association between two things. Am J Psychol 15(1):72–101
Wang X, Guo B, Shen Y, Zhou C, Duan X (2019) Input feature selection method based on feature set equivalence and mutual information gain maximization. IEEE Access 7:151525–151538
Mahdavi S et al. (2019) A knowledge discovery of relationships among dataset entities using optimum hierarchical clustering by de algorithm. In: 2019 IEEE congress on evolutionary computation (CEC). IEEE
Chowdhury S et al (2017) Botnet detection using graph-based feature clustering. J Big Data 4(1):1–23
Gandhi SS, Prabhune SS (2017) Overview of feature subset selection algorithm for high dimensional data. In: 2017 International conference on inventive systems and control (ICISC). IEEE
Saracco J, Chavent M, Kuentz V (2010) Clustering of categorical variables around latent variables. No. 2010–02. Groupe de Recherche en Economie Théorique et Appliquée (GREThA)
Chavent M et al (2011) ClustOfVar: an R package for the clustering of variables. arXiv preprintarXiv:1112.0295
Tran B, Xue B, Zhang M (2017) Using feature clustering for GP-based feature construction on high-dimensional data. European conference on genetic programming. Springer, Cham, pp 210–226
Butterworth R, Piatetsky-Shapiro G, Simovici DA (2005) On feature selection through clustering. In: Fifth IEEE International conference on data mining (ICDM'05). IEEE
Hinton GE, Salakhutdinov RR (2006) Reducing the dimensionality of data with neural networks. Science 313(5786):504–507
Kiarashinejad Y, Abdollahramezani S, Adibi A (2020) Deep learning approach based on dimensionality reduction for designing electromagnetic nanostructures. Comput Mater 6(1):1–12
Xu G et al (2019) Bearing fault diagnosis method based on deep convolutional neural network and random forest ensemble learning. Sensors 19(5):1088
Bouhamed H, Masmoudi A, Lecroq T, Rebaï A (2012) A new learning structure heuristic of Bayesian networks from data. International workshop on machine learning and data mining in pattern recognition. Springer, Berlin, Heidelberg, pp 183–197
Chickering DM (1996) Learning Bayesian networks is NP-complete. Learning from data. Springer, New York, pp 121–130
Yu K, Wu X, Ding W, Mu Y, Wang H (2016) Markov blanket feature selection using representative sets. IEEE Trans Neural Netw Learn Syst 28(11):2775–2788
Cinicioglu EN, Yenilmez T (2016) Determination of variables for a Bayesian network and the most precious one. International conference on information processing and management of uncertainty in knowledge-based systems. Springer, Cham, pp 313–325
Inza I, Larrañaga P, Etxeberria R, Sierra B (2000) Feature subset selection by Bayesian network-based optimization. Artif Intell 123(1–2):157–184
Kuschner KW, Malyarenko DI, Cooke WE, Cazares LH, Semmes OJ, Tracy ER (2010) A Bayesian network approach to feature selection in mass spectrometry data. BMC Bioinform 11(1):1–10
Mourad R, Sinoquet C, Leray P (2011) A hierarchical Bayesian network approach for linkage disequilibrium modeling and data-dimensionality reduction prior to genome-wide association studies. BMC Bioinform 12(1):16
Wang Y, Zhang NL, Chen T (2008) Latent tree models and approximate inference in Bayesian networks. J Artif Intell Res 32:879–900
Zhang Y, Ji L (2009) Clustering of SNPs by a structural EM algorithm. In 2009 International joint conference on bioinformatics, systems biology and intelligent computing, pp. 147–150. IEEE
Hwang KB, Kim BH, Zhang BT (2006) Learning hierarchical Bayesian networks for large-scale data analysis. International conference on neural information processing. Springer, Berlin, Heidelberg, pp 670–679
Zhang NL, Kocka T (2004) Effective dimensions of hierarchical latent class models. J Artif Intell Res (JAIR) 21:1–17
Mourad R et al (2013) A Survey on latent tree models and applications. J Artif Intell Res (JAIR) 47:157–203
Witten IH, Frank E (2002) Data mining: practical machine learning tools and techniques with Java implementations. ACM SIGMOD Rec 31(1):76–77
Njah H, Jamoussi S, Mahdi W, Masmoudi A (2015) A new equilibrium criterion for learning the cardinality of latent variables. In: 2015 IEEE 27th International conference on tools with artificial intelligence (ICTAI). IEEE
Bishop CM, Nasrabadi NM (2006) Pattern recognition and machine learning. Springer, New York
Dougherty J, Kohavi R, Sahami M (1995) Supervised and unsupervised discretization of continuous features. Machine Learning Proceedings. Elsevier, New York, pp 194–202
Bareiss ER, Porter BW (1987) A survey of psychological models of concept representation. Artificial Intelligence Laboratory. University of Texas, Austin
Guvenir HA, Acar B, Demiroz G, Cekin A (1997) A supervised machine learning algorithm for arrhythmia analysis. pp. 433–436
Mertins P et al (2016) Proteogenomics connects somatic mutations to signalling in breast cancer. Nature 534(7605):55
Mesejo P et al (2016) Computer-aided classification of gastrointestinal lesions in regular colonoscopy. IEEE Trans Med Imag 35(9):2051–2063
Coates A et al. (2011) Text detection and character recognition in scene images with unsupervised feature learning. pp. 440–445
Dua D, Graff C (2019) UCI Machine Learning Repository
Dias-Ferreira E et al (2009) Chronic stress causes frontostriatal reorganization and affects decision-making. Science 325(5940):621–625
Tsanas A, Little MA, Fox C, Ramig LO (2014) Objective automatic assessment of rehabilitative speech treatment in Parkinson’s disease. IEEE Trans Neural Syst Rehabil Eng 22(1):181–190
MacQueen JB (1967) Some methods for classification and analysis of multivariate observations. University of California Press, pp. 281–297
Johnson SC (1967) Hierarchical clustering schemes. Psychometrika 32(3):241–254
Balasubramanian M, Schwartz EL (2002) The isomap algorithm and topological stability. Science 295(5552):7–7
Eppstein D, Loffler M, Strash D (2010) Listing all maximal cliques in sparse graphs in near-optimal time. Algorithms and computation. Springer, Berlin, pp 403–414
Liu T et al. (2012) A novel LTM-based method for multi-partition clustering. pp. 203–210
Chen T, Zhang NL, Wang Y (2008) Efficient model evaluation in the search-based approach to latent structure discovery. pp. 57–64
Moon TK (1996) The expectation-maximization algorithm. IEEE Signal Process Magaz 13(6):47–60
Linting M, van der Kooij A (2012) Nonlinear principal components analysis with CATPCA: a tutorial. J Pers Assess 94(1):12–25
Husson F, Josse J (2014) Multiple correspondence analysis. In: Visualization and verbalization of data, pp. 165–184
Weinberger KQ, Saul LK (2006) Unsupervised learning of image manifolds by semidefinite programming. Int J Comput Vision 70(1):77–90
Bartenhagen C et al (2010) Comparative study of unsupervised dimension reduction techniques for the visualization of microarray gene expression data. BMC Bioinform 11(1):567
Sun Y, Todorovic S, Goodison S (2009) Local-learning-based feature selection for high-dimensional data analysis. IEEE Trans Pattern Anal Mach Intell 32(9):1610–1626
Alberto Piatti IDSIA, Marco Zaffalon IDSIA, Marcus Hutter AN (2007) Learning about a categorical latent variable under prior near-ignorance. arXiv preprintarXiv:0705.4312
Scutari M, Ness R (2012) bnlearn: Bayesian network structure learning, parameter learning and inference. R package version, 3
Njah H, Jamoussi S, Mahdi W (2019) Deep Bayesian network architecture for Big Data mining. Concurr Comput Pract Exp 31(2):e4418
Author information
Authors and Affiliations
Multimedia, Information Systems and Advanced Computing Laboratory, Sfax, Tunisia
Hasna Njah, Salma Jamoussi & Walid Mahdi
Higher Institute of Computer Sciences and Multimedia, University of Gabes, Gabes, Tunisia
Hasna Njah
Higher Institute of Computer Sciences and Multimedia, University of Sfax, Sfax, Tunisia
Salma Jamoussi & Walid Mahdi
- Hasna Njah
You can also search for this author inPubMed Google Scholar
- Salma Jamoussi
You can also search for this author inPubMed Google Scholar
- Walid Mahdi
You can also search for this author inPubMed Google Scholar
Corresponding author
Correspondence toHasna Njah.
Ethics declarations
Conflict of interest
The authors declare that there is no conflict of interest regarding the present research paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Njah, H., Jamoussi, S. & Mahdi, W. Interpretable Bayesian network abstraction for dimension reduction.Neural Comput & Applic35, 10031–10049 (2023). https://doi.org/10.1007/s00521-022-07810-4
Received:
Accepted:
Published:
Issue Date:
Share this article
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative