TY - GEN
T1 - On the Potential of Deep Symbolic Models for Classification Problems
AU - Beck, Florian
AU - Fürnkranz, Johannes
AU - Huynh, Van Quoc Phuong
PY - 2025
Y1 - 2025
N2 - With the rise of neural network approaches for machine learning problems, the focus has shifted to learning deep concepts across multiple layers. Although single-hidden-layer networks are sufficient to arbitrarily approximate any function, multilayer networks have proven to be superior in predictive performance for many applications. Unlike for neural networks, in symbolic machine learning approaches like decision tree or rule learning algorithms, the benefits of hidden layers of learned intermediate concepts remain uncertain. In this work, we empirically investigate the potential gains of deep concepts for symbolic approaches from three perspectives. First, we compare the number of possible flat and deep Boolean expressions with limited complexity, underlining the higher expressive power of deep models in such a setting. Second, we use logic minimization algorithms to generate minimal flat and deep Boolean formulas for artificial Boolean classification problems with different numbers of attributes and training examples, showing under which circumstances the use of deep concepts can lead to noticeably less complex models. Third, we compare the predictive performance of flat and deep models with a fixed maximum complexity on these datasets. We interpret these results as evidence that encourages further investigation of algorithms for learning complexity-bounded deep rule sets.
AB - With the rise of neural network approaches for machine learning problems, the focus has shifted to learning deep concepts across multiple layers. Although single-hidden-layer networks are sufficient to arbitrarily approximate any function, multilayer networks have proven to be superior in predictive performance for many applications. Unlike for neural networks, in symbolic machine learning approaches like decision tree or rule learning algorithms, the benefits of hidden layers of learned intermediate concepts remain uncertain. In this work, we empirically investigate the potential gains of deep concepts for symbolic approaches from three perspectives. First, we compare the number of possible flat and deep Boolean expressions with limited complexity, underlining the higher expressive power of deep models in such a setting. Second, we use logic minimization algorithms to generate minimal flat and deep Boolean formulas for artificial Boolean classification problems with different numbers of attributes and training examples, showing under which circumstances the use of deep concepts can lead to noticeably less complex models. Third, we compare the predictive performance of flat and deep models with a fixed maximum complexity on these datasets. We interpret these results as evidence that encourages further investigation of algorithms for learning complexity-bounded deep rule sets.
UR - https://www.scopus.com/pages/publications/105020013725
U2 - 10.1007/978-3-032-05461-6_11
DO - 10.1007/978-3-032-05461-6_11
M3 - Conference proceedings
SN - 9783032054609
T3 - Lecture Notes in Artificial Intelligence
SP - 161
EP - 175
BT - Proceedings of the 28th International Conference on Discovery Science (DS)
A2 - Džeroski, Sašo
A2 - Levatić, Jurica
A2 - Pio, Gianvito
A2 - Simidjievski, Nikola
PB - Springer, Cham
CY - Ljubljana, Slovenia
ER -