TY - JOUR
T1 - Enhancing CNNs via structural intervention with XGBoost
AU - Yavas, Cemil Emre
AU - Chen, Lei
AU - Kadlec, Christopher
AU - Kim, Jongyeop
N1 - Publisher Copyright:
© 2025 IOP Publishing Ltd. All rights, including for text and data mining, AI training, and similar technologies, are reserved.
PY - 2025/4/24
Y1 - 2025/4/24
N2 - This research investigates a novel hybridization strategy between Convolutional Neural Networks (CNNs) and gradient-boosted decision trees to enhance image classification accuracy. While conventional approaches focus on optimizing either CNN architectures or machine learning algorithms independently, we propose that intervening in the architecture itself—by strategically replacing the dense classifier portion of the CNN with a tree-based learner—can yield superior results. In our study, we construct a CNN composed of three convolutional blocks, each followed by ReLU activation, max-pooling, and dropout layers. Instead of proceeding through the final dense layers, we extract features immediately after the Flatten layer and input them into an XGBoost classifier. Our experiments reveal that applying XGBoost to these flattened features results in a higher classification accuracy than the fully optimized CNN. Although other datasets were examined during initial testing, this paper focuses exclusively on CIFAR-10 for clarity and reproducibility. The findings suggest that performance gains can be achieved through structural interventions in model architecture, challenging the prevailing emphasis on end-to-end optimization.
AB - This research investigates a novel hybridization strategy between Convolutional Neural Networks (CNNs) and gradient-boosted decision trees to enhance image classification accuracy. While conventional approaches focus on optimizing either CNN architectures or machine learning algorithms independently, we propose that intervening in the architecture itself—by strategically replacing the dense classifier portion of the CNN with a tree-based learner—can yield superior results. In our study, we construct a CNN composed of three convolutional blocks, each followed by ReLU activation, max-pooling, and dropout layers. Instead of proceeding through the final dense layers, we extract features immediately after the Flatten layer and input them into an XGBoost classifier. Our experiments reveal that applying XGBoost to these flattened features results in a higher classification accuracy than the fully optimized CNN. Although other datasets were examined during initial testing, this paper focuses exclusively on CIFAR-10 for clarity and reproducibility. The findings suggest that performance gains can be achieved through structural interventions in model architecture, challenging the prevailing emphasis on end-to-end optimization.
KW - XGBoost
KW - convolutional neural networks
KW - hybrid models
KW - image classification
KW - statistical analysis
UR - http://www.scopus.com/inward/record.url?scp=105004723297&partnerID=8YFLogxK
U2 - 10.1088/2631-8695/add084
DO - 10.1088/2631-8695/add084
M3 - Article
AN - SCOPUS:105004723297
SN - 2631-8695
VL - 7
JO - Engineering Research Express
JF - Engineering Research Express
IS - 2
M1 - 025230
ER -