Next Article in Journal
Blockchain-Enabled Secure Data Sharing with Honey Encryption and DSNN-Based Key Generation
Previous Article in Journal
Integrating Risk-Averse and Constrained Reinforcement Learning for Robust Decision-Making in High-Stakes Scenarios
Previous Article in Special Issue
GreenNAS: A Green Approach to the Hyperparameters Tuning in Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Novel Ensemble Learning Algorithm for Early Detection of Lower Back Pain Using Spinal Anomalies

1
Institute of Computer Science, Khwaja Fareed University of Engineering and Information Technology, Rahim Yar Khan 64200, Pakistan
2
Department of Software Engineering, University of Lahore, Lahore 54000, Pakistan
3
Department of Computer Science, The Islamia University of Bahawalpur, Bahawalpur 63100, Pakistan
4
Department of Artificial Intelligence and Data Science, Sejong University, Seoul 05006, Republic of Korea
5
Department of Precision Medicine, School of Medicine, Sungkyunkwan University, Suwon 16419, Republic of Korea
*
Authors to whom correspondence should be addressed.
Mathematics 2024, 12(13), 1955; https://doi.org/10.3390/math12131955
Submission received: 19 April 2024 / Revised: 14 June 2024 / Accepted: 21 June 2024 / Published: 24 June 2024
(This article belongs to the Special Issue Machine Learning Theory and Applications)

Abstract

:
Lower back pain (LBP) is a musculoskeletal condition that affects millions of people worldwide and significantly limits their mobility and daily activities. Appropriate ergonomics and exercise are crucial preventive measures that play a vital role in managing and reducing the risk of LBP. Individuals with LBP often exhibit spinal anomalies, which can serve as valuable indicators for early diagnosis. We propose an advanced machine learning methodology for LBP detection that incorporates data balancing and bootstrap** techniques. Leveraging the features associated with spinal anomalies, our method offers a promising approach for the early detection of LBP. Our study utilizes a standard dataset comprising 310 patient records, including spinal anomaly features. We propose an ensemble method called the random forest gradient boosting XGBoost Ensemble (RGXE), which integrates the combined power of the random forest, gradient boosting, and XGBoost methods for LBP detection. Experimental results demonstrate that the proposed ensemble method, RGXE Voting, outperforms state-of-the-art methods, achieving a high accuracy of 0.99. We fine-tuned each method and validated its performance using k-fold cross-validation in addition to determining the computational complexity of the methods. This innovative research holds significant potential to revolutionize the early detection of LBP, thereby improving the quality of life.

1. Introduction

Lower back pain (LBP) is a prevalent musculoskeletal problem that affects millions of people globally [1], leading to discomfort and, in severe cases, debilitating pain. The complexity of LBP stems from factors such as mechanical strain, disc degeneration, muscle imbalance, and psychological stress [2]. Understanding and addressing these root causes present challenges, necessitating comprehensive approaches for effective management. The multifaceted nature of LBP affects daily activities [3] and quality of life [4].
This widespread issue has significant drawbacks [5], including diminished mobility, reduced workplace productivity, and heightened emotional distress. Sleep disturbances [6] and potential reliance on medications further contribute to a decline in the overall quality of life. In severe instances, there is a risk of disability [7], which increases the burden faced by individuals. The associated care costs create financial challenges, emphasizing the need for holistic strategies that consider both well-being and economic implications.
LBP is the primary cause of disability worldwide, affecting approximately 619 million individuals [2]. This prevalence raises significant public health concerns, as the repercussions extend beyond personal suffering to a substantial decrease in work productivity. The financial burden on affected individuals [8] and society underscores the importance of addressing the widespread impact of LBP through comprehensive strategies.
Shifting the focus to healthcare, machine learning [9] transforms the landscape by analyzing extensive patient data for disease diagnosis and treatment personalization [10]. This technology accelerates diagnostics and ensures faster and more efficient healthcare delivery. Its versatility is evident in addressing complex medical challenges and resha** medicine with innovative solutions that prioritize precision, speed, and personalized care. The potential benefits of machine learning extend to the medical community, providing a quick, dependable, and efficient approach for disease detection and diagnosis [11].
In contrast to manual disease detection processes that are prone to human error, machine learning-based algorithms consistently achieve high accuracy in prediction tasks. This reliability offers an efficient alternative for medical diagnoses, minimizing the risks associated with human error. The integration of machine learning has the potential to benefit the medical community [12] by ensuring quick and precise disease detection while advancing the overall effectiveness and accessibility of healthcare services.
Our contributions can be summarized as follows:
  • We propose a novel ensemble method called RGXE that integrates the combined power of the random forest (RF), gradient boosting (GB), and XGBoost (XGB) methods for LBP detection.
  • We implemented advanced classification methods (logistic regression (LR), Gaussian naïve Bayes (GNB), RF, decision tree (DT), support vector machine (SVM), k-nearest neighbors (KNN), GB, and XGB) to evaluate the proposed scheme against state-of-the-art approaches.
  • We improved precision through hyperparameter optimization and k-fold cross-validation and demonstrated exceptional performance compared to existing studies.
The remainder of this paper is organized as follows: Section 2 presents a comparative analysis of the literature. Section 3 elaborates on this novel methodology. The experimental evaluations conducted in this study are described in detail in Section 4. Finally, Section 5 encapsulates the study’s findings and conclusions.

2. Literature Review

We conducted a broad assessment of earlier examinations on the discovery of LBP. This review presents the different scopes of studies that utilize different strategies for distinguishing and surveying LBP. Through a careful examination of the existing literature, we intend to acquire knowledge, distinguish patterns, and pinpoint gaps that currently exist in previous studies. A summary of previous studies is presented in Table 1.
In a previous research [13], feature selection based on a genetic algorithm (GA) was used to identify significant parameters. Two studies compared predictive models that included feature selection with those that did not, assessing performance measures such as accuracy, precision, f1 score, recall, and area under the receiver operating characteristics (ROC) Curve (AUC). A dataset comprising 310 observations and 12 features was obtained from Kaggle. This study aimed to predict early LBP symptoms using machine learning techniques, emphasizing feature selection for enhanced model performance, and achieved an accuracy of 85.2%.
In a previous study [14], the proposed multi-layer perceptron (MLP)-type artificial neural network (ANN) computed the likelihood of surgery based on the identified attributes in a model that mimicked surgical decision-making. Fifty-five criteria were found to be predictive of surgical progression. Each patient (n = 483) who presented with a lumbar spine complaint at a single Australian Tertiary Hospital between 2013 and 2019 had their medical records examined, and relevant information was gathered. The model achieved a remarkable accuracy of 92.1% in predicting surgical candidacy. The excellent discriminative ability (AUC = 0.90) and good data fit in the calibration analysis demonstrated its reliability.
In another study [15], a pioneering approach to LBP classification was introduced, utilizing RF as the classification algorithm. With an impressive accuracy of 85.80%, which significantly surpassed the initial 71.25%, the improvement was attributed to the application of parameter tuning. This marked RF as a leading contender for enhancing LBP classification, showcasing its transformative potential in medical diagnostics. Data were collected from Kaggle. Meticulous methodology and breakthrough incorporation of parameter tuning have set the stage for future advancements in precision medicine for LBP.
In a previous study [16], the proposed GNB machine learning model predicted the risk of chronic LBP. General characteristics, including sex, age, BMI, and physical activity level, were collected from all participants using the Global Physical Activity Questionnaire. Data collected from the CLBP and matched NLBP subjects adhered to ethical standards. Patients with CLBP (n = 20) were recruited from a hospital meeting specific criteria, while patients with NLBP (n = 20) were matched and recruited using the defined exclusion criteria. The model achieved an accuracy of 79% in predicting the risk of LBP.
In a previous study [17], stacked ensemble machine learning was proposed to investigate LBP. This study underlined the importance of early LBP detection and presented a classification system based on various algorithms. The intricate anatomy of the lumbar spine and its vulnerability to pain were highlighted. Kasula et al. used a dataset from Kaggle and employed hyperparameter tuning to optimize the method. The dataset contained 310 observations. The proposed method achieved an accuracy of 76.34%. They proposed a stacking ensemble classifier as an automated tool to predict the LBP tendency of a patient.
Lamichhane et al. [18] used cortical thickness (CT) as a feature to train SVM to accurately classify participants into two groups: LBP and healthy control (HC). Achieving a classification accuracy of 74.51%, an AUC of 0.787, a sensitivity of 74.07%, and a specificity of 75.00%, the model was effective in distinguishing between the two conditions. ROC curves provide a visual representation of classification performance while pinpointing the cortical regions involved in the classification process. These regions are depicted on a brain mesh surface, enriching the depth of the findings. The approach not only showcases the potential of CT in discriminating LBP but also provides insights into the specific neuroanatomical regions associated with the condition, contributing to a comprehensive understanding of the neural correlates of LBP.
Liew et al. [19] utilized functional data boosting to evaluate predictive ensemble models aimed at distinguishing between different subtypes of LBP and HCs during low-load lifting. The study included 49 participants with different LBP statuses. Three models exhibited notable accuracy: Model 1 (control vs. LBP) achieved an AUC of 90.4%, Model 2 (control vs. recurrent LBP) achieved an AUC of 91.2%, and Model 3 (recurrent LBP vs. LBP) demonstrated an impressive AUC of 96.7%. Influential predictors such as the biceps femoris, deltoid, and iliocostalis muscles underscore the potential for targeted interventions in LBP management, marking a significant advancement in predictive modeling accuracy for nuanced subtype classification.
Mao et al. [20] explored the role of the habenula in chronic LBP (cLBP) using resting-state functional connectivity (rsFC) and effective connectivity, revealing enhanced connectivity patterns in patients with cLBP. The combination of rsFC pathways, including the habenula-left superior frontal cortex, habenula-pons, and habenula-thalamus, achieved an accuracy of 75.9% in distinguishing HCs from patients with cLBP using an SVM. The dataset for the study comprised 52 individuals diagnosed with cLBP and an equivalent group of 52 HCs, which were utilized to investigate the rsFC and effective connectivity of the habenula. These findings suggest abnormal habenular connectivity in patients with cLBP, emphasizing the potential of machine learning to discriminate between chronic pain conditions.
Yu et al. [21] highlighted the potential of combining ultrasound and shear wave elastography (SWE) features, particularly highlighting the significant role of SWE elasticity in improving the automatic classification of patients with non-specific LBP (NSLBP). This advancement aids in enhancing diagnostic accuracy and intervention planning. Using a sample of 52 subjects from the University of Hong Kong–Shenzhen Hospital, we employed a SVM model to analyze 48 selected features. The SVM model achieved accuracy, precision, and sensitivity of 0.85, 0.89, and 0.86, respectively, surpassing the previous MRI-based values.
Shim et al. [22] focused on creating machine learning models designed to accurately forecast the likelihood of cLBP. Data were utilized from the Sixth Korea National Health and Nutrition Examination Survey (KNHANES VI-2, 3), including 6119 patients, with 1394 experiencing LBP. The study employed various classification models. These included k-nearest neighbors, RF, naïve Bayes, DT, GB machine, LR, SVM, and ANN. The ANN model emerged as the most effective, with an AUROC of 0.716, surpassing the other algorithms. The study underscores the potential of machine learning, particularly the ANN model, for identifying populations at high risk of cLBP. This offers a promising approach for targeted interventions and preventive strategies.

Research Gap

Following a thorough examination of the current literature, our analysis highlights specific areas of research that require further exploration.
Previously, researchers typically utilized classical machine learning methods to detect LBP. However, a growing need exists for more sophisticated ensemble machine learning approaches. Furthermore, the diagnostic performance scores in recent studies have been less than optimal.

3. Proposed Methodology

Our novel research methodology for detecting LBP in humans is shown in Figure 1. We obtained a dataset of LBP symptoms from Kaggle and conducted comprehensive preprocessing to handle null values. To address class imbalances, we utilized the synthetic minority over-sampling technique (SMOTE) [23] and further augmented the dataset through bootstrap**. After these enhancements, the data was carefully split into an 80%, and 20% for training and testing sets, respectively, to ensure a thorough evaluation of the model’s performance. Using this well-prepared and expanded dataset, we applied a machine learning model to predict LBP symptoms.

3.1. Lower Back Pain Symptoms Data

This study employed a comprehensive dataset [24] featuring 310 rows and 12 distinct features, categorizing instances into two classes: normal and abnormal. The initial 12 columns encapsulated various features, including pelvic incidence, pelvic radius, pelvic tilt, lumbar lordosis angle, sacral slope, degree of spondylolisthesis, pelvic slope, direct tilt, thoracic slope, cervical tilt, sacral angle, and scoliosis slope. Each of these features provides valuable information for the analysis. The last column of the dataset serves as the target variable, indicating whether the case falls into the “normal” or “abnormal” category. This structured dataset forms the foundation for a comprehensive investigation of factors influencing spinal health and abnormalities.

3.2. Synthetic Minority Over-Sampling Technique (SMOTE)-Based Data Resampling

Upon recognizing an imbalance in the dataset, in which 210 instances were labeled abnormal and 100 normal, we took proactive measures to address this issue. We used SMOTE [23] to augment the representation of the minority class, specifically the normal class, by generating synthetic instances (see Figure 2a). This strategic enhancement aims to create a more balanced distribution of abnormal and normal classes, which is a critical step in ensuring unbiased model training [23]. Several previous studies have addressed the issues of imbalanced data by applying data sampling techniques both before [25,26] and during model validation [27,28]. Blagus and Lusa [29] discussed the possibility of overoptimism when applying data sampling techniques. Overoptimism refers to a positive bias in the estimation of the performance of a model. Furthermore, Santos et al. explored two approaches (before and during model validation) for handling imbalanced datasets in more detail [30]. They elaborated on the potential for overoptimism and overfitting when dealing with imbalanced datasets, considering data complexity, cross-validation approaches, and data sampling methods. The study revealed that overfitting might occur when using the oversampling method alone. Regarding overoptimism, the study indicated that this issue might arise in cases of high data complexity, such as overlap** individual feature values, class separability, and the geometry and topology of the data. Additionally, previous work [31] has employed the same techniques applied in this study.
Therefore, we guided the dataset through this transformation in the present study, striving to achieve a fair representation of both classes and enhancing the model’s ability to discern patterns across various instances of spinal health. The resulting balanced dataset is illustrated in Figure 2b, showcasing improved distribution after the application of SMOTE.

3.3. Bootstrap**-Based Data Sampling

Recognizing the limitations of the small dataset, we employed a bootstrap** technique [32] to increase the data volume. By generating multiple resamples from the existing data, we effectively amplified the dataset size, providing a model with a more diverse set of instances to learn from. This approach enhances the performance of the model by exposing it to a broad range of scenarios and patterns. Bootstrap** is a strategic step for optimizing the learning process and improving the overall robustness of the model.

3.4. Data Splitting

To ensure an effective evaluation of the performance of our model, we carefully divided the dataset into a training set that contained 80% of the data and a testing set that contained the remaining 20%. With the help of this division, we trained the model on a significant percentage of the data, which helped identify patterns and relationships efficiently. The distinct testing set evaluated how well the model generalizes to new, untested data and acts as an unbiased standard.

3.5. Novel Proposed Ensemble Method

The design of the proposed RGXE Voting model is shown in Figure 3. By introducing the RGXE Voting Classifier, which is an innovative ensemble model designed to classify LBP, we integrated the strengths of RF, GB, and XGB. By leveraging the unique capabilities of each algorithm, RGXE Voting ensures a robust classification framework that captures the intricate patterns associated with LBP symptoms. This ensemble approach [17] signifies a strategic collaboration of powerful algorithms collectively aimed at achieving heightened predictive accuracy and reliability in discerning patterns within the spinal health realm.
The strategic combination of RF, GB, and XGBoost within the RGXE Voting model underscores the concerted efforts to bolster predictive accuracy and reliability in the realm of spinal health. Each algorithm brings unique strengths: RF adeptly handles high-dimensional data and complex interactions, GB iteratively minimizes errors, and XGBoost efficiently optimizes predictive performance. By harnessing the complementary capabilities of these algorithms, our ensemble method ensures thorough exploration of the feature space, leading to enhanced classification outcomes. Moreover, the ensemble nature of RGXE Voting fortifies against overfitting and augments generalization performance by amalgamating diverse model predictions. This integration of powerful algorithms reflects our dedication to crafting a nuanced classification solution capable of tackling the multifaceted challenges in LBP diagnosis.
Our proposed RGXE model employs a hard voting mechanism for LBP detection, where three classifiers are used to make individual predictions. Each classifier votes for a predicted class, and the class that receives the majority of votes is chosen as the final prediction. This approach leverages the strengths of different classifiers to enhance overall accuracy and robustness for LBP detection.

3.6. Machine Learning (ML) Methods

Following the practices of previous work [32], we briefly describe the machine learning methods utilized in our study as follows:
  • The RF algorithm [33], a potent ensemble-learning technique, was applied to classify LBP within a dataset. This approach builds a group of DTs and uses a voting mechanism to aggregate their output. Each DT was trained using a different subset of the dataset. The RF model makes predictions based on the mode of each DT prediction. The final predicted class represents the culmination of the individual predictions. This algorithm introduces randomness during both feature selection and dataset bootstrap**, thereby promoting diversity among the constituent trees. This diversity enhances the ability of the model to generalize new unseen data and helps prevent overfitting. The RF’s collective decision making through majority voting contributes to a robust classification model adept at identifying patterns related to LBP symptoms.
  • The GB ensemble method [34] for classifying LBP within a dataset involves harnessing the power of the iterative technique for predictive modeling. Applying GB to the LBP dataset involves leveraging the power of the iterative ensemble method for classification. GB creates a DT in a sequence, and each tree focuses on rectifying the previous error. This process optimizes the predictive accuracy of the model by minimizing residual errors. Mathematically, the prediction of a GB model is expressed as the sum of the predictions from all individual trees weighted by the learning rate. By iteratively improving the performance of the model, GB offers a robust approach for classifying LBP instances. The ability of the algorithm to capture complex relationships within data enhances its suitability for discerning nuanced patterns associated with spinal health. Ultimately, the application of GB contributes to the creation of a sophisticated and accurate classifier tailored to the specific challenges posed by LBP classification.
  • The DT model [35] on the LBP dataset involves the deployment of a tree-like structure to partition the data based on features to create a predictive model for classifying instances. The DT serves as a versatile and interpretable tool for classification tasks by dividing the decision-making process into a series of straightforward conditions. The algorithm iteratively selects the most informative features at each node and optimizes the model to effectively discriminate between normal and abnormal lower back conditions. Through its hierarchical structure, DT learns to make decisions by evaluating different feature thresholds, resulting in a clear and interpretable set of rules for classification. This model is particularly useful for medical diagnosis and provides insights into the factors contributing to LBP.
  • SVM [36] is capable of performing both linear and nonlinear classification tasks on the LBP dataset. SVM is particularly suitable for medical diagnosis [31], providing strong performance in detecting patterns within complex datasets. By transforming data points into a higher-dimensional space and identifying the optimal hyperplane for classification, SVM aims to maximize the margin between different classes. This methodology facilitates the creation of a robust classifier for distinguishing between normal and abnormal instances of lower back conditions. SVM is reliable for classifying and analyzing LBP and is suitable for dealing with the complexities of spinal health data.
  • KNN algorithm [37] on the LBP dataset utilizes a versatile and intuitive model for classification tasks. KNN works on the principle of proximity and classifies instances according to the majority class within their nearest neighbors. This method works well with all types of data, such as when looking at spinal health, because it does not need to know the exact distribution of the data. The model computes the distances between the data points and assigns a class label based on the consent of its k-nearest neighbors. In the domain of LBP classification, KNN offers a simple yet effective approach for capturing local patterns within the data. Its adaptability makes it a valuable tool for recognizing patterns and detecting abnormal spinal conditions.
  • LR [38] is well suited for classification tasks, providing a probabilistic framework for discerning patterns within data. By applying a logistic function, the model estimates the probability of an instance belonging to a specific class. This simplicity and interpretability make LR particularly valuable for medical diagnoses, such as classifying LBP instances as normal or abnormal. This method is helpful in healthcare analysis because it provides a good understanding of complex connections and useful information regarding the factors affecting spinal health.
  • XGB algorithm [39] for the LBP dataset utilizes an advanced ensemble-learning technique renowned for its efficiency in classification tasks. XGB builds a robust predictive model through the sequential training of DTs, with each correcting errors made by the previous ones. The ensemble approach combines the strengths of the individual trees, thereby resulting in a highly accurate and resilient classifier. By optimizing a predefined objective function, XGB efficiently handles imbalances in class distribution, making it particularly well suited for medical datasets such as LBP analysis. The ability of this model to capture intricate patterns and exhibit high predictive performance contributes to its prominence in healthcare analytics, thereby providing valuable insights into spinal health conditions.
  • GNB algorithm [40] on the LBP dataset leverages a probabilistic model based on Bayes’ theorem. GNB assumes that features are conditionally independent, allowing for a straightforward and effective classification approach. In the context of LBP classification, GNB models the probability distribution of features for each class and assigns the most likely class based on the observed feature values. Despite its simplicity, GNB has proven particularly valuable in healthcare analytics, offering interpretable insights into the likelihood of instances being classified as normal or abnormal based on the given features. Its efficient computation and ability to handle continuous data make GNB a practical choice for discerning patterns within medical datasets.

3.7. Parameter Settings

The best hyperparameters [41] for the ML methods are listed in Table 2. These hyperparameters were optimized for each method using k-fold cross-validation, including multiple testing and training. Our results show that, by fine-tuning the hyperparameters, we achieved remarkable performance scores in LBP detection.

4. Results and Discussion

Our research findings and discussion demonstrate the outcomes of implementing a novel method for identifying LBP. This section evaluates the performance metrics used to measure the effectiveness of the proposed approach compared with established methods.

4.1. Experimental Settings

In our study, we conducted experiments on the computer with the specified specifications in Table 3. To evaluate the effectiveness of our machine learning models, we relied on metrics; they are accuracy, precision, recall, and F1 score. Additional details related to the experimental setup are listed in Table 3.

4.2. Performance Analysis before Bootstrap**

A comparison of the performance of the applied machine learning models before bootstrap** is presented in Table 4. The accuracy scores revealed that RGXE was the top-performing model with 0.95 accuracy, closely followed by XGB at 0.90. RF and GB also perform well, with accuracies of 0.92 and 0.94, respectively. LR and SVM achieved a solid accuracy of 0.85. However, GNB and KNN exhibited slightly lower accuracies at 0.82 and 0.84, respectively. DT trails, with an accuracy of 0.83. These results provide valuable insight into the relative performance of each model, aiding in informed model selection for classification tasks.
We further analyzed the performance results of our proposed model using the original dataset, as presented in Table 5. The results confirmed that the ensemble learning model achieved moderate success in detecting lower back pain. These findings suggest the need for further research experiments focused on data balancing and boosting techniques.
Figure 4 shows a performance comparison using histogram-based bar charts, contrasting the proposed technique with established machine learning techniques before bootstrap**. The graph accentuates the performance of the new method, shedding light on its effectiveness compared with traditional methods. This analysis offers valuable insights into the relative efficacy of the proposed approach and aids in understanding its potential advantages over conventional techniques.

4.3. Performance Analysis after Bootstrap** and Proposed Approach

Table 6 presents the performance disparities observed across various machine learning methodologies when implemented in conjunction with the bootstrap** technique. Following bootstrap**, our examination of various models on the LBP dataset revealed substantial improvements in classification results. Notably, the RGXE Voting model stands out with an impressive 99% accuracy, demonstrating exceptional performance scores for both normal and abnormal classes. This noteworthy advancement underscores the effectiveness of the RGXE Voting ensemble, which combines RF, GB, and XGB.
Individual models, such as RF and GB, also exhibit heightened accuracy and improved performance after bootstrap**. By contrast, the DT and KNN models also exhibited improved performance. SVM experienced slight improvements in accuracy and precision after bootstrap**, whereas GNB and LR maintained their performance levels.
The RGXE Voting model proposed in this study emerges as the top-performing model after bootstrap**, as shown in Table 7. Table 7 also reveals that for all methods evaluated in this study, the accuracy performance of most methods increased by an average of up to 5%. Its resilience and efficacy in addressing class imbalances and capturing intricate patterns within the LBP dataset were evident through its remarkable accuracy and well-balanced precision, recall, and F1 scores. These attributes make RGXE Voting a robust and reliable choice for classifying LBP.
In Figure 5, the radar chart visualizes the performance of the different methods, showing that our proposed approach consistently leads to superior outcomes across all evaluated techniques. The chart demonstrates the broader spectrum of performance accuracy achieved using our novel method, emphasizing its effectiveness in enhancing the accuracy of the applied methods. This study underscores the success of our innovative approach in significantly improving the accuracy scores across the board for the methods examined.
Figure 6 depicts the outcomes of the evaluation performed using a confusion matrix for the ML methods when employing the bootstrap** technique. The results demonstrate the outstanding performance of the proposed RGXE approach, with a notably high rate of accurate classification and minimum errors compared with alternative methods.
Figure 7 illustrates a performance comparison using histogram-based bar charts, delineating the effectiveness of the newly proposed approach against established machine learning techniques. The graph highlights the performance of the new method and offers insights into its efficacy relative to conventional methods. This visual representation underscores the potential superiority of the proposed approach for achieving desirable outcomes.

4.4. 10-Fold Cross Validations Analysis

Following the application of 10-fold cross-validation, the validation results outlined in Table 8 specifically highlight the superiority of the RGXE method. The analysis underscores that the RGXE model consistently achieves a high k-fold accuracy, surpassing 99%, and shows minimal standard deviation scores. This strong validation ensures the generalizability of our proposed RGXE method, particularly for the identification of LBP.

4.5. Assessment of Computational Complexity Performance

The assessment results of the computational complexities of the implemented ML methods are listed in Table 9. The results revealed that the runtime computations for various methods in our study varied, with RF requiring 0.034 s, GB requiring 0.187 s, and DT performing computations in 0.009 s. SVM had a longer runtime of 1.014 s, whereas the proposed RGXE method exhibited the highest runtime of 2.406 s. These findings highlight the diverse computational efficiencies of the methods, with RGXE demonstrating a longer runtime but potentially offering distinct advantages in detecting LBP.

4.6. Comparison with Previous Studies

In this sub-section, we further analyzed and compared the performance of our proposed method with the results of previous work with the same dataset, as can be seen in Table 10. For a thorough comparison, we evaluated the efficacy of our innovative RGXE model, employing bootstrap**, against established methodologies. Previous studies relied predominantly on classical machine learning, achieving a maximum accuracy of 94%. By contrast, our RGXE model, which leverages a novel bootstrap** approach, exceeds these benchmarks by achieving a high performance score of 99%. This analysis conclusively establishes the superiority of our proposed methodology over existing state-of-the-art methods.

4.7. Performance Validation with Independent Dataset

Additionally, we further validated our proposed model using an independent dataset called “Anemia Types Classification [44]”. This independent dataset comprises more heterogeneous data, featuring 15 broad demographic characteristics across 1281 instances. The performance results of the proposed approach are presented in Table 11. The results demonstrate that the proposed model achieved high performance scores on this independent data, predicting the type of anemia with an accuracy of up to 98%. This additional experiment confirmed that our proposed model performs well even on new datasets.

4.8. Limitations of the Study

The proposed RGXE for the early detection of lower back pain using spinal anomalies demonstrates promising results. However, it has certain limitations. One notable limitation is the relatively high computational time complexity, recorded at 2.41 s. This can be a constraint in scenarios requiring real-time analysis or the rapid processing of large datasets. Future work can focus on optimizing the algorithm to reduce computational time, potentially through techniques such as parallel processing or more efficient data handling methods. Additionally, the current model utilizes original features for building the detection models. While this approach has yielded significant results, there is potential for further enhancement through advanced feature engineering. Incorporating more sophisticated features derived from the raw data could improve the model’s accuracy and robustness, leading to more reliable early detection of lower back pain.

5. Conclusions

This study introduces a novel ML method called RGXE for the early detection of LBP. Using a Kaggle dataset with 310 rows and 12 columns, including spinal anomaly features, we balanced the data using SMOTE and bootstrap** techniques. Subsequently, advanced ML models were applied to the bootstrapped data to evaluate their performance. Validation was conducted via 10-fold cross-validation to ensure the reliability of our results. Furthermore, we present the computational complexity of each method to understand its efficiency in terms of runtime. Our extensive research demonstrates that the proposed RGXE Voting outperforms other ML methods and that of previous work results, achieving an impressive accuracy of 0.99. Fine-tuning was performed for each method to further optimize performance. This study makes a substantial contribution to the early LBP detection field by presenting a robust and efficient approach to revolutionizing healthcare practices.
Furthermore, we aim to advance our method by decreasing its computational runtime complexity and exploring feature engineering methods as well as explainable artificial intelligence to improve not only the performance of the models but also to provide logic and explanation behind the decision being presented. Therefore, this will eventually increase the trust of health practitioners toward the decision being made by the model.

Author Contributions

Conceptualization, M.H., M.S.A.H., N.L.F., M.S. and S.W.L.; methodology, M.H., M.S.A.H., N.L.F., M.S. and S.W.L.; validation, M.H., M.S.A.H., A.R. and M.I.; formal analysis, N.L.F., M.S. and S.W.L.; investigation, M.H., M.S.A.H., A.R. and M.I.; data curation, M.H., M.S.A.H., A.R. and M.I.; writing—original draft preparation, M.H., M.S.A.H., A.R. and M.I.; writing—review and editing, N.L.F., M.S. and S.W.L.; visualization, M.H., M.S.A.H., N.L.F. and M.S.; supervision, M.S.A.H., M.S. and S.W.L.; funding acquisition, M.S. and S.W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Research Foundation of Korea (grant number: NRF2021R1I1A2059735).

Data Availability Statement

The original data presented in this study are openly available from Kaggle at [24,44].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chen, S.; Chen, M.; Wu, X.; Lin, S.; Tao, C.; Cao, H.; Shao, Z.; **ao, G. Global, Regional and National Burden of Low Back Pain 1990–2019: A Systematic Analysis of the Global Burden of Disease Study 2019. J. Orthop. Transl. 2022, 32, 49–58. [Google Scholar] [CrossRef]
  2. Low Back Pain. Available online: https://www.who.int/news-room/fact-sheets/detail/low-back-pain#::text=LBP%20can%20be%20specific%20or,reason%20to%20explain%20the%20pain (accessed on 17 March 2024).
  3. Grabovac, I.; Dorner, T.E. Association between Low Back Pain and Various Everyday Performances: Activities of Daily Living, Ability to Work and Sexual Function. Wien. Klin. Wochenschr. 2019, 131, 541–549. [Google Scholar] [CrossRef] [PubMed]
  4. Agnus Tom, A.; Rajkumar, E.; John, R.; Joshua George, A. Determinants of Quality of Life in Individuals with Chronic Low Back Pain: A Systematic Review. Health Psychol. Behav. Med. 2022, 10, 124–144. [Google Scholar] [CrossRef] [PubMed]
  5. Stefane, T.; dos Santos, A.M.; Marinovic, A.; Hortense, P. Chronic Low Back Pain: Pain Intensity, Disability and Quality of Life. Acta Paul. Enferm. 2013, 26, 14–20. [Google Scholar] [CrossRef]
  6. Alsaadi, S.M.; McAuley, J.H.; Hush, J.M.; Lo, S.; Bartlett, D.J.; Grunstein, R.R.; Maher, C.G. The Bidirectional Relationship between Pain Intensity and Sleep Disturbance/Quality in Patients with Low Back Pain. Clin. J. Pain 2014, 30, 755–765. [Google Scholar] [CrossRef] [PubMed]
  7. Manchikanti, L.; Singh, V.; Datta, S.; Cohen, S.P.; Hirsch, J.A. Comprehensive Review of Epidemiology, Scope, and Impact of Spinal Pain. Pain Physician 2009, 12, E35. [Google Scholar] [CrossRef] [PubMed]
  8. Mathew, J.; Singh, S.B.; Garis, S.; Diwan, A.D. Backing up the Stories: The Psychological and Social Costs of Chronic Low-Back Pain. Int. J. Spine Surg. 2013, 7, e29–e38. [Google Scholar] [CrossRef]
  9. Sarker, M. Revolutionizing Healthcare: The Role of Machine Learning in the Health Sector. J. Artif. Intell. Gen. Sci. (JAIGS) 2024, 2, 35–48. [Google Scholar]
  10. Gill, A.Y.; Saeed, A.; Rasool, S.; Husnain, A.; Hussain, H.K. Revolutionizing Healthcare: How Machine Learning Is Transforming Patient Diagnoses-a Comprehensive Review of AI’s Impact on Medical Diagnosis. J. World Sci. 2023, 2, 1638–1652. [Google Scholar] [CrossRef]
  11. Kasula, B.Y. Machine Learning in Healthcare: Revolutionizing Disease Diagnosis and Treatment. Int. J. Creat. Res. Comput. Technol. Des. 2021, 3, 1–7. [Google Scholar]
  12. Javaid, M.; Haleem, A.; Singh, R.P.; Suman, R.; Rab, S. Significance of Machine Learning in Healthcare: Features, Pillars and Applications. Int. J. Intell. Netw. 2022, 3, 58–73. [Google Scholar] [CrossRef]
  13. Al Imran, A.; Rifatul Islam Rifat, M.; Mohammad, R. Enhancing the Classification Performance of Lower Back Pain Symptoms Using Genetic Algorithm-Based Feature Selection. In Proceedings of the International Joint Conference on Computational Intelligence: IJCCI 2018, Sevile, Spain, 18–20 September 2018; Springer: Singapore, 2018; pp. 455–469. [Google Scholar]
  14. **e, N.; Wilson, P.J.; Reddy, R. Use of Machine Learning to Model Surgical Decision-Making in Lumbar Spine Surgery. Eur. Spine J. 2022, 31, 2000–2006. [Google Scholar] [CrossRef] [PubMed]
  15. Lenka, S.; Victor, N. Lower Back Pain Classification Using Parameter Tuning. Res. J. Pharm. Technol. 2022, 15, 1573–1578. [Google Scholar] [CrossRef]
  16. Thiry, P.; Houry, M.; Philippe, L.; Nocent, O.; Buisseret, F.; Dierick, F.; Slama, R.; Bertucci, W.; Thévenon, A.; Simoneau-Buessinger, E. Machine Learning Identifies Chronic Low Back Pain Patients from an Instrumented Trunk Bending and Return Test. Sensors 2022, 22, 5027. [Google Scholar] [CrossRef]
  17. Bandyopadhyay, S.; Dutta, S. Detecting Lower Back Pain Using Stacked Ensemble Approach. Preprints 2020. [Google Scholar]
  18. Lamichhane, B.; Jayasekera, D.; Jakes, R.; Glasser, M.F.; Zhang, J.; Yang, C.; Grimes, D.; Frank, T.L.; Ray, W.Z.; Leuthardt, E.C.; et al. Multi-Modal Biomarkers of Low Back Pain: A Machine Learning Approach. NeuroImage Clin. 2021, 29, 102530. [Google Scholar] [CrossRef] [PubMed]
  19. Liew, B.X.; Rugamer, D.; De Nunzio, A.M.; Falla, D. Interpretable Machine Learning Models for Classifying Low Back Pain Status Using Functional Physiological Variables. Eur. Spine J. 2020, 29, 1845–1859. [Google Scholar] [CrossRef] [PubMed]
  20. Mao, C.P.; Wu, Y.; Yang, H.J.; Qin, J.; Song, Q.C.; Zhang, B.; Zhou, X.Q.; Zhang, L.; Sun, H.H. Altered Habenular Connectivity in Chronic Low Back Pain: An fMRI and Machine Learning Study. Hum. Brain Mapp. 2023, 44, 4407–4421. [Google Scholar] [CrossRef]
  21. Yu, X.; Xu, X.; Huang, Q.; Zhu, G.; Xu, F.; Liu, Z.; Su, L.; Zheng, H.; Zhou, C.; Chen, Q.; et al. Binary Classification of Non-Specific Low Back Pain Condition Based on the Combination of B-Mode Ultrasound and Shear Wave Elastography at Multiple Sites. Front. Physiol. 2023, 14, 1176299. [Google Scholar] [CrossRef]
  22. Shim, J.-G.; Ryu, K.-H.; Cho, E.-A.; Ahn, J.H.; Kim, H.K.; Lee, Y.-J.; Lee, S.H. Machine Learning Approaches to Predict Chronic Lower Back Pain in People Aged over 50 Years. Medicina 2021, 57, 1230. [Google Scholar] [CrossRef]
  23. Ijaz, M.; Alfian, G.; Syafrudin, M.; Rhee, J. Hybrid Prediction Model for Type 2 Diabetes and Hypertension Using DBSCAN-Based Outlier Detection, Synthetic Minority Over Sampling Technique (SMOTE), and Random Forest. Appl. Sci. 2018, 8, 1325. [Google Scholar] [CrossRef]
  24. Lower Back Pain Symptoms Dataset. Available online: https://www.kaggle.com/datasets/sammy123/lower-back-pain-symptoms-dataset (accessed on 16 March 2024).
  25. Ahmad, J.; Javed, F.; Hayat, M. Intelligent Computational Model for Classification of Sub-Golgi Protein Using Oversampling and Fisher Feature Selection Methods. Artif. Intell. Med. 2017, 78, 14–22. [Google Scholar] [CrossRef] [PubMed]
  26. Awad, A.; Bader-El-Den, M.; McNicholas, J.; Briggs, J. Early Hospital Mortality Prediction of Intensive Care Unit Patients Using an Ensemble Learning Approach. Int. J. Med. Inform. 2017, 108, 185–195. [Google Scholar] [CrossRef]
  27. Sady, C.C.R.; Ribeiro, A.L.P. Symbolic Features and Classification via Support Vector Machine for Predicting Death in Patients with Chagas Disease. Comput. Biol. Med. 2016, 70, 220–227. [Google Scholar] [CrossRef]
  28. Vinodhini, G.; Chandrasekaran, R. A Sampling Based Sentiment Mining Approach for E-Commerce Applications. Inf. Process. Manag. 2017, 53, 223–236. [Google Scholar] [CrossRef]
  29. Blagus, R.; Lusa, L. Joint Use of Over- and under-Sampling Techniques and Cross-Validation for the Development and Assessment of Prediction Models. BMC Bioinform. 2015, 16, 363. [Google Scholar] [CrossRef] [PubMed]
  30. Santos, M.S.; Soares, J.P.; Abreu, P.H.; Araujo, H.; Santos, J. Cross-Validation for Imbalanced Datasets: Avoiding Overoptimistic and Overfitting Approaches [Research Frontier]. IEEE Comput. Intell. Mag. 2018, 13, 59–76. [Google Scholar] [CrossRef]
  31. Raza, A.; Siddiqui, H.U.R.; Munir, K.; Almutairi, M.; Rustam, F.; Ashraf, I. Ensemble Learning-Based Feature Engineering to Analyze Maternal Health during Pregnancy and Health Risk Prediction. PLoS ONE 2022, 17, e0276525. [Google Scholar] [CrossRef] [PubMed]
  32. Yagin, F.H.; Alkhateeb, A.; Raza, A.; Samee, N.A.; Mahmoud, N.F.; Colak, C.; Yagin, B. An Explainable Artificial Intelligence Model Proposed for the Prediction of Myalgic Encephalomyelitis/Chronic Fatigue Syndrome and the Identification of Distinctive Metabolites. Diagnostics 2023, 13, 3495. [Google Scholar] [CrossRef] [PubMed]
  33. Zermane, A.; Tohir, M.Z.M.; Zermane, H.; Baharudin, M.R.; Yusoff, H.M. Predicting Fatal Fall from Heights Accidents Using Random Forest Classification Machine Learning Model. Saf. Sci. 2023, 159, 106023. [Google Scholar] [CrossRef]
  34. Douiba, M.; Benkirane, S.; Guezzaz, A.; Azrour, M. An Improved Anomaly Detection Model for IoT Security Using Decision Tree and Gradient Boosting. J. Supercomput. 2023, 79, 3392–3411. [Google Scholar] [CrossRef]
  35. Hamdi, M.; Hilali-Jaghdam, I.; Elnaim, B.E.; Elhag, A.A. Forecasting and Classification of New Cases of COVID-19 before Vaccination Using Decision Trees and Gaussian Mixture Model. Alex. Eng. J. 2023, 62, 327–333. [Google Scholar] [CrossRef]
  36. Mahmoodi, A.; Hashemi, L.; Jasemi, M.; Mehraban, S.; Laliberté, J.; Millar, R.C. A Developed Stock Price Forecasting Model Using Support Vector Machine Combined with Metaheuristic Algorithms. Opsearch 2023, 60, 59–86. [Google Scholar] [CrossRef]
  37. Lu, D.; Yue, Y.; Hu, Z.; Xu, M.; Tong, Y.; Ma, H. Effective Detection of Alzheimer’s Disease by Optimizing Fuzzy K-Nearest Neighbors Based on Salp Swarm Algorithm. Comput. Biol. Med. 2023, 159, 106930. [Google Scholar] [CrossRef] [PubMed]
  38. Patel, R.K.; Aggarwal, E.; Solanki, K.; Dahiya, O.; Yadav, S.A. A Logistic Regression and Decision Tree Based Hybrid Approach to Predict Alzheimer’s Disease. In Proceedings of the 2023 International Conference on Computational Intelligence and Sustainable Engineering Solutions (CISES), Greater Noida, India, 28–30 April 2023; IEEE: New York, NY, USA, 2023; pp. 722–726. [Google Scholar]
  39. Hoque, R.; Das, S.; Hoque, M.; Haque, E. Breast Cancer Classification Using XGBoost. World J. Adv. Res. Rev. 2024, 21, 1985–1994. [Google Scholar] [CrossRef]
  40. Vedaraj, M.; Anita, C.; Muralidhar, A.; Lavanya, V.; Balasaranya, K.; Jagadeesan, P. Early Prediction of Lung Cancer Using Gaussian Naive Bayes Classification Algorithm. Int. J. Intell. Syst. Appl. Eng. 2023, 11, 838–848. [Google Scholar]
  41. Singh, S.; Patro, S.K.; Parhi, S.K. Evolutionary Optimization of Machine Learning Algorithm Hyperparameters for Strength Prediction of High-Performance Concrete. Asian J. Civ. Eng. 2023, 24, 3121–3143. [Google Scholar] [CrossRef]
  42. Gambo, I.; Mbada, C.; Aina, S.; Ogundare, T.; Ikono, R.; Alimi, O.; Saah, F.; Magreola, M.; Agbonkhese, C. Implementing Decision Support Tool for Low-Back Pain Diagnosis and Prediction Based on the Range of Motions. Indones. J. Electr. Eng. Comput. Sci. 2024, 33, 1302–1314. [Google Scholar] [CrossRef]
  43. Islam, M.S.; Asaduzzaman, M.; Rahman, M.M. Feature Selection and Classification of Spinal Abnormalities to Detect Low Back Pain Disorder Using Machine Learning Approaches. In Proceedings of the 2019 1st International Conference on Advances in Science, Engineering and Robotics Technology (ICASERT), Dhaka, Bangladesh, 3–5 May 2019; IEEE: New York, NY, USA, 2019; pp. 1–4. [Google Scholar]
  44. Anemia Types Classification. Available online: https://www.kaggle.com/datasets/ehababoelnaga/anemia-types-classification/data (accessed on 17 May 2024).
Figure 1. Architecture of our novel proposed research methodology.
Figure 1. Architecture of our novel proposed research methodology.
Mathematics 12 01955 g001
Figure 2. Target class data distribution analysis: (a) before balancing; (b) after balancing.
Figure 2. Target class data distribution analysis: (a) before balancing; (b) after balancing.
Mathematics 12 01955 g002
Figure 3. Structure of our proposed ensemble model.
Figure 3. Structure of our proposed ensemble model.
Mathematics 12 01955 g003
Figure 4. Histogram-based comparisons of all models’ performances before bootstrap**.
Figure 4. Histogram-based comparisons of all models’ performances before bootstrap**.
Mathematics 12 01955 g004
Figure 5. Radar chart-based comparison of ML model performance.
Figure 5. Radar chart-based comparison of ML model performance.
Mathematics 12 01955 g005
Figure 6. Confusion matrix results of ML methods.
Figure 6. Confusion matrix results of ML methods.
Mathematics 12 01955 g006
Figure 7. Histogram-based comparisons of all models’ performances after bootstrap**.
Figure 7. Histogram-based comparisons of all models’ performances after bootstrap**.
Mathematics 12 01955 g007
Table 1. Literature summary of previously published works examined for comparative analysis.
Table 1. Literature summary of previously published works examined for comparative analysis.
Ref. (Year)Dataset UsedData Sample SizeProposed MethodPerformance Accuracy Score
[13] (2019)Data obtained from Kaggle310 observationsGenetic algorithm-based
feature selection
85.2%
[14] (2022)Lumbar records from an Australian hospital483 patientsMulti-layer perceptron-type artificial neural network92.1%
[15] (2022)Data collected from Kaggle49 participantsRandom forest85.80%
[16] (2022)Dataset collected from self-reporteddata set of 1678 cyclesGaussian naïve Bayes79%
[17] (2020)Utilized a dataset from Kaggle310 observationsStacked ensemble machine76.34%
[18] (2021)Dataset collected from self-reported51 patientsSupport vector machine 74.51%
[19] (2020)49 participants across LBP statuses49 participantsThree predictive models96.7%
[20] (2023)52 cLBP patients and 52 healthy controls104 patientsSupport vector machine75.9%
[21] (2023)52 participants from the University of
Hong Kong–Shenzhen Hospital
52 patientsSupport vector machine85%
[22] (2021)KNHANES VI-2, 36119 patientsArtificial neural network71.6%
Table 2. Analysis of hyperparameter tuning of applied machine learning models.
Table 2. Analysis of hyperparameter tuning of applied machine learning models.
MethodsHyperparameter Tuning (Settings)
RFmax_leaf_nodes = None, min_impurity_decrease = 0.0, n_estimators = 10, criterion = ’gini’, max_depth = None, min_weight_fraction_leaf = 0.0
GBmin_impurity_split = None,init = None, random_state = None, max_depth = 3, min_impurity_decrease = 0.0, ccp_alpha = 0.0, loss = ’deviance’, learning_rate = 0.1, n_estimators = 50
DTcriterion = ’gini’, splitter = ’best’, min_samples_split = 2, min_samples_leaf = 1, min_weight_fraction_leaf = 0.0, impurity_decrease = 0.0
SVMC = 1.0, kernel = ’linear’, degree = 3, gamma = ’scale’, coef0 = 0.0, shrinking = True, probability = False
KNNn_neighbors = 3, weights = ’uniform’, algorithm = ’auto’, leaf_size = 30, p = 2, metric = ’minkowski’
LRsolver = ’lbfgs’, max_iter = 100, multi_class = ’auto’, class_weight = None, random_state = None, penalty = ’l2’, dual = False
GNBpriors = None, var_smoothing = 1e-9
XGBmax_depth = 3, min_child_weight = 1, missing = None, reg_alpha = 0, reg_lambda = 1, gamma = 0, learning_rate = 0.1, n_estimators = 100
RGXEXGB(n_estimators = 100, learning_rate = 0.1, random_state = 42), VotingClassifier(voting = Hard), RF(n_estimators = 50, random_state = 42), GB(n_estimators = 100, learning_rate = 0.1, random_state = 42)
Table 3. Specification details of our experimental setup.
Table 3. Specification details of our experimental setup.
Specification ParameterSpecification Value
Computational processorIntel(R) Core(TM) i5-4300M
Central processing unit (CPU)2.60 gigahertz (GHz)
CPU cores2
Logical processors4
Random access memory (RAM)8 gigabytes (GB)
Programming languagePython V3
Table 4. Performance of ML models before bootstrap**.
Table 4. Performance of ML models before bootstrap**.
MethodsAccuracyTargetPrecisionRecallF1 Score
RF0.92Normal0.880.950.91
Abnormal0.950.890.92
GB0.95Normal0.891.000.94
Abnormal1.000.890.94
DT0.83Normal0.820.820.82
Abnormal0.840.840.84
SVM0.85Normal0.830.840.86
Abnormal0.880.840.86
KNN0.84Normal0.760.970.86
Abnormal0.970.730.83
LR0.85Normal0.810.900.85
Abnormal0.900.820.86
GNB0.82Normal0.730.970.84
Abnormal0.970.690.81
XGB0.90Normal0.880.930.90
Abnormal0.930.890.91
Proposed RGXE0.95Normal0.911.000.95
Abnormal1.000.910.95
Table 5. Performance of ML model with original data.
Table 5. Performance of ML model with original data.
Results with Original Data
AccuracyTarget ClassPrecisionRecallF1
0.76Abnormal0.830.810.82
Normal0.600.630.62
Average0.760.760.76
Results with SMOTE after data splitting (train/test)
0.98Abnormal1.000.970.99
Normal0.961.000.98
Average0.980.980.98
Table 6. Performance of ML models after bootstrap**.
Table 6. Performance of ML models after bootstrap**.
MethodsAccuracyTargetPrecisionRecallF1 Score
RF0.98Normal0.970.990.98
Abnormal0.990.970.98
GB0.96Normal0.950.980.96
Abnormal0.970.940.96
DT0.97Normal0.980.980.98
Abnormal0.970.970.97
SVM0.87Normal0.850.890.87
Abnormal0.890.860.87
KNN0.95Normal0.930.980.96
Abnormal0.970.910.94
LR0.85Normal0.850.900.87
Abnormal0.860.800.83
GNB0.82Normal0.750.950.84
Abnormal0.940.710.81
XGB0.98Normal0.980.990.98
Abnormal0.990.980.98
Proposed RGXE0.99Normal1.000.980.99
Abnormal0.991.000.99
Table 7. Performance of ML models before and after bootstrap**.
Table 7. Performance of ML models before and after bootstrap**.
MethodsAccuracy before Bootstrap**Accuracy after Bootstrap**
RF0.920.98
GB0.950.96
DT0.830.97
SVM0.850.87
KNN0.840.95
LR0.850.85
GNB0.820.82
XGB0.900.98
Proposed RGXE0.950.99
Table 8. Performance evaluation based on variations in the 10-fold cross-validation mechanism.
Table 8. Performance evaluation based on variations in the 10-fold cross-validation mechanism.
MethodsAccuracy ± Standard Deviation
RF0.989 ± 0.00863
GB0.962 ± 0.01122
DT0.991 ± 0.00945
SVM0.867 ± 0.03163
KNN0.968 ± 0.01065
LR0.848 ± 0.01746
GNB0.804 ± 0.03974
XGB0.993 ± 0.00759
Proposed RGXE0.995 ± 0.00790
Table 9. Analysis of computational complexity in terms of runtime for the applied approaches.
Table 9. Analysis of computational complexity in terms of runtime for the applied approaches.
MethodsRuntime Computations (in Seconds)
RF0.034183740615844
GB0.186820268630981
DT0.008729457855224
SVM1.013656616210937
KNN0.022443056106567
LR0.145286560058593
GNB0.007312059402465
XGB1.418703794479370
Proposed RGXE2.405933141708374
Table 10. Performance comparison of our proposed RGXE with previous studies with the same dataset.
Table 10. Performance comparison of our proposed RGXE with previous studies with the same dataset.
Ref.Proposed MethodPrecision (%)Accuracy (%)
[13]K-nearest
neighbors
8985
[42]Hybrid
Model
8092
[43]Random forest9094
OurProposed RGXE9999
Table 11. Performance comparison of our proposed RGXE with independent dataset.
Table 11. Performance comparison of our proposed RGXE with independent dataset.
AccuracyTarget ClassPrecisionAccuracyF1
0.98Healthy1.001.001.00
Iron deficiency anemia1.001.001.00
Leukemia0.901.00
Leukemia with thrombocytopenia1.001.001.00
Macrocytic anemia1.001.001.00
Normocytic hypochromic anemia1.001.001.00
Normocytic normochromic anemia1.000.980.99
Other microcytic anemia1.000.890.94
Thrombocytopenia0.851.000.92
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Haider, M.; Hashmi, M.S.A.; Raza, A.; Ibrahim, M.; Fitriyani, N.L.; Syafrudin, M.; Lee, S.W. Novel Ensemble Learning Algorithm for Early Detection of Lower Back Pain Using Spinal Anomalies. Mathematics 2024, 12, 1955. https://doi.org/10.3390/math12131955

AMA Style

Haider M, Hashmi MSA, Raza A, Ibrahim M, Fitriyani NL, Syafrudin M, Lee SW. Novel Ensemble Learning Algorithm for Early Detection of Lower Back Pain Using Spinal Anomalies. Mathematics. 2024; 12(13):1955. https://doi.org/10.3390/math12131955

Chicago/Turabian Style

Haider, Moin, Muhammad Shadab Alam Hashmi, Ali Raza, Muhammad Ibrahim, Norma Latif Fitriyani, Muhammad Syafrudin, and Seung Won Lee. 2024. "Novel Ensemble Learning Algorithm for Early Detection of Lower Back Pain Using Spinal Anomalies" Mathematics 12, no. 13: 1955. https://doi.org/10.3390/math12131955

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop