Adaptive MSD-Splitting: Enhancing C4.5 and Random Forests for Skewed Continuous Attributes

arXiv cs.LG / 4/22/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • Adaptive MSD-Splitting (AMSD) improves discretization of skewed continuous attributes by adjusting the standard-deviation multiplier based on feature skewness, avoiding severe information loss from fixed cutoff rules.
  • Building on MSD-Splitting’s efficiency gains for approximately symmetric data, AMSD narrows bin intervals in dense regions to preserve discriminatory resolution, especially for real-world biomedical and financial datasets.
  • When integrated into ensemble learning, the Random Forest-AMSD (RF-AMSD) framework delivers state-of-the-art accuracy while retaining near-identical O(N) time-complexity improvements compared with the O(N log N) exhaustive discretization search.
  • Experiments on Census Income, Heart Disease, Breast Cancer, and Forest Covertype show AMSD provides a 2–4% accuracy boost over standard MSD-Splitting, with substantial computational cost reductions in random forest settings.

Abstract

The discretization of continuous numerical attributes remains a persistent computational bottleneck in the induction of decision trees, particularly as dataset dimensions scale. Building upon the recently proposed MSD-Splitting technique -- which bins continuous data using the empirical mean and standard deviation to dramatically improve the efficiency and accuracy of the C4.5 algorithm -- we introduce Adaptive MSD-Splitting (AMSD). While standard MSD-Splitting is highly effective for approximately symmetric distributions, its rigid adherence to fixed one-standard-deviation cutoffs can lead to catastrophic information loss in highly skewed data, a common artifact in real-world biomedical and financial datasets. AMSD addresses this by dynamically adjusting the standard deviation multiplier based on feature skewness, narrowing intervals in dense regions to preserve discriminative resolution. Furthermore, we integrate AMSD into ensemble methods, specifically presenting the Random Forest-AMSD (RF-AMSD) framework. Empirical evaluations on the Census Income, Heart Disease, Breast Cancer, and Forest Covertype datasets demonstrate that AMSD yields a 2-4% accuracy improvement over standard MSD-Splitting, while maintaining near-identical O(N) time complexity reductions compared to the O(N log N) exhaustive search. Our Random Forest extension achieves state-of-the-art accuracy at a fraction of standard computational costs, confirming the viability of adaptive statistical binning in large-scale ensemble learning architectures.