Umfassende Service-Einschränkungen im Bereich Ausleihe ab 17. März!

Treffer: Supervised Learning Approaches for Robust Predictive Modelling in Data Science

Title:
Supervised Learning Approaches for Robust Predictive Modelling in Data Science
Source:
International Journal for Research in Applied Science and Engineering Technology. 13:1237-1243
Publisher Information:
International Journal for Research in Applied Science and Engineering Technology (IJRASET), 2025.
Publication Year:
2025
Document Type:
Fachzeitschrift Article
ISSN:
2321-9653
DOI:
10.22214/ijraset.2025.73756
Accession Number:
edsair.doi...........19ef59d0dc2753d85348c6279f8d228d
Database:
OpenAIRE

Weitere Informationen

Supervised learning remains the dominant paradigm for predictive modeling in data science, yet real-world deployments frequently fail due to fragile data pipelines, distributional shift, and optimistic evaluation. This article surveys supervised learning approaches with a focus on robustness—defined as the stability of predictive performance under perturbations to data, environment, or assumptions. We organize the model space into seven families: linear and generalized linear models; tree-based models; kernel methods; instance-based methods; probabilistic generative models; neural networks; and ensemble learning. For each family we discuss inductive biases, optimization, computational complexity, calibration, and typical failure modes. We then synthesize a method-agnostic workflow spanning dataset auditing, leakage prevention, feature engineering, resampling, hyperparameter tuning, model selection, and post-hoc reliability analysis (calibration, uncertainty, and drift monitoring). Robustness strategies—regularization, data augmentation, adversarial training, cost-sensitive learning, resampling for class imbalance, monotonic constraints, conformal prediction, and causal sensitivity analysis—are reviewed with practical guidance. Case vignettes from healthcare, finance, and operations illustrate trade-offs between accuracy, interpretability, and reliability. The paper concludes with open research directions, including integrating causal structure into supervised objectives, leveraging self-supervised pretraining for tabular data, distributionally robust optimization, and aligning evaluation with societal impact.