Pursuing Reliability in Statistical Inference and Machine Learning

Sunday 02 February 2025


Statisticians and machine learning experts have long debated whether their field is more about making accurate predictions or ensuring those predictions are reliable. A new study suggests that the answer lies somewhere in between.


The traditional view of statistical inference, known as frequentism, posits that the accuracy of a prediction method can be judged solely by its performance in different scenarios, without considering the thought process behind it. This approach has been criticized for being too narrow, focusing only on the end result rather than the reasoning that led to it.


In contrast, internalist reliabilism argues that the reliability of a prediction method is not just about how well it performs, but also about the underlying assumptions and principles used to make those predictions. This perspective emphasizes the importance of understanding why a particular approach works or doesn’t work, rather than simply relying on its results.


The study in question explores this debate by examining the foundations of statistical inference, model selection, and machine learning through the lens of internalist reliabilism. The authors argue that these fields are not as disparate as they seem, but rather share a common thread – the pursuit of reliability.


According to the researchers, classical statistics, which has long been seen as the gold standard for data analysis, is actually more closely aligned with internalist reliabilism than previously thought. They point out that frequentist statistics, which focuses on the accuracy of predictions, can be reinterpreted as a form of internalist reliabilism by considering the underlying assumptions and principles used to make those predictions.


The study also examines model selection and machine learning, which are often seen as distinct from classical statistics. However, the authors argue that these fields share a common foundation with classical statistics, rooted in the pursuit of reliability. They demonstrate how the same standards of reliability can be applied to both traditional statistical methods and more modern approaches like cross-validation.


The findings have significant implications for our understanding of statistical inference and machine learning. By recognizing the shared concerns of reliability across different fields, researchers can develop more nuanced and robust approaches to data analysis. This could lead to more accurate predictions, better decision-making, and a deeper understanding of the complex relationships between data, models, and reality.


Ultimately, the study suggests that the pursuit of reliability is not an either-or proposition, but rather a continuous thread that runs through many areas of statistical inference and machine learning. By embracing this perspective, researchers can develop more effective and reliable methods for making predictions and gaining insights from data.


Cite this article: “Pursuing Reliability in Statistical Inference and Machine Learning”, The Science Archive, 2025.


Statistical Inference, Machine Learning, Prediction Accuracy, Reliability, Internalist Reliabilism, Frequentism, Classical Statistics, Model Selection, Data Analysis, Decision-Making


Reference: Hanti Lin, “Internalist Reliabilism in Statistics and Machine Learning: Thoughts on Jun Otsuka’s Thinking about Statistics” (2024).


Leave a Reply