Know When to Abstain: Optimal Selective Classification with Likelihood Ratios

Researchers have applied the Neyman-Pearson lemma to selective classification, framing optimal prediction abstention as a likelihood ratio test. This approach unifies existing methods and inspires novel techniques that outperform current baselines, particularly under covariate shift where test data differs from training data. The work demonstrates improved reliability for AI systems from supervised learning to vision-language models.

Know When to Abstain: Optimal Selective Classification with Likelihood Ratios

Neyman-Pearson Lemma Revives Selective Classification for AI Reliability Under Real-World Shifts

In a significant advancement for trustworthy AI, researchers have revisited a foundational statistical principle—the Neyman-Pearson lemma—to fundamentally improve selective classification. This technique allows machine learning models to abstain from making predictions when they are uncertain, thereby enhancing reliability. The new work demonstrates that framing optimal selection as a likelihood ratio test not only unifies existing methods but also inspires novel, more robust approaches, particularly for the challenging and realistic scenario of covariate shift, where test data differs from training data.

The study, detailed in the preprint arXiv:2505.15008v3, proposes that the optimal rule for a model to reject a prediction is to compare the likelihood of the data under competing hypotheses. This statistical lens provides a rigorous framework for designing selection functions, moving beyond heuristic baselines. The authors show this perspective motivates new methodologies that consistently outperform current techniques across diverse tasks, from supervised learning to modern vision-language models.

Bridging Classical Statistics and Modern AI Reliability

The core innovation lies in applying the decades-old Neyman-Pearson lemma, a cornerstone of statistical hypothesis testing, to the contemporary problem of predictive uncertainty in AI. The lemma formally proves that the most powerful test for distinguishing between two hypotheses is based on a likelihood ratio. By treating the decision to "predict" or "abstain" as a hypothesis test, the researchers derive optimal selection rules.

This approach provides a unifying theory for several post-hoc selection baselines, such as those based on prediction confidence or entropy, revealing they are special cases or approximations of the likelihood ratio test. More importantly, it directly motivates new, theoretically-grounded selection functions that are simpler and more effective, especially when models face unfamiliar data distributions.

Conquering the Covariate Shift Challenge

A central contribution of the work is its focus on covariate shift, a pervasive real-world problem where the input data distribution changes between training and deployment—a scenario that remains underexplored in selective classification. Under such shifts, standard confidence scores can become miscalibrated and unreliable.

The proposed likelihood ratio methods inherently account for distributional differences. By evaluating how "likely" a new input is under the model's understanding versus an alternative, the system can more accurately gauge true uncertainty. The authors rigorously evaluated their methods on a range of vision and language tasks under induced covariate shifts, demonstrating superior performance over existing selective classification baselines.

Experimental Validation and Open-Source Release

The empirical evaluation spans traditional supervised learning settings and contemporary vision-language models (VLMs), testing the robustness of the new selection rules. In experiments, the Neyman-Pearson-informed methods consistently achieved a better trade-off between coverage (the fraction of samples on which a prediction is made) and risk (the error rate on those predictions) compared to existing approaches.

To foster reproducibility and further research, the team has released their code publicly. The repository, available at https://github.com/clear-nus/sc-likelihood-ratios, provides implementations of the proposed methods, facilitating adoption and extension by the broader AI research community working on model reliability and safe deployment.

Why This Matters for AI Deployment

  • Enhances Real-World Reliability: By providing a robust mechanism for abstention under covariate shift, this work directly addresses a major hurdle for deploying AI in dynamic, non-stationary environments.
  • Unifies Theory and Practice: It bridges classical statistical theory with modern machine learning, offering a principled, theoretically-sound foundation for the often-heuristic practice of selective classification.
  • Improves Trust in Complex Models: The methods' effectiveness on large-scale vision-language models is particularly significant, as it offers a path to make these powerful but sometimes unpredictable systems more dependable.
  • Opens New Research Avenues: The likelihood ratio framework provides a clear direction for developing next-generation uncertainty quantification and rejection techniques for AI.

常见问题