What should a model do when it is uncertain: answer anyway, or hand the decision to someone better placed to act? Learning-to-Defer studies exactly this question. Rather than forcing a model to predict on every input, it allows the model to choose between acting and deferring to an expert, a stronger model, or a tool.

In practice, this turns prediction into a routing problem: for each query, decide whether a lightweight model should respond, or whether the query should be routed elsewhere. The objective is not only accuracy, but also the right balance between risk, cost, and reliability.
Where should a model abstain? The toy example below shows a fixed linear decision rule together with a dense mixed pocket in the upper region of the plot.
Draw with your mouse, finger, or pen. Points inside your region are deferred and removed from the accuracy calculation. The goal is to keep coverage high while improving accuracy.
Deferred 0 of 0 points.
My research studies learning-to-defer from both a theoretical and a practical perspective. I am interested in when deferral rules are statistically well-founded, how to design surrogate losses with formal guarantees, and how these systems behave in more realistic settings: multiple experts, limited feedback, robustness constraints, and changing environments.
More concretely, my work asks questions such as: when should a model defer, to whom should it defer, and how should that decision change when experts have different costs, different strengths, or different availability? These questions appear naturally in human-AI systems, model cascades, and resource-constrained decision pipelines.
For concrete results and papers, see the Detailed Publications page.