Skip to main content
Cornell University
Learn about arXiv becoming an independent nonprofit.
We gratefully acknowledge support from the Simons Foundation, member institutions, and all contributors. Donate
arxiv logo > cs > arXiv:2502.10533v3

Help | Advanced Search

arXiv logo
Cornell University Logo

quick links

  • Login
  • Help Pages
  • About

Computer Science > Machine Learning

arXiv:2502.10533v3 (cs)
[Submitted on 14 Feb 2025 (v1), last revised 2 Mar 2026 (this version, v3)]

Title:Identity-Free Deferral For Unseen Experts

Authors:Joshua Strong, Pramit Saha, Yasin Ibrahim, Cheng Ouyang, Alison Noble
View a PDF of the paper titled Identity-Free Deferral For Unseen Experts, by Joshua Strong and 4 other authors
View PDF
Abstract:Learning to Defer (L2D) improves AI reliability in decision-critical environments by training AI to either make its own prediction or defer the decision to a human expert. A key challenge is adapting to unseen experts at test time, whose competence can differ from the training population. Current methods for this task, however, can falter when unseen experts are out-of-distribution (OOD) relative to the training population. We identify a core architectural flaw as the cause: they learn identity-conditioned policies by processing class-indexed signals in fixed coordinates, creating shortcuts that violate the problem's inherent permutation symmetry. We introduce Identity-Free Deferral (IFD), an architecture that enforces this symmetry by construction. From a few-shot context, IFD builds a query-independent Bayesian competence profile for each expert. It then supplies the deferral rejector with a low-dimensional, role-indexed state containing only structural information, such as the model's confidence in its top-ranked class and the expert's estimated skill for that same role, which obscures absolute class identities. We train IFD using an uncertainty-aware, context-only objective that removes the need for expensive query-time expert labels. We formally prove the permutation invariance of our approach, contrasting it with the generic non-invariance of standard population encoders. Experiments on medical imaging benchmarks and ImageNet-16H with real human annotators show that IFD consistently improves generalisation to unseen experts, with gains in OOD settings, all while using fewer annotations than alternative methods.
Comments: Fourteenth International Conference on Learning Representations (ICLR) 2026
Subjects: Machine Learning (cs.LG); Human-Computer Interaction (cs.HC)
Cite as: arXiv:2502.10533 [cs.LG]
  (or arXiv:2502.10533v3 [cs.LG] for this version)
  https://doi.org/10.48550/arXiv.2502.10533
arXiv-issued DOI via DataCite

Submission history

From: Joshua Strong [view email]
[v1] Fri, 14 Feb 2025 19:59:25 UTC (1,349 KB)
[v2] Sat, 24 May 2025 17:50:44 UTC (8,592 KB)
[v3] Mon, 2 Mar 2026 12:59:23 UTC (1,105 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled Identity-Free Deferral For Unseen Experts, by Joshua Strong and 4 other authors
  • View PDF
  • TeX Source
license icon view license
Current browse context:
cs.LG
< prev   |   next >
new | recent | 2025-02
Change to browse by:
cs
cs.HC

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • Click here to contact arXiv Contact
  • Click here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status