The Interpretability Tax on Tabular Foundation Models

The Interpretability Tax on Tabular Foundation Models

Mindful Modeler
Mindful ModelerMar 24, 2026

Key Takeaways

  • PFI works on TFMs but inference cost spikes dramatically
  • Batching predictions cuts TFM interpretability runtime
  • LOCO becomes as cheap as PFI under TFMs
  • Training‑cheap, inference‑expensive flips XAI economics
  • Future XAI may favor model‑specific or batch‑optimized methods

Summary

The post examines how classic model‑agnostic interpretability tools, such as permutation feature importance (PFI) and LOCO, operate on tabular foundation models (TFMs). While these methods function out‑of‑the‑box, TFMs flip the traditional cost balance: training is cheap but inference is expensive, making PFI up to 26‑times slower than on conventional models. By batching predictions and leveraging the training‑cheap nature of TFMs, the author shows that LOCO can match PFI’s runtime, reshaping the economics of explainability. The piece concludes that interpretability for TFMs will likely evolve toward batch‑optimized or model‑specific techniques.

Pulse Analysis

Tabular foundation models represent a paradigm shift in machine‑learning pipelines, moving from heavy training phases to lightweight, in‑context inference. This inversion of costs—where model weights are pre‑trained once and every downstream prediction incurs substantial compute—poses a fresh challenge for explainable AI (XAI). Traditional post‑hoc tools like permutation feature importance (PFI) rely on repeated forward passes, which, on TFMs, can be dozens of times slower than on classic algorithms such as random forests. Practitioners therefore face a trade‑off between model transparency and operational expense, especially when scaling to larger datasets or real‑time applications.

Recent research highlighted in the post demonstrates practical workarounds that align XAI methods with the inference‑heavy nature of TFMs. By aggregating test instances into larger batches, the number of API calls to the model drops dramatically, turning a naïve two‑call PFI routine into a single, memory‑efficient request. This batching strategy not only reduces wall‑clock time but also leverages the parallel processing strengths of modern GPUs. Moreover, the training‑cheap characteristic of TFMs makes re‑training‑based metrics like LOCO surprisingly competitive, as “re‑training” merely becomes another forward pass with a column removed, erasing the traditional overhead gap.

The broader implication is a re‑evaluation of XAI toolkits for foundation‑model‑driven analytics. As inference costs dominate, developers may prioritize batch‑optimized or model‑specific explainers—think TreeSHAP analogues for PFNs—over generic, compute‑intensive approaches. Organizations adopting TFMs must therefore embed cost‑aware interpretability pipelines into their governance frameworks, ensuring that transparency does not become a prohibitive expense. This evolving landscape signals a new era where explainability is engineered alongside model architecture, rather than retrofitted after deployment.

The interpretability tax on tabular foundation models

Comments

Want to join the conversation?