Skip to main content
U.S. flag

An official website of the United States government

Achieving interpretable machine learning by functional decomposition of black-box models into explainable predictor effects

December 10, 2025

Machine learning (ML) models are often based on complex black-box architectures that are difficult to interpret. This interpretability problem can hinder the use of ML in fields like medicine, ecology, and insurance, and has boosted research in interpretable machine learning (IML). Here, we propose a novel approach for the functional decomposition of black-box predictions, which is a core concept of IML. This approach replaces the prediction function with a surrogate model consisting of simpler subfunctions, providing insights into the direction and strength of the main feature contributions and their interactions. Our method is based on a concept termed “stacked orthogonality”, which ensures that the main effects capture as much functional behavior as possible. To compute the subfunctions, we combine neural additive modeling with an efficient post-hoc orthogonalization procedure. Our method yielded plausible results in an analysis of stream biological condition in the Chesapeake Bay watershed (United States).

Publication Year 2025
Title Achieving interpretable machine learning by functional decomposition of black-box models into explainable predictor effects
DOI 10.1038/s44387-025-00033-7
Authors David Kohler, David Rügamer, Lindsey J. Boyle, Kelly O. Maloney, Matthias Schmid
Publication Type Article
Publication Subtype Journal Article
Series Title npj Artificial Intelligence
Index ID 70273156
Record Source USGS Publications Warehouse
USGS Organization Eastern Ecological Science Center
Was this page helpful?