# SHAP - @lundbergUnifiedApproachInterpreting2017 - help users interpret the predictions of complex models - unclear how these methods are related and when one method is preferable over another - unified framework for interpreting predictions - SHAP - SHapley Additive exPlanations - game theoretic approach to explain the output of any machine learning model - connects optimal credit allocation with local explanations using the classic Shapley values from game theory and their related extensions - assigns each feature an importance value for a particular prediction - identification of a new class of additive feature importance measures - theoretical results showing there is a unique solution in this class with a set of desirable properties - notable because several recent methods in the class lack the proposed desirable properties - present new methods that show improved computational performance and/or better consistency with human intuition than previous approaches