Counterfactual explanations are a post-hoc method that reveal how a model's input would need to change to produce a desired, different output. They provide 'what if' scenarios, making opaque machine learning decisions more understandable and actionable for users.
Counterfactual explanations help people understand why an AI model made a certain decision by showing what minimal changes to the input would have led to a different outcome. This provides actionable advice, making AI decisions more transparent and allowing users to understand how to achieve a desired result.
CFEs, what-if explanations, actionable explanations
Was this definition helpful?