Artificial intelligence explanations make complex predictive models more comprehensible. Effective explanations, however, should also anticipate and mitigate possible misinterpretations, e.g., arising when users infer incorrect information that is not explicitly conveyed. To this end, we propose complementary explanations -- a novel method that pairs explanations to compensate for their respective limitations. A complementary explanation adds insights that clarify potential misconceptions stemming from the primary explanation while ensuring their coherence and avoiding redundancy. We also introduce a framework for designing and evaluating complementary explanation pairs based on pertinent qualitative properties and quantitative metrics. Applying our approach allows to construct complementary explanations that minimise the chance of their misinterpretation.
翻译:暂无翻译