Rationale Extraction with Knowledge Distillation (REKD) is a novel approach designed to improve the performance of Rationale Extraction (RE) models, particularly when these models are based on less capable or smaller neural networks. Rationale Extraction itself is an interpretable-by-design framework for deep neural networks, employing a select-predict architecture where two networks jointly learn feature selection (to identify "rationales") and prediction. However, learning to select optimal feature subsets with only remote supervision is computationally challenging, especially for smaller models. REKD addresses this by integrating knowledge distillation: a student RE model learns not only through its own optimization but also by mimicking the rationales and predictions of a more powerful "teacher" or "rationalist" model. This mechanism allows the student to leverage the teacher's expertise, thereby boosting its predictive performance and making interpretable AI more accessible for diverse applications.
REKD improves how smaller AI models explain their decisions by having them learn from a more expert AI model. It combines "rationale extraction," which helps AI show *why* it made a decision, with "knowledge distillation," where a small model learns from a big one, leading to better performance for the smaller, explainable AI.
Rationale Extraction with Knowledge Distillation
Was this definition helpful?