This page depicts the results of the experiments conducted within the scope of my master's thesis which I completed at the Machine Learning Group TU Berlin under the supervision of K. R. Müller and Grégoire Montavon. The abstract and a brief overview of the research process are provided in the following. For a comprehensive overview of the methodologies implemented in this work, please refer to the report. Code and report are linked below.
As nonlinear Machine Learning (ML) models are increasingly used in various real world applications, their black-box nature hinders in-depth model evaluation, apart from performance measures. In response, the field of Explainable Artificial Intelligence (XAI) has made much progress. It aims to reveal the rationale behind complex ML models, often by assigning relevance scores to model parts and input features, e.g., pixels. However, in some domains such as audio processing, where data—like time or time-frequency representations of amplitudes—is of rather unintuitive nature, the extracted explanations can be hard to grasp for humans. Suitably, a novel sub-field in the realm of XAI has emerged in very recent time that aims to decompose explanations into multiple sub-explanations, representing distinct decision concepts. These approaches offer a promising foundation to gain more valuable insights into models and the data domain, especially when dealing with complex data scenarios. This study targets the extraction of concept-based explanations for a neural network applied to audio classification tasks, by utilizing the newly proposed method, Disentangled Relevant Subspace Analysis (DRSA), in combination with Layerwise Relevance Propagation (LRP).