Nolte, GerritSchlüter, MaximilianMurtovi, AlnisSteffen, Bernhard2024-11-112024-11-112023-04-21http://hdl.handle.net/2003/42741http://dx.doi.org/10.17877/DE290R-24573TADS are a novel, concise white-box representation of neural networks. In this paper, we apply TADS to the problem of neural network verification, using them to generate either proofs or concise error characterizations for desirable neural network properties. In a case study, we consider the robustness of neural networks to adversarial attacks, i.e., small changes to an input that drastically change a neural networks perception, and show that TADS can be used to provide precise diagnostics on how and where robustness errors a occur. We achieve these results by introducing Precondition Projection, a technique that yields a TADS describing network behavior precisely on a given subset of its input space, and combining it with PCA, a traditional, well-understood dimensionality reduction technique. We show that PCA is easily compatible with TADS. All analyses can be implemented in a straightforward fashion using the rich algebraic properties of TADS, demonstrating the utility of the TADS framework for neural network explainability and verification. While TADS do not yet scale as efficiently as state-of-the-art neural network verifiers, we show that, using PCA-based simplifications, they can still scale to medium-sized problems and yield concise explanations for potential errors that can be used for other purposes such as debugging a network or generating new training samples.en(rectifier) neural networks(piece-wise) affine functionsdecision treesexplainabilityverificationrobustnessprincipal component analysisdiagnosticsdigit recognitionMNIST004The power of typed affine decision structures: a case studyText