View a PDF of the paper titled The Boundaries of Verifiable Accuracy, Robustness, and Generalisation in Deep Learning, by Alexander Bastounis and 7 other authors
Abstract:In this work, we assess the theoretical limitations of determining guaranteed stability and accuracy of neural networks in classification tasks. We consider classical distribution-agnostic framework and algorithms minimising empirical risks and potentially subjected to some weights regularisation. We show that there is a large family of tasks for which computing and verifying ideal stable and accurate neural networks in the above settings is extremely challenging, if at all possible, even when such ideal solutions exist within the given class of neural architectures.
Submission history
From: Ivan Tyukin [view email]
[v1]
Wed, 13 Sep 2023 16:33:27 UTC (33 KB)
[v2]
Thu, 21 Nov 2024 14:10:03 UTC (33 KB)
Source link
lol