View a PDF of the paper titled On the Local Complexity of Linear Regions in Deep ReLU Networks, by Niket Patel and 1 other authors
Abstract:We define the local complexity of a neural network with continuous piecewise linear activations as a measure of the density of linear regions over an input data distribution. We show theoretically that ReLU networks that learn low-dimensional feature representations have a lower local complexity. This allows us to connect recent empirical observations on feature learning at the level of the weight matrices with concrete properties of the learned functions. In particular, we show that the local complexity serves as an upper bound on the total variation of the function over the input data distribution and thus that feature learning can be related to adversarial robustness. Lastly, we consider how optimization drives ReLU networks towards solutions with lower local complexity. Overall, this work contributes a theoretical framework towards relating geometric properties of ReLU networks to different aspects of learning such as feature learning and representation cost.
Submission history
From: Niket Patel [view email]
[v1]
Tue, 24 Dec 2024 08:42:39 UTC (25,435 KB)
[v2]
Wed, 25 Dec 2024 02:14:07 UTC (25,435 KB)
Source link
lol