An Embedding of ReLU Networks and an Analysis of their Identifiability

07/20/2021
by   Pierre Stock, et al.
0

Neural networks with the Rectified Linear Unit (ReLU) nonlinearity are described by a vector of parameters θ, and realized as a piecewise linear continuous function R_θ: x ∈ℝ^d↦ R_θ(x) ∈ℝ^k. Natural scalings and permutations operations on the parameters θ leave the realization unchanged, leading to equivalence classes of parameters that yield the same realization. These considerations in turn lead to the notion of identifiability – the ability to recover (the equivalence class of) θ from the sole knowledge of its realization R_θ. The overall objective of this paper is to introduce an embedding for ReLU neural networks of any depth, Φ(θ), that is invariant to scalings and that provides a locally linear parameterization of the realization of the network. Leveraging these two key properties, we derive some conditions under which a deep ReLU network is indeed locally identifiable from the knowledge of the realization on a finite set of samples x_i∈ℝ^d. We study the shallow case in more depth, establishing necessary and sufficient conditions for the network to be identifiable from a bounded subset 𝒳⊆ℝ^d.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset