Vocabulary for Universal Approximation: A Linguistic Perspective of Mapping Compositions

05/20/2023
by   Yongqiang Cai, et al.
0

In recent years, deep learning-based sequence modelings, such as language models, have received much attention and success, which pushes researchers to explore the possibility of transforming non-sequential problems into a sequential form. Following this thought, deep neural networks can be represented as composite functions of a sequence of mappings, linear or nonlinear, where each composition can be viewed as a word. However, the weights of linear mappings are undetermined and hence require an infinite number of words. In this article, we investigate the finite case and constructively prove the existence of a finite vocabulary V={ϕ_i: ℝ^d →ℝ^d | i=1,...,n} with n=O(d^2) for the universal approximation. That is, for any continuous mapping f: ℝ^d →ℝ^d, compact domain Ω and ε>0, there is a sequence of mappings ϕ_i_1, ..., ϕ_i_m∈ V, m ∈ℤ_+, such that the composition ϕ_i_m∘ ... ∘ϕ_i_1 approximates f on Ω with an error less than ε. Our results provide a linguistic perspective of composite mappings and suggest a cross-disciplinary study between linguistics and approximation theory.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset
Success!
Error Icon An error occurred

Sign in with Google

×

Use your Google Account to sign in to DeepAI

×

Consider DeepAI Pro