Barron space

From HandWiki

In functional analysis, the Barron space is a function space. It is a Banach space. It originated from the study of universal approximation properties of two-layer neural networks. It has applications in approximation theory and statistical learning theory.

It is named after Andrew R. Barron, who did work on the functional analysis of two-layer neural networks, though he did not define Barron space in his works.[1]

Setup

We quote the following universal approximation theorem:

Universal approximation theorem — Let C(X,m) denote the set of continuous functions from a subset X of a Euclidean n space to a Euclidean space m. Let σC(,). Note that (σx)i=σ(xi), so σx denotes σ applied to each component of x.

Then σ is not polynomial if and only if for every n, m, compact Kn, fC(K,m),ε>0 there exist k, Ak×n, bk, Cm×k such that supxKf(x)g(x)<ε where g(x)=C(σ(Ax+b))

In words, given a subset

Xn

, and a fixed activation function

σ

that is not a polynomial function, then any continuous function of type

Xm

can be approximated as a 2-layered neural network with a linear layer

(A,b)

, followed by the nonlinear activation

σ

, followed by another linear layer

C

. Furthermore, given any compact subset

KX

, the approximation can be arbitrarily good in the uniform norm.

Usually we only consider the case where the neural network has a single output, that is, the case where m=1, since for multiple outputs, the outputs can be separately approximated. We assume that m=1 for the rest of the article.

Number of hidden neurons

In the statement of the theorem, the middle layer is the hidden layer. The number k is the number of neurons in the hidden layer. These neurons are called hidden neurons.

Given a compact set Xn, to approximate a generic continuous function to an accuracy of ϵ in the uniform norm over X, O(ϵd) hidden neurons are needed. This is a manifestation of the curse of dimensionality.

In a 1993 paper, Barron showed that a large class of continuous functions are much more approximable than a generic continuous function. Specifically, he showed that there is a set of continuous functions such that, given any Borel probability measure μ, only O(ϵ2) hidden neurons are needed to approximate f to an accuracy of ϵ in the L2(μ) norm. In this sense, these functions are nice, in that they can be efficiently approximated by a neural network without hitting the curse of dimensionality.[2][3]

Definition

It is natural to consider the infinite width limit, where the summation turns into an integral:f(x):=cσ(aTx+b)ρ(da,db,dc)where a,b,c takes values in n,,, and ρ is a probability distribution over n××.

Different ρ may lead to the same f. That is, the representation of a function as an infinite-width neural network is not unique. However, among these, one may be selected as having lowest regularization loss, as usual in statistical learning theory.

ReLU activation

If σ is the ReLU function, define as follows.

For any p[1,], define the regularization loss of representation ρ asρp:=𝔼(a,b,c)ρ[|c(a1+|b|)|p]1/pif p[1,), andρp:=sup(a,b,c)suppρ|c(a1+|b|)|if p=. This is defined in analogy to Lp spaces, and is motivated by the previous result on the number of hidden neurons.

The special case of p=1 is also called the path norm, since it is interpreted as the path weight c(a1+|b|), averaged across all paths from the inputs to the output of the neural network.

The p-Barron norm of f is defined asfBp:=infρ represents fρpThe p-Barron space over Xn is the set of continuous functions of type X of finite p-Barron norm.

It can be proven that if fBp< for some p[1,], then fBp is the same for all p[1,]. Therefore, all these are the same, and we drop the p and call all of Bp the same Barron norm, and the space of them the Barron space. It is written as [1]

The Barron space is a Banach space.[3]: Thm. 2.3 

Non-ReLU activation

If σ is not the ReLU function, then define the p-extended Barron norm:ρp:=𝔼(a,b,c)ρ[|c(a1+|b|+1)|p]1/pfB~p:=infρ represents fρpSimilarly, define the p-extended Barron spaces.

In general, they are not the same for different values of p.

Multilayer version

There is a generalization for multilayer neural networks with ReLU activations.[4]

Properties

Basic properties

Theorem.[1]: Thm. 1  Given f, then for any positive integer k, there exists a two-layer ReLU network with M hidden neuronsfM(x)=1Mi=1MciReLU(aix+bi)such that fMfL2(Ω)23fB2M, and 1Mi=1M|ci|(ai1+|bi|)2fB.

Theorem.[1]: Thm. 2  (converse to the previous theorem) For any f continuous on Xn, if there exists a sequence of two-layer ReLU networks fM with M=1,2, hidden neurons, converging fMf pointwise, and the Barron norms of these fM are uniformly bounded by a single C:1Mi=1M|ci|(ai1+|bi|)C,M=1,2,3,then f and fBC.

Harmonic analysis

Theorem. For any f continuous on Xn, defineγ(f):=inff^nω12|f^(ω)|dωwhere f^ ranges over Fourier transformations of all possible extensions of f to all of n, then, if γ(f)<, then f.

Furthermore, we have the explicit upper bound:[1]: Prop. 2 f2γ(f)+2f(0)1+2|f(0)|

Statistical learning theory

Let S={z1,z2,,zm}Z be a sample of points and consider a function class of real-valued functions over Z. Then, the empirical Rademacher complexity of given S is defined as:

RadS()=1m𝔼σ[supf|i=1mσif(zi)|]

Theorem.[1]: Thm. 3  For any C>0, let C:={f:fBC}, then RadS(C)2C2ln(2n)|S|, where as a reminder n is the number of dimensions of the domain of f.

This result shows that the space of functions bounded in Barron norm has low Rademacher complexity, which according to statistical learning theory, means they are highly learnable. This rhymes with the fact that they are well approximable by a network with few hidden neurons.

See also

References

  1. 1.0 1.1 1.2 1.3 1.4 1.5 E, Weinan; Ma, Chao; Wu, Lei (2022-02-01). "The Barron Space and the Flow-Induced Function Spaces for Neural Network Models" (in en). Constructive Approximation 55 (1): 369–406. doi:10.1007/s00365-021-09549-y. ISSN 1432-0940. https://doi.org/10.1007/s00365-021-09549-y. 
  2. Barron, A.R. (May 1993). "Universal approximation bounds for superpositions of a sigmoidal function". IEEE Transactions on Information Theory 39 (3): 930–945. doi:10.1109/18.256500. ISSN 0018-9448. Bibcode1993ITIT...39..930B. 
  3. 3.0 3.1 E., Weinan; Wojtowytsch, Stephan (April 2022). "Representation formulas and pointwise properties for Barron functions" (in en). Calculus of Variations and Partial Differential Equations 61 (2). doi:10.1007/s00526-021-02156-6. ISSN 0944-2669. https://link.springer.com/10.1007/s00526-021-02156-6. 
  4. E, Weinan; Wojtowytsch, Stephan (2020-07-30), On the Banach spaces associated with multi-layer ReLU networks: Function representation, approximation theory and gradient descent dynamics, http://arxiv.org/abs/2007.15623