Preprint
Article

Convergence Rate of Regularized Regression Associated with Zonal Translation Networks

Altmetrics

Downloads

73

Views

23

Comments

0

A peer-reviewed article of this preprint also exists.

This version is not peer-reviewed

Submitted:

04 August 2024

Posted:

05 August 2024

You are already at the latest version

Alerts
Abstract
Neural network regularized learning has garnered significant attention in recent years. We give a systematic investigation on the performance of regularized regression associated zonal translation networks. We propose the concept of Marcinkiewicz-Zygmund inequality Setting (MZIS) for the scattered nodes collected from the unit sphere. We show that, under the MZIS, the corresponding convolutional zonal translation network has reproducing property. Based on these facts,we propose a kind of kernel regularized regression learning framework and provide upper bound estimate for the learning rate with the kernel approach. We also give proof for the density of the zonal translation network with spherical Fourier analysis.We provide the approximation error with a K-functional.
Keywords: 
Subject: Computer Science and Mathematics  -   Artificial Intelligence and Machine Learning

MSC:  41A25

1. Introduction

It is known that convolutional neural network provides various models and algorithms to process data model in many fields such as computer vision(see e.g. [1]), natural lagrange processing (see e.g. [2]), and sequence analysis in bio-informatics(see e.g. [3]). The regularized neural network learning has thus become an attractive research topic (see e.g.[4,5,6,7,8,9]). In the present paper, we shall give theory analysis for the convergence rate of regularized regression associated with zonal translation network on the unit sphere.
Let X be a compact subset in the d-dimensional Euclidean space R d with the usual norm x = k = 1 d x k 2 for x = ( x 1 , x 2 , , x d ) R d and Y be a nonempty closed subset contained in [ M , M ] for a given M > 0 . The aim of the regression learning problem is to learn the target function which describes the relationship between the input x X and the output y Y from a hypothesis function space. In most of the cases, the target function is offered as a set of observations z = { z i } i = 1 m = { ( x i , y i ) } i = 1 m Z m which has been drawn independently and identically distributed (i.i.d.) according to a probability joint distribution (measure) ρ ( x , y ) = ρ X ( x ) ρ ( y | x ) on Z = X × Y , where ρ ( y | x ) ( x X ) is the conditional probability of y for a given x and ρ X ( x ) is the marginal probability about x, i.e.,for every integrable function φ : X × Y R there holds
X × Y φ ( x , y ) d ρ = X Y φ ( x , y ) d ρ ( y x ) d ρ X .
For a given normed space ( B , · B ) consisting of real functions on X we define the regularized learning framework with B as
f z , λ : = a r g min f B E z ( f ) + λ 2 f B 2 ,
where λ > 0 is the regularization parameter, E z ( f ) is the empirical mean
E z ( f ) = 1 m i = 1 m ( y i f ( x i ) ) 2 .
The optimal target function is the regression function
f ρ ( x ) = Y y d ρ ( y | x )
satisfying
f ρ = inf f E ρ ( f ) ,
where E ρ ( f ) = Z ( y f ( x ) ) 2 d ρ and the inf is taken over all the ρ X -measurable functions f. Moreover, there holds the famous equality (see e.g.[10])
f f ρ L 2 ( ρ X ) 2 = E ρ ( f ) E ρ ( f ρ ) , f L 2 ( ρ X ) .
The choices for the hypothesis space B in (1) are riches. For example, C.P.An et al choose the algebraic polynomial class as B (see [11,12,13],). In [14], C. De Mol et al chose the dictionary as B. Recently some papers chose the Sobolev space as the hypothesis space B (see [15,16]). By kernel method we mean traditionally replacing B with a reproducing kernel Hilbert space (RKHS) ( H K , · , · K ) which is a Hilbert space consisting of real functions defined on X and there is a Mercer kernel K x ( y ) = K ( x , y ) on X × X (i.e., K x ( y ) is a continuous and symmetric function on X × X and for any n 1 and any { x 1 , x 2 , , x n } X the Mercer matrices ( K ( x i , x j ) ) i , j = 1 , 2 , , n are positive semi-definite) such that
f ( x ) = f , K x K , x X , f H K .
and there hold the embeded inequality
| f ( x ) | c f K , x X , f H K ,
where c is a constant independent of f and x. There are two results about the optimal solution f z , λ . The reproducing property (3) yields the representation
f z , λ ( x ) = k = 1 m c k K x k ( x ) , x X .
The embeded inequality (4) yields the inequality
| f z , λ ( x ) | 2 M λ x X .
Representation (5) is the theory basis for kernel regularized regression (see e.g.[17,18]). Inequality (6) is the key inequality for bounding the learning rate with covering number method (see e.g.[19,20,21]). For other skills of the kernel method one can cite [10,22,23,24] et al.
It is particularly important to mention here that translation networks have recently been used for the hypothesis space of regularized learning (see e.g. [25,26]). From the view of approximation theory, a simple single layer translation network with m neurons is a function space produced by translating a given function ϕ which can be written as
Δ ϕ , X ¯ X = { i = 1 m c i T x i ( ϕ , · ) : c i R , x i X , i = 1 , 2 , , m } ,
where X ¯ = { x i } i = 1 m X is a given node set, and for a given x X T x ( ϕ , y ) is a translation operator corresponding to X. For example, when X = R d and X = [ π , π ] d , we choose T x ( ϕ , y ) as the usual convolution translation operator ϕ ( x y ) for the ϕ defined on R d or an 2 π -periodic function ϕ (see [27,28]. When X = S d 1 = { x R d : x = 1 } is the unit sphere in R d , one can choose T x ( ϕ , y ) as the zonal translation operator ϕ ( x y ) for a given ϕ defined on the interval [ 1 , 1 ] (see [29]). In [30], we defined a kind of translation operator T x ( ϕ , y ) for X = [ 1 , 1 ] .
It is easy to see that the approximation ability and the construction of a translation network depend upon the nodes set X ¯ = { x i } i = 1 m (see e.g. [31,32,33]). On the other hand, according to the view of [34], the quadrature rule and the Marcinkiewicz-Zygmund (M-Z) inequality associated with X ¯ also influence the construction of the translation network Δ ϕ , X ¯ . For a bounded closed set Ω R d with measure d ω satisfying Ω d ω = V < + . We denote by P n L 2 ( Ω ) the linear space of polynomials on Ω of degree at most n, equipped with the L 2 -inner product v , z = Ω v z d ω . The m-point quadrature rule (QR) is
Ω g d ω j = 1 m w j g ( x j ) ,
where X ¯ = { x j } j = 1 m Ω and weights w j are all positive for j = 1 , 2 , , m . We say the QR (7) has polynomial exactness n if
Ω g d ω = j = 1 m w j g ( x j ) , g P n .
The Marcinkiewicz-Zygmund (MZ) inequality based on a set X ¯ = { x j } j = 1 m Ω is
j = 1 m w j | g ( x j ) | 2 1 2 Ω | g ( ω ) | 2 d ω 1 2 , g P n ,
where the weights w j in (9) may be not the same as the w j in (7) and (8).
Accord to the view of [34], the quadrature rule (QR) follows automatically from the Marcinkiewicz-Zygmund (MZ) inequality. H.N.Mhaskar et al gave a method of transition from MZ inequality to polynomial exact GR in [35]. So the MZ inequality (9) is an important features for describing the nodes set X ¯ . Since this reason the node set X ¯ = { x j } j = 1 m Ω which yields an MZ inequality is given a specially terminology called Marcinkiewicz-Zygmynd Family (MZF) (see [34,36,37,38]). However, from these literatures we know that the MZFs do not totally coincide with the Lagrange interpolation nodes in the case of d > 1 . The hyperinterpolations are then developed with the help of exact QR (see [39,40,41,42,43]) and are applied to approximation theory and regularized learning (see e.g.[11,12,13,44]). On the other hand, we find the problem of polynomial exact QR is investigated individually (see e,g, [45,46]). The concept of spherical t-design was first defined in [47] and is given investigations by many papers subsequently, one can see the classical references [48,49]. We say T t = { x i } i = 1 | T t | S d 1 is a spherical t-design if
1 | T t | i = 1 | T t | π ( x i ) = 1 ω d 1 S d 1 π ( x ) d ω ( x ) ,
where ω d 1 is the volume of S d 1 and π ( x ) is a spherical polynomial with degree t. Moreover, in many applications the polynomial exact QR and the MZFS has been used as assumptions. For example, C.P.An et al gave approximation order for the hyperinterpolation approximation under the assumptions that (8), (10) and the MZ inequality (9) hold (see [50,51]). Also, in [25], Lin et al gave investigations on regularized regression associated with zonal translation network by assuming the nodes set X ¯ = { x j } j = 1 m S d 1 is a type of spherical t-design.
Polynomial exact QR is also a good tool in approximation theory. For example, we had ever used it to bound the norm for some Mercer matrices (see [52,53,54,55]). In particular, H.N.Mhaskar et al use polynomial exact QR to construct the first periodic translation operators (see [27]) and the zonal translation network operators (see [29]). Along this line, the translation operators defined on the unit ball, on the Euclidean space R d and on the interval [ 1 , 1 ] are constructed (see [28,30,56]).
Above investigations encourage us to use (8), (10) and (9) as hypothetical conditions to describe the approximation ability of zonal translation networks Δ ϕ , X ¯ X . To ensure the single layer translation network Δ ϕ , X ¯ X can approximation the constant function, Δ ϕ , X ¯ X is modified as
N ϕ , X ¯ X = { i = 1 m c i T x i ( ϕ , · ) + c 0 : c 0 , c i R , x i X , i = 1 , 2 , , m } .
In the case of T y ( ϕ , y ) = σ ( w x + b ) and w , x R d , b R , R.D.Nowak et al used (11) to design regularized learning frameworks (see [57]). An algorithm is provided by S.B.Lin et al [26] for designing such kind of network and is applied to construct regularized learning algorithms. In [5] N ϕ , X ¯ X is used to construct deep neural network learning frameworks. The same type of investigations are given in [58,59] and [60].
In the present paper, we shall design the translation networks N ϕ , X ¯ X by taking X = S d 1 , assuming X ¯ = Ω ( m ) = { x j } j = 1 m S d 1 satisfies equalities (8) and (9), and choosing the zonal translation T x ( ϕ , y ) = ϕ ( x y ) with ϕ being a given integrable function ϕ on [ 1 , 1 ] . Under these assumptions we shall provide a learning framework with N ϕ , Ω ( m ) S d 1 as the hypothesis space and give error analysis.
The contributions of the present paper are two folds. First, after absorbing the ideas of [34,36,37,38] and the successful experience of [11,25,27,29,50,51,61,62], we propose the concept of Marcinkiewicz-Zygmund Inequality Setting (MZIS) for the scattered nodes on the sphere unit, based on this as an assumption we show the reproducing property for the convolutional zonal translation network associated with the scattered nodes Ω ( m ) . Second, we give investigation on the kernel regularized neural network learning by combining classical the kernel approach and the convex analysis method, according to this method, the convergence rate given can be dimensional independent and capacity independent. Since the translation networks are produced by the zonal translations of convolutional kernels, we call them the convolutional zonal translation networks.
The paper is organized as follows. In Section 2 we first show the density for the zonal translations class and then show the reproducing property for the translation network Δ ϕ , Ω ( m ) S d 1 . In Section 3, we shall provide some results in the present paper, for example, a new regression learning framework and a learning setting, the error decomposition for the error analysis,and an estimate for the convergence rate. In Section 4, we shall give some lemmas which are used to prove the main results. The proofs for all the theorems and propositions are given in Section 5.
Throughout the paper, we write A = O ( B ) if there is a positive constant C independent of A and B such that A C B .In particular, by A = O ( 1 ) we show A is a bounded quantity. We write A B if both A = O ( B ) and B = O ( A ) .

2. Some Properties of the Translation Network on the Unit Sphere

Let ϕ L W η 1 = { ϕ : ϕ 1 , W η = 1 1 | ϕ ( x ) | W η ( x ) d x < + } , W η ( x ) = ( 1 x 2 ) η 1 2 , η > 1 2 . Then H.N.Mhaskar et al constructed in [29] a sequence of approximation operators to show that the zonal translation class
Δ ϕ S d 1 = c l { ϕ ( x y ) : y S d 1 } { 1 } = c l { i = 1 m c i ϕ ( x i · ) + c 0 : c 0 , c i R , x i S d 1 , i = 1 , 2 , , m ; m = 1 , 2 , }
is density in L p ( S d 1 ) ( 1 p + ) if a l η ( ϕ ) ^ 0 for all l = 0 , 1 , 2 , , where
a l η ( ϕ ) ^ = c η 1 1 ϕ ( x ) C l η ( x ) C l η ( 1 ) W η ( x ) d x , η = d 2 2
and C n η ( x ) = p n ( η 1 2 , η 1 2 ) ( x ) is the n-th Legendre polynomial satisfies the orthogonal relation
c η 1 1 C n η ( x ) C m η ( x ) W η ( x ) d x = h n η δ n , m .
with c η = Γ ( η + 1 ) Γ ( 1 2 ) Γ ( η + 1 2 ) , h n η = η n + η C n η ( 1 ) and it is known that (see it from (B.2.1),(B.2.2)and (B.5.1) of [63]) C n η ( x ) C n η ( 1 ) = n 2 η 1 , x [ 1 , 1 ] . It follows that
ϕ ( t ) = l = 0 + a l η ( ϕ ) ^ n + η η C l η ( t ) = l = 0 + a l η ( ϕ ) ^ Z l η ( t ) ,
where Z l η ( t ) = l + η η C l η ( t ) , η = d 2 2 .
For a given real number p 1 ,we denote by L p ( ρ X ) the class of ρ X -measurable functions f satisfying f L p ( ρ X ) = ( X | f ( x ) | p d ρ X ) < + .
Let P n d denote the space of all homogeneous polynomials of degree n in d variables. We denote by L p ( S d 1 ) the class of all measurable functions defined on S d 1 with the finite norm
f p , S d 1 = S d 1 | f ( x ) | p d σ ( x ) 1 p , 1 p < + , max x S d 1 | f ( x ) | , p = + ,
and for p = + we assume that L + ( S d 1 ) is the space C ( S d 1 ) of continuous functions on S d 1 with the uniform norm.
For a given integer n 0 , the restriction to S d 1 of a homogeneous harmonic polynomial of degree n is called a spherical harmonics of degree n. If Y P n d , then Y ( x ) = x n Y ( x ) , x = x x , so that Y is determined by its restriction on the unit sphere. Let H n ( S d 1 ) denote the space of spherical harmonics of degree n. Then
dim H n ( S d 1 ) = ( n + d 2 n ) + ( n + d 3 n 1 ) , n = 1 , 2 , 3 , ,
Spherical harmonics of different degrees are orthogonal on the unit sphere. For further properties about harmonics one can refer to [64].
For n = 0 , 1 , 2 , , let { Y l n : 1 l dim H n ( S d 1 ) } be an orthonormal basis of H n ( S d 1 ) . Then
1 ω d 1 S d 1 Y l n ( ξ ) Y l m ( ξ ) d σ ( ξ ) = δ l , l δ m , n ,
where ω d 1 denotes the surface area of S d 1 and ω d 1 = 2 π d 2 Γ ( d 2 ) . Furthermore, by (1.2.8) in [63] we have
j = 1 dim H n ( S d 1 ) Y l n ( x ) Y l n ( y ) = n + η η C n η ( x y ) , x , y S d 1 ,
where C n η ( t ) is the n-th generalized Legendre polynomial the same as in (12). Combining (12) and (13) we have
ϕ ( x y ) = l = 0 + a l η ( ϕ ) ^ l + η η C l η ( x · y ) = l = 0 + a l η ( ϕ ) ^ k = 1 dim H l ( S d 1 ) Y k l ( x ) Y k l ( y ) , x , y S d 1 .

2.1. Density

We first give a general discrimination method for density.
Proposition 2.1. (see Lemma 1 in Chapter 18 of [65]) For a subset V in a normed linear space E, the following two properties are equivalent:
(a) V is fundamental in E (that is, its linear span is dense in E).
(b) V = { 0 } ( that is, 0 is the only element of E * that annihilates V).
Based on this proposition, we can show the density of Δ ϕ S d 1 .
Theorem 2.1. Let ϕ L W η 2 satisfy a l η ( ϕ ) ^ > 0 for all l = 0 , 1 , 2 , . Then Δ ϕ S d 1 is dense in L 2 ( S d 1 ) .
Proof. See the proof in Section 5.

2.2. Reproducing Property

We first restate a proposition.
Proposition 2.2. For any given n 1 there exist a finite subset Ω ( n ) S d 1 and corresponding positive numbers { μ ω : ω Ω ( n ) } such that
S d 1 f ( x ) d σ ( x ) = ω Ω ( n ) μ ω ( n ) f ( ω ) , f H 3 n ( S d 1 ) ,
and for some 1 p < +
S d 1 | f ( x ) | p d σ ( x ) ω Ω ( n ) μ ω ( n ) | f ( ω ) | p , f H n ( S d 1 ) .
Moreover, for any m n and p 1 there exists a constant c p , d > 0 such that
ω Ω ( n ) μ ω ( n ) | f ( ω ) | p c p , d m n d 1 S d 1 | f ( x ) | p d σ ( x ) , f H m ( S d 1 ) .
Proof. (15)-(16) were proved by H.N.Mhaskar et al in [61] and now have been extended to other domains (see e.g.[66,67]). Inequality (17) is proved by [68].
(15)-(16) show the existence of Ω ( n ) S d 1 which has the polynomial exact QR (8) and satisfies MZ inequality (9). Inequality (17) often goes with (15) and (16) and are needed in discussing approximation order(see e.g. [62]).
Based above analysis we propose the following definition.
Definition 2.1(Marcinkiewicz-Zygmund inequalities setting (MZIS)). We say a given finite node set Ω ( n ) S d 1 forms a Marcinkiewicz-Zygmund inequality setting if (15)-(16) and (17) simultaneously hold.
Let ϕ L W η 2 . For a given finite set Ω ( n ) S d 1 satisfy Definition 2.1 and the positive numbers { μ ω : ω Ω ( n ) } are defined as in Proposition 2.2. We define a zonal translation network by
H Ω ( n ) : = c l { f ( x ) = x k Ω ( n ) c k μ k ( n ) T x ( ϕ ) ( x k ) + c 0 : c k R , k = 0 , 1 , 2 , , | Ω ( n ) | } ,
where T x ( ϕ ) ( y ) = ϕ ( x y ) . Then it is easy to see that
H Ω ( n ) = H ϕ ( n ) R ,
where for A , B R and A B = { 0 } we define A B = { a + b : a A , b B } and
H ϕ ( n ) : = c l { f ( x ) = x k Ω ( n ) c k μ k ( n ) T x ( ϕ ) ( x k ) : c k R , k = 0 , 1 , 2 , , | Ω ( n ) | } .
For f ( x ) = x k Ω ( n ) c k μ k ( n ) T x ( ϕ ) ( x k ) , g ( x ) = x k Ω ( n ) d k μ k ( n ) T x ( ϕ ) ( x k ) we define a bivariate operation as
f , g ϕ = x k Ω ( n ) c k d k μ k ( n )
and
f ϕ = x k Ω ( n ) μ k ( n ) | c k | 2 1 2 .
Since (14), we have by Theorem 4 in Chapter 17 of [65] that the matrix T x i ( ϕ , x j ) i , j = 1 , 2 , , | Ω ( n ) | is positive definite for a given n. It follows that the vector c = { μ k ( n ) c k } x k Ω ( n ) is uniqueness. Then ( H ϕ ( n ) , · ϕ ) is a Hilbert space which is isometric isomorphism with l 2 ( Ω ( n ) ) ,where
l 2 ( Ω ( n ) ) = { c = { μ k ( n ) c k } x k Ω ( n ) : c l 2 ( Ω ( n ) ) = x k Ω ( n ) μ k ( n ) | c k | 2 1 2 < + } .
Let C ( [ 1 , 1 ] ) be the set of all continuous functions defined on [ 1 , 1 ] and
ϕ C ( [ 1 , 1 ] ) = sup x [ 1 , 1 ] | ϕ ( x ) | .
We then have the following proposition.
Proposition 2.3. Let ϕ C [ 1 , 1 ] satisfy a n η ( ϕ ) ^ > 0 for all n 1 and Ω ( n ) S d 1 be an MZIS. Then ( H ϕ ( n ) , · ϕ ) is a reproducing kernel Hilbert space associated with kernel
K x * ( ϕ ) ( y ) = K * ( ϕ , x , y ) = x k Ω ( n ) μ k ( n ) T x k ( ϕ , x ) T x k ( ϕ , y ) ,
i.e.,
f ( x ) = f , K x * ( ϕ ) ( · ) ϕ x S d 1 , f H ϕ ( n )
and there is a constant k * > 0 such that
| f ( x ) | k * f ϕ f H ϕ ( n ) , x S d 1 .
Proof. See the proof in Section 5.
Corollary 2.1. Under the assumption of Proposition 2.3, H Ω ( n ) is a reproducing kernel Hilbert space associated with the inner product defined by
f , g H Ω ( n ) = f 1 , g 1 ϕ + c 0 d 0 , w h e r e f ( x ) = f 1 ( x ) + c 0 , g ( x ) = g 1 ( x ) + d 0 , f 1 ( x ) = x k Ω ( n ) c k μ k ( n ) T x ( ϕ ) ( x k ) , g 1 ( x ) = x k Ω ( n ) d k μ k ( n ) T x ( ϕ ) ( x k ) } .
and the corresponding reproducing kernel K x ( ϕ ) ( y ) is
K x ( ϕ ) ( y ) = K x * ( ϕ ) ( y ) + 1 , x , y S d 1 .
Furthermore, there is a constant κ > 0 such that
| f ( x ) | κ f H Ω ( n ) , f H Ω ( n ) , x S d 1 .
Proof. The results can be obtained from Proposition 2.3, Lemma 4.2 and the fact that the real set R is a reproducing kernel Hilbert space whose reproducing kernel is 1 and the inner product is the usual product for two real numbers.

3. Convergence Rate of the Kernel Regularized Regression

We now make a combination of the zonal translation network with the kernel regularized regression.

3.1. Learning Framework

For a set of observations z = { ( x i , y i ) } i = 1 m drawn i.i.d. according to a joint distribution ρ ( x , y ) on Z = S d 1 × Y , Y = [ M , M ] , M > 0 i s a g i v e n r e a l n u m b e r , satisfying ρ ( x , y ) = ρ ( y | x ) ρ S d 1 ( x ) we define a regularized framework as
f z , λ Ω ( m ) : = a r g min f H Ω ( m ) E z ( f ) + λ 2 f H Ω ( m ) 2 ,
where λ = λ ( m ) is the regularization parameter, and
E z ( f ) = 1 m i = 1 m ( y i f ( x i ) ) 2 .
The general integral model of (22) is
f ρ , λ Ω ( m ) : = a r g min f H Ω ( m ) E ρ ( f ) + λ 2 f H Ω ( m ) 2 ,
where
E ρ ( f ) = Z ( y f ( x ) ) 2 d ρ .
To show the convergence analysis for (22) we need to bound the error
f z , λ Ω ( m ) f ρ L 2 ( ρ S d 1 ) ,
which is an approximation problem, whose convergence rate depends upon the approximation ability of H Ω ( m ) . An error decomposition will be given in Section 3.2.

3.2. Error Decomposition

By (2) and the definition of f ρ , λ Ω ( m ) we have the following decompositions
f z , λ Ω ( m ) f ρ L 2 ( ρ S d 1 ) f z , λ Ω ( m ) f ρ , λ Ω ( m ) L 2 ( ρ S d 1 ) + f ρ , λ Ω ( m ) f ρ L 2 ( ρ S d 1 ) = f z , λ Ω ( m ) f ρ , λ Ω ( m ) L 2 ( ρ S d 1 ) + E ρ ( f ρ , λ Ω ( m ) ) E ρ ( f ρ ) + λ 2 f ρ , λ Ω ( m ) H Ω ( m ) 2 f z , λ Ω ( m ) f ρ , λ Ω ( m ) L 2 ( ρ S d 1 ) + D Ω ( m ) ( f ρ , λ ) L 2 ( ρ S d 1 ) ,
where
D Ω ( m ) ( f ρ , λ ) = inf g H Ω ( m ) g f ρ L 2 ( ρ S d 1 ) + λ 2 g H Ω ( m ) .
We have used the fact that for a > 0 , b > 0 , c > 0 and 0 < p 1 there holds
( a + b + c ) p a p + b p + c p .
The K-functional D Ω ( m ) ( f ρ , λ ) shows the approximation error whose decay have been described in [69]. So the main estimate we need to deal with is the sample error
f z , λ Ω ( m ) f ρ , λ Ω ( m ) L 2 ( ρ S d 1 ) .

3.3. The Convergence Rate

To give a capacity independent generalization error for algorithm (22), we need some concepts of convex analysis.
G a ^ t e a u x differentiable.Let ( H , · H ) be a Hilbert space, F ( f ) : H R { } be a real function. We say F is G a ^ t e a u x differentiable at f 0 H if there is an ξ H such that for any g H there holds
lim t 0 F ( f 0 + t g ) F ( f 0 ) t = g , ξ H
and write F G ( f 0 ) = ξ or f F ( f ) = ξ . It known that for a differentiable convex function F ( f ) on H  f 0 = a r g min f H F ( f ) if and only if ( f F ( f ) ) | f = f 0 = 0 . (see Proposition 17.4 in [70]).
In learning theory, to form an explicit learning rate, one often assume the K-functional has a power convergence rate, i.e., we assume that there exists a 0 < β 1 such that
D Ω ( m ) ( f ρ , λ ) = O ( λ β ) , λ 0 + ( m + ) .
Since H Ω ( m ) is a reproducing kernel Hilbert space, the decay of D Ω ( m ) ( f ρ , λ ) can be discussed with the method of [69]. In particular, it is shown in Theorem 2.3 of [71] that one can choose an MZIS Ω ( m ) S d 1 , ϕ L η + and suitable real numbers α , β in such a way that
D Ω ( m ) f ρ , 1 2 α m = O ( 1 2 m β )
if β f ρ L 2 ( S d 1 ) , where we say β f ρ L 2 ( S d 1 ) if there is a function φ L 2 ( S d 1 ) such that
φ ( x ) l = 0 + l β Y l ( f , x ) , Y l ( f , x ) = k = 1 dim H l ( S d 1 ) f ^ k , l Y k l ( x ) , f ^ k , l = S d 1 f ( u ) Y k l ( u ) d σ ( u ) ,
which shows that (25) is reasonable.
Theorem 3.1.Let ϕ C [ 1 , 1 ] satisfy a l η ( ϕ ) ^ > 0 for all l 0 and Ω ( m ) S d 1 be an MZIS. Then for any δ ( 0 , 1 ) , with confidence 1 δ , hold
f z , λ Ω ( m ) f ρ L 2 ( ρ S d 1 ) 4 κ M λ m + D Ω ( m ) ( f ρ , λ ) L 2 ( ρ S d 1 ) λ 3 2 m l o g 4 δ + D Ω ( m ) ( f ρ , λ ) L 2 ( ρ S d 1 ) .
Proof. See the proof in Section 5.

3.4. Conclusion and Discussion

We propose the concept of MZIS for the scattered node set on the unit sphere, with which show that the related convolutional zonal translation network is a reproducing kernel Hilbert space, and show the convergence rate for the kernel regularized least square regression model. We give further analysis to show the advantages of the present paper.
(1)The zonal translation networks which we have chosen is a finite dimensional reproducing kernel Hilbert space, our discussions belong to the scope of kernel method, which is a combination and application of (neural) translation networks to learning theory.
(2)Compared with the existing convergence rate estimate of neural network learning, our upper estimates are dimensional independent (see Theorem in [25], Theorem 3.1 in [72], Theorem 7 in [73], Theorem 1 in [26]).
(3) It is hopeful that the node set in (15)-(16) and (17) may be replaced by a set of uniform distribution observations or a set of random samples (see [72,73]).
(4)The density derivation for the zonal translation network is qualitative, the density is described with the Fourier orthogonal coefficients. We think that this method can be extended to other domains such as the unit ball, the Euclidean space R d and R + = [ 0 , + ) et al.
(5)It is hopeful that with the help of the concept MZIS one may show reproducing property for the deep translation networks and thus give investigations for the performance of deep convolutional translation learning with the kernel method (see e.g.[6,7,59]).

4. Some Lemmas

To prove the main results, we need some lemmas.
Lemma 4.1.Let ( H , · ) be a Hilbert space and ξ be a random variable on ( Z , ρ ) with values in H and { z i } i = 1 m be independent samples drawers of ρ.Assume ξ H M ˜ < + almost surely. Then, for any 0 < δ < 1 , with confidence 1 δ ,it holds
1 m i = 1 m ξ ( z i ) E ( ξ ) H 2 M ˜ log ( 2 δ ) m
Proof. Find it from [74].
Lemma 4.2.Let ( H , · H ) be a Hilbert space over X with respect to kernel K. If ( E , · E ) and ( F , · F ) are closed subspaces of H such that E F and E F = H , then K = L + M , where L and M are the reproducing kernels of E and F respectively. Moreover, for h = e + f , e E , f F , we have
h H = e E 2 + f F 2 1 2 .
Proof. See Corollary 1 in Chapter 31 of [65] or the Theorem in Section 6 in part I of [75].
Lemma 4.3. There hold the following equalities:
f E z ( f ) ( · ) = 2 m i = 1 m ( y i f ( x i ) ) K x i ( ϕ ) ( · ) , f H Ω ( m ) .
and
f E ρ ( f ) ( · ) = 2 Z ( y f ( x ) ) K x ( ϕ ) ( · ) d ρ , f H Ω ( m ) .
Proof of (28). By the equality
a 2 b 2 = 2 ( a b ) b + ( a b ) 2 , a R , b R
we have
lim t 0 E z ( f + t g ) E z ( f ) t = lim t 0 1 m i = 1 m ( 2 t y i f ( x i ) ) g ( x i ) + t 2 g ( x i ) 2 t = 2 m i = 1 m ( y i f ( x i ) ) g ( x i )
Since g ( x ) = g , K x ( ϕ ; · ) H Ω ( m ) and the definition of G a ^ t e a u x derivative, we have by above equality that
lim t 0 E z ( f + t g ) E z ( f ) t = g , 2 m i = 1 m ( y i f ( x i ) ) g ( x i ) K x i ( ϕ ) ( · ) H Ω ( m ) .
We then have (28). By the same way we can have (29).
Lemma 4.4. Let ( H , · H ) be a Hilbert space consisting of real functions on X. Then
f H 2 g H 2 = 2 f g , g H + f g H 2 , f , g H
and
f ( f H 2 ) ( · ) = 2 f ( · ) , f H .
Proof. (31) is the deformation of the parallelogram formula. (32) can be shown with (31).
Lemma 4.5. (22) has a unique solution f z , λ Ω ( m ) and (23) has a unique solution f ρ , λ Ω ( m ) . Moreover,
(i) There hold the bounds
f ρ , λ Ω ( m ) C ( S d 1 ) 2 κ D Ω ( m ) ( f ρ ) λ ,
where κ is defined as in (21).
(ii) There holds the equality
λ f z , λ Ω ( m ) ( · ) = 2 m i = 1 m y i f z , λ Ω ( m ) ( x i ) K x i ( ϕ ) ( · )
and the equality
λ f ρ , λ Ω ( m ) ( · ) = 2 Z y f ρ , λ Ω ( m ) ( x ) K x ( ϕ ) ( · ) d ρ
Proof.Proof of (i). Since E ρ ( f ρ , λ Ω ( m ) ) E ρ ( f ρ ) , we have by (2) that
λ 2 f ρ , λ Ω ( m ) H Ω ( m ) 2 E ρ ( f ρ , λ Ω ( m ) ) E ρ ( f ρ ) + λ 2 f ρ , λ Ω ( m ) H Ω ( m ) 2 = inf g H Ω ( m ) E ρ ( g ) E ρ ( f ρ ) + λ 2 g H Ω ( m ) 2 = inf g H Ω ( m ) g f ρ L 2 ( ρ S d 1 ) 2 + λ 2 g H Ω ( m ) 2 .
By (36) we have (33).
Proof of (ii). By the definition of f z , λ Ω ( m ) and (32) we have
0 = f E z ( f ) + λ 2 f H Ω ( m ) 2 | f = f z , λ Ω ( m ) ,
i.e.,
0 = f E z ( f ) | f = f z , λ Ω ( m ) + λ f ( 1 2 f H Ω ( m ) 2 ) | f = f z , λ Ω ( m ) = 2 m i = 1 m ( y i f z , λ Ω ( m ) ( x i ) ) K x i ( ϕ ) ( · ) + λ f z , λ Ω ( m ) ( · ) .
Hence, (34) holds. We can prove (35) in the same way.
Lemma 4.6. The solutions f z , λ Ω ( m ) and f ρ , λ Ω ( m ) satisfy the inequality
f z , λ Ω ( m ) f ρ , λ Ω ( m ) H Ω ( m ) 2 A ( z ) λ ,
where
A ( z ) = Z y f ρ , λ Ω ( m ) ( x ) K x ( ϕ ) ( · ) d ρ 1 m i = 1 m y i f ρ , λ Ω ( m ) ( x i ) K x i ( ϕ ) ( · ) H Ω ( m ) .
Proof. By (30) we have
a 2 b 2 2 ( a b ) b , a R , b R
It follows that
E z ( f z , λ Ω ( m ) ) E z ( f ρ , λ Ω ( m ) ) = 1 m i = 1 m ( y i f z , λ Ω ( m ) ( x i ) ) 2 ( y i f ρ , λ Ω ( m ) ( x i ) ) 2 2 m i = 1 m ( y i f ρ , λ Ω ( m ) ( x i ) ) × f z , λ Ω ( m ) ( x i ) f ρ , λ Ω ( m ) ( x i ) = f z , λ Ω ( m ) f ρ , λ Ω ( m ) , 2 m i = 1 m y i f ρ , λ Ω ( m ) ( x i ) K x i ( ϕ ) ( · ) H Ω ( m ) ,
where we have used the reproducing property
f z , λ Ω ( m ) ( x i ) f ρ , λ Ω ( m ) ( x i ) = f z , λ Ω ( m ) f ρ , λ Ω ( m ) , K x i ( ϕ ) ( · ) H Ω ( m ) .
By the definition of f z , λ Ω ( m ) we have
0 E z ( f z , λ Ω ( m ) ) + λ 2 f z , λ Ω ( m ) H Ω ( m ) 2 E z ( f ρ , λ Ω ( m ) ) + λ 2 f ρ , λ Ω ( m ) H Ω ( m ) 2 .
On the other hand, by above inequality, (38) and (31) we have
0 E z ( f z , λ Ω ( m ) ) E z ( f ρ , λ Ω ( m ) ) + λ 2 f z , λ Ω ( m ) H Ω ( m ) 2 f ρ , λ Ω ( m ) H Ω ( m ) 2 f z , λ Ω ( m ) f ρ , λ Ω ( m ) , 2 m i = 1 m y i f ρ , λ Ω ( m ) ( x i ) K x i ( ϕ ) ( · ) H Ω ( m ) + f z , λ Ω ( m ) f ρ , λ Ω ( m ) , λ j q ( f ρ , λ Ω ( m ) ) H Ω ( m ) + λ f z , λ Ω ( m ) f ρ , λ Ω ( m ) H Ω ( m ) 2 .
Since (35), we have
0 2 f z , λ Ω ( m ) f ρ , λ Ω ( m ) , Z y f ρ , λ Ω ( m ) ( x ) K x ( ϕ ) ( · ) d ρ 1 m i = 1 m y i f ρ , λ Ω ( m ) ( x i ) K x i ( ϕ ) ( · ) H Ω ( m ) + λ f z , λ Ω ( m ) f ρ , λ Ω ( m ) H Ω ( m ) 2 .
By the Cauchy inequality we have
λ f z , λ Ω ( m ) f ρ , λ Ω ( m ) H Ω ( m ) 2 2 f ρ , λ Ω ( m ) f z , λ Ω ( m ) , Z y f ρ , λ Ω ( m ) ( x ) K x ( ϕ ) ( · ) d ρ 1 m i = 1 m y i f ρ , λ Ω ( m ) ( x i ) K x i ( ϕ ) ( · ) H Ω ( m ) Z y f ρ , λ Ω ( m ) ( x ) K x ( ϕ ) ( · ) d ρ 1 m i = 1 m y i f ρ , λ Ω ( m ) ( x i ) K x i ( ϕ ) ( · ) H Ω ( m ) × 2 f z , λ Ω ( m ) f ρ , λ Ω ( m ) H Ω ( m ) .
We then have (37).
Lemma 4.7. For the solutions f z , λ Ω ( m ) and f ρ , λ Ω ( m ) we have
f z , λ Ω ( m ) f ρ , λ Ω ( m ) H Ω ( m ) 4 κ M λ m + D Ω ( m ) ( f ρ , λ ) L 2 ( ρ S d 1 ) λ 3 2 m l o g 4 δ .
Proof. It is easy to see that
A ( z ) Z y K x ( ϕ ) ( · ) d ρ 1 m i = 1 m y i K x i ( ϕ ) ( · ) H Ω ( m ) + Z f ρ , λ Ω ( m ) ( x ) K x ( ϕ ) ( · ) d ρ S d 1 1 m i = 1 m f ρ , λ Ω ( m ) ( x i ) K x i ( ϕ ) ( · ) H Ω ( m ) = B ( z ) + C ( z ) .
Take ξ ( x , y ) ( · ) = y K x ( ϕ ) ( · ) . Then,
ξ ( x , y ) ( · ) H Ω ( m ) = | y | | K x ( ϕ ) ( x ) | M κ .
Combining (41) with (27) we have, with confidence 1 δ , that
B ( z ) 2 M κ m l o g 2 δ .
Also, take η ( x ) ( · ) = f ρ , λ Ω ( m ) ( x ) K x ( ϕ ) ( · ) . Then
η ( x ) H Ω ( m ) = | f ρ , λ Ω ( m ) ( x ) | | K x ( ϕ ) ( x ) | 4 κ D Ω ( m ) ( f ρ , λ ) L 2 ( ρ S d 1 ) λ
Combing (43) with (27) we have, with confidence 1 δ , that
C ( z ) 4 κ D Ω ( m ) ( f ρ , λ ) L 2 ( ρ S d 1 ) λ m l o g 2 δ .
Collecting (44), (42),(40) and (37), we arrive (39).

5. Proof for Theorems and Propositions

Proof of Theorem 2.1 . If Δ ϕ S d 1 is not dense in L 2 ( S d 1 ) , i.e.,
c l s p a n ( Δ ϕ S d 1 ) L 2 ( S d 1 ) .
Then by Proposition 2.1 we know ( c l s p a n ( Δ ϕ S d 1 ) ) { 0 } and there is a nonzero functional F L 2 ( S d 1 ) such that
F ( f ) = 0 , f c l s p a n ( Δ ϕ S d 1 ) .
It follows that F ( ϕ ( · y ) ) = 0 for all y S d 1 . By the Riesz representation Theorem, F corresponds to a nonzero h L 2 ( S d 1 ) in such a way that
F ( f ) = S d 1 f ( x ) h ( x ) d σ ( x ) , f L 2 ( S d 1 ) .
Consequently, S d 1 ϕ ( x y ) h ( y ) d σ ( y ) = 0 , x S d 1 , which gives
S d 1 S d 1 ϕ ( x y ) h ( y ) d σ ( y ) h ( x ) d σ ( x ) = 0 .
Combining (45) with (14) we have
l = 0 + a l η ( ϕ ) ^ k = 1 dim H l ( S d 1 ) S d 1 h ( y ) Y k l ( y ) d σ ( y ) 2 = 0 .
It follows that S d 1 h ( y ) Y k l ( y ) d σ ( y ) = 0 for all l 0 . Therefore, h = 0 . We have obtained a contradiction.
Proof of Proposition 2.3. By the the definition of · , · ϕ and the definition kernel K x * ( ϕ , y ) in (18) we have for any f ( x ) = x k Ω ( n ) c k μ k ( n ) T x ( ϕ ) ( x k ) that
f , K x * ( ϕ , · ) ϕ = x k Ω ( n ) c k μ k ( n ) T x ( ϕ ) ( x k ) = f ( x ) .
(19) then holds. We now show (20). In fact, by the Cauchy’s inequality we have
| f ( x ) | = | x k Ω ( n ) c k μ k ( n ) T x ( ϕ ) ( x k ) | c L 2 ( Ω ( n ) ) x k Ω ( n ) μ k ( n ) | T x ( ϕ ) ( x k ) | 2 1 2 = f ϕ x k Ω ( n ) μ k ( n ) | ϕ ( x k · x ) | 2 1 2 .
On the other hand, by the localized kernel theory about the Jacobi polynomials in [76] (see Lemma 2.5 in [76]), for a given n, we have an operator V n : L W η 1 Π n [ 1 , 1 ] (where Π n [ 1 , 1 ] is the set of all the algebraic polynomials of order n defined on [ 1 , 1 ] and η = d 2 2 ) such that V n ( ϕ ) Π 2 n [ 1 , 1 ] , V n ( g ) = g for any g Π n [ 1 , 1 ] and there is constant c > 0 such that
V n ( ϕ ) C ( [ 1 , 1 ] ) c ϕ C ( [ 1 , 1 ] ) , V n ( ϕ ) ϕ C ( [ 1 , 1 ] ) c E n ( ϕ ) C ( [ 1 , 1 ] ) ,
where
E n ( ϕ ) C ( [ 1 , 1 ] ) = inf p Π n [ 1 , 1 ] ϕ p C ( [ 1 , 1 ] ) .
It follows by the Minkowski inequality and inequality (17) that
x k Ω ( n ) μ k ( n ) | ϕ ( x k · x ) | 2 1 2 x k Ω ( n ) μ k ( n ) | ( ϕ V n ( ϕ ) ) ( x k · x ) | 2 1 2 + x k Ω ( n ) μ k ( n ) | V n ( ϕ ) ( x k · x ) | 2 1 2 c E n ( ϕ ) C ( [ 1 , 1 ] ) ( x k Ω ( n ) μ k ( n ) ) 1 2 + O ( 1 ) S d 1 | V n ( ϕ ) ( x · y ) | 2 d σ ( x ) 1 2 = O ( 1 ) E n ( ϕ ) C ( [ 1 , 1 ] ) + O ( 1 ) 1 1 | V n ( ϕ ) ( u ) | 2 W η ( u ) d u 1 2 = O ( 1 ) .
(20) thus holds, where we have used the Funk-Hecke formula (see (1.2.11) in [63])
S d 1 f ( x · y ) Y n ( y ) d σ ( y ) = λ n ( f ) Y n ( x ) , Y n H n ( S d 1 )
and
λ n ( f ) = ω d 1 1 1 f ( t ) C η ( t ) C η ( 1 ) d t , η = d 1 2 .
Proof of Theorem 3.2. Since (39) we have by (21) that
f z , λ Ω ( m ) f ρ , λ Ω ( m ) L 2 ( ρ S d 1 ) f z , λ Ω ( m ) f ρ , λ Ω ( m ) C ( S S d 1 ) κ f z , λ Ω ( m ) f ρ , λ Ω ( m ) H Ω ( m ) 4 κ M λ m + D Ω ( m ) ( f ρ , λ ) L 2 ( ρ S d 1 ) λ 3 2 m l o g 4 δ .
Taking (46) into (24) we have (26).

Author Contributions

Conceptualization, Ran. X.X.; methodology, Sheng B.H.; validation, Ran. X.X.; formal analysis, Ran. X.X.; resources, Sheng B.H. and Wang S.H.; writing—original draft preparation, Ran. X.X. and Sheng B.H.; writing—review and editing, Sheng B.H. and Wang S.H.; supervision, Sheng B.H.. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China under Grants No.61877039, the NSFC/RGC Joint Research Scheme (Project No. 12061160462 and N_CityU102/20) of China and Natural Science Foundation of Jiangxi Province of China (20232BAB201021).

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Communications of the ACM., 2017, 60, 84–90. [Google Scholar] [CrossRef]
  2. Wu, Y.; Schuster, M.; Chen, Z.; Le, Q.V.; Norouzi, M.; Macherey, W.; Cao, Y.; Gao, Q. Google’s neural machine translation system: Bridging the gap between human and machine translation. arXiv 2016, arXiv:1609.08144v2 [cs.CL]. [Google Scholar]
  3. Alipanahi, B.; Delong, A.; Weirauch, M.T.; Frey, B.J. Predicting the sequence specificities of DNA-and RNA-binding proteins by deep learning. Nature Biotechnology. 2015, 33, 831–838. [Google Scholar] [CrossRef] [PubMed]
  4. Chui, C.K.; Lin, S.B.; Zhou, D. X. Construction of neural networks for realization of localized deep learning. arXiv 2018, arXiv:1803.03503 [cs.LG]. [Google Scholar] [CrossRef]
  5. Chui, C. K.; Lin, S.B.; Zhou, D.X. Deep neural networks for rotation-invariance approximation and learning. Anal. and Appl. 2019, 17, 737–772. [Google Scholar] [CrossRef]
  6. Fang, Z.Y.; Feng, H.; Huang, S.; Zhou, D.X. Theory of deep convolutional neural networks II: spherical analysis. Neural Networks. 2020, 131, 154–162. [Google Scholar] [CrossRef] [PubMed]
  7. Feng, H.; Huang, S.; Zhou, D.X. Generalization analysis of CNNs for classification on spheres. IEEE Transactions on Neural Networks and Learning Systems. [CrossRef] [PubMed]
  8. Zhou, D. X. Deep distributed convolutional neural networks: universality. Anal. Appl. 2018, (16), 895–919. [Google Scholar] [CrossRef]
  9. Zhou, D. X. Universality of deep convolutional neural networks. Appl. Comput. Harmon. Anal. 2020, 48, 787–794. [Google Scholar] [CrossRef]
  10. Cucker, F.; Smale, S. On the mathematical foundations of learning. Bull.Amer.Math.Soc. 2001, 39, 1–49. [Google Scholar] [CrossRef]
  11. An, C.P.; Chen, X.J.; Sloan, I.H.; Womersley, R.S. Regulatized least squares approximations on the sphere using spherical designs. SIAM J.Numer.Anal. 2012, 50, 1513–1534. [Google Scholar] [CrossRef]
  12. An, C.P.; Wu, H.N. Lasso hyperinterpolation over general regions. SIAM J. Sci. Comput. 2021, 43, A3967–A3991. [Google Scholar] [CrossRef]
  13. An, C.P.; Ran, J.S. Hard thresholding hyperinterpolation over general regions. arXiv 2023, arXiv:2209.14634v2 [math.NA]. [Google Scholar]
  14. De Mol, C.; De Vito, E.; Rosasco, L. Elastic-net regularization in learning theory. J. Complexity, 2009, 25, 201–230. [Google Scholar] [CrossRef]
  15. Fischer, S.; Steinwart, I. Sobolev norm learning rates for regularized least-squares algorithms. J.Mach.Learn.Res. 2020, 21, 8464–8501. [Google Scholar]
  16. Lai, J.F.; Li, Z.F.; Huang, D.G.; Lin, Q. The optimality of kernel classifiers in Sobolev space. arXiv 2024, arXiv:2402.01148v1 [math.ST]. [Google Scholar]
  17. Sun, H.W.; Wu, Q. Least square regression with indefinite kernels and coefficient regularization. Appl. Comput. Harmon. Anal. 2011, 30, 96–109. [Google Scholar] [CrossRef]
  18. Wu, Q.; Zhou, D.X. Learning with sample dependent hypothesis spaces. Comput. Math. Appl. 2008, 56, 2896–2907. [Google Scholar] [CrossRef]
  19. Chen, H.; Wu, J.T.; Chen, D.R. Semi-supervised learning for regression based on the diffusion matrix (in Chinese). Sci Sin Math. 2014, 44, 399–408. [Google Scholar]
  20. Sun, X.J.; Sheng, B.H. The learning rate of kernel regularized regression associated with a correntropy-induced loss. Adv. in Math.(Beijing) 2024, 53, 633–652. [Google Scholar]
  21. Wu, Q.; Zhou, D.X. Analysis of support vector machine classification. J. Comput.Anal. Appl. 2006, 8, 99–119. [Google Scholar]
  22. Cucker, F.; Zhou, D. X. Learning Theory: An Approximation Theory Viewpoint. Cambridge University Press, New York, 2007.
  23. Sheng, B. H. Reproducing property of bounded linear operators and kernel regularized least square regressions. Int. J. Wavelets Multiresolution Inf. Process. 2024, 2450013. [Google Scholar] [CrossRef]
  24. Steinwart, I.; and Christmann, A. Support Vector Machines,Springer-Verlag, New York, 2008.
  25. Lin, S.B.; Wang, D.; Zhou, D.X. Sketching with spherical designs for noisy data fitting on spheres. SIAM J.Sci.Comput. 2024, 46, A313–A337. [Google Scholar] [CrossRef]
  26. Lin, S.B.; Zeng, J.S.; Zhang, X.Q. Constructive neural network learning. IEEE Trans. on Cybernetics. 2019, 49, 221–232. [Google Scholar] [CrossRef] [PubMed]
  27. Mhaskar, H.N.; Micchelli, C.A. Degree of approximation by neural and translation networks with single hidden layer. Adv. Appl. Math. 1995, 16, 151–183. [Google Scholar] [CrossRef]
  28. Sheng, B.H.; Zhou, S.P.; Li, H.T. On approximation by tramslation networks in Lp(Rk) spaces. Adv. in Math.(Beijing) 2007, 36, 29–38. [Google Scholar]
  29. Mhaskar, H. N.; Narcowich, F. J.; Ward, J. D. Approximation properties of zonal function networks using scattered data on the sphere. Adv. Comput. Math. 1999, 11, 121–137. [Google Scholar] [CrossRef]
  30. Sheng, B. H. On approximation by reproducing kernel spaces in weighted Lp-spaces. J. Syst. Sci. Complex. 2007, 20(4), 623–638. [Google Scholar] [CrossRef]
  31. Narcowich, F.J.; Ward, J.D.; Wendland, H. Sobolev error estimates and a Bernstein inequality for scattered data interpolation via radial basis functions. Constr.Approx. 2006, 24, 175–186. [Google Scholar] [CrossRef]
  32. Narcowich, F.J.; Ward, J.D. Scattered data interpolation on spheres: error estimates and locally supported basis functions. SIAM J.Math.Anal. 2002, 33, 1393–1410. [Google Scholar] [CrossRef]
  33. Narcowich, F.J.; Sun, X.P.; Ward, J.D.; Wendland, H. Direct and inverse Sobolev error estimates for scattered data interpolation via spherical basis functions. Found.Comput.Math. 2007, 7, 369–390. [Google Scholar] [CrossRef]
  34. Gröchenig, K. Sampling, Marcinkiewicz-Zygmund inequalities, approximation and quadrature rules. J.Approx.Theory 2020, 257, 105455. [Google Scholar] [CrossRef]
  35. Mhaskar, H.N.; Narcowich, F.J.; Sivakumar, N.; Ward, J.D. Approximation with interpolatory constraints. Proc.Amer.Math. Soc. 2001, 130, 1355–1364. [Google Scholar] [CrossRef]
  36. Marzo, J. Marcinkiewicz-Zygmund inequalities and interpolation by spherical harmonics. J. Funct. Anal. 2007, 250, 559–587. [Google Scholar] [CrossRef]
  37. Marzo, J.; Pridhnani, B. Sufficiant conditions for sampling and interpolation on the sphere. Constr. Approx. 2014, 40, 241–257. [Google Scholar] [CrossRef]
  38. Wang, H.P. Marcinkiewicz-Zygmund inequalities and interpolation by spherical polynomials with respect to doubling weights. J.Math.Anal.Appl. 2015, 423, 1630–1649. [Google Scholar] [CrossRef]
  39. Gia, T.L.; Slon, I.H. The nuiform norm of hyperinterpolation on the unit sphere in an arbitrary number of dimensions em Constr. Approx. 2001, 17, 249–265. [Google Scholar] [CrossRef]
  40. Sloan, I.H. Polynomial interpolation and hyperinterpolation over general regions. J.Approx.Theory 1995, 83, 238–254. [Google Scholar] [CrossRef]
  41. Sloan, I.H.; Womersley, R.S. Constructive polynomial approximation on the sphere. J.Approx. Theory 2000, 103, 91–118. [Google Scholar] [CrossRef]
  42. Wang, H.P. Optimal lower estimates for the worst case cubature error and the approximation by hyperinterpolation operators in the Sobolev space sertting on the sphere. Int. J. Wavelets Multiresolution Inf. Process. 2009, 7, 813–823. [Google Scholar] [CrossRef]
  43. Wang, H.P.; Wang, K.; Wang, X.L. On the norm of the hyperinterpolation operator on the d-dimensional cube. Comput. Appl. 2014, 68, 632–638. [Google Scholar]
  44. Sloan, I.H.; Womersley, R.S. Filtered hyperinterpolation: a constructive polynomial approximation on the sphere. Int.J.Geomath. 2012, 3, 95–117. [Google Scholar] [CrossRef]
  45. Bondarenko, A.; Radchenko, D.; Viazovska, M. Well-seperated spherical designs. Constr. Approx. 2015, 41, 93–112. [Google Scholar] [CrossRef]
  46. Hesse, K.; Womersley, R.S. Numerical integration with polynomial exactness over a spherical cap. Adv.Math.Math. 2012, 36, 451–483. [Google Scholar] [CrossRef]
  47. Delsarte, P.; Goethals, J.M.; Seidel, J.J. Spherical codes and designs. Geom.Dedicata. 1977, 6, 363–388. [Google Scholar] [CrossRef]
  48. An, C.P.; Chen, X.J.; Sloan, I.H.; Womersley, R.S. Well conditioned spherical designs for integration and interpolation on the two-sphere. SIAM J. Numer.Anal. 2010, 48, 2135–2157. [Google Scholar] [CrossRef]
  49. Chen, X.; Frommer, A.; Lang, B. Computational existence proof for spherical t-designs. Numer. Math. 2010, 117, 289–305. [Google Scholar] [CrossRef]
  50. An, C.P.; Wu, H.N. Bypassing the quadrature exactness assumption of hyperinterpolation on the sphere. J.Complexity. 2024, 80, 101789. [Google Scholar] [CrossRef]
  51. An, C.P.; Wu, H.N. On the quadrature exactness in hyperinterpolation. BIT Numer. Math. 2022, 62, 1899–1919. [Google Scholar] [CrossRef]
  52. Sheng, B.H. The covering numbers for some periodic reproducing kernel spaces Acta Math. Scientia 2009, 29A(6), 1590–1600. [Google Scholar]
  53. Sheng, B.H. Estimate the norm of the Mercer kernel matrices with discrete orthogonal transforms. Acta Math. Hungar. 2009, 122, 339–355. [Google Scholar] [CrossRef]
  54. Sheng, B.H.; Wang, J.L.; Li, P. On the covering number of some Mercer kernel Hilbert spaces. J. Complexity 2008, 24, 241–258. [Google Scholar]
  55. Zhang, C.P.; Sheng, B.H.; Chen, Z.X. Applications of Bernstein -Durrmeyer operators in estimating the norm of Mercer kernel matrices Anal. Theory and Appl. 2008, 24, 74–86. [Google Scholar]
  56. Sun, X. J.; Sheng, B. H.; Liu, L.; Pan, X. L. On the density of translation networks defined on the unit ball. Math. Found. Comput. 2023. [Google Scholar] [CrossRef]
  57. Parhi, R.; Nowak, R.D. Banach space representer theorems for neural networks and ridge splines. J. Mach. Learn Res. 2021, 22, 1–40. [Google Scholar]
  58. Oono, K.; Suzuki, Y.J. Approximation and non-parameteric estimate of ResNet-type convolutional neural networks. arXiv 2023, arXiv:1903.10047v4 [stat.ML]. [Google Scholar]
  59. Shen, G.H.; Jiao, Y.L.; Lin, Y.Y.; Huang, J. Non-asymptotic excess risk bounds for classification with deep convolutional neural networks. arXiv 2021, arXiv:2105.00292 [cs.LG]. [Google Scholar]
  60. Mallat, S. Understanding deep convolutional networks. Phil. Trans. R. Soc. 2016, 374A, 20150203. [Google Scholar] [CrossRef]
  61. Mhaskar, H.N.; Narcowich, F.J.; Ward, J.D. Spherical Marcinkiewicz-Zygmund inequalities and positive quadratue. Math.Comput. 2001, 70, 1113–1130, Corrigendum: Math.Comp. 2001, (71), 453-454.). [Google Scholar] [CrossRef]
  62. Wang, H.P.; Wang, K. Optimal recovery of Besov classes of generalized smoothness and Sobolev class on the sphere. J. Complexity 2016, 32, 40–52. [Google Scholar] [CrossRef]
  63. Dai, F.; Xu, Y. Approximation Theory and Harmonic Analysis on Spheres and Balls. Springer, New York, 2013.
  64. Müller, C. Spherical Harmonic, Springer-Verlag, Berlin, 1966.
  65. Cheney, W.; Light, W. A Course in Approximation Theory, China Machine Press, Beijing, 2004.
  66. Brown, G.; Dai, F. Approximation of smooth functions on compact two-point homogeneous spaces. J. Funct. Anal. 2005, 220, 401–422. [Google Scholar] [CrossRef]
  67. Dai, F.; Wang, H.P. Positive cubature formulas and Marcinkiewicz-Zygmund inequalities on spherical caps. Constr. Approx. 2010, 31, 1–36. [Google Scholar] [CrossRef]
  68. Dai, F. On generalized hyperinterpolation on the sphere. Proc. Amer. Math. Soc. 2006, 134, 2931–2941. [Google Scholar] [CrossRef]
  69. Sheng, B. H.; Wang, J. L. Moduli of smoothness, K-functionals and Jackson-type inequalities associated with kernel function approximation in learning theory. Anal. Appl. 2024. [Google Scholar] [CrossRef]
  70. Bauschke, H. H.; Combettes, P. L. Convex Analysis and Monotone Operator Theory in Hilbert Spaces, Springer, New York, 2010.
  71. Sheng, B.H.; Xiang, D.H. The convergence rate for a K-functional in learning theory. J. Inequal. Appl. 2010, 249507. [Google Scholar] [CrossRef]
  72. Lin, S.B.; Wang, Y.G.; Zhou, D.X. Distributed filtered hyperinterpolation for noisy data on the sphere. SIAM J. Numer. Anal. 2021, 59, 634–659. [Google Scholar] [CrossRef]
  73. Montúfar, G.; Wang, Y.G. Distributed learning via filtered hyperinterpolation on manifolds. Found.Comput.Math. 2022, 22, 1219–1271. [Google Scholar] [CrossRef]
  74. Smale, S.; Zhou, D.X. Learning theory estimates via integral operators and their applications. Constr. Approx. 2007, 26, 153–172. [Google Scholar] [CrossRef]
  75. Aronszajn, N. Theory of reproducing kernels. Trans. Amer. Math. Soc. 1950, 68, 337–404. [Google Scholar] [CrossRef]
  76. Kyriazis, G.; Petrushev, P.; Xu, Y. Jacobi decomposition of weighted Triebel-Lizorkin and Besov spaces. arXiv 2006, arXiv:math/0610624 [math.CA]. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated