Preprint
Article

Deep Radial Basis Function Networks

Altmetrics

Downloads

132

Views

107

Comments

0

A peer-reviewed article of this preprint also exists.

This version is not peer-reviewed

Submitted:

18 February 2024

Posted:

21 February 2024

You are already at the latest version

Alerts
Abstract
Radial basis function (RBF) networks are often viewed as instable when used in multi-layered architectures and therefore are mostly used in a single-layered manner. Universal approximation theorems for single-layered RBF networks further render deeper architectures useless. However, deep neural networks have proven their effectiveness on many different tasks. We show that deeper RBF architectures with multiple radial basis function layers are achievable and able to learn. We introduce an initialization scheme for deep RBF networks based on k-means clustering and covariance estimation. We further show how to make use of convolutions to speed up the calculation of a Mahalanobis distance in a partially-connected fashion, similar to convolutional neural networks (CNNs). Finally, we evaluate our approach on image classification as well as speech emotion recognition tasks. Our results show that deep RBF networks perform similar to simple CNNs.
Keywords: 
Subject: Computer Science and Mathematics  -   Artificial Intelligence and Machine Learning

1. Introduction

The reconstruction of an unknown function based on a finite set of data - typical given as pairs of sensory input and target output - is a major goal in applications of numerical analysis, such as function approximation or pattern classification. In real-world applications, the unknown functions must be modelled using multivariate approximation schemes. Several approaches can be applied for multivariate approximation, such as finite elements techniques or spline functions together with triangulation, just to name a few. Typically the known data set is scattered, where it is assumed that the data set does not have any special properties in terms of density or spacing or other regularities. In [1] Franke discussed the problem of function interpolation given a set of scattered data, and he introduced the basis function approach for such tasks of function interpolation and approximation.
Radial Basis Functions (RBFs) are a special class of basis functions used for multivariate interpolation of scattered data, where the final interpolating function is obtained by a linear combination of multiple RBF kernel outputs. Kernel functions are of a fixed type, e.g. the Gaussian density function to mention the most common one. In this approach each RBF kernel calculates its output based on the distance between an input vector and the kernel center by leveraging some pre-defined proximity measure. In terms of artificial neural networks - this approximation scheme is a network with a single layer of RBF kernels, followed by a linear weighting layer. Primary, the Euclidean norm is used for this distance calculation, see Figure 1. More general distance measures such as the Mahalanobis distance have been introduced as proximity measures, this allows for more complex contributions of the input features to the kernel activation. Fundamental mathematical results on RBF can be found in the books by Powell [2], Buhmann [3] or Fasshauer [4], and in the papers authored by Micchelli [5], Dyn [6] or Schaback [7].
Neural network interpretation of RBF networks goes back to Broomhead and Lowe [8] as well as to Moody and Darken [9]. Park and Sandberg [10] proved that RBF networks with one hidden layer are universal approximators, theoretically rendering deeper architectures irrelevant. Broomhead and Lowe uniformly sampled RBF centers from the training data or input space, if prior domain knowledge was absent [8]. They optimized the linear output layer directly via the pseudo-inverse solution [8]. Due to the two-layered architecture the existence of a solution for the optimization problem was guaranteed [8]. Moody and Darken [9] further refined the initialization process by using vector-quantization approaches such as k-means clustering to determine data-based initial center locations [9]. Additionally, they introduced a p-nearest Neighbour (PNN) heuristic to calculate the width parameters used in the Gaussian RBF [9]. The PNN heuristic was chosen to create an overlap between neighboring RBF kernels in order to fully cover the input space [9]. Besides unsupervised clustering, supervised learning vector quantization (LVQ) [11] can be applied to initalize the centers of the RBF centers [12], also ensemble methods [13,14] and semi-supervised learning techniques [15] can be be considered to pre-train an initial RBF architecture. In order to improve the performance of the RBF network, a third optimization phase can be introduced [16]. In this third phase, all network parameters are simultaneously trained by supervised backpropagation training, see Schwenker et al. [16] for details.
Besides this general RBF training procedure a lot of different heuristics have been developed over the years, see for instance [17] or [18]. Research was mostly focused on parameter initialization and finding better radial basis functions [19]. However, the same overall network architecture was maintained. More recently, in [19,20] RBF networks were used as classifiers in deep convolutional neural networks while achieving comparable performance to commonly used MLP classifiers on the considered benchmark tasks [19]. The success of deep learning shows that neural networks can benefit greatly from deeper network architectures. However, the biggest advantage of RBF networks over other neural network approaches lies in the data driven parameter initialization, providing good initial model performance. In this work, we propose to apply the deep learning paradigm to RBF networks while incorporating the data driven initialization.
The main contributions of the paper can be summarized in the following statements:
  • The paper is not in the mainstream of deep neural network architectures. It is a paper in which we want to describe a new perspective on RBF networks in the context of modern deep learning. This is done by proposing an approach that allows the construction and the training of deep networks consisting of RBF units – that means of units that are based on distance computation and kernel functions. In contrast to related work on standard RBF networks, the idea of designing multilayered RBF architectures is considered. Thus, this paper must be considered as an attempt to introduce RBF networks into the field of modern deep learning, with the aim of initiating a discussion on this type of deep and perhaps recurrent type of RBF networks.
  • Moreover, we present in our work a partially connected architecture based on such distance-based RBF units, following the same idea of standard Convolutional Neural Network (CNN) architectures. Thus, this type of architecture may be of interest in applications where typically CNNs are used, for instance to extract patterns from images.
  • From the implementation point of view we describe how such a distance computation for partiallly connected units can be implemented in modern deep learning frameworks.
  • In a series of experiments, we investigated the use of RBF-based deep architectures on benchmark applications.
  • This paper is a first proof of concept. Of course such deep RBF networks need more research, in particular research on different types of activation functions, distance functions as well as parameter initialization and parameter optimization is necessary. Accordingly, this paper is just the very first step and may lead to a new direction of research in deep RBF architectures.
The remainder of this work is structured as follows. In Section 2 we give a short introduction to shallow RBF networks and describe the main components and methods necessary to train deep RBF networks. Then we briefly describe our experimental framework as well as report the results of our experiments in Section 3. Finally we discuss our results in Section 4, followed by a conclusion and a proposal for possible future research in Section 5.

2. Deep RBF networks

In this work we focus on multi-layer RBF networks, which we call deep RBF networks. We keep the overall RBF architecture consisting of an input layer and a linear output layer. In contrast to regular RBF networks we allow for multiple hidden layers with RBF kernels. Finally, we distinguish between fully connected and partially connected RBF layers. In the latter we follow a similar approach to CNNs, resulting in a patch-wise distance calculation for the RBF kernels. We use partially connected layers to extract features, followed by a regular (shallow) fully connected RBF network for classification.

2.1. Shallow RBF networks

A typical shallow RBF network consists of a single layer of RBF kernels, followed by a fully connected linear layer. The linear layer combines the output of the RBF kernels based on the weight matrix W :
y j ( x ) = i = 1 k w i j · h i ( x )
Figure 1. RBF network consisting of a layer of k distance based RBF kernels as well as a linear weighted output of the kernels activations. The network consists of n input neurons and m output neurons.
Figure 1. RBF network consisting of a layer of k distance based RBF kernels as well as a linear weighted output of the kernels activations. The network consists of n input neurons and m output neurons.
Preprints 99242 g001
The overall architecture is shown in Figure 1. The kernel activation h i ( x ) of neuron i is calculated by the squared Mahalanobis distance between center c i and the input vector x , followed by an activation function f:
h i ( x ) = f x c i R i 2
The Mahalanobis distance is specified by a positive-definite matrix R i with
x c i R i = ( x c i ) T R i ( x c i )
We call R i the Mahalanobis distance matrix in the following. In contrast to other neural networks, RBF networks use a distance based activation provided by the RBF kernels. Commonly the Gaussian function is used as activation function. A selection of activation functions used for RBF networks is given below, where r is the squared Mahalanobis distance:
(4) Gaussian h ( r ) = e r (5) Quadratic h ( r ) = 1 r (6) Multiquadric h ( r ) = 1 + r (7) Inverse quadratic h ( r ) = 1 1 + r (8) Inverse multiquadric h ( r ) = 1 1 + r

2.2. Partially connected RBF networks

We consider the context of image classification with 2D-inputs. Given an input image X R W × H × C where W and H denote the image width and height and C the channel dimension. For partially connected layers we follow the idea of the weight sharing approach as used in CNNs. The activation map h ( X ) R W × H × 1 of a single RBF neuron can be calculated as follows:
h ( X ) = f 1 C c = 1 C X c c c R c 2
With center c R N × N × C and Mahalanobis distance matrix R c R N 2 × N 2 × C . The ⊖ operation denotes patch-wise subtraction in a sliding window similar to a convolution operation. Figure 2 illustrates the general idea. A more detailed mathematical description is given below.
Note that in the above formula padding and strides are omitted. As higher numbers of input channels increase the sum, we normalize the sum by 1 C before applying the RBF. Otherwise the output activations tend to zero with increasing number of channels.

2.3. Cholesky decomposition

We use the Mahalanobis distance as distance metric between centers and data points in the RBF kernel. This metric, represented by a Mahalanobis distance matrix, is adapted to the dataset at hand. To ensure at least positive semi-definiteness during training we use the Cholesky decomposition on the Mahalanobis distance matrix R j via a lower triangular matrix G j in the following way
R j = G j G j T
and then optimize only the triangular matrix G j . This results in an alternative formulation of the squared Mahalanobis distance between input x and center c j :
x c j G j 2 = ( x c j ) T G j G j T ( x c j ) = [ G j T ( x c j ) ] T G j T ( x c j ) = G j T ( x c j ) 2 2

2.4. Patch-wise distance calculation as convolution

Calculating the Mahalanobis distance between image patches and centers is an expensive operation. We reformulate the problem in terms of convolutions, which are efficiently implemented in most machine learning frameworks. In the following equations the indices indicating individual neurons and channels are omitted. Additionally, the convolution operator ∗ is used.

2.4.1. Euclidean distance

The Euclidean distance between an image patch x and a center c can be reformulated in the following way:
x c 2 = i = 1 N ( x i c i ) 2 = i = 1 N x i 2 2 x i c i + c i 2 = i = 1 N x i 2 + i = 1 N c i 2 2 x , c
The last dot product translates into a convolution when performing the calculation in a sliding window fashion. This results in an efficient implementation of the scaled Euclidean distance. In this case we can assume that G is a diagonal matrix with diagonal elements σ reshaped to a matrix σ R N × N :
X c G 2 = σ X c 2 = ( X 2 σ ) 2 ( X ( σ c ) ) + i = 1 N j = 1 N c i j 2 σ i j
Where ⊙2 denotes element-wise squaring, ⊙ is the Hadamard product and ⊖ is the patch-wise subtraction, performed over a sliding window.

2.4.2. Mahalanobis distance

The calculation of the Euclidean distance is simple due to the decomposition into a dot product and square terms. This is a direct result of the Euclidean distance being characterized by a diagonal distance matrix. For the Mahalanobis distance with arbitrary distance matrix the covariance between feature dimensions has to be considered. We derived the following simplification, consisting of multiple stages of convolution operations. Let X R W × H denote a single channel input image, G R N 2 × N 2 the Cholesky decomposed distance matrix and c R N × N the corresponding neuron center. An image patch p i j at location ( i , j ) is defined as follows:
p i j = x i j x i , j + N 1 x i + N 1 , j x i + N 1 , j + N 1 R N × N
In the following, the centers and image patches are used in their flattened representation, were all rows of the corresponding matrix are concatenated. This is denoted as c ¯ = c 11 , , c 1 N , c 21 , , c N N . For the Mahalnobis distance to a single image patch it holds that:
G T ( p i j ¯ c ¯ ) = g 11 ( x i j c 11 ) + + g 1 , N 2 ( x i + N 1 , j + N 1 c N N ) g N 2 , 1 ( x i j c 11 ) + + g N 2 , N 2 ( x i + N 1 , j + N 1 c N N ) = p i j ¯ T g 1 * c ¯ T g 1 * p i j ¯ T g N 2 * c ¯ T g N 2 *
For the whole image, the dot products are replaced by convolution operations, resulting in the following patch-wise distance calculation:
X c G 2 = i = 0 N X * g ^ i * c ¯ T g i * 2
Where g ^ i * R N × N denotes the i-th row of G , viewed as a square matrix. Note that each row of the Cholesky decomposed matrix G has to be convolved with the input image. With increasing kernel size N and larger input images, this calculation may be slow and memory intensive.

2.5. Parameter initialization

We follow commonly used strategies to initialize the centers and matrices. The centers are initialized by k-means clustering while the distance matrices are initialized by a PNN heuristic. To initialize subsequent layers we use a cascading scheme where we initialize the parameters of a layer with the output of the previous layer. To initialize the whole network, the input data is propagated through all layers. During initialization, the last classification layer can be viewed as a single linear transformation with fixed input. Thus, we can make use of the pseudo-inverse solution to initialize the weights of the linear output layer.

2.5.1. K-means clustering

The centers c j of a single neuron j are initialized by k-means clustering on the corresponding input data. Clustering can be performed in an unsupervised manner, where global statistics of the input are extracted as centers. In this case, non-labeled data can be used to find good initial parameters. Additionally, class labels can be leveraged in the clustering process by performing the clustering process on data of a given class. This results in more class specific statistics as initial centers. We refer to this approach as class-specific k-means clustering in contrast to global k-means clustering, where the centers are selected by clustering the whole input data. An advantage of class-specific k-means may be that the number of centers per class can be adjusted to e.g. counteract underrepresented classes. However, a heuristic is needed to determine the number of selected centers per class. For very small data sets the clustering approach is not possible when using more neurons than distinct clusters, limiting the number of obtainable initial centers. To speed up the clustering process we use mini-batch k-means clustering as proposed in [21].

2.5.2. Patch-wise k-means clustering

For the partially connected networks we initialize the centers by extracting image patches from all training samples and performing global or class-specific k-means clustering on those patches. Patches are extracted by using a sliding window over the training samples. Each input image generates several image patches, resulting in a large number of image patches.
Additionally, processing of individual input channels has to be considered. One possibility to address this issue is to perform clustering for each channel individually. However, with increasing number of input channels this approach is not feasible. We propose to view an input image together with its channels as a volume and perform clustering over this volume. This approach is fast but leads to correlated channels, which may reduce model flexibility.

2.5.3. Mahalanobis distance matrix initialization

Initially, the distance matrix is approximated by the following p nearest neighbour PNN heuristic on the centers:
σ j = α p · i N c j c i 2
Where N contains the p nearest neighbors of center c j , based on the Euclidean distance. The parameter α is a scaling factor which has to be set heuristically. The goal of this heuristic is to achieve a good coverage of the input space by adjusting the kernel widths σ k as proposed by [9]. This PNN heuristic is fast to calculate as it is solely based on the centers and does not depend on the number of training samples. However, the heuristic fails for non-distinct or too similar centers, where the distance approaches zero. In this case we set σ j = 1 , which results in an Euclidean distance matrix. This approach yields a diagonal matrix with width parameters for each input dimension on the diagonal:
G j = 1 σ j · I

2.6. Training procedure

After initialization, the full network is trained by back-propagation. We use a softmax activation in the final linear layer in combination with the cross-entropy loss for classification.
For the Mahalanobis distance matrices, training can be done either on the full Cholesky decomposed matrix i.e on the lower triangular matrix G j , or to reduce complexity only on the diagonal entries of the triangular distance matrix. The latter implicitly forces the assumption that the feature dimensions are uncorrelated by ignoring the covariance, resulting in a feature-wise scaled version of the Euclidean distance. The number of parameters used for the Cholesky decomposed matrix (which is triangular) and a single neuron with input dimensionality C is C ( C + 1 ) 2 . Using the diagonal entries of the Mahalanobis distance matrix reduces the number of parameters to C, which improves training speed.
Additionally, we propose a supervised approach to pre-train subsequent layers by temporarily adding a fully connected softmax layer for classification and minimizing cross entropy loss, virtually resulting in training a single layer RBF network for each layer.
To get an impression of how good individual classifiers perform we further combine the prediction of multiple classifiers as follows: Each classifier makes an individual prediction, which is selected by its certainty. The certainty is based on the output of the softmax layer. Predictions with a certainty above 95 % are summed and the final prediction is obtained by an argmax operation on the summed output distribution.

3. Experimental Evaluation

Similar to CNNs, our architecture makes use of data which is strictly arranged on a uniform grid like images. Thus, we evaluate our approach only on 2D structured data. Namely we use common image classification tasks as well as speech emotion recognition tasks where we pre-process the speech data into MEL spectrograms of fixed size.
For both types of tasks we use the network architecture depicted in Figure 3, consisting of partially as well as fully connected RBF kernels. Task dependent details and hyperparameters are given in the corresponding section.

3.1. Datasets

In the following we describe the used datasets in more detail. For image classification we include the common benchmark datasets MNIST [22] and CIFAR10 [23]. For both datasets we first scale the images to the interval [ 0 , 1 ] . Besides classic image classification benchmarks, we also include the following two speech emotion recognition datasets for evaluation: RAVDESS [24] and EmoDB [25].
MNIST is a gray-scale image dataset often used to evaluate machine learning models. The dataset consists of 28 × 28 pixel sized images of hand-written digits from zero to nine, resulting in a 10-class classification problem. In total there are 60 , 000 training samples and 10 , 000 test samples.
CIFAR10 consists of RGB images from ten different classes, containing different objects and animals. Each image has 32 × 32 pixels, resulting in 3072 features per image. The whole dataset is similarly sized to the MNIST dataset, containing 50 , 000 training samples and 10 , 000 test samples.
EmoDB is a small emotion recognition dataset where neutral sentences are spoken by ten different actors in seven different emotions. The following emotions are contained in the data: anger, boredom, anxiety, happiness, sadness, disgust and neutral. In total the dataset consists of 535 speech recordings. As the dataset is rather small, we also consider a simpler binary classification task based on arousal. For this task we define anger, anxiety, happiness, sadness and disgust as emotions with high arousal. Boredom, sadness and neutral are defined as emotions with low arousal. We reserve 20 % of the samples for testing and split the remaining data into 20 % validation and 80 % training data.
RAVDESS is an acted emotion dataset consisting of speech, song and videos with speech. In total there are 24 actors which speak and sing sentences of the form "Kids are talking by the door." in eight different emotions. Additionally, portrait videos of the actors speaking the same sentences are available. The following eight emotions are included: neutral, calm, happy, sad, angry, fearful, disgust and surprised. Note that there are different emotions than in the EmoDB dataset. More precisely, boredom and anxiety is missing in the RAVDESS dataset, while calm and surprised occurs only in the RAVDESS dataset. In total the dataset consists of 2452 sound files, where 1440 are spoken sentences files and the remaining 1012 are sung sentences. The video files are omitted for our tasks as they contain the same audio as the spoken sentences. We use 20 % of the data for testing and split the remaining data into 20 % validation and 80 % training data.

3.2. Image classification

Architecture Our image classification architecture is composed as follows. First we build blocks of two partially connected layers each, followed by a max pooling operation to reduce the spatial dimensions. In total we use three such blocks with 20 RBF kernels in each block. All kernels use the full Mahalanobis distance matrix and are initialized with class-specific k-means clustering. Finally, a RBF network with 128 neurons is used for classification. The classification network uses the Euclidean distance, which allows for more individiual neurons compared to the Mahalanobis distance, while keeping the number of parameters low. This is necessary as the input dimension to the classifier is rather large, resulting in a large number of parameters when using the Mahalanobis distance.
Training We use the Adam optimizer in its standard configuration for all of our experiments. We train our models on augmented data, where we shift, scale, flip and rotate the input images to generate more training data. The whole network is trained for 200 epochs, individual layers are not pre-trained.
Results The results for the MNIST dataset are summarized in Table 1. In average, our RBF architecture reaches a test accuracy of 99.5 % with a low standard deviation. Using 5 classifiers as ensemble further increases the test accuracy to 99.67 % . Nonetheless, the performance is not on-par with state-of-the art CNN architectures, achieving over 99.8 % with a similar number of parameters.
In Table 2 the results for the CIFAR10 dataset are summarized. Our RBF approach reaches a mean test accuracy of 80.72 % with a rather high standard deviation of 0.63 % . Performing an ensemble classification with 5 classifiers improves the performance to 85.01 % , similar to early CNN architectures like [26]. Note that more recent architectures are able to reach test accuracies beyond 99 % on CIFAR10 [28,29].

3.3. Speech Emotion Recognition

Architecture Our emotion recognition RBF networks consist of four partially connected layers with 8 neurons each, followed by a RBF classifier with 128 neurons. Again, the RBF network uses the Euclidean distance while all partially connected layers use the full Mahalanobis distance matrix. The centers are initialized with global k-means clustering. We use global k-means clustering as we only use 8 kernels, while distinguishing 7 different classes. Due to the high input dimensions, more kernels would increase the number of parameters which we aim to keep low.
Training We first extract the MEL spectrogram based features from the audio files for both datasets. We limit the samples to a duration of three seconds and use use 128 filters for the MEL spectrogram. In total our pre-processing yields a 128 × 256 feature map for each speech sample. As both emotion datasets are rather small, we augment the spectrograms by randomly shifting the whole spectrogram in the time domain as well as masking random parts of the frequency and time domain as proposed by [31]. The network is optimized by using the Adam optimizer with default parameters. The whole network is trained for 200 epochs without pre-training individual layers.
Table 3. Test accuracy comparison on EmoDB.
Table 3. Test accuracy comparison on EmoDB.
Model Test accuracy Parameters
CNN (MEL spectrograms) [32] 72.06 -
DeepRBF 72.15 ± 1.24 281,040
DeepRBF Ensemble a 74.76 1,405,200
GMM (MFCC) [33] 79.8 -
DeepRBF (binary) b 96.07 ± 0.11 280,400
DeepRBF Ensemble (binary) ab  96.26 1,402,000
a Ensemble of 5 trained classifiers; b Problem reduced to binary classification task.
Table 4. Test accuracy comparison on RAVDESS.
Table 4. Test accuracy comparison on RAVDESS.
Model Test accuracy Parameters
DeepRBF (speech) 67.36 ± 0.91 281,168
DeepRBF (speech + song) 71.41 ± 1.09 281,168
CNN14 (speech) [34] 72.10  b 79,690,184
DeepRBF Ensemble (speech) a 74.30 1,405,840
ERANN (MEL spectrograms, speech) [35] 74.80  b 24,023,562
DeepRBF Ensemble (speech + song) a 78.82 1,405,840
a Ensemble of 5 trained classifiers; b 4-fold cross-validated results.
Results In all three instances of the two multi-class emotion recognition tasks we observe a high standard deviation concerning the test accuracy. For the EmoDB dataset we achieve a mean test accuracy of 72.15 % and 74.76 % for a single classifier and ensemble classification respectively. Constraining the task to a binary problem gives more consistent results with a mean test accuracy of over 96 % and a lower standard deviation. Those results seem consistent with a CNN approach on the same MEL spectrograms [32] reaching 72.06 % test accuracy. More sophisticated approaches like [33] are able to achieve even better results.
Regarding the RAVDESS dataset we observe similar performance as the considered CNN architectures [34,35], while using significantly less trainable parameters. Namely we reach a test accuracy of 71.41 % and 74.30 % for a single classifier and ensemble classification respectively. Again we observe comparable results to a CNN architecture based on MEL spectrograms [35].
Including the song samples for the RAVDESS dataset further improves the testing accuracy on the combined dataset to 78.82 % .

4. Discussion

A common observation between all evaluated datasets is the increased performance when considering ensemble classification of multiple RBF classifiers. This indicates that the full potential of our approach is not yet reached when considering single classifiers. This assumption is also reinforced by the high variance in test accuracy of single classifiers.
For the emotion recognition datasets the variance in test accuracy is higher than for the image classification tasks. We attribute this behaviour to the low number of training samples in those datasets.
Note that for the emotion recognition tasks the comparison between different models is complicated due to inconsistent feature selection and evaluation procedures. CNN14 and ERANN use 4-fold cross validated results [34,35] while we only report the mean test accuracy over a fixed test set (containing 20 % of the samples). Overall we observe similar performance between our approach and the well established CNN architectures on all considered datasets which is a promising.

5. Conclusion

This work provides a new way of using RBF kernels in deep neural networks by introducing an initialization scheme suitable for multi-layered RBF architectures. Additionally, we introduce a partially-connected RBF layer similar to CNNs. By utilizing the Cholesky decomposition we guarantee positive semi-definiteness of the learnable distance matrix used in the Mahalanobis distance calculation. We showed that our proposed architecture performs well on 2D structured data like images and MEL spectrograms. In contrast to CNNs, the distance based activation used in our RBF kernels favors interpretability of the underlying calculation. Nonetheless the flexibility of the proposed deep RBF approach - for instance due to the Mahalanobis distances offering complex approximations of data densities - can be a disadvantage in other contexts, such as limited scalability to larger network architectures. All in all the proposed deep RBF architecture is just the very first step in this direction of research.

Author Contributions

Conceptualization, F.W. and F.S.; methodology, F.W. and F.S.; software, F.W.; validation, F.W. and F.S.; formal analysis, F.W. and F.S.; investigation, F.W.; writing—original draft preparation, F.W.; writing—review and editing, F.W. and F.S; supervision, F.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
RBF Radial Basis Function
CNN Convolutional Neural Network
PNN p-nearest Neighbour

References

  1. Franke, R. Scattered data interpolation: tests of some methods. Mathematics of computation 1982, 38, 181–200. [Google Scholar]
  2. Powell, M.J.D., Radial Basis Functions for Multivariable Interpolation: A Review. In Algorithms for Approximation; Clarendon Press: USA, 1987; pp. 143–167.
  3. Buhmann, M.D. Radial Basis Functions: Theory and Implementations; Cambridge Monographs on Applied and Computational Mathematics, Cambridge University Press, 2003. [CrossRef]
  4. Fasshauer, G.E. Meshfree approximation methods with MATLAB; Vol. 6, World Scientific, 2007.
  5. Micchelli, C.A. Interpolation of scattered data: Distance matrices and conditionally positive definite functions. Constructive Approximation 1986, 2, 143–145. [Google Scholar] [CrossRef]
  6. Dyn, N. Interpolation and approximation by radial and related functions. Approximation Theory 1989, 6, 211–234. [Google Scholar]
  7. A practical guide to radial basis functions. Electonic Resource 2007, 11, 1–12.
  8. Broomhead, D.S.; Lowe, D. Multivariable Functional Interpolation and Adaptive Networks. Complex Syst. 1988, 2. [Google Scholar]
  9. Moody, J.; Darken, C.J. Fast Learning in Networks of Locally-Tuned Processing Units. Neural Computation 1989, 1, 281–294. [Google Scholar] [CrossRef]
  10. Park, J.; Sandberg, I.W. Universal Approximation Using Radial-Basis-Function Networks. Neural Computation 1991, 3, 246–257. [Google Scholar] [CrossRef] [PubMed]
  11. Schwenker, F.; Kestler, H.; Palm, G.; Hoher, M. Similarities of LVQ and RBF learning-a survey of learning rules and the application to the classification of signals from high-resolution electrocardiography. Proceedings of IEEE International Conference on Systems, Man and Cybernetics. IEEE, 1994, Vol. 1, pp. 646–651.
  12. Schwenker, F.; Kestler, H.A.; Palm, G. Unsupervised and supervised learning in radial-basis-function networks. In Self-Organizing neural networks: recent advances and applications; Springer, 2002; pp. 217–243.
  13. Glodek, M.; Schels, M.; Schwenker, F. Ensemble Gaussian mixture models for probability density estimation. Computational statistics 2013, 28, 127–138. [Google Scholar] [CrossRef]
  14. Han, Z.; Qian, X.; Huang, H.; Huang, T. Efficient design of multicolumn RBF networks. Neurocomputing 2021, 450, 253–263. [Google Scholar] [CrossRef]
  15. Hady, M.F.A.; Schwenker, F.; Palm, G. Semi-supervised learning for tree-structured ensembles of RBF networks with co-training. Neural Networks 2010, 23, 497–509. [Google Scholar] [CrossRef] [PubMed]
  16. Schwenker, F.; Kestler, H.A.; Palm, G. Three learning phases for radial-basis-function networks. Neural Networks 2001, 14, 439–458. [Google Scholar] [CrossRef] [PubMed]
  17. Zhou, M.; Gao, F.; Chao, J.; Liu, Y.X.; Song, H. Application of radial basis functions neutral networks in spectral functions. Physical Review D 2021, 104, 076011. [Google Scholar] [CrossRef]
  18. Sun, L.; Li, S.; Liu, H.; Sun, C.; Qi, L.; Su, Z.; Sun, C. A Brand-New Simple, Fast, and Effective Residual-Based Method for Radial Basis Function Neural Networks Training. IEEE Access 2023, 11, 28977–28991. [Google Scholar] [CrossRef]
  19. Amirian, M.; Schwenker, F. Radial Basis Function Networks for Convolutional Neural Networks to Learn Similarity Distance Metric and Improve Interpretability. IEEE Access 2020, 8, 123087–123097. [Google Scholar] [CrossRef]
  20. Sinha, T.; Verma, B. Fast Convolutional Neural Network with iterative and non-iterative learning. Applied Soft Computing 2022, 125, 109197. [Google Scholar] [CrossRef]
  21. Sculley, D. Web-Scale k-Means Clustering. Proceedings of the 19th International Conference on World Wide Web; Association for Computing Machinery: New York, NY, USA, 2010; WWW ’10; pp. 1177–1178. [Google Scholar] [CrossRef]
  22. LeCun, Y.; Cortes, C.; Burges, C. MNIST handwritten digit database. ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist 2010, 2.
  23. Krizhevsky, A.; Nair, V.; Hinton, G. CIFAR-10 (Canadian Institute for Advanced Research).
  24. Livingstone, S.R.; Russo, F.A. The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS), 2018. [CrossRef]
  25. Burkhardt, F.; Paeschke, A.; Rolfes, M.; Sendlmeier, W.; Weiss, B. A database of German emotional speech. 2005, Vol. 5, pp. 1517–1520.
  26. Zeiler, M.D.; Fergus, R. Stochastic Pooling for Regularization of Deep Convolutional Neural Networks. 1st International Conference on Learning Representations, ICLR 2013, Scottsdale, Arizona, USA, May 2-4, 2013, Conference Track Proceedings; Bengio, Y.; LeCun, Y., Eds., 2013.
  27. Byerly, A.; Kalganova, T.; Dear, I. No Routing Needed Between Capsules, 2021, [arXiv:cs.CV/2001.09136].
  28. Tan, M.; Le, Q.V. EfficientNetV2: Smaller Models and Faster Training. CoRR 2021, abs/2104.00298, [2104.00298].
  29. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. CoRR 2020, abs/2010.11929, [2010.11929].
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. CoRR 2015, abs/1512.03385, [1512.03385].
  31. Park, D.S.; Chan, W.; Zhang, Y.; Chiu, C.C.; Zoph, B.; Cubuk, E.D.; Le, Q.V. SpecAugment: A Simple Data Augmentation Method for Automatic Speech Recognition. Interspeech 2019 2019. [Google Scholar] [CrossRef]
  32. Chauhan, K.; Sharma, K.K.; Varma, T. Speech Emotion Recognition Using Convolution Neural Networks. 2021 International Conference on Artificial Intelligence and Smart Systems (ICAIS), 2021, pp. 1176–1181. [CrossRef]
  33. Chaspari, T.; Dimitriadis, D.; Maragos, P. Emotion classification of speech using modulation features. 2014 22nd European Signal Processing Conference (EUSIPCO), 2014, pp. 1552–1556.
  34. Kong, Q.; Cao, Y.; Iqbal, T.; Wang, Y.; Wang, W.; Plumbley, M.D. PANNs: Large-Scale Pretrained Audio Neural Networks for Audio Pattern Recognition, 2020, [arXiv:cs.SD/1912.10211].
  35. Verbitskiy, S.; Berikov, V.; Vyshegorodtsev, V. ERANNs: Efficient Residual Audio Neural Networks for Audio Pattern Recognition, 2021, [arXiv:cs.SD/2106.01621].
Figure 2. Patch-wise distance calculation, followed by a RBF.
Figure 2. Patch-wise distance calculation, followed by a RBF.
Preprints 99242 g002
Figure 3. Full RBF architecture used for image classification and emotion recognition.
Figure 3. Full RBF architecture used for image classification and emotion recognition.
Preprints 99242 g003
Table 1. Test accuracy comparison on MNIST.
Table 1. Test accuracy comparison on MNIST.
Model Test accuracy Parameters
DeepRBF 99.50 ± 0.07 229,160
Stochastic Pooling b [26] 99.53 -
DeepRBF Ensemble a 99.67 1,145,800
CNN + Vector Capsules b [27] 99.87 1,514,187
a Ensemble of 5 trained classifiers; b CNN architecture.
Table 2. Test accuracy comparison on CIFAR10.
Table 2. Test accuracy comparison on CIFAR10.
Model Test accuracy Parameters
DeepRBF 80.72 ± 0.63 268,600
Stochastic Pooling b [26] 84.87 -
DeepRBF Ensemble a 85.01 1,343,000
ResNet110 b [30] 93.57 1,7000,000
EfficientNetV2 b [28] 99.10 121,000,000
ViT-H/14 c [29] 99.50 ± 0.06 632,000,000
a Ensemble of 5 trained classifiers; b CNN architecture; c Transformer architecture.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated