1. Introduction
Sea clutter, backscatter from the ocean surface, is a challenge for radars to detect and classify objects at sea, and therefore it is important for radars to deal with sea clutter in various sea conditions for accurate detection and performance evaluation [
1]. However, due to the expensive and time-consuming nature of acquiring the desired sea clutter data, intensive studies have been focused on mathematical modeling of sea clutter data based on the random process model to generate abundant sea clutter data in a short time [
2,
3,
4,
5,
6].
Several distributions have been considered for modeling sea clutter amplitude. Early models adopted the Rayleigh distribution to characterize sea clutter [
7], but measurements have shown deviations from this model and other alternatives have emerged such as log-normal [
8], K-distribution [
9], Weibull [
10]. Among them, K-distribution is most widely used due to its accuracy [
11,
12] and its modified versions, KA [
13] and KK [
14] distributions, have been proposed to improve the accuracy on tail distribution at low grazing angles [
15], To reduce the model complexity of KA, KK distributions, relatively simple Pareto distribution [
16,
17] is considered to fit both high and low grazing angle. In addition, the Memoryless Nonlinear Transform (MNLT) method [
18], which allows the generation of samples with a specific correlation function, was used to model the spatiotemporal correlation. Although these parametric methods based on mathematical modeling have shown success in finding an appropriate distribution to fit the sea clutter data under ideal assumptions, they are still insufficient to model the full range of geometry and sea conditions in practice, especially for the time correlation [
19].
Recently, the use of deep learning for data augmentation has become widespread in many fields. The advantages of deep learning approaches in data augmentation are that data with approximately the same statistical properties as sample data, such as distribution and correlation, can be generated without explicit mathematical models, and several generative techniques can be used to generate unlearned data with interpolated features [
20,
21]. There are a few attempts to generate sea clutter data using deep learning [
22,
23]. In [
22], a Generative Adversarial Network (GAN) with two generators that independently generate the real and imaginary components of sea clutter data is proposed, and in [
23], a so-called Auxiliary Classifier GAN (AC-VAEGAN) is proposed to generate one-dimensional (1-D) sea-land clutter amplitude data to deal with imbalanced data sets in classification tasks.
These existing methods are based on one-dimensional data generation, and therefore multi-dimensional structures such as real-imaginary dependency are difficult to implement. More importantly, since GAN independently generates data block by block, the length of the coherently generated data is limited to the length of the training data. However, the non-blockwise generative algorithms based on RNN and its variants tend to be inaccurate as the length of the sample sequence increases [
24,
25]. Therefore, it is desirable to design a deep generative algorithm capable of generating arbitrarily long multidimensional data from relatively short length of training data.
In this paper, it is proposed a novel sea clutter data augmentation scheme based on conditional FLOW [
26], a variant of FLOW [
27] which is one of the most robust deep generative methods. The proposed algorithm, named globally conditioned conditional flow (GCC-FLOW ), is a modified version of conditional flow that introduces a global condition variable to reflect the sea state condition. The autoregressive use of GCC-FLOW is able to generate arbitrarily long length of sea clutter sequences with the approximately the same statistical characteristics of the given sea clutter data. Furthermore, by using the global condition variable of the GCC-FLOW for different sea states, the proposed method is capable of generating diverse sea clutter data with the interpolated sea state features.
The rest of this paper is organized as follows: In
Section 2, the architecture of the proposed GCC-FLOW is described in detail. In
Section 3, the training sea clutter data, hyperparameter settings for the GCC-FLOW to generate sea clutter data are described. In
Section 4, the experimental results are presented in terms of ensemble distribution and time correlations of the generated sea clutter compared to those of the training data. Finally, the conclusion is drawn in
Section 5.
2. Methods
2.1. Autoregressive Data Generation Using GCC-FLOW
We assume that the sea clutter data is a stationary random process as assumed in most of the literature [
2,
3,
4,
5,
28]. Existing generative methods based on the GAN sample sea clutter sequence
following the joint distribution
learned from the train data. In order to increase
L, the data size should be increased. The concatenation of two data
and
by generated GAN cannot be used as a sequence of the sampled random process of length
, since
and
are independent due to the block-by-block data generation of GAN.
Figure 1 illustrates the autoregressive use of the proposed Globally Conditioned Conditional FLOW (GCC-FLOW ) in order to generate long data sequences. GCC-FLOW generates a sequence
given a conditional sequence
following a conditional distribution
parameterized by a global conditional variable
. The conditional pdf
is learned by GCC-FLOW to approximate the conditional distribution of the successive data
given
for
. Given the initial sequence
from the data,
is generated by the GCC-FLOW and
is used as the condition sequence for
for
. The joint distribution of the concatenated sample sequence
satisfies
assuming the block Markov property,
Since each conditional distribution is approximated by learned from GCC-FLOW , the joint distribution generated by the GCC-FLOW autoregressively approximates the true data distribution as desired.
2.2. Architecture of GCC-FLOW
The structure of GCC-FLOW is similar to conditional-FLOW [
26] except that the global condition variable
is fed into the coupling layer in the FLOW step as shown in
Figure 2. GCC-FLOW consists of a pair of FLOW networks, the condition part and the output part, which basically has two identically independent deep network structures that are used repeatedly in the multi-scale frame;
and
, parameterized by the network weights
and
respectively, where
is the index for multi-scale stages and
is the index for flow steps in a multi-scale stage, and
for the condition part and
for the output part.
In each flow step,
and
are updated as the following
where ⊙ is element-wise multiplication and
,
are the result of halving the channel of
, respectively.
The structure of the coupling layer function
and the conditional coupling layer function
is described in
Figure 3. The input tensors
and
, both of size
, are converted to the
tensors, denoted by
,
, and
, by convolution networks. The global variable
is multiplied by a linear weight
of size
(
). These two tensors,
and
, are properly scaled and added channel-wise via actnorm. For example, in the coupling layer part,
where
denote the intermediate output tensor of the size
and
is the
i-th channel vector of
and
are the actnorm parameters for
and
are the actnorm parameters for
. Finally,
is processed through a convolution network to restore the original size
and split it in half channel-wise to produce
and
.
At each multi-scale stage, the output is split in half channel-wise, and the first half becomes the m-th stage output, , and the remaining half is squeezed, i.e., the length is reduced by half and the channel size is doubled, and fed into the next flow step. Note that in this split-squeeze setting the length of at each multi-scale stage becomes , where L is the input length of GCC-FLOW .
In the training phase, networks are trained for
to be i.i.d. Gaussian,
, and in the generation phase the sampled
from i.i.d. Gaussian
are used as input to the inverse network of GCC-FLOW , which is canonically given by the structure of FLOW [
27].
3. Experiment Settings
3.1. Data Set
The sea clutter data used in the experiments come from the
Dartmouth and
Grimbsby databases collected by the IPIX radar [
29]. The following two data datasets are used:
- 1)
High sea state data: The 3rd range bin of the data file #269 at Dartmouth labeled as high. The average wave height is 1.8m (max 2.9m)
- 2)
Low sea state data: The 5th range bin of the data file #287 at Dartmouth labeled as low. The average wave height is 0.8m (max 1.3m)
All the data samples are complex-valued and the length of the high sea state data and the low sea state data is
. The detailed specification of the data, such as resolution and sampling rate, can be found in [
29].
3.2. GCC-FLOW Settings
The length of and is set to 512, i.e. , and the height of the clutter data and is set to 2, i.e., , since the data represent single range bin complex data. The number of multiscale stages of GCC-FLOW is set to 4, i.e., and in a multiscale stage the number of FLOW iterations is set to . The channel size in the coupling layer is . The model is trained on NVIDIA RTX 3090 with a learning rate of , batch size of 4, and trained on 512 randomly selected sequences with length of 1024 by randomly sampled starting points.
4. Experiment Results
4.1. Sea Clutter Augmentation Using GCC-FLOW
Figure 4 compares the magnitude histogram of the generated sea clutter data by the GCC-FLOW trained on the high sea state data and on the low sea state data, respectively, with the magnitude histograms of the original data.
Figure 5 compares the magnitude autocorrelation of the generated sea clutter data by the GCC-FLOW trained on the high sea state data and on the low sea state data, respectively, with that of the original high sea state and the low sea sate data, where the magnitude autocorrelation is defined as
where
,
, and
L is set to 512.
The GCC-FLOW is trained using 256 blocks of 512 consecutive samples (131,072 training samples in total) with the global condition variable set to 0. GCC-FLOW generates 262,144 consecutive samples and the first 131,072 samples are used to plot the histogram and the autocorrelation function.
To confirm that the generated samples have the same statistical properties at each time point, consecutive
samples are taken from the beginning (
), from the middle (
), and from the end (
).
Figure 6 and
Figure 7 show the magnitude histogram and the magnitude autocorrelation of the generated sea clutter data using GCC-FLOW trained from the high sea state data and the low sea state data, respectively, at the three time points. Note that the characteristic of the high sea data (concentrated distribution and high frequency fluctuation in correlation) and the characteristic of the low sea data (spread distribution and low frequency fluctuation in correlation) are preserved in all time points.
4.2. Sea Clutter generation using GCC-FLOW with global condition
Using the global condition variable
, a single GCC-FLOW can be trained on two different datasets; high sea state data with
and low sea state data with
.
Figure 8 and
Figure 9 plot the magnitude histogram and magnitude autocorrelation of the generated sea clutter data using GCC-FLOW with
and
, respectively, at the three different time points (beginning, middle, and end) from the
samples.
The benefit of this single mode training on two different datasets is the feature interpolation, i.e., by adjusting
between 0 and 1, the GCC-FLOW can generate data with interpolated statistical characteristics between the high and low sea state.
Figure 10 shows the magnitude histogram of generated data for
at three different time points (beginning, middle, and end) of
samples.
Figure 11 presents the autocorrelation of generated data for
at three different time points (beginning, middle, and end) of
samples. In both figures, as
increases from 0 to 1 the characteristics of magnitude distribution and autocorrelation transforms from the low sea state characteristics to the high sea state characteristics. This suggests that by using GCC-FLOW with a global condition variable tuned to sea state, various sea state sea clutter can be generated without the need for the dedicated training datasets for different sea states.
5. Conclusions
A novel work, Globally Conditioned Conditional FLOW (GCC-FLOW ), has been proposed for sea clutter data augmentation due to the expensive and time-consuming nature of acquiring the desired sea clutter data. Experimental results show that GCC-FLOW successfully generates sequences of arbitrary length with the same statistical properties of the given sea clutter data, in contrast to the existing block-by-block methods. Moreover, by introducing a global state variable, the proposed GCC-FLOW can generate sea clutter with different sea state by feature interpolation without requiring the training data with a target feature.
Funding
This study is the result of research funded by the Agency for Defense Development (contract number UI220077JD).
Institutional Review Board Statement
Not Applicable
Informed Consent Statement
Not Applicable
Data Availability Statement
Conflicts of Interest
The authors declare no conflicts of interest.
References
- K. D. Ward, R.J.A.T.; Watts, S. Sea clutter: Scattering, the K distribution and radar performance. Waves in Random and Complex Media 2007, 17, 233–234. [CrossRef]
- E. Conte, M.L. Characterisation of radar clutter as a spherically invariant random process. IEE Proceedings F (Communications, Radar and Signal Processing) 1987, 134, 191–197(6). [CrossRef]
- Rangaswamy, M.; Weiner, D.; Ozturk, A. Computer generation of correlated non-Gaussian radar clutter. IEEE Transactions on Aerospace and Electronic Systems 1995, 31, 106–116. [Google Scholar] [CrossRef]
- Barnard, T.; Weiner, D. Non-Gaussian clutter modeling with generalized spherically invariant random vectors. IEEE Transactions on Signal Processing 1996, 44, 2384–2390. [Google Scholar] [CrossRef]
- Armstrong, B.; Griffiths, H. Modelling spatially correlated K-distributed clutter. Electronics Letters 1991, 27, 1355–1356. [Google Scholar] [CrossRef]
- Wang, J.; Xu, X. Simulation of Correlated Low-Grazing-Angle Sea Clutter Based on Phase Retrieval. IEEE Transactions on Geoscience and Remote Sensing 2015, 53, 3917–3930. [Google Scholar] [CrossRef]
- Goldstein, H. Sea echo. In Propagation of Short Radio Waves; McGraw-Hill New York, 1951; Vol. 13, pp. 481–527.
- Trunk, G.V.; George, S.F. Detection of Targets in Non-Gaussian Sea Clutter. IEEE Transactions on Aerospace and Electronic Systems 1970, AES-6, 620–628. [Google Scholar] [CrossRef]
- Jakeman, E.; Pusey, P. A model for non-Rayleigh sea echo. IEEE Transactions on Antennas and Propagation 1976, 24, 806–814. [Google Scholar] [CrossRef]
- Matsuo Sekine, Y.H.M. Weibull Radar Clutter; Radar, Sonar and Navigation, Institution of Engineering and Technology, 1990.
- Ishikawa, Y.; Sekine, M.; Musha, T. Observation of k-distributed sea clutter via an x-band radar. Electronics and Communications in Japan (Part I: Communications) 1994, 77, 72–82. [Google Scholar] [CrossRef]
- Bouvier, C.; Martinet, L.; Favier, G.; Artaud, M. Simulation of radar sea clutter using autoregressive modelling and K-distribution. Proceedings International Radar Conference, 1995, pp. 425–430. [CrossRef]
- Middleton, D. New physical-statistical methods and models for clutter and reverberation: the KA-distribution and related probability structures. IEEE Journal of Oceanic Engineering 1999, 24, 261–284. [Google Scholar] [CrossRef]
- Rosenberg, L.; Crisp, D.J.; Stacy, N.J. Analysis of the KK-distribution with X-band medium grazing angle sea-clutter. 2009 International Radar Conference "Surveillance for a Safer World" (RADAR 2009), 2009, pp. 1–6.
- Dong, Y. Distribution of X-band high resolution and high grazing angle sea clutter; Citeseer, 2006.
- FARSHCHIAn, M.; Posner, F.L. The Pareto distribution for low grazing angle and high resolution X-band sea clutter. 2010 IEEE Radar Conference. IEEE, 2010, pp. 789–793.
- Bocquet, S. Simulation of correlated Pareto distributed sea clutter. 2013 International Conference on Radar, 2013, pp. 258–261. [CrossRef]
- Tough, R.J.A.; Ward, K.D. The correlation properties of gamma and other non-Gaussian processes generated by memoryless nonlinear transformation. Journal of Physics D: Applied Physics 1999, 32, 3075. [Google Scholar] [CrossRef]
- Watts, S.; Rosenberg, L. Challenges in radar sea clutter modelling. IET Radar, Sonar & Navigation 2022, 16, 1403–1414. [Google Scholar]
- Lample, G.; Zeghidour, N.; Usunier, N.; Bordes, A.; Denoyer, L.; Ranzato, M. Fader networks: Manipulating images by sliding attributes. Advances in neural information processing systems 2017, 30.
- He, Z.; Zuo, W.; Kan, M.; Shan, S.; Chen, X. AttGAN: Facial Attribute Editing by Only Changing What You Want. IEEE Transactions on Image Processing 2019, 28, 5464–5478. [Google Scholar] [CrossRef] [PubMed]
- Bin DING, Xue XIA, X.L. Sea Clutter Data Augmentation Method Based on Deep Generative Adversarial Network. Journal of Electronics and Information Technology 2021, 43, 1985. [Google Scholar]
- Zhang, X.; Wang, Z.; Lu, K.; Pan, Q.; Li, Y. Data Augmentation and Classification of Sea-Land Clutter for Over-the-Horizon Radar Using AC-VAEGAN. IEEE Transactions on Geoscience and Remote Sensing 2023. [Google Scholar] [CrossRef]
- Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Transactions on Neural Networks 1994, 5, 157–166. [Google Scholar] [CrossRef] [PubMed]
- Pascanu, R.; Mikolov, T.; Bengio, Y. On the difficulty of training recurrent neural networks. Proceedings of the 30th International Conference on Machine Learning; Dasgupta, S.; McAllester, D., Eds.; PMLR: Atlanta, Georgia, USA, 2013; Vol. 28, Proceedings of Machine Learning Research, pp. 1310–1318.
- Pumarola, A.; Popov, S.; Moreno-Noguer, F.; Ferrari, V. C-flow: Conditional generative flow models for images and 3d point clouds. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 7949–7958.
- Kingma, D.P.; Dhariwal, P. Glow: Generative flow with invertible 1x1 convolutions. Advances in neural information processing systems 2018, 31.
- Carretero-Moya, J.; Gismero-Menoyo, J.; Blanco-del Campo, A.; Asensio-Lopez, A. Statistical Analysis of a High-Resolution Sea-Clutter Database. IEEE Transactions on Geoscience and Remote Sensing 2010, 48, 2024–2037. [Google Scholar] [CrossRef]
- Ipix radar website. http://soma.ece.mcmaster.ca/ipix/dartmouth/index.html.
Figure 1.
Autoregressive generation using GCC-FLOW
Figure 1.
Autoregressive generation using GCC-FLOW
Figure 2.
Architecture of GCC-FLOW
Figure 2.
Architecture of GCC-FLOW
Figure 3.
Architecture of coupling networks in GCC-FLOW
Figure 3.
Architecture of coupling networks in GCC-FLOW
Figure 4.
Magnitude distribution the generated data by GCC-FLOW in comparison with the true data: high sea state(left), low sea state(right)
Figure 4.
Magnitude distribution the generated data by GCC-FLOW in comparison with the true data: high sea state(left), low sea state(right)
Figure 5.
Autocorrelation of the generated data by GCC-FLOW in comparison with the true data: high sea state(left), low sea state(right)
Figure 5.
Autocorrelation of the generated data by GCC-FLOW in comparison with the true data: high sea state(left), low sea state(right)
Figure 6.
Distribution(1) and Autocorrelation(2) of the generated data trained on high sea state data: beginning(a), middle(b), end(c)
Figure 6.
Distribution(1) and Autocorrelation(2) of the generated data trained on high sea state data: beginning(a), middle(b), end(c)
Figure 7.
Distribution(1) and Autocorrelation(2) of the generated data trained on low sea state data: beginning(a), middle(b), end(c)
Figure 7.
Distribution(1) and Autocorrelation(2) of the generated data trained on low sea state data: beginning(a), middle(b), end(c)
Figure 8.
Distribution(1) and Autocorrelation(2) of the generated data by GCC-FLOW with : beginning(a), middle(b), end(c)
Figure 8.
Distribution(1) and Autocorrelation(2) of the generated data by GCC-FLOW with : beginning(a), middle(b), end(c)
Figure 9.
Distribution(1) and Autocorrelation(2) of the generated data by GCC-FLOW with : beginning(a), middle(b), end(c)
Figure 9.
Distribution(1) and Autocorrelation(2) of the generated data by GCC-FLOW with : beginning(a), middle(b), end(c)
Figure 10.
Interpolated distribution by the global condition variable 0.2, 0.5, 0.8 : beginning(a), middle(b), end(c)
Figure 10.
Interpolated distribution by the global condition variable 0.2, 0.5, 0.8 : beginning(a), middle(b), end(c)
Figure 11.
Interpolated autocorrelation by the global condition variable 0.2(1), 0.5(2), 0.8(3) : beginning(a), middle(b), end(c)
Figure 11.
Interpolated autocorrelation by the global condition variable 0.2(1), 0.5(2), 0.8(3) : beginning(a), middle(b), end(c)
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).