1. Introduction
Adaptive filtering algorithms play a pivotal role in signal processing, encompassing tasks such as system identification, channel estimation, feedback cancellation, and noise removal [
1]. While literature commonly assumes Gaussian distribution for system noise, real-world scenarios, including underwater acoustics [
2,
3,
4,
5], low-frequency atmospheric disturbances [
6], and artificial interference [
7,
8,
9], often exhibit sudden changes in signal or noise intensity [
10]. These abrupt variations can disrupt algorithms, serving as external solid interference or outliers [
11,
12].
Recently, the sparse quasi-Newton least-mean mixed-norm (SQNLMMN) algorithm [
13] has emerged as a potential solution to mitigate the impact of both Gaussian and non-Gaussian noises on the convergence behavior of adaptive algorithms [
14]. This algorithm introduces a novel cost function incorporating a linear combination of
and
norms while promoting sparsity. Despite its promise, the SQNLMMN algorithm exhibits specific weaknesses. Firstly, it overlooks the presence of input noise at the adaptive filter inputs, leading to biased coefficient estimates. Secondly, the fixed mixing parameter
, governing the balance between the two norms, fails to adapt dynamically. This rigidity in parameter choice trades off convergence rate and mean squared deviation (MSD) concerning the weight coefficients. Notably, the approach of employing a mixed step size for least mean fourth (LMF) and normalized LMF algorithms [
15] to address such trade-offs [
16] differs from the concept of a variable mixing parameter [
17,
18].
Based on the unbiased criteria, several methods have been reported to compensate the biases caused by the noisy inputs, such as the bias compensation normalized least mean square algorithm (BC-NLMS) [
19], bias compensation proportional normalized least mean square algorithm (BC-PNLMS) [
20], bias compensation normalized least mean fourth algorithm (BC-NLMF) [
21], and bias compensation least mean mixed norm algorithm (BC-LMMN) [
22]. However, the BC-LMMN algorithm used a fixed mixing factor, which resulted in a higher misadjustment. In [
23], the authors proposed using a biased-compensated generalized mixed norm algorithm and cooperating with correntropy-induced metric (CIM) as the sparse penalty constraint for sparse system identification problems. Unlike conventional mixed norm approach, they mixed norm with
and
to better combat non-Gaussian noise as well as impulse noise. A modified cost function that considered the cost caused by the input noise was used to compensate for the bias [
23]. Hereinafter, we refer to it as the BC-CIM-LGMN algorithm. Yet, estimating the input noise power might be adversely affected by impulse noise present at the output of the unidentified system. The same trick that adopted CIM as the sparse penalty constraint is applied in [
20], which is referred to as BC-CIM-PNLMS hereinafter. On the other hand, an
norm cost function was used to accelerate the convergence for the sparse systems. In [
24], the authors combined it with an improved BC-NLMS, referred to as the BC-ZA-NLMS algorithm. Unfortunately, the BC-ZA-NLMS algorithm fails to consider the impact of the impulse noise.
However, few studies have comprehensively addressed all impairments, including noisy input, impulse noise in observations (measurements), and sparse unknown systems. Building upon the SQNLMMN algorithm, this paper introduces a robust bias compensation method for the sparse normalized quasi-Newton least-mean with variable-mixing-norm (BC-SNQNLM-VMN) adaptive filtering algorithm. The key contributions of this research are as follows. Firstly, we introduce a normalized variant of the SQNLMN algorithm and incorporate it with the Huber function to alleviate the impact of impulse noise. Secondly, we develop a bias compensation method to counteract the influence of noisy input on the weight coefficients of the adaptive filter. Thirdly, we introduce a convex combination approach concerning the mixing parameter, enabling the utilization of the variable mixing parameter in the mixed norm approach. Consequently, our proposed method can simultaneously achieve rapid convergence and low misadjustment.
The rest of this paper is organized as follows.
Section 2 describes the system model we considered in this work.
Section 3 briefly reviews the SQNLMMN algorithm and outlines the proposed BC-SNQNLM-VMN adaptive filtering algorithm.
Section 4 validates the effectiveness of our proposed BC-SNQNLM-VMN algorithm by using computer simulations. Conclusions are drawn in
Section 5.
2. System Models
The system with finite impulse represented by the vector
to be identified that considers both input noise and observation noise is depicted in
Figure 1. The outputs from this system are subject to corruption by two types of noise. The observable desired signal
can be mathematically defined as:
where
is the transpose symbol; the weight vector
, with components
, arranged in column form, represents the unknown system to be identified.
represents the weight vector of the unknown system;
denotes the input regressor vector. Note that the measurement noise is assumed to consist of two components: background additive white Gaussian noise (AWGN) denoted as
and impulse noise denoted as
. The AWGN noise
has a zero mean and variance
. This variance is related to the signal-to-noise ratio (SNR) as follows:
where
represents the variance of
. In addition, impulse noise is accounted for in the system model. Two conventional models are employed in this study. The first one is the Bernoulli Gaussian (BG) model [
25], defined as:
where,
takes the value of one with a probability of
and zero with a probability of
. Additionally,
represents a white Gaussian process characterized by a mean of zero and a variance of
. The strength of this impulse noise is
is used to quantify its strength as follows:
Another model utilized is the alpha-stable impulse noise model [
13], which can be characterized by the parameter vector
. Here
represents the characteristic factor,
denotes the symmetry parameter,
stands for the dispersion parameter, and
indicates the location parameter. A reduced
value signifies a heightened presence of impulse noise.
In this paper, we consider the noisy input case, i.e., the input of the adaptive filter
differs from that of the unknown system. We assume an AWGN input noise
with zero-mean and variance
is added to the original input
, i.e.,
. The strength of
is determined by the
as follows:
where
denotes the variance of
. The weights of the adaptive filter, denoted by
, are updated iteratively through an adaptive algorithm, which computes correction terms based on
. These corrections rely on the error signal, expressed as:
where
denotes the input regressor vector linked to the adaptive filter.
3. Proposed BC-SNQNLM-VMN Adaptive Filtering Algorithm
3.1. Review of SQNLMMN Algorithm [13]
The cost function of the SQNLMMN algorithm is expressed as:
where
and
are the cost functions for least mean square (LMS) and LMF algorithms, respectively; a fixed mixing parameter
is used to control the mixture of the two cost functions;
denotes the sparsity-promoting term, which is regulated by a positive parameter
. According to [
13], we have the resulting updating recursion of the sparse quasi-Newton least-mean mixed-norm (SQNLMMN) algorithm as follows:
where the step size is chosen as
and
that controls the convergence rate for LMS and LMF algorithms, respectively. Note that
is a common step size;
denotes the sparsity penalty term, and
p denotes the parameter that controls zero-attraction [
13]. The matrix
that approximates the inverse of the Hessian matrix of the cost function can be expressed as follows:
where
is described as follows:
with
and
. Note that
is the Hessian matrix for
. Let
be the
norm and approximate as follows:
where the parameter
is used to determine the region of zero attraction [
26].
The derivation of the gradient for this penalty term is as follows:
where
The operator
denotes the sign function. In order to streamline Eq. (
13), we utilize the first-order Taylor approximation of the exponential function in the following manner:
Therefore, we can approximate Eq. (
13) as follows:
The Hessian of Eq. (
11) can be derived as:
with
3.2. Normalized SQNLMMN
Inspired by the design of normalized LMS and normalized LMF [
27], we propose a normalized version of SQNNLMMN algorithm by modifying Eq. (
8) that considers the noisy inputs
as follows:
where
denotes the noisy input regressor vector and
represents the input noise vector. The noisy error signal
is calculated as follows:
Note that the matrix
is a contaminated version of
(see Eq. (
9)) defined as follows:
where
governs the impact of the penalty term;
denotes a forgetting factor [
13]; the matrix
is a contaminated version of
(see Eq. (
10)) defined as follows:
Note that the difference between and , i.e., , results in the biases during the weight updating process.
3.3. Bias Compensation Design
To compensate for the bias of the normalized SQNLMMN algorithm, we introduce a bias compensation vector
into the weight-updating recursion and rewrite Eq. (
18) as follows:
with
We further define the weight estimation error vector as follows:
By combining Eqs. (
22) and (
24), we then have the following recursion
It has been reported that the sparsity terms in Eq. (
23), i.e.,
, should be ignored when deriving the bias compensation term
; otherwise the derived vector
will compensate for the bias caused by both the input noise and this term [
24]. Hence, the recursion for weight updating can be formulated as follows:
Given the noisy input vector
, we then derive
based on the unbiased criterion as follows:
In addition, two commonly used assumptions have been used in the following analysis [
28]:
Assumption A1. The input noise and background noise are zero-mean AWGN noises and the ratio is a prior knowledge.
Assumption A2. The signals , , , , and are statistically independent.
By taking expectation on both sides of Eq. (
26) for the given
and assuming
, we have
By replacing
and
in the first term of the right-hand side (RHS) of Eq. (
28), we have
where
and
Note that as the condition
and the deviation of
being small hold, the the second term of the RHS of Eq. (
28) can be approximated as follows [
28]:
Thus, we can rewrite the nominator of Eq. (
32) as follows:
Furthermore, we can rewrite the denominator of the second term of the RHS of Eq. (
32) as follows:
where
and
Combining the results Eqs. (
29) to (
36) and substituting them into Eq. (
28), we obtain the following results:
By using the stochastic approximation [
29], we derive the bias-compensation vector as follows:
with
where
3.4. Variable Mixing Parameter Design
For the conventional SQNLMMN algorithm, it was suggested to use a fixed mixing parameter to achieve the best performance in terms of the convergence rate. However, a small mixing parameter, say , could slowly achieve a lower misadjustment in the steady state than that with a large . This inspires using a variable mixing parameter to attain fast convergence rate and small misadjustment simultaneously.
Figure 2 depicts the block diagram of the variable mixing parameter scheme design. Two adaptive filters are combined as follows:
where
and
are the fast and slow filters, respectively, i.e.,
;
is the smoothed combination factor. Referring to Eq. (
18), the adaptation recursion for each filter can be expressed as follows:
with
and
where the scaling factor
in Eq. (
42) is used to mitigate the interaction between
and
;
and
; the matrix
is a contaminated version of
(see Eq. (
10)) defined as follows:
where
and
. The bias compensation vector
associated with
can be expressed as follows:
with
where
The smoothed combination equation for
is given by
where
C is the length used to smooth the combination factor
, which can be calculated as follows [
30]:
with
where
is the step size for adjusting the recursion of
;
and
denote the sigmoid and sign function, respectively. Note that we confine
, and we check if the condition holds every
iterations. We force
when
and set
when
[
31].
3.5. Robustness Consideration
To obtain the impact of impulse noise
on the convergence of the adaptive filter
in the proposed BC-SNQNLM-VMN algorithm, we propose applying the modified Huber function
(·) on
as follows [
32]:
where
is a threshold as follows:
with
.
where
denotes the median operation;
is an observation vector for
with length
defined as follows:
Note that we choose , , and in the computer simulations.
Furthermore, the estimation of
is required to calculate the bias compensation vector
. For robustness concern, it has been reported that
can be estimated as follows [
19,
21]:
where
, which is the ratio of
and
, is assumed to be available as in [
33].
4. Simulation Results
4.1. Setup
Computer simulations evaluated the effectiveness of the proposed algorithm. The unknown sparse system
comprises 32 taps (
M=32), which has
nonzero taps, i.e., the sparsity is
. We randomly choose the positions of the nonzero taps among
M taps, and their values follow a standard Gaussian distribution. A standard AWGN models the input signal. The signal-to-noise ratio (SNR) for the input signal
dB (see Eq. (
5)) and the SNR of the observed signal
dB (see Eq. (
2)). For the BG impulse noise model, we choose the SNR of the additive impulse noise
dB (see Eq. (
4)). Moreover, we designate the occurrence probability of BG impulse as
for weak BG noise and
for strong BG noise. For the alpha-stable impulse noise model, we define
for weak alpha stable noise and
for strong alpha stable noise [
11]. Other main parameters are setting as follows:
,
,
,
,
,
,
,
, and
.
The performance metric is the normalized mean-square deviation (NMSD), which can be calculated as follows:
We plotted the NMSD learning curves and the evolution of mixing parameters in the simulation results by averaging over 100 independent Monte Carlo trials. The comparative works were BC-NLMS [
19], BC-NLMF [
21], BC-LMMN [
22], BC-CIM-LGMN [
23], and BC-CIM-PNLMS [
20] algorithms.
4.2. Results
4.2.1. Baseline: No Impulse Noise
We compare the NMSD learning curves concerning various bias-compensated adaptive filtering algorithms in the absence of impulse noise. As shown in
Figure 3, our proposed method can achieve the lowest NMSD during the steady state, outperforming the comparative works by
dB. Note that
(see. Eq. (
49)) implies no smoothing was applied. The results confirmed that combining two filters with a smoothed factor (see Eq. (
41)) exhibits a more smooth convergence behavior in the transient stage (
) during the simulation. Note that: 1) the step sizes are chosen so that the convergence rate is at the same level for all algorithms to have a fair comparison. 2) the NMSD loss is about 12 dB without bias compensation.
4.2.2. Evaluation of Variable Mixing Parameter Method
As shown in
Figure 4, we evaluate the effectiveness of the proposed variable mixing parameter scheme. In
Figure 4a,b, the additive impulse noise corresponds to the weak and strong BG impulse noise, respectively. The results exhibit that our proposed combining mixing parameter
scheme worked well in both types of additive impulse noise. Referring to
Figure 4(a), we can observe the variation of the mixing parameter in the weak BG impulse noise case as follows:
First, keeps at 1 in the early stage (), which implies the proposed method behaves like .
Then, decreases gradually during the transient stage (), which implies the proposed method behaves changing from to .
Finally, keeps around 0 in the steady-state stage ().
Similar results were observed in the strong BG impulse noise case (see
Figure 4(b)).
In
Figure 5, we evaluated the impact of the additive alpha-stable impulse noise in both weak (
Figure 5(a)) and strong (
Figure 5(b)) cases. In this scenario, we observe that the strong alpha stable impulse noise makes the NMSD learning curves exhibit more fluctuations in the steady state. In addition, we observe that the evolution of
shows more fluctuations at the beginning of the transient stage and quickly reaches its steady state. This implies the proposed method behaves by changing from
to
earlier than in the case of weak impulse noise. Compared with the baseline, the smoothed combination method exhibits fewer fluctuations, especially in the strong impulse noise cases. Thus, we choose
without an explicit statement in the following simulation. The results have shown the impact of the noisy input on the resulting NMSD. Without bias compensation, the NMSD loss in the steady status is about 12 dB, the same as the baseline. Therefore, we can confirm the robustness of our proposed method.
4.2.3. Performance Comparisons in the Presence of Impulse Noise
As shown in
Figure 6, we compare our proposed method with other comparative works for the BG impulse noise case. In the weak BG impulse noise case (see
Figure 6(a)), our proposed method achieves the lowest NMSD and improves by 3 dB to 15 dB compared to comparative works. Note that the BC-CIM-PNLMS did not consider the impact of impulse noise, which resulted in the worst NMSD performance. However, BC-CIM-LGMN, BC-LMMN, and BC-NLMF algorithms diverged in the strong BG impulse noise case (see
Figure 6(b)). In this case, only the BC-NLMS and our proposed method still function well, and our method improves the NMSD by 3 dB in the steady state.
As shown in
Figure 7, we compare our proposed method with other comparative works for the alpha stable impulse noise case. In the weak alpha stable impulse noise case (see
Figure 7(a)), our proposed method achieves the lowest NMSD and improves by 4.5 dB to 7 dB compared to comparative works. Note that the BC-CIM-PNLMS did not consider the impact of impulse noise, exhibiting some NMSD learning curve spikes. In addition, the comparative works exhibit poor performance in the strong alpha stable impulse noise case (see
Figure 7(b)). In this case, the BC-CIM-PNLMS exhibits stronger spikes than that in the weak alpha stable impulse noise case. However, our proposed method shows the lowest NMSD loss (about 0.9 dB) compared to other comparative works in the steady state.
5. Conclusions
The noisy input signals result in a significant NMSD loss even in the absence of an impulse noise scenario. In this paper, we have presented a robust bias compensation method for the SNQNLM algorithm. Furthermore, we have proposed a variable-mixing-norm method to attain a high convergence rate and low misadjustment during adaptation. Simulation results have confirmed that our proposed BC-SNQNLM-VMN algorithm outperforms the comparative works by 3 to 15 dB for BG impulse noise and 4.5 to 7 dB for alpha stable impulse noise in terms of NMSD, respectively.
Author Contributions
Conceptualization, Y.-R. C., H.-E. H., and G. Q.; methodology, Y.-R. C. and H.-E. H.; software, Y.-R. C. and H.-E. H.; validation, Y.-R. C. and G. Q.; formal analysis, Y.-R. C. and H.-E. H.; investigation, H.-E. H.; resources, Y.-R. C. and H.-E. H.; data curation, Y.-R. C. and H.-E. H.; supervision, Y.-R.C.; writing—original draft preparation, Y.-R. C. and H.-E. H.; writing—review and editing, Y.-R. C., H.-E. H., and G. Q.; project administration, Y.-R.C.; funding acquisition, Y.-R.C.
Funding
This work was supported in part by the Ministry of Science and Technology of Taiwan under Grant 112-2221-E-197-022.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Diniz, P.S.R. Adaptive Filtering Algorithms and Practical Implementation, 5th ed.; Springer Cham: 1 W 42nd St Fl 15 New York, NY, 2020. [Google Scholar]
- Tan, G.; Yan, S.; Yang, B. Impulsive Noise Suppression and Concatenated Code for OFDM Underwater Acoustic Communications. 2022 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC);, 2022; pp. 1–6. [CrossRef]
- Wang, J.; Li, J.; Yan, S.; Shi, W.; Yang, X.; Guo, Y.; Gulliver, T.A. A Novel Underwater Acoustic Signal Denoising Algorithm for Gaussian/Non-Gaussian Impulsive Noise. IEEE Trans. Veh. Technol. 2021, 70, 429–445. [Google Scholar] [CrossRef]
- Diamant, R. Robust Interference Cancellation of Chirp and CW Signals for Underwater Acoustics Applications. IEEE Access 2018, 6, 4405–4415. [Google Scholar] [CrossRef]
- Ge, F.X.; Zhang, Y.; Li, Z.; Zhang, R. Adaptive Bubble Pulse Cancellation From Underwater Explosive Charge Acoustic Signals. IEEE J. Oceanic Eng. 2011, 36, 447–453. [Google Scholar] [CrossRef]
- Fieve, S.; Portala, P.; Bertel, L. A new VLF/LF atmospheric noise model. Radio Science 2007, 42, 1–14. [Google Scholar] [CrossRef]
- Rehman, I.u.; Raza, H.; Razzaq, N.; Frnda, J.; Zaidi, T.; Abbasi, W.; Anwar, M.S. A Computationally Efficient Distributed Framework for a State Space Adaptive Filter for the Removal of PLI from Cardiac Signals. Mathematics 2023, 11. [Google Scholar] [CrossRef]
- Peng, L.; Zang, G.; Gao, Y.; Sha, N.; Xi, C. LMS Adaptive Interference Cancellation in Physical Layer Security Communication System Based on Artificial Interference. 2018 2nd IEEE Advanced Information Management,Communicates,Electronic and Automation Control Conference (IMCEC);, 2018; pp. 1178–1182. [CrossRef]
- Chen, Y.E.; Chien, Y.R.; Tsao, H.W. Chirp-Like Jamming Mitigation for GPS Receivers Using Wavelet-Packet-Transform-Assisted Adaptive Filters. 2016 International Computer Symposium (ICS);, 2016; pp. 458–461. [CrossRef]
- Yu, H.C.; Chien, Y.R.; Tsao, H.W. A study of impulsive noise immunity for wavelet-OFDM-based power line communications. 2016 International Conference On Communication Problem-Solving (ICCP);, 2016; pp. 1–2. [CrossRef]
- Chien, Y.R.; Wu, S.T.; Tsao, H.W.; Diniz, P.S.R. Correntropy-Based Data Selective Adaptive Filtering. IEEE Trans. Circuits Syst. I Regular Papers 2024, 71, 754–766. [Google Scholar] [CrossRef]
- Chien, Y.R.; Xu, S.S.D.; Ho, D.Y. Combined boosted variable step-size affine projection sign algorithm for environments with impulsive noise. Digital Signal Process. 2023, 140, 104110. [Google Scholar] [CrossRef]
- Maleki, N.; Azghani, M. Sparse Mixed Norm Adaptive Filtering Technique. Circuits Syst Signal Process 2020, 39, 5758–5775. [Google Scholar] [CrossRef]
- Shao, M.; Nikias, C. Signal processing with fractional lower order moments: stable processes and their applications. Proc. IEEE 1993, 81, 986–1010. [Google Scholar] [CrossRef]
- Walach, E.; Widrow, B. The least mean fourth (LMF) adaptive algorithm and its family. IEEE Trans. Inform. Theory 1984, 30, 275–283. [Google Scholar] [CrossRef]
- Balasundar, C.; Sundarabalan, C.K.; Santhanam, S.N.; Sharma, J.; Guerrero, J.M. Mixed Step Size Normalized Least Mean Fourth Algorithm of DSTATCOM Integrated Electric Vehicle Charging Station. IEEE Trans. Ind. Inform. 2023, 19, 7583–7591. [Google Scholar] [CrossRef]
- Papoulis, E.; Stathaki, T. A normalized robust mixed-norm adaptive algorithm for system identification. IEEE Signal Process. Lett. 2004, 11, 56–59. [Google Scholar] [CrossRef]
- Zerguine, A. A variable-parameter normalized mixed-norm (VPNMN) adaptive algorithm. Eurasip J. Adv. Signal Process. 2012, 2012, 55. [Google Scholar] [CrossRef]
- Jung, S.M.; Park, P. Normalised least-mean-square algorithm for adaptive filtering of impulsive measurement noises and noisy inputs. Electron. Lett. 2013, 49, 1270–1272. [Google Scholar] [CrossRef]
- Jin, Z.; Guo, L.; Li, Y. The Bias-Compensated Proportionate NLMS Algorithm With Sparse Penalty Constraint. IEEE Access 2020, 8, 4954–4962. [Google Scholar] [CrossRef]
- Lee, M.; Park, T.; Park, P. Bias-Compensated Normalized Least Mean Fourth Algorithm for Adaptive Filtering of Impulsive Measurement Noises and Noisy Inputs. 2019 12th Asian Control Conference (ASCC);, 2019; pp. 220–223.
- Lee, M.; Park, I.S.; Park, C.e.; Lee, H.; Park, P. Bias Compensated Least Mean Mixed-norm Adaptive Filtering Algorithm Robust to Impulsive Noises. 2020 20th International Conference on Control, Automation and Systems (ICCAS), 2020, pp. 652–657. [CrossRef]
- Ma, W.; Zheng, D.; Zhang, Z. Bias-compensated generalized mixed norm algorithm via CIM constraints for sparse system identification with noisy input. 2017 36th Chinese Control Conference (CCC);, 2017; pp. 5094–5099. [CrossRef]
- Yoo, J.; Shin, J.; Park, P. An Improved NLMS Algorithm in Sparse Systems Against Noisy Input Signals. IEEE Trans. Circuits Syst II, Exp. Briefs 2015, 62, 271–275. [Google Scholar] [CrossRef]
- Kim, S.R.; Efron, A. Adaptive robust impulse noise filtering. IEEE Transa. Signal Process. 1995, 43, 1855–1866. [Google Scholar] [CrossRef]
- Gu, Y.; Jin, J.; Mei, S. l0 Norm Constraint LMS Algorithm for Sparse System Identification. IEEE Signal Processing Letters 2009, 16, 774–777. [Google Scholar] [CrossRef]
- Eweda, E. Global Stabilization of the Least Mean Fourth Algorithm. IEEE Trans. Signal Process. 2012, 60, 1473–1477. [Google Scholar] [CrossRef]
- Zheng, Z.; Liu, Z.; Zhao, H. Bias-Compensated Normalized Least-Mean Fourth Algorithm for Noisy Input. Circuits Syst. Signal Process. 2017, 36, 3864–3873. [Google Scholar] [CrossRef]
- Zhao, H.; Zheng, Z. Bias-compensated affine-projection-like algorithms with noisy input 2016. 52, 712–714. [CrossRef]
- Arenas-Garcia, J.; Figueiras-Vidal, A.; Sayed, A. Mean-square performance of a convex combination of two adaptive filters. IEEE Trans. Signal Process. 2006, 54, 1078–1090. [Google Scholar] [CrossRef]
- Lu, L.; Zhao, H.; Li, K.; Chen, B. A Novel Normalized Sign Algorithm for System Identification Under Impulsive Noise Interference. Circuits Syst. Signal Process. 2015, 35, 3244–3265. [Google Scholar] [CrossRef]
- Zhou, Y.; Chan, S.C.; Ho, K.L. New Sequential Partial-Update Least Mean M-Estimate Algorithms for Robust Adaptive System Identification in Impulsive Noise. IEEE Trans. Ind. Electron. 2011, 58, 4455–4470. [Google Scholar] [CrossRef]
- Jung, S.M.; Park, P. Stabilization of a Bias-Compensated Normalized Least-Mean-Square Algorithm for Noisy Inputs. IEEE Trans. Signal Process. 2017, 65, 2949–2961. [Google Scholar] [CrossRef]
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).