Reduction of the complexity of the NLMS algorithm has received attention in the area of adaptive filtering. A processing cost reduction method, in which the component of the weight vector is updated when the absolute value of the sample is greater than or equal to the average of the absolute values of the input samples, has been proposed. The convergence analysis of the processing cost reduction method has been derived from a low-pass filter expression. However, in this analysis the effect of the weignt vector components whose adaptations are skipped is not considered in terms of the direction of the gradient estimation vector. In this paper, we use an arbitrary value instead of the average of the absolute values of the input samples as a threshold level, and we derive the convergence characteristics of the processing cost reduction method with arbitrary threshold level for zero-mean white Gaussian samples. From the analytical results, it is shown that the range of the gain constant to insure convergence and the misadjustment are independent of the threshold level. Moreover, it is shown that the convergence rate is a function of the threshold level as well as the gain constant. When the gain constant is small, the processing cost is reduced by using a large threshold level without a large degradation of the convergence rate.
The copyright of the original papers published on this site belongs to IEICE. Unauthorized use of the original or translated papers is prohibited. See IEICE Provisions on Copyright for details.
Copy
Kiyoshi TAKAHASHI, Shinsaku MORI, "Convergence Analysis of Processing Cost Reduction Method of NLMS Algorithm" in IEICE TRANSACTIONS on Fundamentals,
vol. E77-A, no. 5, pp. 825-832, May 1994, doi: .
Abstract: Reduction of the complexity of the NLMS algorithm has received attention in the area of adaptive filtering. A processing cost reduction method, in which the component of the weight vector is updated when the absolute value of the sample is greater than or equal to the average of the absolute values of the input samples, has been proposed. The convergence analysis of the processing cost reduction method has been derived from a low-pass filter expression. However, in this analysis the effect of the weignt vector components whose adaptations are skipped is not considered in terms of the direction of the gradient estimation vector. In this paper, we use an arbitrary value instead of the average of the absolute values of the input samples as a threshold level, and we derive the convergence characteristics of the processing cost reduction method with arbitrary threshold level for zero-mean white Gaussian samples. From the analytical results, it is shown that the range of the gain constant to insure convergence and the misadjustment are independent of the threshold level. Moreover, it is shown that the convergence rate is a function of the threshold level as well as the gain constant. When the gain constant is small, the processing cost is reduced by using a large threshold level without a large degradation of the convergence rate.
URL: https://global.ieice.org/en_transactions/fundamentals/10.1587/e77-a_5_825/_p
Copy
@ARTICLE{e77-a_5_825,
author={Kiyoshi TAKAHASHI, Shinsaku MORI, },
journal={IEICE TRANSACTIONS on Fundamentals},
title={Convergence Analysis of Processing Cost Reduction Method of NLMS Algorithm},
year={1994},
volume={E77-A},
number={5},
pages={825-832},
abstract={Reduction of the complexity of the NLMS algorithm has received attention in the area of adaptive filtering. A processing cost reduction method, in which the component of the weight vector is updated when the absolute value of the sample is greater than or equal to the average of the absolute values of the input samples, has been proposed. The convergence analysis of the processing cost reduction method has been derived from a low-pass filter expression. However, in this analysis the effect of the weignt vector components whose adaptations are skipped is not considered in terms of the direction of the gradient estimation vector. In this paper, we use an arbitrary value instead of the average of the absolute values of the input samples as a threshold level, and we derive the convergence characteristics of the processing cost reduction method with arbitrary threshold level for zero-mean white Gaussian samples. From the analytical results, it is shown that the range of the gain constant to insure convergence and the misadjustment are independent of the threshold level. Moreover, it is shown that the convergence rate is a function of the threshold level as well as the gain constant. When the gain constant is small, the processing cost is reduced by using a large threshold level without a large degradation of the convergence rate.},
keywords={},
doi={},
ISSN={},
month={May},}
Copy
TY - JOUR
TI - Convergence Analysis of Processing Cost Reduction Method of NLMS Algorithm
T2 - IEICE TRANSACTIONS on Fundamentals
SP - 825
EP - 832
AU - Kiyoshi TAKAHASHI
AU - Shinsaku MORI
PY - 1994
DO -
JO - IEICE TRANSACTIONS on Fundamentals
SN -
VL - E77-A
IS - 5
JA - IEICE TRANSACTIONS on Fundamentals
Y1 - May 1994
AB - Reduction of the complexity of the NLMS algorithm has received attention in the area of adaptive filtering. A processing cost reduction method, in which the component of the weight vector is updated when the absolute value of the sample is greater than or equal to the average of the absolute values of the input samples, has been proposed. The convergence analysis of the processing cost reduction method has been derived from a low-pass filter expression. However, in this analysis the effect of the weignt vector components whose adaptations are skipped is not considered in terms of the direction of the gradient estimation vector. In this paper, we use an arbitrary value instead of the average of the absolute values of the input samples as a threshold level, and we derive the convergence characteristics of the processing cost reduction method with arbitrary threshold level for zero-mean white Gaussian samples. From the analytical results, it is shown that the range of the gain constant to insure convergence and the misadjustment are independent of the threshold level. Moreover, it is shown that the convergence rate is a function of the threshold level as well as the gain constant. When the gain constant is small, the processing cost is reduced by using a large threshold level without a large degradation of the convergence rate.
ER -