Abstract
In equalization and deconvolution tasks, the correlated nature of the input signal slows the convergence speeds of stochastic gradient adaptive filters. Prewhitening techniques have been proposed to improve the convergence performance, but the additional coefficient memory and updates for the prewhitening filter can be prohibitive in some applications. We present two simple algorithms that employ the equalizer as a prewhitening filter within the gradient updates. These self-whitening algorithms provide quasi-Newton convergence locally about the optimum coefficient solution for deconvolution and equalization tasks. Multichannel extensions of the techniques are also described.
Keywords
Affiliated Institutions
Related Publications
HOGWILD!: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent
Stochastic Gradient Descent (SGD) is a popular algorithm that can achieve state-of-the-art performance on a variety of machine learning tasks. Several researchers have recently ...
Publication Info
- Year
- 1999
- Type
- article
- Volume
- 47
- Issue
- 4
- Pages
- 1161-1165
- Citations
- 35
- Access
- Closed
External Links
Social Impact
Social media, news, blog, policy document mentions
Citation Metrics
Cite This
Identifiers
- DOI
- 10.1109/78.752617