Complex Factor Analysis and Extensions

Journal Article (2018)
Author(s)

Ahmad Mouri Sardarabadi (TU Delft - Signal Processing Systems)

AJ van der Veen (TU Delft - Signal Processing Systems)

Research Group
Signal Processing Systems
Copyright
© 2018 A. Mouri Sardarabadi, A.J. van der Veen
DOI related publication
https://doi.org/10.1109/TSP.2017.2780047
More Info
expand_more
Publication Year
2018
Language
English
Copyright
© 2018 A. Mouri Sardarabadi, A.J. van der Veen
Research Group
Signal Processing Systems
Issue number
4
Volume number
66
Pages (from-to)
954-967
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Many subspace-based array signal processing algorithms assume that the noise is spatially white. In this case, the noise covariance matrix is a multiple of the identity and the eigenvectors of the data covariance matrix are not affected by it. If the noise covariance is an unknown arbitrary diagonal (e.g., for an uncalibrated array), the eigenvalue decomposition leads to incorrect subspace estimates and it has to be replaced by a more general “factor analysis” decomposition (FA), which then reveals all relevant information. We consider this data model and several extensions where the noise covariance matrix has a more general structure, such as banded, sparse, block diagonal, and cases, where we have multiple data covariance matrices that share the same noise covariance matrix. Starting from a nonlinear weighted least squares formulation, we propose new estimation algorithms for both classical FA and its extensions. The optimization is based on Gauss–Newton gradient descent. Generally, this leads to an iteration involving the inversion of a very large matrix. Using the structure of the problem, we show how this can be reduced to the inversion of a matrix with dimension equal to the number of unknown noise covariance parameters. This results in new algorithms that have faster numerical convergence and lower complexity compared to several maximum-likelihood based algorithms that could be considered state of the art. The new algorithms scale well to large dimensions and can replace eigenvalue decompositions in many applications even if the noise can be assumed to be white.

Files

Complex_Factor_Analysis_and_Ex... (pdf)
(pdf | 0.86 Mb)
- Embargo expired in 03-12-2021
License info not available