Greedy Capon Beamformer

We provide an efficient function to compute the Greedy Capon Beamformer (GCB) which is a (sparse) greedy version of the standard Capon Beamformer (SCB). GCB is fast to compute, often outperforms SCB with a clear margin, and unlike the SCB it is not affected by coherence of the signals. In the Github repo, we also include a demo example that can be used to reproduces Figures in the paper.
For more information, see: Greedy Capon Beamformer, Esa Ollila arXiv:2404.15329 [eess.SP], Apr. 7, 2024.
[Github]


Covariance Learning Orthogonal Matching Pursuit (CL-OMP)

Code (both MATLAB and python functions) for the Covariance Learning Orthogonal Matching Pursuit (CL-OMP) method. Also a demo simulation that reproduces Figure 1 in the paper is provided. Comparison in the simulation study is also made to SOMP, and a function to compute it is also provided.
For more information, see: Matching Pursuit Covariance Learning, Esa Ollila Proc. 32nd European Signal Processing Conference (EUSIPCO'24), Lyon, August 26-30, 2024, pp. 2447-2451.
[Github] [PDF] [Slides]


Regularized tapered sample covariance matrix

Tabasco toolbox provfilides an easy to use function for computing the TABASCO (TApered or BAnded Shrinkage COvariance) estimator. Tabasco is a covariance matrix estimation method that combines tapering or banding of the sample covariance matrix with shrinkage towards scaled identity matrix. It enjoyes benefits of both approaches: tapering/banding leverages on the assumed structure of the covariance matrix while optimal shrinkage guarantees that the obtained ovariance matrix estimator is positive definite, while still preserving the structure applied using banding or tapering template. Optimal parameters, e.g., the bandwidth of banding matrix, or the shinkage intensity coefficient minimizing the MSE are adaptively estimated from the data. We also include a demo example that reproduces Figure 2 (upper left corner plot) in our paper.
For more information, see: Regularized Tapered Sample Covariance Matrix, Esa Ollila and Arnaud Breloy, IEEE Transactions on Signal Processing, vol. 70, pp. 2306-2320, 2022.
[Matlab toolbox]


Linear pooling of sample covariance matrices

Linpool toolbox provides an easy to use function for computing an estimator of the population covariance matrix in the setting where there are data from K populations available. We propose estimating each class covariance matrix as a distinct linear combination of all class sample covariance matrices. This approach is shown to reduce the estimation error when the sample sizes are limited, and the true class covariance matrices share a somewhat similar structure. We provide an effective data adaptive method for computing the optimal coefficients in the linear combination that minimize the mean squared error under the general assumption that the samples are drawn from (unspecified) elliptically symmetric distributions. An application in global minimum variance portfolio optimization using real stock data is also given.
For more information, see: Linear Pooling of Sample Covariance Matrices, Elias Raninen, David E. Tyler, and Esa Ollila,IEEE Transactions on Signal Processing, vol. 70, pp. 659-672, 2022.
[Matlab toolbox] [ Slides] [youtube video]


Coupled regularized sample covariance matrix estimator for multiple classes

CoupledRSCM toolbox for Matlab/R/python provides an easy to use function for computing regularized SCM (RSCM) estimators for multiclass problems that couple together two different target matrices for regularization: the pooled (average) SCM of the classes and the scaled identity matrix. Regularization toward the pooled SCM is beneficial when the population covariances are similar, whereas regularization toward the identity matrix guarantees that the estimators are positive definite. We derive the MSE optimal tuning parameters for the estimators as well as propose a method for their estimation, avoiding commonly used computer intensive cross-validation.
For more information, see: Coupled Regularized Sample Covariance Matrix Estimator for Multiple Classes, Elias Raninen, David E. Tyler, and Esa Ollila,IEEE Transactions on Signal Processing, vol. 69, pp. 5681-5692, 2021
[toolbox]


Bias adjusted sign covariance matrix

BASIC toolbox for Matlab provides an easy to use function for computing the BASIC (Bias Adjusted SIgn Covariance) estimator as well as BASIC Shrinkage estimator (BASICS), which is better suited for high dimensional problems, where the dimension can be larger than the sample size. It is well-known that the widely popular spatial sign covariance matrix (SSCM) does not provide consistent estimates of the eigenvalues of the shape matrix (trace normalized covariance matrix). BASIC alleviates this problem by performing an approximate bias correction to the eigenvalues of the SSCM. For more information, see: Bias adjusted sign covariance matrix, Elias Raninen and Esa Ollila,IEEE Signal Processing Letters, vol. 28, pp. 2092-2096, 2021.
[toolbox]


Shrinking the eigenvalues of M-estimators of covariance

shrinkM MATLAB toolbox is a collection of functions to compute an M-estimator of scatter (covariance) matrix whose eigenvalues are shrinked optimally towards the grand mean of the eigenvalues. Optimality means that the shrinkage parameter is chosen to minimize the mean squared error data-adaptively without computationally prohibitive cross-validation procedure. The aforementioned shrinkage (or regularized) M-estimators of covariance are described in detail in reference below. We also include a demo example of using these covariance matrix estimators on simulated data. As can be noted, the proposed robust shrinkage M-estimators (RMVT and RHub) outperform conventional non-robust estimators with a large margin when data is non-Gaussian:
For more information, see Shrinking the eigenvalues of M-estimators of covariance matrix, Esa Ollila, Daniel P. Palomar and Frederic Pascal, IEEE Transactions on Signal Processing, vol. 69, pp. 256-269, 2021.
[Matlab toolbox] [ Slides] [youtube video]


Pedestrian detection using Riemannian classification algorithms

We have developed a python toolbox for pedestrian detection from images using classification algorithms that utilize Riemannian geometry.
For more information, see A Comparative Study of Supervised Learning Algorithms for Symmetric Positive Definite Features, Ammar Mian, Elias Raninen and Esa Ollila, Proc. European Signal Processing Conference (EUSIPCO'20), Amsterdam, Netherlands, August 24-28, pp. 950-954. 2020.
[Python toolbox] [pdf]


Block-wise Minimization-Majorization algorithm for Huber's criterion

Huber's criterion can be used for robust joint estimation of regression and scale parameters in the linear model. The codes use the block-wise minimization majorization framework with novel data-adaptive step sizes proposed in our paper for both the location and scale. Sparse signal recovery using Huber's criterion can be performed using the normalized iterative hard thresholding approach which is also implemented in the toolbox. We illustrate the usefulness of the algorithms, hubreg and hubniht in an image denoising application and simulation studies. For image denoising, we provide an easy-to-use function hubniht_denoising. An example of HUBNIHT denoising performance of Lena image in salt-and-peper noise is shown below:
and the Matlab script that produced the results is shown here.
For more information about the method, see Block-wise minimization-majorization algorithm for Huber's criterion: sparse learning and applications by Esa Ollila and Ammar Mian, in Proc. IEEE International Workshop on Machine Learning for Signal Processing (MLSP'20), Sep 21-24, 2020, Espoo, Finland, pp. 1-6.
[Toolbox] [Slides]


Compressive Regularized Discriminant Analysis

CompressiveRDA MATLAB toolbox contains function to compute the CRDA classifiers introduced in the reference below. The proposed classifiers can be applied in high-dimensional sparse data scenarios, where the sample size is smaller than the dimension of the data, which is potentially very large. Our method is especially designed for feature selection purposes and can be used as a gene selection method in genomic studies. The toolbox also contains demo examples of classification of real-world data sets used in our paper. One reason for success of CRDA can be its ability to find patterns in data as shown in this gene heat map: For more information, see A Compressive Classification Framework for High-Dimensional Data, Muhammad N. Tabassum and Esa Ollila, IEEE Open Journal of Signal Processing, Vol 1., pp. 177-186, 2020.
[Matlab toolbox]


Regularized Sample Covariance Matrix (RSCM) Estimator

RegularizedSCM MATLAB toolbox contains function to compute the Ell-RSCM estimators of the covariance matrix introduced in the reference below. The proposed estimators can be applied in high-dimensional sparse data scenarios, where the sample size is smaller than the dimension of the data, which is potentially very large. The toolbox also contains demo examples arising from finance and classification. In the former, the estimators are applied in portfolio optimization problem of real data, where we use historical (daily) stock returns from the Hong Kong's Hang Seng Index (HSI) and Standard and Poor's 500 (S&P 500) index. In the latter, the estimators are used in classification problem of real data sets using the framework of regularized discriminant analysis. In both cases, the proposed estimators are able to yield better performance than the state-of-the-art methods. See example below:
For more information, see Optimal shrinkage covariance matrix estimation under random sampling from elliptical distributions, Esa Ollila and Elias Raninen, IEEE Transactions on Signal Processing, vol. 67, no. 10, pp. 2707 - 2719, 2019.
[Matlab toolbox] [ Slides] [youtube video]


Novel FastICA power iteration algorithm

This MATLAB toolbox contains function to compute the fastICA power iteration estimator, called PowerICA, that was proposed in the reference below. PowerICA method is efficient alternative to conventional FastICA estimator in the cases when the sample size is not orders of magnitude larger than the data dimensionality. This is the finite-sample regime in which the (fixed-point) FastICA algorithm is often reported to have convergence problems. The estimator is thus useful in many real data problems where FastICA is commonly used such as blind-source separation of fMRI data. This is illustrated in the simulation example below:
For more information, see Alternative derivation of FastICA with novel power iteration algorithm, Shahab Basiri, Esa Ollila and Visa Koivunen, IEEE Signal Processing Letters, vol. 24, no 9, 2017.
[Matlab toolbox]


Reliable and fast bootstrap-based inference in the ICA model

This software contains function to compute the low complexity and stable bootstrap procedure for FastICA estimators, referred to as enhanced fast and robust bootstrap (EFRB) method for FastICA, which was introduced in the reference below. Our bootstrapping techniques allow for performing cost efficient and reliable bootstrap-based statistical inference in the ICA model. Performing statistical inference is needed to quantitatively assess the quality of the estimators and testing hypotheses on mixing coefficients in the ICA model. EFRB is shown to provice accurate description of the distribution of FastICA mixing matrix estimator:
For more information, see Enhanced bootstrap method for statistical inference in the ICA model
Shahab Basiri, Esa Ollila and Visa Koivunen, Signal Processing, vol. 138, pp. 53-62, 2017
[Matlab code]


Sparse signal recovery in the multiple measurements vector model

Download the MATLAB function that computes the HUB-SNIHT estimator introduced in the reference below. The estimator solves the multichannel sparse recovery problem of complex-valued measurements where the objective is to find good recovery of jointly sparse unknown signal vectors from the given multiple measurement vectors which are different linear combinations of the same known elementary basis vectors or dictionary.

For more information, see Multichannel sparse recovery of complex-valued signals using Huber's criterion Esa Ollila, in Proc. Int'l Workshop on Compressed Sensing Theory and its Applications to Radar, Sonar and Remote Sensing (CoSeRa'15), Pisa, Italy, June 16-19, 2015, pp. 32-36.
[Matlab code]


Tests of circularity

Software to compute the tests of circularity introduced in Adjusting the generalized likelihood ratio test of circularity robust to non-normality, Esa Ollila and Visa Koivunen, in Proc. 10th IEEE Int.'l Workshop on Signal Processing Advances in Wireless Communications (SPAWC'09), Perugia, Italy, June 21-24, 2009, pp. 558 - 562.
[Matlab code]
Also includes code for GLRT of circularity introduced in Generalized complex elliptical distributions


Tools for (non-circular) complex elliptically symmetric distributions

This software package contains a collection of tools for random number generation from complex elliptically symmetric (CES) distributions (including complex normal (Gaussian), complex t-, complex generalized Gaussian, (isotropic) symmetric alpha-stable distribution) and Wald's type circularity test.
For more information, see Complex elliptically symmetric random variables - generation, characterization and circularity tests, Esa Ollila, Jan Eriksson and Visa Koivunen, IEEE Transactions on signal processing, vol. 59, no.1, pp. 58-69, 2011.
[Matlab code]


Robust regression based on Oja ranks

Software to compute the regression coefficent estimates based on Oja ranks introduced in Estimates of regression coefficients based on lift rank covariance matrix, Esa Ollila, Hannu Oja and Visa Koivunen, Journal of the American Statistical Association, vol. 98, no. 461, pp. 90-98, 2003.
[C code and R code]