Abstract
Factor analysis is a well known statistical method to describe the variability among
observed variables in terms of a smaller number of unobserved latent variables called
factors. While dealing with multivariate time series, the temporal correlation structure
of data may be modeled by including correlations in latent factors, but a crucial choice
is the covariance function to be implemented. We show that analyzing multivariate time
series in terms of latent Gaussian processes, which are mutually independent but with
each of them being characterized by exponentially decaying temporal correlations, leads
to an efficient implementation of the expectation–maximization algorithm for the maximum
likelihood estimation of parameters, due to the properties of block-tridiagonal matrices.
The proposed approach solves an ambiguity known as the identifiability problem, which
renders the solution of factor analysis determined only up to an orthogonal transformation.
Samples with just two temporal points are sufficient for the parameter estimation: hence
the proposed approach may be applied even in the absence of prior information about the
correlation structure of latent variables by fitting the model to pairs of points with varying
time delay. Our modeling allows one to make predictions of the future values of time series
and we illustrate our method by applying it to an analysis of published gene expression data
from cell culture HeLa.
| Lingua originale | Inglese |
|---|---|
| Rivista | PHYSICA. A |
| DOI | |
| Stato di pubblicazione | Pubblicato - 2012 |
Keywords
- Inverse problems
- Latent variables
- factor models