An Acceleration Scheme for Mini-batch, Streaming PCA

Salaheddin Alakkari (Trinity College Dublin), John Dingliana (Trinity College Dublin)

Abstract
In this paper, we propose an acceleration scheme for mini-batch streaming PCA methods that are based on the Stochastic Gradient Approximation. Our scheme converges to the first $k>1$ eigenvectors in a single data pass even when using a very small batch size. We provide empirical convergence results of our scheme based on the spiked covariance model. Our scheme does not require any prior knowledge of the data distribution and hence is well suited for streaming data scenarios. Furthermore, based on empirical evaluations using the spiked covariance model and large-scale benchmark datasets, we find that our acceleration scheme outperforms related state-of-the-art online PCA approaches including SGA, Incremental PCA and Candid Covariance-free Incremental PCA.

DOI
10.5244/C.33.168
https://dx.doi.org/10.5244/C.33.168

Files
Paper (PDF)

BibTeX
@inproceedings{BMVC2019,
title={An Acceleration Scheme for Mini-batch, Streaming PCA},
author={Salaheddin Alakkari and John Dingliana},
year={2019},
month={September},
pages={168.1--168.12},
articleno={168},
numpages={12},
booktitle={Proceedings of the British Machine Vision Conference (BMVC)},
publisher={BMVA Press},
editor={Kirill Sidorov and Yulia Hicks},
doi={10.5244/C.33.168},
url={https://dx.doi.org/10.5244/C.33.168}
}