Affiliation:
1. Carnegie Mellon University, PA, USA
2. University of Minnesota, Minneapolis, MN
Abstract
How can we efficiently decompose a tensor into sparse factors, when the data do not fit in memory? Tensor decompositions have gained a steadily increasing popularity in data-mining applications; however, the current state-of-art decomposition algorithms operate on main memory and do not scale to truly large datasets. In this work, we propose P
ar
C
ube
, a new and highly parallelizable method for speeding up tensor decompositions that is well suited to produce sparse approximations. Experiments with even moderately large data indicate over 90% sparser outputs and 14 times faster execution, with approximation error close to the current state of the art irrespective of computation and memory requirements. We provide theoretical guarantees for the algorithm’s correctness and we experimentally validate our claims through extensive experiments, including four different real world datasets (E
nron
, L
bnl
, F
acebook
and N
ell
), demonstrating its effectiveness for data-mining practitioners. In particular, we are the first to analyze the very large N
ell
dataset using a sparse tensor decomposition, demonstrating that P
ar
C
ube
enables us to handle effectively and efficiently very large datasets. Finally, we make our highly scalable parallel implementation publicly available, enabling reproducibility of our work.
Publisher
Association for Computing Machinery (ACM)
Cited by
23 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献