Deep learning based matrix completion

Research output: Journal Publications and Reviews (RGC: 21, 22, 62)21_Publication in refereed journal

15 Scopus Citations
View graph of relations

Related Research Unit(s)


Original languageEnglish
Pages (from-to)540-549
Journal / PublicationNeurocomputing
Online published2 Jun 2017
Publication statusPublished - 29 Nov 2017


Previous matrix completion methods are generally based on linear and shallow models where the given incomplete matrices are of low-rank and the data are assumed to be generated by linear latent variable models. In this paper, we first propose a novel method called AutoEncoder based matrix completion (AEMC). The main idea of AEMC is to utilize the partially observed data to learn and construct a nonlinear latent variable model in the form of AutoEncoder. The hidden layer of the AutoEncoder has much fewer units than the visible layers do. Meanwhile, the unknown entries of the data are recovered to fit the nonlinear latent variable model. Based on AEMC, we further propose a deep learning based matrix completion (DLMC) method. In DLMC, AEMC is used as a pre-training step for both the missing entries and network parameters; the hidden layer of AEMC is then used to learn stacked AutoEncoders (SAEs) with greedy layer-wise training; finally, fine-tuning is carried out on the deep network formed by AEMC and SAEs to obtain the missing entries of the data and the parameters of the network. In addition, we also provide out-of-sample extensions for AEMC and DLMC to recover online incomplete data. AEMC and DLMC are compared with state-of-the-art methods in the tasks of synthetic matrix completion, image inpainting, and collaborative filtering. The experimental results verify the effectiveness and superiority of the proposed methods.

Research Area(s)

  • Matrix completion, AutoEncoder, deep learning, out-of-sample extension, mage inpainting, collaborative filtering