Matrix completion is the task of filling in the missing entries of a partially observed matrix. A wide range of datasets are naturally organized in matrix form. One example is the movie-ratings matrix, as appears in the Netflix problem: Given a ratings matrix in which each entry {\displaystyle (i,j)}(i,j) represents the rating of movie {\displaystyle j}j by customer {\displaystyle i}i, if customer {\displaystyle i}i has watched movie {\displaystyle j}j and is otherwise missing, we would like to predict the remaining entries in order to make good recommendations to customers on what to watch next. Another example is the term-document matrix: The frequencies of words used in a collection of documents can be represented as a matrix, where each entry corresponds to the number of times the associated term appears in the indicated document.

Without any restrictions on the number of degrees of freedom in the completed matrix this problem is underdetermined since the hidden entries could be assigned arbitrary values. Thus we require some assumption on the matrix to create a well-posed problem, such as assuming it has maximal determinant, is positive definite, or is low-rank.[1][2]

For example, one may assume the matrix has low-rank structure, and then seek to find the lowest rank matrix or, if the rank of the completed matrix is known, a matrix of rank {\displaystyle r}r that matches the known entries. The illustration shows that a partially revealed rank-1 matrix (on the left) can be completed with zero-error (on the right) since all the rows with missing entries should be the same as the third row. In the case of the Netflix problem the ratings matrix is expected to be low-rank since user preferences can often be described by a few factors, such as the movie genre and time of release. Other applications include computer vision, where missing pixels in images need to be reconstructed, detecting the global positioning of sensors in a network from partial distance information, and multiclass learning. The matrix completion problem is in general NP-hard, but under additional assumptions there are efficient algorithms that achieve exact reconstruction with high probability.

In statistical learning point of view, the matrix completion problem is an application of matrix regularization which is a generalization of vector regularization. For example, in the low-rank matrix completion problem one may apply the regularization penalty taking the form of a nuclear norm {\displaystyle R(X)=\lambda \|X\|_{*}}{\displaystyle R(X)=\lambda \|X\|_{*}}

1. Johnson, Charles R. (1990). "Matrix completion problems: a survey". Matrix Theory and Applications. Proceedings of Symposia in Applied Mathematics. 40: 171–198. doi:10.1090/psapm/040/1059486. ISBN 9780821801543.

2. Laurent, Monique (2008). "Matrix Completion Problems". Encyclopedia of Optimization. 3: 221–229. doi:10.1007/978-0-387-74759-0_355. ISBN 978-0-387-74758-3.

Link: https://en.wikipedia.org/wiki/Matrix_completion