Computational Mathematics and Scientific Computing Seminar

A geometrical connection between sparse and low-rank matrices (and its uses for machine learning)

Time and Location:

Sept. 29, 2023 at 10AM; Warren Weaver Hall, Room 1302

Speaker:

Lawrence Saul, Flatiron Institute

Abstract:

Many problems in high dimensional data analysis can be formulated as a search for structure in large matrices. One important type of structure is sparsity; for example, when a matrix is sparse, with a large number of zero elements, it can be stored in a highly compressed format. Another type of structure is linear dependence; when a matrix is low-rank, it can be expressed as the product of two smaller matrices. It is well known that neither one of these structures implies the other. But can one find more subtle connections by looking beyond the canonical decompositions of linear algebra?

In this talk, I will consider when a sparse nonnegative matrix can be recovered from a real-valued matrix of significantly lower rank. I will describe an algorithm for this problem based on a nonlinear low-rank decomposition of sparse matrices. Then I will discuss various settings in manifold learning and deep learning where this problem arises in a natural way. Arguably the most popular matrix decompositions in machine learning are those--such as principal component analysis, or nonnegative matrix factorization--that have a simple geometric interpretation. I will show that these nonlinear low-rank decompositions also have a simple geometric interpretation, but one that is entirely different in character.