Special Lecture in Celebration of Marsha Berger and Margaret Wright

Generalized Tensor Decomposition: Utility for Data Analysis and Mathematical Challenges

Time and Location:

April 21, 2023 at 2PM; Warren Weaver Hall, Room 1302

Speaker:

Tamara Kolda, MathSci.ai

Abstract:

Tensor decomposition is an unsupervised learning methodology that has applications in a wide variety of domains, including chemometrics, criminology, and neuroscience. We focus on low-rank tensor decomposition using  canonical polyadic or CANDECOMP/PARAFAC format. A low-rank tensor decomposition is the minimizer according to some nonlinear program. The usual objective function is the sum of squares error (SSE) comparing the data tensor and the low-rank model tensor. This leads to a nicely-structured problem with subproblems that are linear least squares problems which can be solved efficiently in closed form. However, the SSE metric is not always ideal. Thus, we consider using other objective functions. For instance, KL divergence is an alternative metric is useful for count data and results in a nonnegative factorization. In the context of nonnegative matrix factorization, for instance, KL divergence was popularized by Lee and Seung (1999). We can also consider various objectives such as logistic odds for binary data, beta-divergence for nonnegative data, and so on. We show the benefits of alternative objective functions on real-world data sets. We consider the computational of generalized tensor decomposition based on other objective functions, summarize the work that has been done thus far, and illuminate open problems and challenges.