UWSpace is currently experiencing technical difficulties resulting from its recent migration to a new version of its software. These technical issues are not affecting the submission and browse features of the site. UWaterloo community members may continue submitting items to UWSpace. We apologize for the inconvenience, and are actively working to resolve these technical issues.
 

Geodesic Convex Analysis of Group Scaling for the Paulsen Problem and the Tensor Normal Model

Loading...
Thumbnail Image

Date

2021-11-18

Authors

Ramachandran, Akshay

Journal Title

Journal ISSN

Volume Title

Publisher

University of Waterloo

Abstract

The framework of scaling problems has recently had much interest in the theoretical computer science community due to its variety of applications, from algebraic complexity to machine learning. In this thesis, our main motivation will be two new applications: the Paulsen problem from frame theory, and the tensor normal model in statistical estimation. In order to give new results for these problems, we provide novel convergence analyses for matrix scaling and tensor scaling. Specifically, we will use the framework of geodesic convex optimization presented in Bürgisser et al. [20] and analyze two sufficient conditions (called strong convexity and pseudorandomness) for fast convergence of the natural gradient flow algorithm in this setting. This allows us to unify and improve many previous results [62], [63], and [36] for special cases of tensor scaling. In the first half of the thesis, we focus on the Paulsen problem where we are given a set of n vectors in d dimensions that ε-approximately satisfy two balance conditions, and asked whether there is a nearby set of vectors that exactly satisfy those balance conditions. This is an important question from frame theory [24] for which very little was known despite considerable attention. We are able to give optimal distance bounds for the Paulsen problem in both the worst-case and the average-case by improving the smoothed analysis approach of Kwok et al. [62]. Specifically, we analyze certain strong convergence conditions for frame scaling, and then show that a random perturbation of the input frame satisfies these conditions and can be scaled to a nearby solution. In the second half of the thesis, we study the matrix and tensor normal models, which are a family of Gaussian distributions on tensor data where the covariance matrix respects this tensor product structure. We are able to generalize our scaling results to higher- order tensors and give error bounds for the maximum likelihood estimator (MLE) of the tensor normal model with a number of samples only a single dimension factor above the existence threshold. This result relies on some spectral properties of random Gaussian tensors shown by Pisier [80]. We also give the first rigorous analysis of the Flip-Flop algorithm, showing that it converges exponentially to the MLE with high probability. This explains the empirical success of this well-studied heuristic for computing the MLE.

Description

Keywords

geodesic convexity, group scaling, tensor normal model, Paulsen problem, covariance estimation, frame theory

LC Keywords

Citation