Printable PDF
Department of Mathematics,
University of California San Diego

****************************

Math 278C - Optimization and Data Science

Hanbaek Lyu

University of Wisconsin – Madison

Convergence and Complexity of Stochastic Block Majorization-Minimization

Abstract:

Stochastic majorization-minimization (SMM) is an online extension of the classical principle of majorization-minimization, which consists of sampling i.i.d. data points from a fixed data distribution and minimizing a recursively defined majorizing surrogate of an objective function. In this paper, we introduce stochastic block majorization-minimization, where the surrogates can now be only block multi-convex and a single block is optimized at a time within a diminishing radius. Relaxing the standard strong convexity requirements for surrogates in SMM, our framework gives wider applicability including online CANDECOMP/PARAFAC (CP) dictionary learning and yields greater computational efficiency especially when the problem dimension is large. We provide an extensive convergence analysis on the proposed algorithm, which we derive under possibly dependent data streams, relaxing the standard i.i.d. assumption on data samples. We show that the proposed algorithm converges almost surely to the set of stationary points of a nonconvex objective under constraints at a rate $O(  \frac{  (\log n)^{1+\epsilon}  }{  n^{1/2}  }  )$ for the empirical loss function and $O(  \frac{  (\log n)^{1+\epsilon}  }{  n^{1/4}  }  )$ for the expected loss function, where n denotes the number of data samples processed. Under some additional assumption, the latter convergence rate can be improved to $O(  \frac{  (\log n)^{1+\epsilon}  }{  n^{1/2}  }  )$. Our results provide first convergence rate bounds for various online matrix and tensor decomposition algorithms under a general Markovian data setting.

Host: Jiawang Nie

March 2, 2022

3:00 PM

https://ucsd.zoom.us/j/94927846567

Meeting ID: 949 2784 6567
Password: 278CWN22

****************************