Printable PDF
Department of Mathematics,
University of California San Diego

****************************

Math 278A - Center for Computational Mathematics Seminar

Jeb Runnoe

UCSD

Second-Derivative SQP Methods for Large-Scale Nonconvex Optimization

Abstract:

The class of sequential quadratic programming (SQP) methods solve a nonlinearly constrained optimization problem by solving a sequence of related quadratic programming (QP) subproblems.  Each subproblem involves the minimization of a quadratic model of the Lagrangian function subject to the linearized constraints. In contrast to the quasi-Newton approach, which maintains a positive-definite approximation of the Hessian of the Lagrangian, second-derivative SQP methods use the exact Hessian of the Lagrangian. In this context, we will discuss a dynamic convexification strategy with two main features. First, the method makes minimal matrix modifications while ensuring that the iterates of the QP subproblem are bounded. Second, the solution of the convexified QP is a descent direction for a merit function that is used to force convergence from any starting point. This talk will focus on the dynamic convexification of a class of primal-dual SQP methods.  Extensive numerical results will be presented.

March 12, 2024

11:00 AM

APM 2402 and Zoom ID 990 3560 4352
 

****************************