2025/2026 SEMINARS

FALL

WINTER

SPRING

Math 208 - Algebraic Geometry

Oprea, Dragos

Oprea, Dragos

Oprea, Dragos

Math 209 - Number Theory

Bucur, Alina

Bucur, Alina

Bucur, Alina

Math 211A - Algebra

Golsefidy, Alireza

Golsefidy, Alireza

Golsefidy, Alireza

Math 211B - Group Actions

Frisch, Joshua

Frisch, Joshua

Frisch, Joshua

Math 218 - Biological Systems

Miller, Pearson

Miller, Pearson

Miller, Pearson

Math 243 - Functional Analysis

Ganesan, Priyanga & Vigdorovich, Itamar

Ganesan, Priyanga & Vigdorovich, Itamar

Vigdorovich, Itamar

Math 248 - Real Analysis

Bejenaru, Ioan

Bejenaru, Ioan

Bejenaru, Ioan

Math 258 - Differential Geometry

Spolaor, Luca

Spolaor, Luca

Spolaor, Luca

Math 268 - Logic

TBD

TBD

TBD

Math 269 - Combinatorics

Rhoades, Brendon & Warnke, Lutz

Rhoades, Brendon & Warnke, Lutz

Rhoades, Brendon & Warnke, Lutz

Math 278A - CCoM

Cheng, Li-Tien

Cheng, Li-Tien

Cheng, Li-Tien

Math 278B - Math of Info, Data

Cloninger, Alexander

Cloninger, Alexander

Cloninger, Alexander

Math 278C - Optimization

Nie, Jiawang

Nie, Jiawang

Nie, Jiawang

Math 288A - Probability

Peca-Medlin, John

Peca-Medlin, John

Peca-Medlin, John

Math 288B - Statistics

TBD

TBD

TBD

Math 292 - Topology Seminar

Chow, Bennett

Chow, Bennett

Chow, Bennett

Tue, Dec 9 2025
  • 12:00 pm
    Tongtong Liang - UCSD
    Rethinking Generalization in Deep Learning: The Role of Data Geometry

    Advancement to Candidacy

    APM 6402

    We study how data geometry shapes generalization in overparameterized neural networks. The analysis focuses on solutions reached under stable training dynamics and the induced, data-dependent form of regularization. We link capacity to geometric features of the input distribution. This view explains when training prefers shared representations versus memorization. We present a decomposition based on depth-type notions to separate regions where learning is data-rich from regions where activation is scarce. For the uniform distribution on the ball, the framework predicts the curse of dimensionality. For mixtures supported on low-dimensional subspaces, it predicts adaptation to the intrinsic dimension. Experiments on synthetic data and MNIST support these trends. The results provide a unified account of how stability and geometry interact to govern effective capacity of GD-trained neural networks.

  • 1:00 pm
    Siddharth Vishwanath - University of California, San Diego
    A Statistical Framework for Multidimensional Scaling From Noisy Data

    APM 6402 & Zoom

    Multidimensional scaling (MDS) extracts meaningful information from pairwise dissimilarity data (e.g., distances between sensors or disagreement scores between individuals) by embedding these relationships into a Euclidean space. However, in practice, the observed dissimilarities are often noisy subject to measurement errors and/or corrupted by noise, but the resulting embeddings are typically interpreted without accounting for this variation. This talk presents recent work on developing a principled statistical framework for MDS. We show that the classical MDS algorithm achieves minimax-optimal performance across a wide range of noise models and loss functions. Building on this, we develop a framework for constructing valid confidence sets for the embedded points obtained via MDS, enabling formal uncertainty quantification for geometric structure inferred from noisy relational data.

Thu, Dec 11 2025
  • 3:00 pm
    Juan Felipe Ariza Mejia - University of Iowa
    McDuff superrigidity for group $II_1$ factors

    Math 243: Seminar in Functional Analysis

    APM 6402

    Developing new techniques at the interface of geometric group theory and von Neumann algebras, we identify the first examples of ICC groups $G$ whose von Neumann algebras are McDuff and exhibit a new rigidity phenomenon, termed McDuff superrigidity: any arbitrary group $H$ satisfying $LG\cong LH$ must decomposes as $H=G \times A$ for some ICC amenable group $A$.  Our groups appear as infinite direct sums of $W^*$-superrigid wreath-like product groups with bounded cocycle.  In this talk I will introduce this class of groups and a natural array into a weakly-$\ell^2$ representation of the group that witnesses the bound on the 2-cocycle. I will then show how this array leads to an interplay between two deformations of the group von Neumann algebra and how these can be used to prove this class of groups satisfies infinite product rigidity.  This is joint work with Ionut Chifan, Denis Osin and Bin Sun.

Tue, Jan 6 2026
  • 11:00 am
    Junchen Zhao - Texas A&M University
    TBA

    Math 243: Seminar in Functional Analysis

    APM 6402

    TBA

Thu, Feb 12 2026
  • 4:00 pm
    Dave Penneys - Ohio State University
    TBA

    Math 295: Colloquium Seminar

    APM 6402

    TBA