Systems | Information | Learning | Optimization
 

SILO: Theory for Diffusion Models

Abstract: In this talk I will survey our recent efforts to develop a rigorous theory for understanding diffusion generative modeling. The first part will cover discretization analyses that prove that diffusion models can approximately sample from arbitrary probability distributions provided one can have a sufficiently accurate estimate for the score …

SILO: Learning Dynamics for Nash and Coarse Correlated Equilibria in Bimatrix Games

Abstract: In this talk, we will focus on learning in two-player games. First, we will provide a brief introduction to the possible behaviors of learning algorithms and mention various techniques that have been extensively used to guarantee convergence to Nash equilibria in zero-sum games. Finally, we will demonstrate how these …

SILO: Beyond Decoder-Only Next Token Prediction

Abstract: This talk presents two distinct approaches that expand the potential of Transformer architectures beyond the traditional decoder-only, causal-attention models for next-token prediction. In the first half, we will examine looped Transformers with an adaptive iteration mechanism, demonstrating that these models can learn highly length-generalizable solutions for algorithmic tasks. The …

SILO: Recent Advances in Min-max Optimization: Convergence Guarantees and Practical Performance

Abstract: Min-max optimization plays a prominent role in game theory, statistics, economics, finance, and engineering. It has recently received significant attention, especially in the machine learning community, where adversarial training of neural networks, multi-agent reinforcement learning, and distributionally robust learning are formulated as structured min-max optimization problems. Stochastic Gradient Descent Ascent …

SILO: Confidence Sequences via Online Learning

Abstract: Confidence sequence provides ways to characterize uncertainty in stochastic environments, which is a widely-used tool for interactive machine learning algorithms and statistical problems including A/B testing, Bayesian optimization, reinforcement learning, and offline learning.  In these problems, constructing confidence sequences that are tight without losing correctness is crucial since it …

SILO: Hidden Convexity of Deep Neural Networks: Exact and Transparent Lasso Formulations via Geometric Algebra

Abstract: In this talk, we introduce an analysis of deep neural networks through convex optimization and geometric (Clifford) algebra. We begin by introducing exact convex optimization formulations for ReLU neural networks. This approach demonstrates that deep networks can be globally trained through convex programs, offering a globally optimal solution. Our …