MOCAO lectures 2026
Title: Model-Based Derivative-Free Optimization: Algorithms and Approximation Theory
Abstract:
Derivative-free optimization (DFO) refers to nonlinear optimization algorithms that do not rely on the availability of gradient or Hessian information. It is primarily designed for settings when functions are black-box, expensive to evaluate and/or noisy, relevant to applications including climate science and quantum computing. A widely used and studied class of DFO methods for local optimization is model-based DFO (MBDFO), where the general principles from nonlinear optimization algorithms are followed, but with local approximations to the objective constructed by polynomial interpolation (rather than, e.g. Taylor series). In these lectures, we will cover the foundational approximation theory and algorithms behind MBDFO, and look at recent extensions to constrained and noisy problems. The lecture by Zaikun Zhang (Sun Yat-sen University) will consider MBDFO in subspaces to handle high-dimensional problems, and Sara Shashaani (North Carolina State University) will discuss adaptive sampling MBDFO methods for stochastic problems.
Other presenters: Dr. Lindon Roberts, The University of Melbourne
Registration (for urgent updates): https://docs.google.com/forms/d/e/1FAIpQLScW2PQtKLcxsRyZSzdZBv1lPGzEkD2aFeiwHVUP9CbRLSWRTw/viewform?usp=publish-editor
July 27-31, 2026, the time will be communicated later.