MoCaO Lectures: Data Science

July 11-15, 2022

The MoCaO Lectures in Computation and Optimisation for 2022 we are focusing on Data science and in particular machine learning, its algorithms, mathematical foundations and applications. These lectures are designed to be accessible to novices to the field who have a mathematics and computational background, such as PhD students, postdoc and/or inquisitive academics who wish to have a better understanding of recent advances in this dynamic field. These lectures will be given online via Zoom. Please read the notice below regarding the registration.


Speakers:

Prof. Stephen Wright: is the George B. Dantzig Professor of Computer Sciences at the University of Wisconsin-Madison. He is a past chair of the Mathematical Optimization Society and a SIAM Fellow. Currently he directs the Institute for Foundations of Data Science at the University of Wisconsin Madison. Steve is a world-renowned expert in optimization and the author of several highly cited books in this field.

Prof. Guoyin Li: is a professor in the School of Mathematics and Statistics at University of New South Wales. He was awarded an Australian Research Council Future Fellowship (for mid-career researchers) during 2014-2018. His research interests include optimisation, variational analysis, machine learning and tensor computations.

Dr. Quoc Thong Le Gia: is a Senior Lecturer in the School of Mathematics and Statistics, UNSW, Sydney. His research interests include Numerical Analysis, Approximation Theory; Partial Differential Equations; Machine Learning and Stochastic Processes.


Dates:
The 11th , 12th and 13th of July 12noon-1pm: Speaker Prof Stephen Wright
The 14th July 12noon-1pm: Speaker Dr. Quoc Thong Le Gia
The 15th July 12.30pm-2pm: Prof. Guoyin Li


IMPORTANT: Website and Registration:
Due to unforeseen problems with the registration system, all registrations up till until the date 29/06/2022 have been lost. If you did not receive a Zoom link or you have not yet registered, please contact Dr Quoc Thong Le Gia (qlegia@unsw.edu.au) .

ICM2022 Down Under @SMRI, 6–8 July 2022

Live talks are given by Regina Burachik (University of South Australia, Section 16 – Control Theory and Optimisation) and George Willis (University of Newcastle, Section 2 – Algebra).

The ICM Down Under will conclude with a twilight talk by SMRI Director Geordie Williamson on Friday evening (8 July 20:00 AEST). The talk will be recorded and broadcasted by the London Mathematical Society (virtual ICM public lecture, 8 July 12:00 BST).

For further information and registration, please visit: https://mathematical-research-institute.sydney.edu.au/news/icm2022-down-under/

MoCaO Lectures: Data Science – Second Announcement

July 11-15, 2022

The MoCaO Lectures in Computation and Optimisation for 2022 we are focusing on Data science and in particular machine learning, its algorithms, mathematical foundations and applications. These lectures are designed to be accessible to novices to the field who have a mathematics and computational background, such as PhD students, postdoc and/or inquisitive academics who wish to have a better understanding of recent advances in this dynamic field. These lectures will be given online via Zoom. Please read the notice below regarding the registration.


Speakers:

Prof. Stephen Wright: is the George B. Dantzig Professor of Computer Sciences at the University of Wisconsin-Madison. He is a past chair of the Mathematical Optimization Society and a SIAM Fellow. Currently he directs the Institute for Foundations of Data Science at the University of Wisconsin Madison. Steve is a world-renowned expert in optimization and the author of several highly cited books in this field.

Prof. Guoyin Li: is a professor in the School of Mathematics and Statistics at University of New South Wales. He was awarded an Australian Research Council Future Fellowship (for mid-career researchers) during 2014-2018. His research interests include optimisation, variational analysis, machine learning and tensor computations.

Dr. Quoc Thong Le Gia: is a Senior Lecturer in the School of Mathematics and Statistics, UNSW, Sydney. His research interests include Numerical Analysis, Approximation Theory; Partial Differential Equations; Machine Learning and Stochastic Processes.


Dates:
The 11th , 12th and 13th of July 12noon-1pm: Speaker Prof Stephen Wright
The 14th July 12noon-1pm: Speaker Dr. Quoc Thong Le Gia
The 15th July 12.30pm-2pm: Prof. Guoyin Li


IMPORTANT: Website and Registration:
Due to unforeseen problems with the registration system, all registrations up till until the date 29/06/2022 have been lost. We encourage those who have already registered to reregister using the new google form the bottom of the webpage (so you may receive the zoom details)


MoCaO Lectures:  Data Science – Mathematics of Computation and Optimisation


We apologies for any inconvenience this issue may cause. If you have any enquiries, please send an email to MoCaO@austms.org.au. Please check the website prior to the lectures for last minute information or announcements.

VA & OPT: Alberto De Marchi 

Title: Constrained Structured Optimization and Augmented Lagrangian Proximal Methods

Speaker: Alberto De Marchi (Universität der Bundeswehr München)

Date and Time: Wed May 25 2022, 17:00 AEST (Register here for remote connection via Zoom)

Abstract:

In this talk we discuss finite-dimensional constrained structured optimization problems and explore methods for their numerical solution. Featuring a composite objective function and set-membership constraints, this problem class offers a modeling framework for a variety of applications. A general and flexible algorithm is proposed that interlaces proximal methods and safeguarded augmented Lagrangian schemes. We provide a theoretical characterization of the algorithm and its asymptotic properties, deriving convergence results for fully nonconvex problems. Adopting a proximal gradient method with an oracle as a formal tool, it is demonstrated how the inner subproblems can be solved by off-the-shelf methods for composite optimization, without introducing slack variables and despite the appearance of set-valued projections. Illustrative examples show the versatility of constrained structured programs as a modeling tool and highlight benefits of the implicit approach developed. A preprint paper is available at arXiv:2203.05276.

VA & OPT: Mareike Dressler 

Title: Algebraic Perspectives on Signomial Optimization

Speaker: Mareike Dressler (University of New South Wales)

Date and Time: Wed May 11 2022, 17:00 AEST (Register here for remote connection via Zoom)

Abstract:

Signomials are obtained by generalizing polynomials to allow for arbitrary real exponents. This generalization offers great expressive power, but has historically sacrificed the organizing principle of “degree” that is central to polynomial optimization theory. In this talk, I introduce the concept of signomial rings that allows to reclaim that principle and explain how this leads to complete convex relaxation hierarchies of upper and lower bounds for signomial optimization via sums of arithmetic-geometric exponentials (SAGE) nonnegativity certificates. In the first part of the talk, I discuss the Positivstellensatz underlying the lower bounds. It relies on the concept of conditional SAGE and removes regularity conditions required by earlier works, such as convexity of the feasible set or Archimedeanity of its representing signomial inequalities. Numerical examples are provided to illustrate the performance of the hierarchy on problems in chemical engineering and reaction networks. In the second part, I provide a language for and basic results in signomial moment theory that are analogous to those in the rich moment-SOS literature for polynomial optimization. That theory is used to turn (hierarchical) inner-approximations of signomial nonnegativity cones into (hierarchical) outer-approximations of the same, which eventually yields the upper bounds for signomial optimization. This talk is based on joint work with Riley Murray.

VA & OPT: Lars Grüne

Title: The turnpike property: a classical feature of optimal control problems revisited

Speaker: Lars Grüne (University of Bayreuth)

Date and Time: Wed May 04 2022, 17:00 AEST (Register here for remote connection via Zoom)

Abstract:

The turnpike property describes a particular behavior of optimal control problems that was first observed by Ramsey in the 1920s and by von Neumann in the 1930s. Since then it has found widespread attention in mathematical economics and control theory alike. In recent years it received renewed interest, on the one hand in optimization with partial differential equations and on the other hand in model predictive control (MPC), one of the most popular optimization based control schemes in practice. In this talk we will first give a general introduction to and a brief history of the turnpike property, before we look at it from a systems and control theoretic point of view. Particularly, we will clarify its relation to dissipativity, detectability, and sensitivity properties of optimal control problems in both finite and infinite dimensions. In the final part of the talk we will explain why the turnpike property is important for analyzing the performance of MPC.

VA & OPT: Andreas Lohne

Title: Approximating convex bodies using multiple objective optimization

Speaker: Andreas Löhne (Friedrich Schiller University Jena)

Date and Time: Wed Apr 27 2022, 17:00 AEST (Register here for remote connection via Zoom)

Abstract:

The problem to compute a polyhedral outer and inner approximation of a convex body can be reformulated as a problem to solve approximately a convex multiple objective optimization problem. This extends a previous result showing that multiple objective linear programming is equivalent to compute a $V$-representation of the projection of an $H$-polyhedron. These results are also discussed with respect to duality, solution methods and error bounds.

VA & OPT Webinar: Héctor Ramírez

Title: Extensions of Constant Rank Qualification Constrains condition to Nonlinear Conic Programming

Speaker: Héctor Ramírez (Universidad de Chile)

Date and Time: Wed Apr 13 2022, 11:00 AEST (Register here for remote connection via Zoom)

Abstract:

We present new constraint qualification conditions for nonlinear conic programming that extend some of the constant rank-type conditions from nonlinear programming. As an application of these conditions, we provide a unified global convergence proof of a class of algorithms to stationary points without assuming neither uniqueness of the Lagrange multiplier nor boundedness of the Lagrange multipliers set. This class of algorithms includes, for instance, general forms of augmented Lagrangian, sequential quadratic programming, and interior point methods. We also compare these new conditions with some of the existing ones, including the nondegeneracy condition, Robinson’s constraint qualification, and the metric subregularity constraint qualification. Finally, we propose a more general and geometric approach for defining a new extension of this condition to the conic context. The main advantage of the latter is that we are able to recast the strong second-order properties of the constant rank condition in a conic context. In particular, we obtain a second-order necessary optimality condition that is stronger than the classical one obtained under Robinson’s constraint qualification, in the sense that it holds for every Lagrange multiplier, even though our condition is independent of Robinson’s condition.

VA & OPT Webinar: Sorin-Mihai Grad

Title: Extending the proximal point algorithm beyond convexity

Speaker: Sorin-Mihai Grad (ENSTA Paris)

Date and Time: Wed Apr 06 2022, 17:00 AEST (Register here for remote connection via Zoom)

Abstract:

Introduced in in the 1970’s by Martinet for minimizing convex functions and extended shortly afterwards by Rockafellar towards monotone inclusion problems, the proximal point algorithm turned out to be a viable computational method for solving various classes of (structured) optimization problems even beyond the convex framework. In this talk we discuss some extensions of proximal point type algorithms beyond convexity. First we propose a relaxed-inertial proximal point type algorithm for solving optimization problems consisting in minimizing strongly quasiconvex functions whose variables lie in finitely dimensional linear subspaces, that can be extended to equilibrium functions involving such functions. Then we briefly discuss another generalized convexity notion for functions we called prox-convexity for which the proximity operator is single-valued and firmly nonexpansive, and see that the standard proximal point algorithm and Malitsky’s Golden Ratio Algorithm (originally proposed for solving convex mixed variational inequalities) remain convergent when the involved functions are taken prox-convex, too. The talk contains joint work with Felipe Lara and Raúl Tintaya Marcavillaca (both from University of Tarapacá).

VA & OPT Webinar: Pham Ky Anh

Title: Regularized dynamical systems associated with structured monotone inclusions

Speaker: Pham Ky Anh (Vietnam National University)

Date and Time: Wed Mar 30 2022, 17:00 AEST (Register here for remote connection via Zoom)

Abstract:

In this report, we consider two dynamical systems associated with additively structured monotone inclusions involving a multi-valued maximally monotone operator A and a single-valued operator B in real Hilbert spaces. We established strong convergence of the regularized forward-backward and regularized forward – backward–forward dynamics to an “optimal” solution of the original inclusion under a weak assumption on the single-valued operator B. Convergence estimates are obtained if the composite operator A + B is maximally monotone and strongly (pseudo)monotone. Time-discretization of the corresponding continuous dynamics provides an iterative regularization forward-backward method or an iterative regularization forward-backward-forward method with relaxation parameters. Some simple numerical examples were given to illustrate the agreement between analytical and numerical results as well as the performance of the proposed algorithms.

1 2 3 4 10