Back
Machine learning has become a popular application domain for modern optimization techniques, pushing its algorithmic frontier. The need for large scale optimization algorithms which can handle millions of dimensions or data points, typical for the big data era, have brought a resurgence of interest for first order algorithms, making us revisit the venerable stochastic gradient method [Robbins-Monro 1951] as well as the Frank-Wolfe algorithm [Frank-Wolfe 1956]. In this talk, I will review recent improvements on these algorithms which can exploit the structure of modern machine learning approaches. I will explain why the Frank-Wolfe algorithm has become so popular lately; and present a surprising tweak on the stochastic gradient method which yields a fast linear convergence rate. Motivating applications will include weakly supervised video analysis and structured prediction problems.
Simon Lacoste-Julien ( Department of Computer Science and Operations Research, Université de Montréal)
Assistant Professor
Simon is an expert on optimization, and has also contributed notably to the theory of approximate inference. After positions in the labs of Mike Jordan, Zoubin Ghahramani (where I briefly shared an office with him), and Francis Bach, he now runs his own show in Montréal.