Speaker:HE Niao,University of Illinois at Urbana-Champaign
Time:2016-07-22,16:00-17:00
Place:Room 1318, School of Mathematical Sciences
Detail:
The era of Big Data is introducing big challenges for machine learning to accommodate to rapidly growing data and increasingly complex models. Many traditional learning methods fail to work due to the lack of scalability or theoretical guidance. In this talk, I am going to show how simple optimization algorithms such as stochastic gradient descent, when combined with newly developed techniques, could make a huge difference. I will discuss some of our recent works that advance three important sub-domains of machine learning, including kernel machines, Bayesian inference, and reinforcement learning. For all these cases, we developed simple new algorithms that allow to train bigger models, learn better yet faster, come with provable guarantees, and improve significantly over previous state-of-the-arts. These advances are proven to be useful in a wide range of machine learning applications on large-scale real world datasets.
She is an Assistant Professor in the Department of Industrial and Enterprise Systems Engineering at University ofIllinois at Urbana-Champaign. She completed her PhD in Operations Research under the supervision of Professor ArkadiNemirovski in the Department of Mathematics at Georgia Institute of Technology in 2015.Meanwhile she got a Master degree in Computational Science and Engineering. She has been working on developing fast algorithms and theoretical analysis for large-scale convex/stochastic/robust/distributed optimization with applications to finance, machine learning, and decision-making under uncertainty. She also has a particular interest in the integration of optimization, machine learning and statistics.
Organizer: School of Mathematical Sciences