科学研究
报告题目:

On the (Non)smoothness of Neural Network Training

报告人:

张景昭(清华大学交叉信息科学院)

报告时间:

报告地点:

ZOOMID:561 420 9883 密码:tmcc2022

报告摘要:

In this talk, we will discuss the following questions―why is neural network training non-smooth from an optimization perspective, and how should we analyze convergence in nonsmooth settings? We will start by showing that the non-smoothness is essential to standard neural network training procedures, and that network training converges in an unstable manner. We then provide theoretical models for understanding why optimization in neural network is unstable. We conclude by showing that new definitions of convergence in the nonsmooth settings can reconcile theory with practice.