科学研究
报告题目:

On the (Non)smoothness of Neural Network Training

报告人:

张景昭(清华大学交叉信息科学院)

报告时间:

报告地点:

ZOOMID:561 420 9883 密码:tmcc2022

报告摘要:

In this talk, we will discuss the following questions―why is neural network training non-smooth from an optimization perspective, and how should we analyze convergence in nonsmooth settings? We will start by showing that the non-smoothness is essential to standard neural network training procedures, and that network training converges in an unstable manner. We then provide theoretical models for understanding why optimization in neural network is unstable. We conclude by showing that new definitions of convergence in the nonsmooth settings can reconcile theory with practice.

Baidu
sogou

世界杯买球赛平台「2022卡塔尔世界杯官方合作伙伴」