Stochastic Gradient Methods: Bias, Stability and Generalization

發布者:曹玲玲發布時間:2024-12-23浏覽次數:10

報告人:Yunwen Lei 助理教授 香港大學

時間:2024年12月24日(周二)下午14:00-15:00

地點:#騰訊會議 576 320 840

報告摘要:Recent developments of stochastic optimization often suggest biased gradient estimators to improve either the robustness, communication efficiency or computational speed. Representative biased stochastic gradient methods (BSGMs) include Zeroth-order stochastic gradient descent (SGD), Clipped-SGD and SGD with delayed gradients. In this talk, we present the first framework to study the stability and generalization of BSGMs for convex and smooth problems. We apply our general result to develop the first stability bound for Zeroth-order SGD with reasonable step size sequences, and the first stability bound for Clipped-SGD. While our stability analysis is developed for general BSGMs, the resulting stability bounds for both Zeroth-order SGD and Clipped-SGD match those of SGD under appropriate smoothing/clipping parameters.

報告人簡介:Yunwen Lei is currently an Assistant Professor at the Department of Mathematics, The University of Hong Kong. His main research interests include machine learning, statistical learning theory and stochastic optimization. He has published papers in prestigious journals and conference proceedings, including IEEE TIT, JMLR, COLT, ICLR, ICML and NeurIPS.

  • 聯系方式
  • 通信地址:南京市江甯區bet356手机版唯一官网路2号bet356手机版唯一官网九龍湖校區計算機學院
  • 郵政編碼:211189
  • ​辦公地點:bet356手机版唯一官网九龍湖校區計算機樓
  • 學院微信公衆号
Baidu
sogou