Presenter: Hanseul Cho (OptiML Lab @KAIST AI)
Papers presented:
A mean field view of the landscape of two-layer neural networks (Mei et al., PNAS 2018)
Mean-field theory of two-layers neural networks: dimension-free bounds and kernel limit (Mei et al., COLT 2019)
Presenter (offline): Hanseul Cho (OptiML Lab @KAIST AI)
Papers presented:
Hausdorff dimension, heavy tails, and generalization in neural networks (Şimşekli et al., NeurIPS 2020; Journal of Statistical Mechanics 2021)
Generalization Bounds using Lower Tail Exponents in Stochastic Optimizers (Hodgkinson et al., ICML 2022)
Presenter (offline): Junghyun Lee (OSI Lab & OptiML @KAIST AI)
Papers presented:
Entropy-SGD: biasing gradient descent into wide valleys (Chaudhari et al., ICLR 2017; Journal of Statistical Mechanics 2018)
Entropic gradient descent algorithms and wide flat minima (Pittorino et al., ICLR 2021; Journal of Statistical Mechanics 2021)