Author of the publication

BN-invariant Sharpness Regularizes the Training Model to Better Generalization.

, , , , and . IJCAI, page 4164-4170. ijcai.org, (2019)

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed. You can also use the button next to the name to display some publications already assigned to the person.

 

Other publications of authors with the same name

Asynchronous Stochastic Gradient Descent with Delay Compensation, , , , , , and . (2016)cite arxiv:1609.08326Comment: 20 pages, 5 figures.BN-invariant Sharpness Regularizes the Training Model to Better Generalization., , , , and . IJCAI, page 4164-4170. ijcai.org, (2019)Provable Adaptivity in Adam., , , , , , and . CoRR, (2022)A Group Symmetric Stochastic Differential Equation Model for Molecule Multi-modal Pretraining., , , , and . ICML, volume 202 of Proceedings of Machine Learning Research, page 21497-21526. PMLR, (2023)On the Convergence of Adam under Non-uniform Smoothness: Separability from SGDM and Beyond., , , , , and . CoRR, (2024)A Generative Approach to Control Complex Physical Systems., , , , , , , , , and . CoRR, (2024)Provable Adaptivity of Adam under Non-uniform Smoothness., , , , , , , , and . KDD, page 2960-2969. ACM, (2024)Non-Asymptotic Analysis of Excess Risk via Empirical Risk Landscape., , and . CoRR, (2020)Uncertainty Calibration for Ensemble-Based Debiasing Methods., , , , , and . NeurIPS, page 13657-13669. (2021)Reweighting Augmented Samples by Minimizing the Maximal Expected Loss., , , , , and . ICLR, OpenReview.net, (2021)