Adam Can Converge Without Any Modification On Update Rules Y Zhang, C Chen, N Shi, R Sun, ZQ Luo Advances in Neural Information Processing Systems, 2022, 2022 | 39 | 2022 |
Provable adaptivity in adam B Wang, Y Zhang, H Zhang, Q Meng, ZM Ma, TY Liu, W Chen arXiv preprint arXiv:2208.09900, 2022 | 20 | 2022 |
HyperDQN: A randomized exploration method for deep reinforcement learning Z Li, Y Li, Y Zhang, T Zhang, ZQ Luo International Conference on Learning Representations, 2021 | 13 | 2021 |
When Expressivity Meets Trainability: Fewer than Neurons Can Work J Zhang, Y Zhang, M Hong, R Sun, ZQ Luo Advances in Neural Information Processing Systems, 2021, 2021 | 8 | 2021 |
Remax: A simple, effective, and efficient method for aligning large language models Z Li, T Xu, Y Zhang, Y Yu, R Sun, ZQ Luo arXiv preprint arXiv:2310.10505, 2023 | 3 | 2023 |
Fast QLB algorithm and hypothesis tests in logistic model for ophthalmologic bilateral correlated data YQ Lin, YS Zhang, GL Tian, CX Ma Journal of Biopharmaceutical Statistics 31 (1), 91-107, 2021 | 3 | 2021 |
Why Transformers Need Adam: A Hessian Perspective Y Zhang, C Chen, T Ding, Z Li, R Sun, ZQ Luo arXiv preprint arXiv:2402.16788, 2024 | | 2024 |
Does Adam Converge and When? Y Zhang, C Chen, ZQ Luo ICLR 2022 Blog Track, 2022 | | 2022 |