Follow
Ming Yin
Title
Cited by
Cited by
Year
Near-optimal provable uniform convergence in offline policy evaluation for reinforcement learning
M Yin, Y Bai, YX Wang
(AISTATS oral) International Conference on Artificial Intelligence and …, 2021
83*2021
Towards instance-optimal offline reinforcement learning with pessimism
M Yin, YX Wang
(NeurIPS) Advances in neural information processing systems 34, 4065-4078, 2021
722021
Asymptotically efficient off-policy evaluation for tabular reinforcement learning
M Yin, YX Wang
(AISTATS) International Conference on Artificial Intelligence and Statistics …, 2020
692020
Near-optimal offline reinforcement learning with linear representation: Leveraging variance information with pessimism
M Yin, Y Duan, M Wang, YX Wang
(ICLR) Internation Conference on Learning Representations, 2022, 2022
682022
Near-optimal offline reinforcement learning via double variance reduction
M Yin, Y Bai, YX Wang
(NeurIPS) Advances in neural information processing systems 34, 7677-7688, 2021
672021
Mmmu: A massive multi-discipline multimodal understanding and reasoning benchmark for expert agi
X Yue, Y Ni, K Zhang, T Zheng, R Liu, G Zhang, S Stevens, D Jiang, ...
(CVPR) Conference on Computer Vision and Pattern Recognition, 2024
472024
Theoremqa: A theorem-driven question answering dataset
W Chen, M Yin, M Ku, P Lu, Y Wan, X Ma, J Xu, X Wang, T Xia
(EMNLP) Empirical Methods in Natural Language Processing, 2023
372023
Optimal uniform ope and model-based offline reinforcement learning in time-homogeneous, reward-free and task-agnostic settings
M Yin, YX Wang
(NeurIPS) Advances in neural information processing systems 34, 12890-12903, 2021
252021
Sample-efficient reinforcement learning with loglog (t) switching cost
D Qiao, M Yin, M Min, YX Wang
(ICML) International Conference on Machine Learning, 18031-18061, 2022
242022
Offline reinforcement learning with differentiable function approximation is provably efficient
M Yin, M Wang, YX Wang
(ICLR) Internation Conference on Learning Representations, 2023, 2023
132023
On Instance-Dependent Bounds for Offline Reinforcement Learning with Linear Function Approximation
T Nguyen-Tang, M Yin, S Gupta, S Venkatesh, R Arora
(AAAI) AAAI Conference on Artificial Intelligence, 2023, 2023
112023
Logarithmic switching cost in reinforcement learning beyond linear mdps
D Qiao, M Yin, YX Wang
arXiv preprint arXiv:2302.12456, 2023
62023
Offline Stochastic Shortest Path: Learning, Evaluation and Towards Optimality
M Yin, W Chen, M Wang, YX Wang
(UAI) The 38th Conference on Uncertainty in Artificial Intelligence, 2022
52022
Non-stationary Reinforcement Learning under General Function Approximation
S Feng, M Yin, R Huang, YX Wang, J Yang, Y Liang
(ICML) International Conference on Machine Learning, 2023
42023
No-Regret Linear Bandits beyond Realizability
C Liu, M Yin, YX Wang
(UAI) The 39th Conference on Uncertainty in Artificial Intelligence, 2023
32023
Why quantization improves generalization: Ntk of binary weight neural networks
K Zhang, M Yin, YX Wang
arXiv preprint arXiv:2206.05916, 2022
32022
Offline Policy Evaluation for Reinforcement Learning with Adaptively Collected Data
S Madhow, D Xiao, M Yin, YX Wang
3rd Offline RL Workshop: Offline RL as a''Launchpad'', 2022
22022
Posterior Sampling with Delayed Feedback for Reinforcement Learning with Linear Function Approximation
M Yin*, NL Kuang*, M Wang, YX Wang, YA Ma
(NeurIPS) Advances in neural information processing systems, 2023, 2023
12023
Model-free algorithm with improved sample efficiency for zero-sum markov games
S Feng, M Yin, YX Wang, J Yang, Y Liang
arXiv preprint arXiv:2308.08858, 2023
12023
On the Data Complexity of Problem-Adaptive Offline Reinforcement Learning
M Yin
UC Santa Barbara, 2023
1*2023
The system can't perform the operation now. Try again later.
Articles 1–20