Seguir
Yu Bai
Yu Bai
OpenAI
Email confirmado em openai.com - Página inicial
Título
Citado por
Citado por
Ano
The landscape of empirical risk for nonconvex losses
S Mei, Y Bai, A Montanari
The Annals of Statistics 46 (6A), 2747-2774, 2018
3762018
Provable self-play algorithms for competitive reinforcement learning
Y Bai, C Jin
International conference on machine learning, 551-560, 2020
1872020
Policy finetuning: Bridging sample-efficient offline and online reinforcement learning
T Xie, N Jiang, H Wang, C Xiong, Y Bai
Advances in neural information processing systems 34, 27395-27407, 2021
1722021
Transformers as statisticians: Provable in-context learning with in-context algorithm selection
Y Bai, F Chen, H Wang, C Xiong, S Mei
Advances in neural information processing systems 36, 2024
1622024
Near-Optimal Reinforcement Learning with Self-Play
Y Bai, C Jin, T Yu
Advances in Neural Information Processing Systems, 2020, 2020
1602020
A sharp analysis of model-based reinforcement learning with self-play
Q Liu, T Yu, Y Bai, C Jin
International Conference on Machine Learning, 7001-7010, 2021
1562021
Beyond linearization: On quadratic and higher-order approximation of wide neural networks
Y Bai, JD Lee
International Conference on Learning Representations (ICLR) 2020, 2019
1382019
Proxquant: Quantized neural networks via proximal operators
Y Bai, YX Wang, E Liberty
International Conference on Learning Representations (ICLR) 2019, 2018
1262018
Provably Efficient Q-Learning with Low Switching Cost
Y Bai, T Xie, N Jiang, YX Wang
Advances in Neural Information Processing Systems, 2019, 2019
1162019
When can we learn general-sum Markov games with a large number of players sample-efficiently?
Z Song, S Mei, Y Bai
International Conference on Learning Representations (ICLR) 2022, 2021
1132021
Near-optimal provable uniform convergence in offline policy evaluation for reinforcement learning
M Yin, Y Bai, YX Wang
International Conference on Artificial Intelligence and Statistics, 1567-1575, 2021
95*2021
How important is the train-validation split in meta-learning?
Y Bai, M Chen, P Zhou, T Zhao, J Lee, S Kakade, H Wang, C Xiong
International Conference on Machine Learning, 543-553, 2021
882021
Approximability of discriminators implies diversity in GANs
Y Bai, T Ma, A Risteski
International Conference on Learning Representations (ICLR) 2019, 2018
872018
Sample-efficient learning of stackelberg equilibria in general-sum games
Y Bai, C Jin, H Wang, C Xiong
Advances in Neural Information Processing Systems 34, 25799-25811, 2021
782021
Near-optimal offline reinforcement learning via double variance reduction
M Yin, Y Bai, YX Wang
Advances in neural information processing systems 34, 7677-7688, 2021
722021
The role of coverage in online reinforcement learning
T Xie, DJ Foster, Y Bai, N Jiang, SM Kakade
arXiv preprint arXiv:2210.04157, 2022
712022
Subgradient descent learns orthogonal dictionaries
Y Bai, Q Jiang, J Sun
International Conference on Learning Representations (ICLR) 2019, 2018
592018
Negative preference optimization: From catastrophic collapse to effective unlearning
R Zhang, L Lin, Y Bai, S Mei
arXiv preprint arXiv:2404.05868, 2024
572024
Don't Just Blame Over-parametrization for Over-confidence: Theoretical Analysis of Calibration in Binary Classification
Y Bai, S Mei, H Wang, C Xiong
International Conference on Machine Learning, 566-576, 2021
532021
Towards understanding hierarchical learning: Benefits of neural representations
M Chen, Y Bai, JD Lee, T Zhao, H Wang, C Xiong, R Socher
Advances in Neural Information Processing Systems, 2020, 2020
532020
O sistema não pode efectuar a operação agora. Tente mais tarde.
Artigos 1–20