Superhuman AI for multiplayer poker N Brown, T Sandholm Science 365 (6456), 885-890, 2019 | 1100 | 2019 |
Superhuman AI for heads-up no-limit poker: Libratus beats top professionals N Brown, T Sandholm Science 359 (6374), 418-424, 2018 | 1008 | 2018 |
Openai o1 system card A Jaech, A Kalai, A Lerer, A Richardson, A El-Kishky, A Low, A Helyar, ... arXiv preprint arXiv:2412.16720, 2024 | 526 | 2024 |
Human-level play in the game of Diplomacy by combining language models with strategic reasoning Meta Fundamental AI Research Diplomacy Team (FAIR)†, A Bakhtin, ... Science 378 (6624), 1067-1074, 2022 | 307 | 2022 |
Deep counterfactual regret minimization N Brown, A Lerer, S Gross, T Sandholm International Conference on Machine Learning, 2019 | 305 | 2019 |
Safe and nested subgame solving for imperfect-information games N Brown, T Sandholm Neural Information Processing Systems, 2017 | 246* | 2017 |
Solving imperfect-information games via discounted regret minimization N Brown, T Sandholm Proceedings of the AAAI Conference on Artificial Intelligence 33 (01), 1829-1836, 2019 | 201 | 2019 |
Combining deep reinforcement learning and search for imperfect-information games N Brown, A Bakhtin, A Lerer, Q Gong Advances in neural information processing systems 33, 17057-17069, 2020 | 188 | 2020 |
Libratus: The Superhuman AI for No-Limit Poker. N Brown, T Sandholm IJCAI, 5226-5228, 2017 | 163 | 2017 |
Depth-limited solving for imperfect-information games N Brown, T Sandholm, B Amos Advances in neural information processing systems 31, 2018 | 108 | 2018 |
Improving Policies via Search in Cooperative Partially Observable Games A Lerer, H Hu, J Foerster, N Brown AAAI Conference on Artificial Intelligence, 2020 | 95 | 2020 |
Hierarchical Abstraction, Distributed Equilibrium Computation, and Post-Processing, with Application to a Champion No-Limit Texas Hold'em Agent. N Brown, S Ganzfried, T Sandholm AAAI Workshop: Computer Poker and Imperfect Information 15, 07, 2015 | 89 | 2015 |
Off-belief learning H Hu, A Lerer, B Cui, L Pineda, N Brown, J Foerster International Conference on Machine Learning, 4369-4379, 2021 | 80 | 2021 |
A unified approach to reinforcement learning, quantal response equilibria, and two-player zero-sum games S Sokota, R D'Orazio, JZ Kolter, N Loizou, M Lanctot, I Mitliagkas, ... arXiv preprint arXiv:2206.05825, 2022 | 70 | 2022 |
Modeling strong and human-like gameplay with KL-regularized search AP Jacob, DJ Wu, G Farina, A Lerer, H Hu, A Bakhtin, J Andreas, N Brown International Conference on Machine Learning, 9695-9728, 2022 | 67 | 2022 |
Dream: Deep regret minimization with advantage baselines and model-free learning E Steinberger, A Lerer, N Brown arXiv preprint arXiv:2006.10410, 2020 | 62 | 2020 |
Human-level performance in no-press diplomacy via equilibrium search J Gray, A Lerer, A Bakhtin, N Brown arXiv preprint arXiv:2010.02923, 2020 | 60 | 2020 |
Mastering the game of no-press diplomacy via human-regularized reinforcement learning and planning A Bakhtin, DJ Wu, A Lerer, J Gray, AP Jacob, G Farina, AH Miller, ... arXiv preprint arXiv:2210.05492, 2022 | 59 | 2022 |
Dynamic thresholding and pruning for regret minimization N Brown, C Kroer, T Sandholm Proceedings of the AAAI conference on artificial intelligence 31 (1), 2017 | 56 | 2017 |
No-press diplomacy from scratch A Bakhtin, D Wu, A Lerer, N Brown Advances in Neural Information Processing Systems 34, 18063-18074, 2021 | 53 | 2021 |