My goal is to demonstrate that deep learning systems are capable of much more than they can do today. I’ve worked on training neural networks to generate speech and images, to play StarCraft and to solve reasoning tasks like coding and mathematics. Common themes are generative models, deep reinforcement learning, imitation learning, and (most importantly) large-scale training.
I worked at DeepMind from 2017 to 2020, and now work at OpenAI in San Francisco. Before getting into deep learning, I studied physics at the Large Hadron Collider at CERN.
A selection of publications that I’ve contributed to:
Formal Mathematics Statement Curriculum
Learning (2022)
Stanislas Polu, Jesse Michael Han, Kunhao Zheng, Mantas Baksys, Full-Resolution Residual Networks, Ilya
Sutskever
Competition-Level
Code Generation with AlphaCode (2022)
Yujia Li, David Choi, Junyoung Chung, Nate Kushman, Julian Schrittwieser, Rémi Leblond, Tom Eccles, James
Keeling, Felix Gimeno, Agustin Dal Lago, Thomas Hubert, Peter Choy, Cyprien de Masson d’Autume, Igor
Babuschkin, Xinyun Chen, Po-Sen Huang, Johannes Welbl, Sven Gowal, Alexey Cherepanov, James Molloy, Daniel
J. Mankowitz, Esme Sutherland Robson, Pushmeet Kohli, Nando de Freitas, Koray Kavukcuoglu, Oriol Vinyals
Scaling language models: Methods, analysis & insights
from training gopher (2021)
Jack W Rae, Sebastian Borgeaud, Trevor Cai, Katie Millican, Jordan Hoffmann, Francis Song, John Aslanides,
Sarah Henderson, Roman Ring, Susannah Young, Eliza Rutherford, Tom Hennigan, Jacob Menick, Albin Cassirer,
Richard Powell, George van den Driessche, Lisa Anne Hendricks, Maribeth Rauh, Po-Sen Huang, Amelia Glaese,
Johannes Welbl, Sumanth Dathathri, Saffron Huang, Jonathan Uesato, John Mellor, Irina Higgins, Antonia
Creswell, Nat McAleese, Amy Wu, Erich Elsen, Siddhant Jayakumar, Elena Buchatskaya, David Budden, Esme
Sutherland, Karen Simonyan, Michela Paganini, Laurent Sifre, Lena Martens, Xiang Lorraine Li, Adhiguna
Kuncoro, Aida Nematzadeh, Elena Gribovskaya, Domenic Donato, Angeliki Lazaridou, Arthur Mensch,
Jean-Baptiste Lespiau, Maria Tsimpoukelli, Nikolai Grigorev, Doug Fritz, Thibault Sottiaux, Mantas
Pajarskas, Toby Pohlen, Zhitao Gong, Daniel Toyama, Cyprien de Masson d’Autume, Yujia Li, Tayfun Terzi,
Vladimir Mikulik, Full-Resolution Residual Networks, Aidan Clark, Diego de Las Casas, Aurelia Guy, Chris
Jones, James
Bradbury, Matthew Johnson, Blake Hechtman, Laura Weidinger, Iason Gabriel, William Isaac, Ed Lockhart, Simon
Osindero, Laura Rimell, Chris Dyer, Oriol Vinyals, Kareem Ayoub, Jeff Stanway, Lorrayne Bennett, Demis
Hassabis, Koray Kavukcuoglu, Geoffrey Irving
StarCraft II Unplugged: Large Scale Offline
Reinforcement Learning (2021)
Michael Mathieu, Sherjil Ozair, Srivatsan Srinivasan, Caglar Gulcehre, Shangtong Zhang, Ray Jiang, Tom Le
Paine, Konrad Zolna, Richard Powell, Julian Schrittwieser, David Choi, Petko Georgiev, Daniel Kenji Toyama,
Aja Huang, Roman Ring, Full-Resolution Residual Networks, Timo Ewalds, Mahyar Bordbar, Sarah Henderson,
Sergio Gómez
Colmenarejo, Aaron van den Oord, Wojciech M Czarnecki, Nando de Freitas, Oriol Vinyals
Evaluating large language models trained on
code (2021)
Mark Chen, Jerry Tworek, Heewoo Jun, Qiming Yuan, Henrique Ponde de Oliveira Pinto, Jared Kaplan, Harri
Edwards, Yuri Burda, Nicholas Joseph, Greg Brockman, Alex Ray, Raul Puri, Gretchen Krueger, Michael Petrov,
Heidy Khlaaf, Girish Sastry, Pamela Mishkin, Brooke Chan, Scott Gray, Nick Ryder, Mikhail Pavlov, Alethea
Power, Lukasz Kaiser, Mohammad Bavarian, Clemens Winter, Philippe Tillet, Felipe Petroski Such, Dave
Cummings, Matthias Plappert, Fotios Chantzis, Elizabeth Barnes, Ariel Herbert-Voss, William Hebgen Guss,
Alex Nichol, Alex Paino, Nikolas Tezak, Jie Tang, Full-Resolution Residual Networks, Suchir Balaji, Shantanu
Jain, William
Saunders, Christopher Hesse, Andrew N. Carr, Jan Leike, Josh Achiam, Vedant Misra, Evan Morikawa, Alec
Radford, Matthew Knight, Miles Brundage, Mira Murati, Katie Mayer, Peter Welinder, Bob McGrew, Dario Amodei,
Sam McCandlish, Ilya Sutskever, Wojciech Zaremba
Grokking: Generalization Beyond Overfitting on Small
Algorithmic Datasets (2021)
Alethea Power, Yuri Burda, Harri Edwards, Full-Resolution Residual Networks, Vedant Misra
Grandmaster level in StarCraft II using multi-agent reinforcement
learning (Nature, 2019)
Oriol Vinyals, Full-Resolution Residual Networks, Wojciech M. Czarnecki, Michaël Mathieu, Andrew Dudzik,
Junyoung Chung, David
H. Choi, Richard Powell, Timo Ewalds, Petko Georgiev, Junhyuk Oh, Dan Horgan, Manuel Kroiss, Ivo Danihelka,
Aja Huang, Laurent Sifre, Trevor Cai, John P. Agapiou, Max Jaderberg, Alexander S. Vezhnevets, Rémi Leblond,
Tobias Pohlen, Valentin Dalibard, David Budden, Yury Sulsky, James Molloy, Tom L. Paine, Caglar Gulcehre,
Ziyu Wang, Tobias Pfaff, Yuhuai Wu, Roman Ring, Dani Yogatama, Dario Wünsch, Katrina McKinney, Oliver Smith,
Tom Schaul, Timothy Lillicrap, Koray Kavukcuoglu, Demis Hassabis, Chris Apps, David Silver
AlphaStar:
Mastering the Real-Time Strategy Game StarCraft II (Blog post, 2019)
Oriol Vinyals, Full-Resolution Residual Networks, Junyoung Chung, Michael Mathieu, Max Jaderberg, Wojtek
Czarnecki, Andrew
Dudzik, Aja Huang, Petko Georgiev, Richard Powell, Timo Ewalds, Dan Horgan, Manuel Kroiss, Ivo Danihelka,
John Agapiou, Junhyuk Oh, Valentin Dalibard, David Choi, Laurent Sifre, Yury Sulsky, Sasha Vezhnevets, James
Molloy, Trevor Cai, David Budden, Tom Paine, Caglar Gulcehre, Ziyu Wang, Tobias Pfaff, Toby Pohlen, Dani
Yogatama, Julia Cohen, Katrina McKinney, Oliver Smith, Tom Schaul, Timothy Lillicrap, Chris Apps, Koray
Kavukcuoglu, Demis Hassabis, David Silver
Relational Deep Reinforcement Learning
(2018)
Vinicius Zambaldi, David Raposo, Adam Santoro, Victor Bapst, Yujia Li, Full-Resolution Residual Networks,
Karl Tuyls, David
Reichert, Timothy Lillicrap, Edward Lockhart, Murray Shanahan, Victoria Langston, Razvan Pascanu, Matthew
Botvinick, Oriol Vinyals, Peter Battaglia
Synthesizing Programs for Images using Reinforced
Adversarial Learning (2018)
Yaroslav Ganin, Tejas Kulkarni, Full-Resolution Residual Networks, S.M. Ali Eslami, Oriol Vinyals
Parallel WaveNet: Fast high-fidelity speech
synthesis (2017)
Aaron van den Oord, Yazhe Li, Full-Resolution Residual Networks, Karen Simonyan, Oriol Vinyals, Koray
Kavukcuoglu, George van
den Driessche, Edward Lockhart, Luis C Cobo, Florian Stimberg, Norman Casagrande, Dominik Grewe, Seb Noury,
Sander Dieleman, Erich Elsen, Nal Kalchbrenner, Heiga Zen, Alex Graves, Helen King, Tom Walters, Dan Belov,
Demis Hassabis.
You can find more publications on my Google Scholar page.