A scalable laplace approximation for neural networks H Ritter, A Botev, D Barber 6th international conference on learning representations, ICLR 2018 …, 2018 | 406 | 2018 |
Online structured laplace approximations for overcoming catastrophic forgetting H Ritter, A Botev, D Barber Advances in Neural Information Processing Systems 31, 2018 | 293 | 2018 |
Practical Gauss-Newton optimisation for deep learning A Botev, H Ritter, D Barber International Conference on Machine Learning, 557-565, 2017 | 223 | 2017 |
Hamiltonian generative networks P Toth, DJ Rezende, A Jaegle, S Racanière, A Botev, I Higgins arXiv preprint arXiv:1909.13789, 2019 | 210 | 2019 |
Nesterov's accelerated gradient and momentum as approximations to regularised update descent A Botev, G Lever, D Barber 2017 International joint conference on neural networks (IJCNN), 1899-1903, 2017 | 181 | 2017 |
Better, faster fermionic neural networks JS Spencer, D Pfau, A Botev, WMC Foulkes arXiv preprint arXiv:2011.07125, 2020 | 43 | 2020 |
Complementary Sum Sampling for Likelihood Approximation in Large Scale Classification A Botev, B Zheng, D Barber AISTATS 54, 1030-1038, 2017 | 33 | 2017 |
International Conference on Learning Representations P Toth, DJ Rezende, A Jaegle, S Racanière, A Botev, I Higgins OpenReview, 2020 | 31 | 2020 |
Disentangling by subspace diffusion D Pfau, I Higgins, A Botev, S Racanière Advances in Neural Information Processing Systems 33, 17403-17415, 2020 | 31 | 2020 |
Deep learning without shortcuts: Shaping the kernel with tailored rectifiers G Zhang, A Botev, J Martens arXiv preprint arXiv:2203.08120, 2022 | 27 | 2022 |
Which priors matter? benchmarking models for learning latent dynamics A Botev, A Jaegle, P Wirnsberger, D Hennes, I Higgins arXiv preprint arXiv:2111.05458, 2021 | 23 | 2021 |
Deep transformers without shortcuts: Modifying self-attention for faithful signal propagation B He, J Martens, G Zhang, A Botev, A Brock, SL Smith, YW Teh arXiv preprint arXiv:2302.10322, 2023 | 18 | 2023 |
Sampling QCD field configurations with gauge-equivariant flow models R Abbott, MS Albergo, A Botev, D Boyda, K Cranmer, DC Hackett, ... Sissa Medialab, 2022 | 18 | 2022 |
Aspects of scaling and scalability for flow-based sampling of lattice QCD R Abbott, MS Albergo, A Botev, D Boyda, K Cranmer, DC Hackett, ... The European Physical Journal A 59 (11), 257, 2023 | 12 | 2023 |
Normalizing flows for lattice gauge theory in arbitrary space-time dimension R Abbott, MS Albergo, A Botev, D Boyda, K Cranmer, DC Hackett, ... arXiv preprint arXiv:2305.02402, 2023 | 9 | 2023 |
Griffin: Mixing Gated Linear Recurrences with Local Attention for Efficient Language Models S De, SL Smith, A Fernando, A Botev, G Cristian-Muraru, A Gu, R Haroun, ... arXiv preprint arXiv:2402.19427, 2024 | 7 | 2024 |
Symetric: Measuring the quality of learnt hamiltonian dynamics inferred from vision I Higgins, P Wirnsberger, A Jaegle, A Botev Advances in Neural Information Processing Systems 34, 25591-25605, 2021 | 7 | 2021 |
The Gauss-Newton matrix for deep learning models and its applications A Botev UCL (University College London), 2020 | 6 | 2020 |
Dealing with a large number of classes--Likelihood, Discrimination or Ranking? D Barber, A Botev arXiv preprint arXiv:1606.06959, 2016 | 5 | 2016 |
Overdispersed variational autoencoders H Shah, D Barber, A Botev 2017 International Joint Conference on Neural Networks (IJCNN), 1109-1116, 2017 | 2 | 2017 |