ΑΙhub.org
 

Congratulations to the NeurIPS 2021 award winners!


by
02 December 2021



share this:
trophy

The thirty-fifth Conference on Neural Information Processing Systems (NeurIPS 2021) will be held from Monday 6 December to Tuesday 14 December. This week, the awards committees announced the winners of the outstanding paper award, the test of time award and – for the first time – the best paper award in the new datasets and benchmarks track.

Outstanding paper award

Six articles received outstanding paper awards this year. The winners are:

A Universal Law of Robustness via Isoperimetry
Sébastien Bubeck and Mark Sellke
The authors propose a theoretical model to explain why many state-of-the-art deep networks require many more parameters than are necessary to smoothly fit the training data.

On the Expressivity of Markov Reward
David Abel, Will Dabney, Anna Harutyunyan, Mark K. Ho, Michael Littman, Doina Precup and Satinder Singh
This paper provides a clear exposition of when Markov rewards are, or are not, sufficient to enable a system designer to specify a task, in terms of their preference for a particular behaviour, preferences over behaviours, or preferences over state and action sequences.

Deep Reinforcement Learning at the Edge of the Statistical Precipice
Rishabh Agarwal, Max Schwarzer, Pablo Samuel Castro, Aaron Courville and Marc G. Bellemare
This work presents practical approaches to improve the rigor of deep reinforcement learning algorithm comparison: specifically, that the evaluation of new algorithms should provide stratified bootstrap confidence intervals, performance profiles across tasks and runs, and interquartile means.

MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers
Krishna Pillutla, Swabha Swayamdipta, Rowan Zellers, John Thickstun, Sean Welleck, Yejin Choi and Zaid Harchaoui
This paper presents MAUVE, a divergence measure to compare the distribution of model-generated text with the distribution of human-generated text.

Continuized Accelerations of Deterministic and Stochastic Gradient Descents, and of Gossip Algorithms
Mathieu Even, Raphaël Berthier, Francis Bach, Nicolas Flammarion, Pierre Gaillard, Hadrien Hendrikx, Laurent Massoulié and Adrien Taylor
This paper describes a “continuized” version of Nesterov’s accelerated gradient method in which the two separate vector variables evolve jointly in continuous-time, but uses gradient updates that occur at random times determined by a Poisson point process. This new approach leads to a (randomized) discrete-time method.

Moser Flow: Divergence-based Generative Modeling on Manifolds
Noam Rozen, Aditya Grover, Maximilian Nickel and Yaron Lipman
In this work, the authors propose a method for training continuous normalizing flow (CNF) generative models over Riemannian manifolds.

Test of time award

This year, the test of time award goes to a paper from 2010:

Online Learning for Latent Dirichlet Allocation
Matthew Hoffman, David Blei and Francis Bach
This paper introduced a stochastic variational gradient based inference procedure for training Latent Dirichlet Allocation (LDA) models on very large text corpora. The idea has had a significant impact on the ML community. It provided the first stepping stone for general stochastic gradient variational inference procedures for a much broader class of models.

Datasets & benchmarks best paper award

There were two awards in this category:

Reduced, Reused and Recycled: The Life of a Dataset in Machine Learning Research
Bernard Koch, Emily Denton, Alex Hanna, Jacob Gates Foster
This work analyzes thousands of papers, and studies the evolution of dataset use within different machine learning subcommunities. It finds that, in most communities, there is an evolution towards using fewer different datasets over time, and that these datasets come from a handful of elite institutions.

ATOM3D: Tasks on Molecules in Three Dimensions
Authors: Raphael John Lamarre Townshend, Martin Vögele, Patricia Adriana Suriana, Alexander Derry, Alexander Powers, Yianni Laloudakis, Sidhika Balachandar, Bowen Jing, Brandon M. Anderson, Stephan Eismann, Risi Kondor, Russ Altman, Ron O. Dror
The authors introduce a collection of benchmark datasets with 3D representations of small molecules and/or biopolymers for solving a wide range of problems, from single molecular structure prediction to design and engineering tasks.

You can find out more about the awards in this blog post.

More information about the talks, workshops and tutorials can be found here.



tags: ,


Lucy Smith is Senior Managing Editor for AIhub.
Lucy Smith is Senior Managing Editor for AIhub.

            AIhub is supported by:



Subscribe to AIhub newsletter on substack



Related posts :

Top AI ethics and policy issues of 2025 and what to expect in 2026

, and   04 Mar 2026
In the latest issue of AI Matters, a publication of ACM SIGAI, Larry Medsker summarised the year in AI ethics and policy, and looked ahead to 2026.

The greatest risk of AI in higher education isn’t cheating – it’s the erosion of learning itself

  03 Mar 2026
Will AI hollow out the pipeline of students, researchers and faculty that is the basis of today’s universities?

Forthcoming machine learning and AI seminars: March 2026 edition

  02 Mar 2026
A list of free-to-attend AI-related seminars that are scheduled to take place between 2 March and 30 April 2026.
monthly digest

AIhub monthly digest: February 2026 – collective decision making, multi-modal learning, and governing the rise of interactive AI

  27 Feb 2026
Welcome to our monthly digest, where you can catch up with AI research, events and news from the month past.

The Good Robot podcast: the role of designers in AI ethics with Tomasz Hollanek

  26 Feb 2026
In this episode, Tomasz argues that design is central to AI ethics and explores the role designers should play in shaping ethical AI systems.

Reinforcement learning applied to autonomous vehicles: an interview with Oliver Chang

  25 Feb 2026
In the third of our interviews with the 2026 AAAI Doctoral Consortium cohort, we hear from Oliver Chang.

The Machine Ethics podcast: moral agents with Jen Semler

In this episode, Ben and Jen Semler talk about what makes a moral agent, the point of moral agents, philosopher and engineer collaborations, and more.

Extending the reward structure in reinforcement learning: an interview with Tanmay Ambadkar

  23 Feb 2026
Find out more about Tanmay's research on RL frameworks, the latest in our series meeting the AAAI Doctoral Consortium participants.



AIhub is supported by:







Subscribe to AIhub newsletter on substack




 















©2026.02 - Association for the Understanding of Artificial Intelligence