ΑΙhub.org
 

#NeurIPS2020 invited talks round-up: part one


by
11 December 2020



share this:
NeurIPS logo

There were seven interesting and varied invited talks at NeurIPS this year. Here, we summarise the first three, which were given by Charles Isbell (Georgia Tech), Jeff Shamma (King Abdullah University of Science and Technology) and Shafi Goldwasser (UC Berkeley, MIT and Weizmann Institute of Science).

Charles Isbell: You can’t escape hyperparameters and latent variables: machine learning as a software engineering enterprise

The invited talks kicked off in style with a presentation from Charles Isbell. He had posted a teaser on Twitter indicating that he was trying something new with the format, and it certainly did not disappoint. The talk received rave reviews during both the live chat channel and afterwards on social media.

Machine learning has reached the point where it is pervasive in our lives and, like other successful technological fields, must take responsibility for avoiding the harms associated with what it is producing.

Charles’ Christmas Carol-themed talk made full use of the virtual format and saw him visit Michael Littmann, who gave a consummate performance in the role of Scrooge (a blinkered machine learning theorist), to discuss algorithmic bias and how researchers can take steps to identify and combat this bias in their work. The ghosts of machine learning’s past, present and future came in the form of interviews with many researchers working across the discipline. Their insights included approaches that might help us to develop more robust machine-learning systems.

summary slide from Isbell talk at NeurIPS
The summary slide from Charles Isbell’s talk

Watch the talk here.


Jeff Shamma: Feedback control perspectives on learning

Jeff started with a definition of feedback control: real-time decision making in dynamic and uncertain environments. Feedback systems are a never-ending loop, with the steps of: sensing what is happening, deciding what to do, and acting. Areas where feedback control has been applied include: manufacturing, energy, biomedical, transportation, logistics, and communication. As feedback control is an enabling technology it does not perhaps receive the attention warranted for such ubiquity.

Feedback control is, of course, present in learning systems. For example, in reinforcement learning, an agent is in feedback with its environment.

During his talk, Jeff presented three benefits of feedback: 1) it can be used to stabilise and shape behaviour, 2) it can provide robustness to variation, 3) it enables tracking of command signals. These concepts were related to specific research problems in evolutionary game theory, no-regret learning, and multi-agent learning.

Summary slide from Shamma talk at NeurIPS
Concluding slide from Jeff Shamma’s talk

Watch the talk here.


Shafi Goldwasser: Robustness, verification, privacy: addressing machine learning adversaries

To start her talk, Shafi noted that her background is in cryptography. This field has had a big impact on the world, from electronic commerce to cryptocurrencies, from cloud computing to quantum computing. Experience in this area has allowed her to approach machine learning problems from a cryptographic standpoint.

Shafi’s talk focussed on three recent works, covering the topics of privacy, verification and robustness. For each case, cryptography inspired models were detailed; an important aim in each instance being to address the challenges presented by adversaries.

The three works presented in the talk were based on these topics:
1) Verifiability of machine learning models
2) Privacy during training: from prototype to large scale data
3) Robustness – accurate predictions when test distribution deviates from training distribution

Shafi Goldwasser talk intro slide
The three studies presented by Shafi Goldwasser in her talk

Watch the talk here.




tags: ,


Lucy Smith is Senior Managing Editor for AIhub.
Lucy Smith is Senior Managing Editor for AIhub.




            AIhub is supported by:


Related posts :



Interview with Eden Hartman: Investigating social choice problems

  24 Apr 2025
Find out more about research presented at AAAI 2025.

The Machine Ethics podcast: Co-design with Pinar Guvenc

This episode, Ben chats to Pinar Guvenc about co-design, whether AI ready for society and society is ready for AI, what design is, co-creation with AI as a stakeholder, bias in design, small language models, and more.

Why AI can’t take over creative writing

  22 Apr 2025
A large language model tries to generate what a random person who had produced the previous text would produce.

Interview with Amina Mević: Machine learning applied to semiconductor manufacturing

  17 Apr 2025
Find out how Amina is using machine learning to develop an explainable multi-output virtual metrology system.

Images of AI – between fiction and function

“The currently pervasive images of AI make us look somewhere, at the cost of somewhere else.”

Grace Wahba awarded the 2025 International Prize in Statistics

  16 Apr 2025
Her contributions laid the foundation for modern statistical techniques that power machine learning algorithms such as gradient boosting and neural networks.

Repurposing protein folding models for generation with latent diffusion

  14 Apr 2025
The awarding of the 2024 Nobel Prize to AlphaFold2 marks an important moment of recognition for the of AI role in biology. What comes next after protein folding?




AIhub is supported by:






©2024 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association