ΑΙhub.org
 

#AAAI2023 workshops round-up 2: health intelligence and privacy-preserving AI


by
21 March 2023



share this:
AAAI banner, with Washington DC view and AAAI23 text

As part of the 37th AAAI Conference on Artificial Intelligence (AAAI2023), 32 different workshops were held, covering a wide range of different AI topics. We continue our round-up of these workshops with summaries from the organisers of two of the workshops, who tell us their key takeaways from their respective events.


Health Intelligence (W3PHIAI)

Organisers: Arash Shaban-Nejad, Martin Michalowski, Simone Bianco, Szymon Wilk, David L. Buckeridge, John S. Brownstein.

This workshop included contributions spanning theory, methods and systems, for application to web-based healthcare, with a focus on applications in population and personalized health. The main takeaways from the workshop were as follows:

  • Explainability in AI-based health applications is gaining traction. While advances have been made methodologically to improve diagnosis, prognosis, and treatment, there is a need to explain decisions being made in the context of care. Several interesting statistical, textual, and visual methods were presented with compelling applications to health.
  • The identification of biomarkers to describe, define, and predict biological age is a very active topic of research. Aging affects organisms differently, and chronological age does not always coincide with biological age. As part of the workshop’s hackathon dedicated to this problem, very novel ideas were presented on how to use multimodal health data to predict biological age.
  • The set of papers presented at the workshop spanned a range of topics. Past workshops were more heavily focused on deep learning methods. This iteration of the workshop saw a more diverse set of research presented on topics including natural language processing, explainability, classification, fairness/ethics, amongst others.

Privacy Preserving AI (PPAI)

Organisers: Ferdinando Fioretto, Catuscia Palamidessi, Pascal Van Hentenryck.

This workshop focussed on both the theoretical and practical challenges related to the design of privacy-preserving AI systems, including multidisciplinary components, such as policy, legal issues, and societal impact of privacy in AI. The three main takeaways from the event were:

  • Privacy and policy in AI applications
    During the workshop, the invited talk by Kobbi Nissim, a renowned privacy expert, provided valuable insights into the compliance of machine learning (ML) systems with existing privacy legal requirements. The discussion revolved around the need for collaborations between AI practitioners, privacy experts, and policy makers. Participants also debated the effectiveness of current privacy-preserving techniques in meeting legal requirements and the potential consequences of non-compliance.
  • Differential privacy for data release tasks
    The workshop also included a comprehensive discussion on the role of differential privacy in data release tasks. The value of synthetic dataset generators was a major topic, as these generators can create datasets that maintain the statistical properties of the original data while ensuring privacy. Further, a need of comparing traditional data anonymization techniques, such as k-anonymity and cell suppression with differentially private algorithms.
  • Challenges in auditing differential privacy ML models
    Another key discussion point during the workshop was the challenges faced when auditing differential privacy ML models. The discussion emphasized the difficulties in verifying the privacy guarantees of differential privacy models and the potential trade-offs between privacy and model performance. A discussion around the need for standardized metrics, tools, and methodologies to assess the privacy-preserving properties of AI models was also underscored.

You can read the first round up in our series of AAAI workshop summaries here.



tags: ,


AIhub is dedicated to free high-quality information about AI.
AIhub is dedicated to free high-quality information about AI.




            AIhub is supported by:


Related posts :



monthly digest

AIhub monthly digest: January 2025 – artists’ perspectives on generative AI, biomedical knowledge graphs, and ML for studying greenhouse gas e

  29 Jan 2025
Welcome to our monthly digest, where you can catch up with AI research, events and news from the month past.

Public competition for better images of AI – winners announced!

  28 Jan 2025
See the winning images from the Better Images of AI and Cambridge Diversity Fund competition.

Translating fiction: how AI could assist humans in expanding access to global literature and culture

  27 Jan 2025
Dutch publishing house Veen Bosch & Keuning (VBK) has confirmed plans to experiment using AI to translate fiction.

Interview with Yuki Mitsufuji: Improving AI image generation

  23 Jan 2025
Find out about two pieces of research tackling different aspects of image generation.

The Good Robot podcast: Using feminist chatbots to fight trolls with Sarah Ciston

  22 Jan 2025
Eleanor and Kerry chat to Sarah Ciston about the difficult labor of content moderation, chatbots to combat trolls, and more.

An open-source training framework to advance multimodal AI

  22 Jan 2025
EPFL researchers have developed 4M, a next-generation, framework for training versatile and scalable multimodal foundation models.

Optimizing LLM test-time compute involves solving a meta-RL problem

  20 Jan 2025
By altering the LLM training objective, we can reuse existing data along with more test-time compute to train models to do better.

Generating a biomedical knowledge graph question answering dataset

  17 Jan 2025
Introducing PrimeKGQA - a scalable approach to dataset generation, harnessing the power of large language models.




AIhub is supported by:






©2024 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association