ΑΙhub.org
 

European parliament approves draft EU AI act

by
16 June 2023



share this:
EU flag

An important milestone in the process of EU AI legislation was taken on 14 June when the European parliament voted in favour of adopting the proposed AI act (with 499 votes in favour, 28 against and 93 abstentions). The next step will involve talks with EU member states on the final form of the law. The aim is to reach an agreement by the end of this year.

At the core of the proposed act is a risk-based approach, which establishes obligations for providers and those deploying AI systems depending on the level of risk posed.

AI systems deemed to present an “unacceptable risk” would be completely prohibited. In the draft act, this includes “real-time” biometric identification systems (when deployed in publicly accessible spaces), systems that deploy harmful manipulative “subliminal techniques”, systems that exploit specific vulnerable groups, and systems used by public authorities, or on their behalf, for social scoring purposes.

Systems classified as “high risk” would be subject to new regulations including registration of these systems by the providers in an EU-wide database before releasing to the market, and the necessity to comply with a range of requirements including those relating to risk management, testing, technical robustness, data training and data governance, transparency, human oversight, and cybersecurity. Such high-risk applications will include AI systems that pose significant harm to people’s health, safety, fundamental rights or the environment.

AI systems presenting “limited risk” would be subject to a limited set of transparency obligations. All other AI systems presenting only low or minimal risk could be developed and used in the EU without conforming to any additional legal obligations.

On the subject of generative AI, systems based on such models, like ChatGPT, would have to comply with transparency requirements (disclosing that the content was AI-generated, also helping distinguish deep-fake images from real ones) and ensure safeguards against generating illegal content. Detailed summaries of the copyrighted data used for their training would also have to be made publicly available.

You can read more details about the proposed AI act in this document.



tags:


Lucy Smith , Managing Editor for AIhub.
Lucy Smith , Managing Editor for AIhub.




            AIhub is supported by:


Related posts :



The Machine Ethics podcast: Socio-technical systems with Lisa Talia Moretti

In this episode Ben chats to Lisa about data and AI literacy, data governance, ethical frameworks, and more.
01 November 2024, by

Building trust in AI: Transparent models for better decisions

AI is becoming a part of our daily lives, from approving loans to diagnosing diseases. But if we can't understand the decisions the models output, how can we trust them?
31 October 2024, by

Congratulations to the #ECAI2024 outstanding paper award winners

Find out which articles won the ECAI and PAIS 2024 awards.
30 October 2024, by

AIhub monthly digest: October 2024 – Nobel Prizes, the AI Song Contest, and towards safe and reliable AI agents

Welcome to our monthly digest, where you can catch up with AI research, events and news from the month past.
29 October 2024, by

Congratulations to the winners of the #AIES2024 best paper awards

Find out who won the best paper award and the best student paper award.
28 October 2024, by

Tweet round up from #ECAI2024: part 2

We took a look at what participants got up to over the second half of the European Conference on Artificial Intelligence.
25 October 2024, by




AIhub is supported by:






©2024 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association