ΑΙhub.org
 

European parliament approves draft EU AI act

by
16 June 2023



share this:
EU flag

An important milestone in the process of EU AI legislation was taken on 14 June when the European parliament voted in favour of adopting the proposed AI act (with 499 votes in favour, 28 against and 93 abstentions). The next step will involve talks with EU member states on the final form of the law. The aim is to reach an agreement by the end of this year.

At the core of the proposed act is a risk-based approach, which establishes obligations for providers and those deploying AI systems depending on the level of risk posed.

AI systems deemed to present an “unacceptable risk” would be completely prohibited. In the draft act, this includes “real-time” biometric identification systems (when deployed in publicly accessible spaces), systems that deploy harmful manipulative “subliminal techniques”, systems that exploit specific vulnerable groups, and systems used by public authorities, or on their behalf, for social scoring purposes.

Systems classified as “high risk” would be subject to new regulations including registration of these systems by the providers in an EU-wide database before releasing to the market, and the necessity to comply with a range of requirements including those relating to risk management, testing, technical robustness, data training and data governance, transparency, human oversight, and cybersecurity. Such high-risk applications will include AI systems that pose significant harm to people’s health, safety, fundamental rights or the environment.

AI systems presenting “limited risk” would be subject to a limited set of transparency obligations. All other AI systems presenting only low or minimal risk could be developed and used in the EU without conforming to any additional legal obligations.

On the subject of generative AI, systems based on such models, like ChatGPT, would have to comply with transparency requirements (disclosing that the content was AI-generated, also helping distinguish deep-fake images from real ones) and ensure safeguards against generating illegal content. Detailed summaries of the copyrighted data used for their training would also have to be made publicly available.

You can read more details about the proposed AI act in this document.



tags:


Lucy Smith , Managing Editor for AIhub.
Lucy Smith , Managing Editor for AIhub.




            AIhub is supported by:


Related posts :



The Machine Ethics podcast: New forms of story telling with Guy Gadney

In this episode, Ben chats to Guy Gadney about new forms of story telling, placing people inside a story, natural language in games, and more.
04 March 2024, by

#AAAI2024 invited talk: Milind Tambe – using ML for social good

Winner of the 2024 AAAI Award for Artificial Intelligence for the Benefit of Humanity, Milind spoke about recent projects.
01 March 2024, by

AIhub monthly digest: February 2024 – causal relations in text, applied reinforcement learning, and AAAI 2024

Welcome to our monthly digest, where you can catch up with AI research, events and news from the month past.
29 February 2024, by

#AAAI2024 in tweets: part two

Find out what the conference participants got up to during the second half of the event.
28 February 2024, by

Unlocking the potential of entity-centric knowledge graphs: transforming healthcare and beyond

The concept of entity-centric knowledge graphs holds promise in reshaping how we organize, access, and leverage data.
27 February 2024, by and

Congratulations to the #AAAI2024 outstanding paper winners

The winners of the outstanding papers were announced at the conference during the opening ceremony.
26 February 2024, by





©2024 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association