ΑΙhub.org
 

European parliament approves draft EU AI act


by
16 June 2023



share this:
EU flag

An important milestone in the process of EU AI legislation was taken on 14 June when the European parliament voted in favour of adopting the proposed AI act (with 499 votes in favour, 28 against and 93 abstentions). The next step will involve talks with EU member states on the final form of the law. The aim is to reach an agreement by the end of this year.

At the core of the proposed act is a risk-based approach, which establishes obligations for providers and those deploying AI systems depending on the level of risk posed.

AI systems deemed to present an “unacceptable risk” would be completely prohibited. In the draft act, this includes “real-time” biometric identification systems (when deployed in publicly accessible spaces), systems that deploy harmful manipulative “subliminal techniques”, systems that exploit specific vulnerable groups, and systems used by public authorities, or on their behalf, for social scoring purposes.

Systems classified as “high risk” would be subject to new regulations including registration of these systems by the providers in an EU-wide database before releasing to the market, and the necessity to comply with a range of requirements including those relating to risk management, testing, technical robustness, data training and data governance, transparency, human oversight, and cybersecurity. Such high-risk applications will include AI systems that pose significant harm to people’s health, safety, fundamental rights or the environment.

AI systems presenting “limited risk” would be subject to a limited set of transparency obligations. All other AI systems presenting only low or minimal risk could be developed and used in the EU without conforming to any additional legal obligations.

On the subject of generative AI, systems based on such models, like ChatGPT, would have to comply with transparency requirements (disclosing that the content was AI-generated, also helping distinguish deep-fake images from real ones) and ensure safeguards against generating illegal content. Detailed summaries of the copyrighted data used for their training would also have to be made publicly available.

You can read more details about the proposed AI act in this document.



tags:


Lucy Smith is Senior Managing Editor for AIhub.
Lucy Smith is Senior Managing Editor for AIhub.




            AIhub is supported by:



Related posts :



EU proposal to delay parts of its AI Act signal a policy shift that prioritises big tech over fairness

  27 Nov 2025
The EC has proposed delaying parts of the act until 2027 following intense pressure from tech companies and the Trump administration.

Better images of AI on book covers

  25 Nov 2025
We share insights from Chrissi Nerantzi on the decisions behind the cover of the open-sourced book ‘Learning with AI’, and reflect on the significance of book covers.

What is AI poisoning? A computer scientist explains

  24 Nov 2025
Poisoning is a growing problem in the world of AI – in particular, for large language models.

New AI technique sounding out audio deepfakes

  21 Nov 2025
Researchers discover a smarter way to detect audio deepfakes that is more accurate and adaptable to keep pace with evolving threats.

Learning robust controllers that work across many partially observable environments

  20 Nov 2025
Exploring designing controllers that perform reliably even when the environment may not be precisely known.

ACM SIGAI Autonomous Agents Award 2026 open for nominations

  19 Nov 2025
Nominations are solicited for the 2026 ACM SIGAI Autonomous Agents Research Award.

Interview with Mario Mirabile: trust in multi-agent systems

  18 Nov 2025
We meet ECAI Doctoral Consortium participant, Mario, to find out more about his research.



 

AIhub is supported by:






 












©2025.05 - Association for the Understanding of Artificial Intelligence