ΑΙhub.org
 

European parliament approves draft EU AI act


by
16 June 2023



share this:
EU flag

An important milestone in the process of EU AI legislation was taken on 14 June when the European parliament voted in favour of adopting the proposed AI act (with 499 votes in favour, 28 against and 93 abstentions). The next step will involve talks with EU member states on the final form of the law. The aim is to reach an agreement by the end of this year.

At the core of the proposed act is a risk-based approach, which establishes obligations for providers and those deploying AI systems depending on the level of risk posed.

AI systems deemed to present an “unacceptable risk” would be completely prohibited. In the draft act, this includes “real-time” biometric identification systems (when deployed in publicly accessible spaces), systems that deploy harmful manipulative “subliminal techniques”, systems that exploit specific vulnerable groups, and systems used by public authorities, or on their behalf, for social scoring purposes.

Systems classified as “high risk” would be subject to new regulations including registration of these systems by the providers in an EU-wide database before releasing to the market, and the necessity to comply with a range of requirements including those relating to risk management, testing, technical robustness, data training and data governance, transparency, human oversight, and cybersecurity. Such high-risk applications will include AI systems that pose significant harm to people’s health, safety, fundamental rights or the environment.

AI systems presenting “limited risk” would be subject to a limited set of transparency obligations. All other AI systems presenting only low or minimal risk could be developed and used in the EU without conforming to any additional legal obligations.

On the subject of generative AI, systems based on such models, like ChatGPT, would have to comply with transparency requirements (disclosing that the content was AI-generated, also helping distinguish deep-fake images from real ones) and ensure safeguards against generating illegal content. Detailed summaries of the copyrighted data used for their training would also have to be made publicly available.

You can read more details about the proposed AI act in this document.



tags:


Lucy Smith is Senior Managing Editor for AIhub.
Lucy Smith is Senior Managing Editor for AIhub.




            AIhub is supported by:



Related posts :



#AAAI2026 social media round up: part 1

  23 Jan 2026
Find out what participants have been getting up to during the first few of days at the conference

Congratulations to the #AAAI2026 outstanding paper award winners

  22 Jan 2026
Find out who has won these prestigious awards at AAAI this year.

3 Questions: How AI could optimize the power grid

  21 Jan 2026
While the growing energy demands of AI are worrying, some techniques can also help make power grids cleaner and more efficient.

Interview with Xiang Fang: Multi-modal learning and embodied intelligence

  20 Jan 2026
In the first of our new series of interviews featuring the AAAI Doctoral Consortium participants, we hear from Xiang Fang.

An introduction to science communication at #AAAI2026

  19 Jan 2026
Find out more about our session on Wednesday 21 January.

Guarding Europe’s hidden lifelines: how AI could protect subsea infrastructure

  15 Jan 2026
EU-funded researchers are developing AI-powered surveillance tools to protect the vast network of subsea cables and pipelines that keep the continent’s energy and data flowing.

What’s coming up at #AAAI2026?

  14 Jan 2026
Find out what's on the programme at the annual AAAI Conference on Artificial Intelligence.



 

AIhub is supported by:






 












©2025.05 - Association for the Understanding of Artificial Intelligence