• news-banner

    Expert Insights

EU AI Act: Key provisions now in force

The beginning of February, specifically 2 February 2025, marked the first of many deadlines to come imposed by the European Union’s (EU) Regulation (EU) 2024/1689, commonly known as the EU AI Act (the Act).

The first provisions of the Act to take effect include both a positive and a negative requirement:

  • positive, in that providers and deployers of AI Systems must proactively ensure that individuals, (which include but are not limited to their employees), have sufficient AI literacy to operate AI Systems; and
  • negative, in so far as those looking to leverage AI must not use AI Systems for certain AI practices deemed to pose unacceptable risk (Prohibited AI Systems).

Published two days later, on 4 February 2025, the EU Commission Guidelines, C(2025) 884 (the Guidelines), provide much-needed perimeter guidance as firms address the cost and complexity of complying with the Act. The Guidelines provide clarity pertaining to Prohibited AI Systems and touch upon instances where such systems may interact with the AI literacy requirements.

AI literacy requirements

Article 4 of the Act (Article 4) mandates that entities involved in the provision and deployment of AI Systems must take steps to ensure that those operating AI Systems possess a sufficient degree of AI literacy.

The Act defines AI literacy as having the requisite skills, knowledge, and understanding to appreciate the potential benefits and risks inherent in the use of AI, including any potential for harm.

When assessing the necessary level of AI literacy, firms must consider the experience, educational background and training of the relevant staff, as well as the context in which a given AI system will be used. Note that the obligation to ensure AI literacy potentially extends beyond staff (subject to the guidance below).

Article 4 now having come into effect means that AI literacy must be embedded into firms by design and by default across all aspects of a firm’s business model, rather than be limited to a select few such as a firm’s IT team or chief technology officer.

Firms’ compliance must also be comprehensively documented. Firms will need to have implemented robust AI governance policies and CPD AI training programs for all relevant staff.

Firms will do well to consider that merely having in place a generic AI governance policy is highly unlikely to satisfy the AI literacy requirement. As a minimum, any such policy would, in any event, have to be tailored to the firm's specific AI use cases.

While Article 4 does not prescribe specific penalties or fines for non-compliance, regulatory authorities are likely to consider violations of Article 4 when determining the appropriate sanctions for other infringements of an organisation's duties under the Act.

Prohibited AI Systems

Article 5 of the Act (Article 5) affects a ban on Prohibited AI Systems. This is predicated on the idea that certain uses for AI pose unacceptable risk, including:

  • AI Systems that implement social scoring to assess or categorise individuals or groups based on their conduct or actual, assumed, or anticipated traits;
  • AI Systems that use subliminal techniques to manipulate or deceive with a view to altering a person's or group's behaviour by impairing their capacity to make informed choices;
  • AI Systems that take advantage of the vulnerabilities of certain individuals or groups to alter their behaviour;
  • AI Systems that apply profiling or personality trait evaluations to predict criminal conduct;
  • AI Systems that generate or enlarge databases of facial recognition through untargeted scraping of facial images from the internet or from CCTV;
  • AI Systems that decipher someone’s emotions in a workplace or educational setting;
  • Biometric identification systems in real-time that are used in publicly accessible areas for law enforcement purposes; and
  • Biometric classification AI Systems that use biometric data to infer characteristics such as political views, race, trade union membership, beliefs, or sexual orientation.

To contextualise Article 5, these Prohibited AI Systems are deemed to be in conflict with European Union values, the rule of law, and fundamental rights.

Exceptions to certain Prohibited AI Systems are available in limited circumstances – more on this below.

The Act provisions setting out the punishment for breaches of Article 5 are set to come into force on 2 August 2025. Infringements of Article 5 could cost firms either €35 million or 7% of their total global annual turnover for the previous year, whichever is greater. Compounding this will be, of course, the potentially damning reputational damage of being amongst the cohort of non-compliant firms.

Interaction between Articles 4 and 5 of the Act

The Guidelines illustrate instances the nuanced interplay between the Articles detailed above.

For example, in the context of real-time Remote Biometric Identification (RBI) systems (the use of which may be prohibited under Article 5(1)(h) of the Act), the Guidelines confirm that no decision that would adversely affect an individual may be taken solely based on the output of the real-time RBI system. Here, the firm seeking to deploy the real-time RBI system should, in its Fundamental Right Impact Assessment (FRIA), clarify the role of a human agent in verifying and interpreting the output and provide training to operate the system. The Guidelines further prescribe that the individual in charge of human oversight must have “sufficient AI literacy, training and authority to understand how the system functions and when it underperforms or malfunctions.”

Emotional recognition AI Systems

The Guidelines provide the following clarifications in relation to the Article 5(1)(f) prohibition on emotional recognition AI Systems in workplaces or educational settings:

  • Firms looking to use AI Systems to decipher an individual’s emotions in a work or educational context for medical or safety reasons will be able to do so, subject to the other requirements of the Act. However, this exception is caveated; the Guidelines explicitly confine the definition of “safety reasons” within this exception to “the protection of life and health and not to protect other interests, for example property against theft or fraud.”
  • Firms with employees likely to be exposed to specific heightened risks, such as professional pilots or drivers, will benefit from Recital 18 of the AI Act, which excludes physical states such as pain or fatigue from the definition of emotion recognition AI Systems.
  • Even if an emotion recognition system does not fulfil the conditions for it to be subject to the prohibition in Article 5(1)(f) of the Act, such AI Systems will be categorised as high-risk AI Systems (HRAI Systems) under Article 6(2) and Annex III, point (1)(c) of the Act. Firms using emotional recognition AI Systems classified as high risk will be subject to enhanced compliance obligations in respect of, inter alia, data governance, transparency and human oversight.

Subliminal techniques

The Guidelines also provide further detail on the ban on the use of subliminal techniques found in Article 5(1)(a) of the Act. This Article does not represent a blanket ban on the use of all subliminal techniques, with the Guidelines confirming that such techniques are not necessarily prohibited unless all other conditions listed in Article 5(1)(a) are met. The Guidelines provide the following non-exhaustive examples of subliminal techniques:

Subliminal Visual Messaging

AI can subtly flash images or text during videos – such messaging is too quick for individuals to consciously register it but is still capable of influencing attitudes and behaviours.

Subliminal Audio Messaging

AI might introduce sounds or spoken words quietly or amidst other noises, affecting listeners subconsciously.

Subliminal Tactile Feedback

AI could induce barely noticeable physical sensations, influencing emotions or actions without conscious perception.

Undetectable Cues

AI could present visual or auditory stimuli, such as ultra-fast flashes or ultra-low volumes, that escape normal human detection.

Embedded Imagery

AI can embed images within visuals that elude conscious recognition but can still subconsciously impact behaviour.

Misdirection

AI may focus user attention on certain elements in order to obscure others, leveraging cognitive biases and vulnerabilities impacting individuals’ attention; and

Temporal Manipulation

AI may modify the perceived flow of time during interactions, potentially affecting user behaviour, fostering either impatience or reliance.

Whether these techniques will be subject to the prohibition will be largely fact dependent. The potential for perimeter uncertainty here is exacerbated by the lack of a definition of AI Systems deploying ‘purposefully manipulative techniques.’ In addressing this point, a distinction is drawn in the Guidelines between non-AI-enabled AI Systems which may seek to manipulate or influence human behaviour, and AI Systems which are able to adapt and respond well to individuals’ individual circumstances and vulnerabilities; the rationale being that it is this adaptability that heightens AI’s potential for harm in this context.

Article 5(1)(a) is particularly pertinent to advertisers. The Guidelines give the example of a “chatbot that is designed… to use subliminal messaging techniques, such as flashing brief visual cues and embedding inaudible auditory signals or to exploit emotional dependency or specific vulnerabilities of users in advertisements.” Here, the Guidelines confirm that if the other conditions in Article 5(1)(a) of the Act were fulfilled, in particular regarding the significant harm, this system would be likely to fall within the scope of the prohibition.

Firms seeking to create efficiencies via the use of AI chatbots as part of, for example, the sales pipeline, ought to familiarise themselves with the Guidelines, particularly as there is no one-size-fits all assessment prescribed. This emphasises the need for a case-by-case approach throughout its examples.

For further information, please see our Beginners Guide’ to the Act here.

Let’s talk

If you would like to understand how your firm may be affected by the EU AI Act, please contact Partners, Racheal Muldoon and Mark Bailey.  

Our thinking

  • IBA Annual Conference 2025

    Simon Ridpath

    Events

  • Next Gen Rural Professionals Drinks Reception

    Events

  • Key aspects of the FCA’s PISCES Sourcebook

    Jodie Dennis

    Insights

  • Mike Barrington and Mary Perham write for Tax Adviser on what the proposed changes to business property relief mean for investors and entrepreneurs, and for their businesses

    Mike Barrington

    In the Press

  • Bloomberg quotes Catrin Harrison on the recent exodus of non-doms from the UK

    Catrin Harrison

    In the Press

  • Trusts and Matrimonial Disputes in England

    Tom Watts

    Insights

  • The Financial Times and Daily Mail quote Emma Humphreys on the impact of the UK Government's Spending Review on housebuilding targets

    Emma Humphreys

    In the Press

  • Alumni Drinks Reception

    Events

  • Consultation on Private International Law and Digital Assets Law Commission Proposes Landmark Reforms

    Racheal Muldoon

    Insights

  • Navigating International M&A Disputes: Insights and Strategies for 2025

    Stephen Burns

    Quick Reads

  • Bridging Differences: The Role of Mediation in Resolving Cross-Border Trust Disputes

    Tamasin Perkins

    Insights

  • Rachel Warren writes for Solicitors Journal on the new failure prevent fraud offence

    Rachel Warren

    In the Press

  • MoneyWeek quotes Mary Perham on whether business property relief can be claimed on a furnished holiday let

    Mary Perham

    In the Press

  • Anti-greenwashing in the UK, EU and the US: the outlook for 2025 and best practice guidance

    Caroline Greenwell

    Insights

  • Landmark rulings from the Italian Revenue Agency on income tax exemption on gains from Italian shares held in trust

    Nicola Saccardo

    Quick Reads

  • Sowing doubt: slashing green farm funding is a risk we can't afford

    Maddie Dunn

    Quick Reads

  • Can a contractor adjudicate to recover outstanding retention monies from the employer’s assignee?

    Kate Knox

    Insights

  • HS2 - still no sign of a train leaving the station

    Richard Flenley

    Quick Reads

  • Tamasin Perkins writes for IFA Magazine on what financial advisers need to know about The Law Commission’s recent report: Modernising Wills Law

    Tamasin Perkins

    In the Press

  • London International Disputes Week: Trusts hurt: the fraud lawyer, the trust, and the avenues of attack (and defence)

    Tamasin Perkins

    Events

Back to top