Securiti launches Gencore AI, a holistic solution to build Safe Enterprise AI with proprietary data - easily

View

Article 5: Prohibited Artificial Intelligence Practices | EU AI Act

Contributors

Anas Baig

Product Marketing Manager at Securiti

Syed Tatheer Kazmi

Associate Data Privacy Analyst, Securiti

CIPP/Europe

Article 5 of the AI Act contains detailed information on various activities and practices that are expressly prohibited.

The AI Act prohibits the following practices:

Subliminal Techniques

No AI system or model shall be made available on the market that uses subliminal techniques to influence the users’ consciousness. This extends to the use of possibly manipulative and deceptive techniques that may result in the distortion of a person’s ability to make an informed decision.

Exploitation of a Vulnerability

No AI system or model shall be made available on the market that exploits any vulnerabilities of a natural person or a specific group of persons, including their age, disability, social/economic situation, or association with a group in a manner that may cause harm to that person or someone else.

Social Evaluation

No AI system or model shall be made available on the market whose purpose is to evaluate or classify a natural person or group of persons based on their social behavior, inferred to predicted personality characteristics, or a social score that may lead to:

  • Unfavorable treatment for the natural persons in a social context that is unrelated to the context for which the data was initially generated or collected;
  • Unfavorable treatment of natural persons or groups of persons that is unjustified or disproportionate to their social behavior.

Risk Assessment

No AI system or model shall be made available on the market whose purpose is to make risk assessments of natural persons related to the likelihood of that person committing a criminal offense based solely on the profiling of that purpose. However, this prohibition does not apply to AI systems used to support human assessments related to the involvement of a person in a criminal activity, where such assessments rely on factual evidence directly associated with criminal conduct.

Facial Recognition

No AI system or model shall be made available on the market whose purpose is to create and expand facial recognition databases through the untargeted scraping of facial images using the internet or CCTV footage.

Employee Emotions

No AI system or model shall be made available on the market whose purpose is to assess the emotions of natural persons in a workplace or educational institute. However, this prohibition does not apply to AI systems where it is used for medical or safety reasons.

Biometric Categorization

No AI system or model shall be made available on the market whose purpose is to use biometric categorization systems to categorize natural persons based on their biometric data to deduce their race, political opinions, trade union membership, religious or philosophical beliefs, sex life, or sexual orientation. However, this prohibition does not apply to the labeling and filtration of lawfully acquired biometric datasets by law enforcement agencies (LEAs).

Real-Time Remote Biometric Identification by LEA

No AI system or model shall be made available on the market whose purpose is to put real-time remote biometric identification systems in use for law enforcement purposes unless it is necessary to:

  • Conduct a targeted search for specific victims of abduction, trafficking in human beings, or sexual exploitation of human beings, as well as searching for missing persons;
  • Prevent a specific and substantial threat to the life and safety of a natural person from an imminent terrorist attack;
  • Identify a person suspected to have committed a criminal offense;
  • Conduct a criminal investigation;
  • Execute a criminal penalty for a natural person found to have committed a criminal offense.

LEAs using real-time remote biometric identification in a publicly accessible space must ensure the use is in accordance with the aforementioned purposes and take into account the following considerations:

  • The nature of the possible usage, as well as the seriousness, probability, and scale of harm that may occur if the AI system is not used;
  • The consequences of the AI system’s usage to the rights and freedoms of natural persons involved, as well as the seriousness and scale of these consequences.

Furthermore, the use of real-time remote biometric identification in publicly accessible spaces will only be authorized if the LEA concerned conducts a fundamental rights impact assessment as required under the AI Act while also ensuring such a system is appropriately registered in the EU database. In cases of extreme emergency, such systems may be used without registration, provided the LEA completes the registration process without undue delay.

The use of real-time remote biometric identification in publicly accessible space will be subject to prior authorization to be granted by a judicial authority or relevant independent administrative authority whose decision is binding on the Member State in which the use is to take place. In cases of duly justified emergency, such systems may be used without the necessary authorization provided the LEA requests and gain the authorization without undue delay within 24 hours.

If the request is rejected, its use must be stopped immediately, and any collected data must be disposed of in addition to the generated results and outputs.

Each use of real-time remote biometric identification in a publicly accessible space should l be communicated to a relevant market surveillance authority and national data protection authority in accordance with the national rules. The notification must, at least, encompass the details outlined in Article 5(6). Such communication must not contain any sensitive operational data.

Join Our Newsletter

Get all the latest information, law updates and more delivered to your inbox


Share


More Stories that May Interest You

Videos

View More

Mitigation OWASP Top 10 for LLM Applications 2025

Generative AI (GenAI) has transformed how enterprises operate, scale, and grow. There’s an AI application for every purpose, from increasing employee productivity to streamlining...

View More

DSPM vs. CSPM – What’s the Difference?

While the cloud has offered the world immense growth opportunities, it has also introduced unprecedented challenges and risks. Solutions like Cloud Security Posture Management...

View More

Top 6 DSPM Use Cases

With the advent of Generative AI (GenAI), data has become more dynamic. New data is generated faster than ever, transmitted to various systems, applications,...

View More

Colorado Privacy Act (CPA)

What is the Colorado Privacy Act? The CPA is a comprehensive privacy law signed on July 7, 2021. It established new standards for personal...

View More

Securiti for Copilot in SaaS

Accelerate Copilot Adoption Securely & Confidently Organizations are eager to adopt Microsoft 365 Copilot for increased productivity and efficiency. However, security concerns like data...

View More

Top 10 Considerations for Safely Using Unstructured Data with GenAI

A staggering 90% of an organization's data is unstructured. This data is rapidly being used to fuel GenAI applications like chatbots and AI search....

View More

Gencore AI: Building Safe, Enterprise-grade AI Systems in Minutes

As enterprises adopt generative AI, data and AI teams face numerous hurdles: securely connecting unstructured and structured data sources, maintaining proper controls and governance,...

View More

Navigating CPRA: Key Insights for Businesses

What is CPRA? The California Privacy Rights Act (CPRA) is California's state legislation aimed at protecting residents' digital privacy. It became effective on January...

View More

Navigating the Shift: Transitioning to PCI DSS v4.0

What is PCI DSS? PCI DSS (Payment Card Industry Data Security Standard) is a set of security standards to ensure safe processing, storage, and...

View More

Securing Data+AI : Playbook for Trust, Risk, and Security Management (TRiSM)

AI's growing security risks have 48% of global CISOs alarmed. Join this keynote to learn about a practical playbook for enabling AI Trust, Risk,...

Spotlight Talks

Spotlight 46:02

Building Safe Enterprise AI: A Practical Roadmap

Watch Now View
Spotlight 13:32

Ensuring Solid Governance Is Like Squeezing Jello

Watch Now View
Spotlight 40:46

Securing Embedded AI: Accelerate SaaS AI Copilot Adoption Safely

Watch Now View
Spotlight 10:05

Unstructured Data: Analytics Goldmine or a Governance Minefield?

Viral Kamdar
Watch Now View
Spotlight 21:30

Companies Cannot Grow If CISOs Don’t Allow Experimentation

Watch Now View
Spotlight 2:48

Unlocking Gen AI For Enterprise With Rehan Jalil

Rehan Jalil
Watch Now View
Spotlight 13:35

The Better Organized We’re from the Beginning, the Easier it is to Use Data

Watch Now View
Spotlight 13:11

Securing GenAI: From SaaS Copilots to Enterprise Applications

Rehan Jalil
Watch Now View
Spotlight 47:02

Navigating Emerging Technologies: AI for Security/Security for AI

Rehan Jalil
Watch Now View
Spotlight 59:55

Building Safe
Enterprise AI

Watch Now View

Latest

View More

Building Safe, Enterprise-grade AI with Securiti’s Gencore AI and NVIDIA NIM

Businesses are rapidly adopting generative AI (GenAI) to boost efficiency, productivity, innovation, customer service, and growth. However, IT & AI executives—particularly in highly regulated...

Automating EU AI Act Compliance View More

Automating EU AI Act Compliance: A 5-Step Playbook for GRC Teams

Artificial intelligence is revolutionizing industries, driving innovation in healthcare, finance, and beyond. But with great power comes great responsibility—especially when AI decisions impact health,...

Navigating Data Regulations in India’s Telecom Sector View More

Navigating Data Regulations in India’s Telecom Sector: Security, Privacy, Governance & AI

Gain insights into the key data regulations in India’s telecom sector and how they impact your business. Learn how Securiti helps ensure swift compliance...

India Strengthens Spam Rules View More

India Strengthens Spam Rules: Key Takeaways from TRAI’s Latest Amendment

India tightens spam regulations with TRAI’s latest amendment. Explore obligations on telecom providers, telemarketers, noncompliance penalties, and how Securiti can help.

5-Step AI Compliance Automation Playbook View More

EU AI Act: 5-Step AI Compliance Automation Playbook

Download the whitepaper to learn about the EU AI Act & its implication on high-risk AI systems, 5-step framework for AI compliance automation and...

A 6-Step Automation Guide View More

Say Goodbye to ROT Data: A 6-Step Automation Guide

Eliminate redundant obsolete and trivial (ROT) data with a strategic 6-step automation guide. Download the whitepaper today to discover how to streamline data management...

Kenya DPA Mapping View More

Kenya DPA Mapping: A Step-by-Step Compliance Roadmap

Download the Kenya DPA Mapping infographic for a step-by-step compliance roadmap. Simplify your compliance journey with Securiti DPA Compliance.

Texas Data Privacy and Security Act (TDPSA) View More

Navigating the Texas Data Privacy and Security Act (TDPSA): Key Details

Download the infographic to learn key details about Texas’ Data Privacy and Security Act (TDPSA) and simplify your compliance journey with Securiti.

Gencore AI and Amazon Bedrock View More

Building Enterprise-Grade AI with Gencore AI and Amazon Bedrock

Learn how to build secure enterprise AI copilots with Amazon Bedrock models, protect AI interactions with LLM Firewalls, and apply OWASP Top 10 LLM...

DSPM Vendor Due Diligence View More

DSPM Vendor Due Diligence

DSPM’s Buyer Guide ebook is designed to help CISOs and their teams ask the right questions and consider the right capabilities when looking for...

What's
New