Veeam Completes Acquisition of Securiti AI to Create the Industry’s First Trusted Data Platform for Accelerating Safe AI at Scale

View

Understanding Article 4 of the EU AI Act: Roadmap to AI Literacy

Contributors

Syed Tatheer Kazmi

Data Privacy Analyst

CIPP/Europe

Muhammad Faisal Sattar

Director of Product Legal & Global Data Compliance

FIP, CIPT, CIPM, CIPP/Asia

Under the European Union’s Artificial Intelligence Act (AI Act), organizations are required to ensure that all personnel involved in the development, deployment, or operation of AI systems possess a robust understanding of AI. Effective February 2, 2025, the AI literacy obligation under Article 4 of the AI Act is not a mere legal formality but a strategic commitment to responsible, informed, and ethical AI usage.

Consider a financial institution using an AI-driven loan approval system. Without adequate AI literacy among its staff, loan officers might fail to recognize biases embedded in the algorithm, leading to discriminatory lending practices and potential legal repercussions. This scenario highlights the critical need for comprehensive AI literacy across all sectors. Therefore, organizations should have a strategic, long-term approach to ensure a mature level of AI literacy.

Although formal public enforcement of penalties for non-compliance with the AI Act begins on August 2, 2025, private enforcement (individuals or organizations initiating legal actions or complaints independently) is already active, allowing individuals to raise concerns through legal channels even before a formal oversight body is established.

Before we delve further into AI literacy, it is important to understand what AI literacy means.

What is AI literacy?

Article 4 of the AI Act defines AI literacy as:

“Skills, knowledge, and understanding that allow providers, deployers and affected persons, taking into account their respective rights and obligations in the context of the AI Act, to make an informed deployment of AI systems, as well as to gain awareness about the opportunities and risks of AI and possible harm it can cause.” 

This definition is quite broad, generic, and flexible. The AI Act does not prescribe a specific level or quality of expertise but emphasizes the importance of being able to make well-informed decisions when deploying or developing AI systems. This includes not just developers and deployers, but also 'affected persons,' such as consumers whose loan applications are processed by AI, or job applicants screened by AI-powered tools.

In essence, the culmination of AI literacy lies in having a holistic understanding of how best to use AI in that particular context. This understanding ensures that individuals are well-equipped to navigate its complex and ever-evolving landscape. It also encompasses ethical considerations, ensuring that AI systems are used responsibly and aligned with your organization's core values.

AI Literacy Obligations

Article 4 of the Act sets forth the requirements for AI literacy.

It states that:

“Providers and deployers of AI systems shall take measures to ensure, to their best extent, a sufficient level of AI literacy of their staff and other persons dealing with the operation and use of AI systems on their behalf.” 

The scope of Article 4 is critical. It applies to providers and developers. This means that most organizations deploying AI will be subject to these literacy obligations. Providers and deployers must ensure that their employees and all persons involved in operating these systems achieve a “sufficient level” of AI literacy. The term other persons could mean the affected person, i.e., contractor, service provider, or user (client). Although the AI Act uses terms like “to their best extent” and “sufficient level,” it offers little specific guidance on these criteria, leaving them to be interpreted contextually based on the technical skills, experience, education, and training of their staff, along with the specific type and risk of AI systems used, the size and resources of the organization, and the groups that are impacted by these systems.

There is currently no comprehensive guidance on AI literacy mandated by the European Commission. However, based on our understanding and guidelines from the Dutch Data Protection Authority, organizations can adopt the following steps to ensure compliance with the AI literacy obligation.

Steps for Organizations to Ensure AI Literacy

Below are practical steps to help organizations establish effective AI literacy programs.

1. Identify AI Systems and Assess Risks

Organizations should begin by mapping where, how, and by whom AI is used within their operations while carefully evaluating the risks these systems may pose to employees, customers, and society at large. Creating a comprehensive inventory by leveraging existing records of processing activities (RoPA) is a practical first step. It is equally important to assess the broader impact of these AI systems and review any existing policies, initiatives, or frameworks that support AI literacy.

Example: A project manager in a technology firm may be responsible for cataloging all AI systems. In doing so, the manager evaluates the risk level associated with each system, considers the potential effects on various stakeholders, and identifies the roles employees play in managing these systems.

2. Define AI Literacy Goals by Role and Risk Level

Not every role requires the same depth of AI expertise. Individuals involved in AI development or critical decision-making need advanced training, while others may benefit from a general overview. Training programs should be prioritized based on the risk profile of each AI system. Key decision-makers such as executives and managers must have the insights necessary to oversee AI deployment responsibly, and employees who work directly with AI must be aware of the risks and potential outcomes. Tailored initiatives across different functions, such as HR, IT, and operations, ensure that all parts of the organization are equipped to manage AI risks effectively.

Example: In a company, a content manager using generative AI to create marketing materials must understand the underlying processes of AI, including potential biases and inaccuracies. Similarly, HR personnel responsible for recruitment need deeper insights into AI systems, especially if these systems are used for profiling or assessment, as such tools can significantly influence hiring decisions.

This example shows that knowledge and skill requirements vary among employees within an organization, influenced by the specific context and the risk level of the AI system.

3. Develop and Implement Comprehensive AI Literacy Programs

Once knowledge gaps are identified, organizations should design training programs that address these specific needs. These programs may include specialized training sessions, courses on AI ethics and legal compliance, and targeted instruction for staff involved in procuring or making decisions about AI systems. Training should be offered at multiple levels, from basic overviews to advanced courses, and should cater to various learning styles. Larger organizations might consider appointing an AI officer to coordinate training efforts and drive continuous improvement. Maintaining detailed records of all training activities is essential to demonstrate compliance and support periodic reviews.

Example: An organization might develop a guiding document titled “Approach to AI” that outlines the principles for responsible AI use and serves as a reference for all employees.

4. Monitor, Evaluate, and Enhance AI Literacy Initiatives

Given the rapid evolution of AI technology and regulatory standards, ongoing evaluation is critical. Organizations should implement regular internal or external audits, conduct employee surveys, and establish baseline measurements to gauge the effectiveness of their training programs. Adopting a multi-year AI literacy action plan can help ensure that training remains relevant, adaptive, and aligned with emerging risks and opportunities.

Example: Conducting annual employee surveys can provide valuable feedback on the effectiveness of AI training programs, allowing organizations to refine their approach and ensure that skill development keeps pace with technological advancements.

Conclusion

Ensuring a sufficient level of AI literacy within an organization is not just a regulatory requirement under the AI Act but a competitive edge and strategic imperative for managing AI-related risks responsibly. By tailoring AI training programs to the specific roles and risk levels within an organization, they can have a well-informed workforce capable of ethical and effective AI deployment.

To see what other organizations are doing on AI literacy, you can review the Commission's "living repository" of AI literacy practices from various organizations across different sectors and sizes. While replicating the practices in this repository does not automatically guarantee compliance with Article 4 of the AI Act, it encourages learning and exchange among providers and deployers of AI systems. The Commission does not endorse or evaluate these practices but aims to support a collaborative approach to improving AI literacy across industries.

Join Our Newsletter

Get all the latest information, law updates and more delivered to your inbox


Share

More Stories that May Interest You
Videos
View More
Mitigating OWASP Top 10 for LLM Applications 2025
Generative AI (GenAI) has transformed how enterprises operate, scale, and grow. There’s an AI application for every purpose, from increasing employee productivity to streamlining...
View More
Top 6 DSPM Use Cases
With the advent of Generative AI (GenAI), data has become more dynamic. New data is generated faster than ever, transmitted to various systems, applications,...
View More
Colorado Privacy Act (CPA)
What is the Colorado Privacy Act? The CPA is a comprehensive privacy law signed on July 7, 2021. It established new standards for personal...
View More
Securiti for Copilot in SaaS
Accelerate Copilot Adoption Securely & Confidently Organizations are eager to adopt Microsoft 365 Copilot for increased productivity and efficiency. However, security concerns like data...
View More
Top 10 Considerations for Safely Using Unstructured Data with GenAI
A staggering 90% of an organization's data is unstructured. This data is rapidly being used to fuel GenAI applications like chatbots and AI search....
View More
Gencore AI: Building Safe, Enterprise-grade AI Systems in Minutes
As enterprises adopt generative AI, data and AI teams face numerous hurdles: securely connecting unstructured and structured data sources, maintaining proper controls and governance,...
View More
Navigating CPRA: Key Insights for Businesses
What is CPRA? The California Privacy Rights Act (CPRA) is California's state legislation aimed at protecting residents' digital privacy. It became effective on January...
View More
Navigating the Shift: Transitioning to PCI DSS v4.0
What is PCI DSS? PCI DSS (Payment Card Industry Data Security Standard) is a set of security standards to ensure safe processing, storage, and...
View More
Securing Data+AI : Playbook for Trust, Risk, and Security Management (TRiSM)
AI's growing security risks have 48% of global CISOs alarmed. Join this keynote to learn about a practical playbook for enabling AI Trust, Risk,...
AWS Startup Showcase Cybersecurity Governance With Generative AI View More
AWS Startup Showcase Cybersecurity Governance With Generative AI
Balancing Innovation and Governance with Generative AI Generative AI has the potential to disrupt all aspects of business, with powerful new capabilities. However, with...

Spotlight Talks

Spotlight 50:52
From Data to Deployment: Safeguarding Enterprise AI with Security and Governance
Watch Now View
Spotlight 11:29
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Watch Now View
Spotlight 11:18
Rewiring Real Estate Finance — How Walker & Dunlop Is Giving Its $135B Portfolio a Data-First Refresh
Watch Now View
Spotlight 13:38
Accelerating Miracles — How Sanofi is Embedding AI to Significantly Reduce Drug Development Timelines
Sanofi Thumbnail
Watch Now View
Spotlight 10:35
There’s Been a Material Shift in the Data Center of Gravity
Watch Now View
Spotlight 14:21
AI Governance Is Much More than Technology Risk Mitigation
AI Governance Is Much More than Technology Risk Mitigation
Watch Now View
Spotlight 12:!3
You Can’t Build Pipelines, Warehouses, or AI Platforms Without Business Knowledge
Watch Now View
Spotlight 47:42
Cybersecurity – Where Leaders are Buying, Building, and Partnering
Rehan Jalil
Watch Now View
Spotlight 27:29
Building Safe AI with Databricks and Gencore
Rehan Jalil
Watch Now View
Spotlight 46:02
Building Safe Enterprise AI: A Practical Roadmap
Watch Now View
Latest
View More
DataAI Security: Why Healthcare Organizations Choose Securiti
Discover why healthcare organizations trust Securiti for Data & AI Security. Learn key blockers, five proven advantages, and what safe data innovation makes possible.
View More
The Anthropic Exploit: Welcome to the Era of AI Agent Attacks
Explore the first AI agent attack, why it changes everything, and how DataAI Security pillars like Intelligence, CommandGraph, and Firewalls protect sensitive data.
Network Security: Definition, Challenges, & Best Practices View More
Network Security: Definition, Challenges, & Best Practices
Discover what network security is, how it works, types, benefits, and best practices. Learn why network security is core to having a strong data...
View More
What is Cybersecurity Management?
Discover what cybersecurity management is, its importance, the CISO’s role, types, and best practices for effective cybersecurity management. Learn more.
Montana Privacy Amendment on Notices: What to Change by Oct 1 View More
Montana Privacy Amendment on Notices: What to Change by Oct 1
Download the whitepaper to learn about the Montana Privacy Amendment on Notices and what to change by Oct 1. Learn how Securiti helps.
2026 Privacy Law Updates: Key Developments You Need to Know View More
2026 Privacy Law Updates: Key Developments You Need to Know
Access the whitepaper to learn about key privacy law updates in 2026. Discover key developments you need to know. Learn how Securiti can help.
India’s DPDPA Rules View More
India’s DPDPA Rules
Access the infographic to learn about India’s DPDPA 2025 basics. Discover phased timelines, what the rules require, when they apply, key obligations, and much...
View More
DataAI Security For Healthcare
Secure PHI and sensitive clinical data, govern access, prevent breaches, and enable trusted AI with HIPAA-aligned DataAI Security for healthcare organizations.
View More
Navigating HITRUST: A Guide to Certification
Securiti's eBook is a practical guide to HITRUST certification, covering everything from choosing i1 vs r2 and scope systems to managing CAPs & planning...
The DSPM Architect’s Handbook View More
The DSPM Architect’s Handbook: Building an Enterprise-Ready Data+AI Security Program
Get certified in DSPM. Learn to architect a DSPM solution, operationalize data and AI security, apply enterprise best practices, and enable secure AI adoption...
What's
New