AI’s New Reality: Accountability, Transparency, and Control
This month signaled a turning point in AI regulation worldwide. Governments are no longer experimenting, they are legislating. From California’s expansive accountability laws to Europe’s operational rollout of the AI Act and Asia’s draft frameworks emphasizing risk tiers and human oversight, AI is moving firmly into a rules-based era.
The trend is unmistakable: regulators are treating AI less as a novelty and more as infrastructure demanding provenance, explainability, and human liability. This signals the start of “compliance-driven innovation,” where responsible design, transparent datasets, and auditable systems become strategic assets.
As 2026 approaches, expect AI governance to tighten further, not only around safety and ethics but around operational transparency, supply-chain accountability, and real-world impact. The age of voluntary AI ethics is closing; the age of enforceable AI integrity has begun.
North & South America Jurisdiction
1. AAIP Leads Meeting on Advancing Data Protection Standards
October 15, 2025 Argentina
Argentina’s Agency for Access to Public Information (AAIP) led the 64th Bureau meeting of the Council of Europe’s Convention 108 Committee, reviewing progress toward ratification of Convention 108+, which needs just five more signatories to enter into force.
Members advanced work on new guidelines addressing data protection in large-scale language models and neuroscience, and outlined the 2026–2029 work plan. Priorities include promoting Convention 108+ adoption, strengthening oversight of cross-border data flows, and enhancing cooperation on AI and privacy governance.
The meeting reaffirmed Convention 108+ as the core global framework for safeguarding privacy in the age of AI and emerging technologies.
2. Brazil’s ANPD Launches First AI and Data Protection Regulatory Sandbox
October 14, 2025 Brazil
Brazil’s National Data Protection Authority (ANPD) has announced the final results of its AI Regulatory Sandbox, selecting three projects: Metatext AI, Synapse AI, and IA Greenworld to develop privacy-aligned AI solutions under ANPD supervision through December 2026.
The next phase involves training participants in regulation, AI ethics, and experimental governance, ensuring a consistent understanding of data protection principles.
The initiative aims to foster innovation with algorithmic transparency and privacy-by-design, reinforcing Brazil’s commitment to responsible AI development within a structured, monitored environment.
3. Louisiana Bans AI Tools from Foreign Adversaries
October 13, 2025 Louisiana, United States
Louisiana’s Governor has issued an Executive Order banning the use of AI tools developed by foreign adversaries, including entities tied to the Chinese Communist Party
The order applies across state agencies, universities, and public schools, requiring immediate review of existing AI systems and vendor contracts to remove prohibited tools such as DeepSeek from use.
The move highlights a growing national-security focus in the U.S. AI governance, emphasizing data sovereignty and resilience against foreign technological influence.
4. California Expands AI Transparency Act, Delays Operative Date to 2026
October 13, 2025 California, United States
Governor Gavin Newsom has signed AB 853, amending the California AI Transparency Act (CAITA) to broaden compliance obligations and extend its start date to August 2, 2026, aligning with the EU AI Act.
The law now covers not only generative AI developers but also large online platforms, AI hosting services, and capture device manufacturers like smartphone makers. These entities will be required to implement provenance detection tools, enable user transparency on AI-generated or altered content, and embed disclosure data in media.
By expanding CAITA’s reach, California reinforces its position as a regulatory frontrunner in AI transparency and content authenticity, setting the stage for deeper alignment with global AI governance standards.
5. California Enacts Nation’s First Law Regulating Companion Chatbots
October 13, 2025 California, United State
California has enacted SB 243, the first U.S. law imposing design, disclosure, and safety obligations on operators of AI “companion chatbots.” Effective January 1, 2026, the law requires operators to embed clear AI disclosures, display suitability warnings, and implement harm prevention protocols, including mandatory crisis referral mechanisms and bans on sexual or self-harm content for minors.
From July 1, 2027, operators must also file annual reports detailing crisis response statistics and safety measures. The law introduces a private right of action, allowing individuals harmed by violations to seek damages and injunctive relief.
SB 243 signals California’s next regulatory phase in AI accountability and child online safety, extending beyond privacy to mental health and human-interaction ethics..
6. California Enacts AB 489 on Deceptive Healthcare Terms
October 13, 2025 California, United States
California has enacted Assembly Bill 489, effective January 1, 2026, prohibiting AI systems from using healthcare titles or terminology that could falsely suggest professional medical licensure or certification. The law aims to curb deceptive AI health communications, particularly in AI-generated health advice, reports, and assessments.
Violations fall under healthcare licensing boards’ oversight, with each misuse of a protected term treated as a separate offense. The measure underscores California’s growing focus on AI transparency and consumer protection in digital health, pushing developers to include clear disclaimers and avoid misleading branding in healthcare-related AI products. This aligns with California’s broader regulatory trend of tightening AI accountability across high-risk sectors, complementing recent laws on chatbot safeguards and AI transparency.
7. California Governor Signs AB 316, Ensuring Human Accountability for AI-Caused Harm
October 13, 2025 California, United States
California has enacted Assembly Bill 316, introducing a first-of-its-kind legal framework to ensure human accountability for AI-caused harm. The law prohibits defendants from arguing that an AI system acted autonomously or independently as a defense in civil or criminal proceedings, reinforcing that liability remains with the individuals or entities deploying or developing the AI.
Effective January 1, 2026, AB 316 represents a major step toward clarifying responsibility in AI-related incidents, aligning with California’s broader efforts to set ethical and legal standards for artificial intelligence across sectors.
8. California Governor Signs Assembly Bill 325 Targeting Algorithmic Price-Fixing
October 6, 2025 California, United States
Governor Gavin Newsom has signed Assembly Bill 325, authored by Majority Leader Cecilia Aguiar-Curry, modernizing California’s antitrust laws to address algorithmic price fixing and safeguard consumers and small businesses from corporate collusion driven by AI and automation.
The new law strengthens the Cartwright Act, clarifying that using algorithms to coordinate or inflate prices or depress wages constitutes illegal conduct, even when done through technology rather than direct communication. It also introduces safe harbors for businesses that use pricing algorithms lawfully or unknowingly.
Effective January 1, 2026, AB 325 positions California as the first U.S. state to explicitly outlaw algorithmic collusion, reflecting a growing regulatory trend targeting AI-enabled market manipulation and protecting fair competition across key sectors like housing, groceries, and healthcare. Read More
9. IAB Canada Releases AI Use Case Map For Digital Advertising
October 2, 2025 Canada
IAB Canada has published its AI Use Case Map for Digital Advertising, offering a practical framework for understanding and implementing AI across the ad ecosystem.
The map outlines applications across five key functions: Audience & Identity, Creative & Content, Media Buying & Optimization, Measurement & Attribution, and Privacy & Governance, reflecting the full digital advertising lifecycle from audience targeting to performance analysis.
The initiative aims to help organizations integrate AI responsibly and strategically, strengthening internal workflows and governance as AI-driven marketing becomes increasingly central to the advertising industry.
10. Montana HB 178 Takes Effect, Limiting Government Entities’ Use of AI
October 1, 2025 Montana, United States
Montana’s House Bill 178 has officially taken effect, introducing one of the strictest state-level frameworks governing how public authorities may use Artificial Intelligence.
The law prohibits AI use for cognitive or behavioral manipulation, discriminatory classification, malicious intent, or mass public surveillance. By setting clear ethical and operational limits, HB 178 aims to safeguard citizens’ rights and prevent government misuse of AI technologies, marking a significant milestone in responsible AI governance at the state level.
11. Regulations to Protect Against Employment Discrimination Related to Artificial Intelligence Take Effect
October 1, 2025 California, United States
California’s Regulations to Protect Against Employment Discrimination Related to Artificial Intelligence took effect on October 1, 2025, marking the state’s first comprehensive framework to address algorithmic bias in hiring and workplace decisions.
The rules clarify that algorithmic bias violates state anti-discrimination laws, require employers to retain automated-decision data for four years, and prohibit tools that elicit disability-related information. They also introduce formal definitions for “automated-decision systems” and related terms to support consistent enforcement.
By modernizing its civil rights framework, California aims to ensure fairness and accountability in AI-assisted employment practices, reinforcing its position as a leader in responsible technology regulation.
12. European Data Protection Supervisor Releases Guidance For Use Of GenAI By EU Institutions
October 28, 2025
The European Data Protection Supervisor (EDPS) has issued updated guidance on the use of Generative AI (GenAI) by EU institutions, agencies, and offices. The revised document reflects the rapid evolution of AI technologies and reinforces the need for strong safeguards when personal data is processed.
Key updates include a refined definition of GenAI, a practical compliance checklist, and clarified roles to determine whether entities act as controllers, joint controllers, or processors. It also outlines how to establish a lawful basis, ensure purpose limitation, and uphold data subject rights.
The guidance aims to help EU bodies adopt GenAI responsibly, balancing innovation with the EU’s commitment to data protection and accountability.
13. Netherlands’ AP Releases New AI Literacy Guide
October 23, 2025 Netherlands
The Autoriteit Persoonsgegevens (AP) has released the ‘Building AI Literacy’ guide, which is a follow-up to its earlier 'Get Started with AI Literacy' guide. It helps organizations comply with the upcoming AI Act by promoting responsible AI use. The guide stresses that all staff involved in AI should understand both technical and ethical aspects of AI systems.
The guide includes practical examples for assessing AI impact, mitigating risks, and maximizing societal and business benefits. By linking literacy with compliance, the AP highlights that awareness and accountability are essential pillars of trustworthy AI implementation ahead of the AI Act’s full enforcement.
14. Denmark’s Issues Guidance on Prohibited AI Uses under EU AI Act
October 22, 2025 Denmark
The Danish Agency for Digital Government (Digitaliseringsstyrelsen) has published six guides on AI practices that are prohibited under the EU AI Act. Each guide focuses on a specific prohibition: harmful manipulation, exploiting vulnerabilities, social scoring, non-targeted facial image collection, emotion recognition in workplaces or schools, and a main guide outlining all key prohibitions.
The agency recommends reading these guides together with the European Commission’s Guidelines on Prohibited AI Practices to ensure full compliance with EU rules.
15. European Commission Launches COMPASS-AI to Advance AI in Healthcare
October 21, 2025
The European Commission has launched COMPASS-AI, a flagship initiative under its Apply AI Strategy, to promote safe and effective use of AI in healthcare. The program will bring together experts and provide a digital platform to support responsible AI integration in clinical settings. It will focus on key areas such as cancer care and remote healthcare, offering guidelines and training for both professionals and patients.
COMPASS-AI aims to improve precision medicine, enhance diagnostics, and enable more personalized care. The initiative also supports the European Health Data Space and seeks to build trust in AI-driven healthcare.
16. Germany’s DSK Issues Guidance on AI Systems Using Retrieval Augmented Generation (RAG)
October 17, 2025 Germany
Germany’s Conference of Independent Data Protection Authorities (DSK) has published guidance for organizations using Retrieval Augmented Generation (RAG) to improve the accuracy and reliability of Large Language Model (LLM) outputs. RAG connects LLMs to internal knowledge sources, producing more context-specific and precise responses.
The guidance highlights RAG’s benefits, including supporting digital sovereignty by enabling local model use, promoting data protection by design, and reducing LLM errors and hallucinations.
However, DSK warns that RAG does not eliminate legal risks from unlawfully trained models. Organizations must still conduct thorough data protection assessments and implement strong technical and organizational safeguards to protect transparency and data rights.
Italy’s AI Law (Law No. 132 of 2025) officially took effect, making it the first comprehensive national AI law in the EU. The legislation establishes a human-centric framework for ethical, transparent, and accountable AI, complementing the EU AI Act and GDPR.
The law mandates fair and responsible data processing, clear user communication, and parental consent for users under 14. In critical sectors like healthcare, AI systems cannot replace professional judgment. Oversight is shared between the Agency for Digital Italy (AgID) and the National Cybersecurity Agency, ensuring both compliance and security.
It also introduces criminal penalties for illegal deepfakes, clarifies that copyright applies only to human-authored works, and requires the National AI Strategy to be updated biennially, signaling Italy’s commitment to privacy-aligned innovation and robust AI governance.
18. EU Launches AI Act Support Tools to Guide Implementation
October 8, 2025
The European Commission has rolled out two new resources to help organizations implement the AI Act across Europe: the AI Act Service Desk and the AI Act Single Information Platform. These tools are designed to give businesses practical guidance and legal clarity for using AI responsibly. The platform serves as a central hub, offering tailored advice for all AI Act stakeholders.
It includes the Compliance Checker, which shows whether your organization needs to comply and how; the AI Act Explorer, which makes it easy to navigate the Act’s chapters, annexes, and recitals; and a Question Submission Form to get answers directly from the AI Act Service Desk.
The AI Act Service Desk is staffed by expert professionals. They work closely with the AI Office to answer stakeholder questions and provide specialized guidance. Organizations should explore these tools to ensure smooth compliance with the AI Act requirements.
19. Czechia’s MIT Introduces Draft AI Law Aligning it with EU AI Act
October 1, 2025 Czechia
The Czech Ministry of Industry and Trade has proposed a draft AI law closely mirroring the EU Artificial Intelligence Act, focusing on national institutional and enforcement mechanisms.
The law designates the Czech Telecommunications Office as the main supervisory authority, with the National Bank overseeing financial AI, and the Office for Personal Data Protection supervising AI involving personal data. It also establishes a regulatory sandbox for testing high-risk AI systems and outlines penalties of up to €35 million or 7% of global turnover for violations.
Expected to take effect 15 days after publication in 2026, the legislation aims to ensure consistent AI governance, promote responsible innovation, and harmonize Czech oversight with the EU’s broader regulatory framework.
20. India’s MeitY Proposed Draft Amendments to the IT Rules to Regulate Synthetic Content
October 22, 2025 India
India’s Ministry of Electronics and Information Technology (MeitY) has proposed amendments to the IT Rules to address synthetically generated content, algorithmically created information that appears authentic.
The draft rules would require intermediaries to clearly label synthetic content using unique metadata or identifiers, while Significant Social Media Intermediaries (SSMIs) must obtain user declarations and verify such content before dissemination.
Aimed at enhancing transparency and accountability, the proposal marks India’s first step toward a structured framework for AI-generated content governance, ensuring users can distinguish between authentic and synthetic media.
21. Australia’s NAIC Releases New Tools to Advance Responsible AI Governance
October 21, 2025 Australia
Australia’s National AI Centre (NAIC) has launched several initiatives to promote transparent and accountable AI governance nationwide. Key releases include, AI Systems Register Template for tracking and managing AI systems, the AI Screening Tool with seven questions to determine appropriate oversight levels, and the AI Policy Guide and Template to help organizations create aligned governance policies.
These initiatives strengthen responsible AI adoption across sectors, reinforcing Australia’s position as a leader in ethical and trustworthy AI governance. NAIC also introduced two tiers of Responsible AI Guidance: Foundations for beginners and Implementation Practices for advanced users, focused on accountability, transparency, and continuous monitoring.
22. Australia’s ACSC Issues Guidance on Securing AI and ML Supply Chains
October 16, 2025 Australia
The Australian Cyber Security Centre (ACSC), part of the Australian Signals Directorate, has published new guidance to help organizations manage cybersecurity and privacy risks within Artificial Intelligence (AI) and Machine Learning (ML) supply chains.
The advisory identifies six key risk areas: data, model, software, infrastructure, hardware, and third-party dependencies. It recommends a risk-based approach that includes vetting and assessing suppliers, embedding cybersecurity requirements in contracts, using trusted and sanitized data sources, validating AI/ML models, and maintaining secure configurations across systems.
This guidance also establishes a national framework for safeguarding AI/ML ecosystems, urging organizations to embed supply chain cybersecurity into their broader AI governance and risk management strategies.
23. Taiwan Evaluates Draft AI Basic Act to Strengthen Governance
October 9, 2025 Taiwan
Taiwan’s Legislative Yuan has reviewed the Draft Artificial Intelligence Basic Act, proposing stronger governance and clearer oversight structures. The evaluation recommends naming the Ministry of Digital Affairs (MODA) as the central AI authority, aligning definitions of AI and high-risk systems with international standards, and embedding human oversight throughout the AI lifecycle.
It also calls for SME support, AI talent development, and initiatives to attract global experts, reinforcing Taiwan’s commitment to responsible innovation and ethical AI deployment while enhancing its global competitiveness.
24. Vietnam's AI Draft Law: Basic Principles and Prohibitions
October 7, 2025 Vietnam
Vietnam’s Ministry of Science and Technology has released the Draft Law on Artificial Intelligence for public consultation. Comprising nine chapters and 70 articles, the draft establishes a human-centered, risk-based framework for AI governance.
It introduces seven guiding principles, including ensuring human control, safety, fairness, and transparency, and adopts a four-tier risk classification system ranging from “unacceptable” to “low risk.”
The law also lists nine prohibited AI practices, such as cognitive manipulation, exploiting vulnerable groups, social credit scoring, mass facial recognition, and harmful deepfake creation.
If enacted, Vietnam would become one of the first nations with a comprehensive, risk-based AI law, balancing innovation with ethical safeguards and citizen protection.
25. Kazakhstan Nears Adoption of First Comprehensive AI Law
October 2, 2025 Kazakhstan
Kazakhstan is close to passing its first comprehensive Artificial Intelligence Law, following approval by the Mazhilis and pending Senate review. The draft law introduces a unified framework for AI governance, establishing transparency, safety, and accountability standards while defining roles and obligations for developers, users, and regulators.
Oversight will be shared between the Ministry of Artificial Intelligence and Digital Development and the National Security Committee. Once enacted, Kazakhstan will become the first Central Asian nation with a broad AI legal regime aimed at fostering responsible innovation and protecting citizens’ rights.
House Bill 5764, the AI for Mainstreet Act, was introduced in the U.S. House of Representatives, aims to empower small businesses with AI adoption support, potentially setting a precedent for national SME-focused AI policy.
The European Ombudswoman has launched an inquiry into the Commission's monitoring of private bodies (CEN and CENELEC) developing the AI Act's technical standards, which could reshape compliance transparency across the EU.
From November 21- 23, 2025, Fuzhou will host the Asia-Pacific Artificial Intelligence Education Conference. The event will spotlight AI’s role in transforming education, providing a platform for international academic exchange, industry collaboration, and practical applications.
Generative AI (GenAI) has transformed how enterprises operate, scale, and grow. There’s an AI application for every purpose, from increasing employee productivity to streamlining...
With the advent of Generative AI (GenAI), data has become more dynamic. New data is generated faster than ever, transmitted to various systems, applications,...
Accelerate Copilot Adoption Securely & Confidently Organizations are eager to adopt Microsoft 365 Copilot for increased productivity and efficiency. However, security concerns like data...
As enterprises adopt generative AI, data and AI teams face numerous hurdles: securely connecting unstructured and structured data sources, maintaining proper controls and governance,...
What is CPRA? The California Privacy Rights Act (CPRA) is California's state legislation aimed at protecting residents' digital privacy. It became effective on January...
Balancing Innovation and Governance with Generative AI Generative AI has the potential to disrupt all aspects of business, with powerful new capabilities. However, with...
Spotlight Talks
Spotlight
50:52
From Data to Deployment: Safeguarding Enterprise AI with Security and Governance
Discover why healthcare organizations trust Securiti for Data & AI Security. Learn key blockers, five proven advantages, and what safe data innovation makes possible.
Explore the first AI agent attack, why it changes everything, and how DataAI Security pillars like Intelligence, CommandGraph, and Firewalls protect sensitive data.
Securiti’s latest blog walks you through all the important information and guidance you need to ensure your AI systems are compliant with GDPR requirements.
Access the whitepaper to learn about what businesses need to know about Australia’s Guidance for AI Adoption. Discover how Securiti helps ensure compliance.
Extend Microsoft Purview with Securiti to discover, classify, and reduce data & AI risk across hybrid environments with continuous monitoring and automated remediation. Learn...
Securiti's eBook is a practical guide to HITRUST certification, covering everything from choosing i1 vs r2 and scope systems to managing CAPs & planning...
Get certified in DSPM. Learn to architect a DSPM solution, operationalize data and AI security, apply enterprise best practices, and enable secure AI adoption...