Announcing Agent Commander - The First Integrated solution from Veeam + Securiti.ai enabling the scaling of safe AI agents

View

EDPB Guidance on Deceptive Patterns in Social Media Interfaces

Published March 8, 2023
Author

Maria Khan

Data Privacy Legal Manager at Securiti

FIP, CIPT, CIPM, CIPP/E

Listen to the content

On 14 March 2022, the European Data Protection Board (EDPB) released guidelines titled Dark Patterns in Social Media Platform Interfaces: How to recognize and avoid them. These Guidelines provided best practice recommendations to designers and social media platform providers on how they can assess and avoid dark / deceptive patterns in social media interfaces that violate the requirements of the GDPR. Subsequently, the EDPB adopted version 2.0 of these Guidelines on 14 February 2023, titled Deceptive design patterns in social media platform interfaces: how to recognise and avoid them.

Deceptive patterns are interfaces and user experiences that lead social media users into making unwilling and potentially harmful choices, which are often geared towards the benefit of a social media platform and the detriment of the user, in relation to the processing of their personal data. Deceptive patterns aim to hinder users’ ability to make a conscious choice with respect to their personal data and ultimately take away users’ control for the protection of their personal data.

The EPDB has recognized six major categories of deceptive patterns

Overloading

Users are provided with too much information to push them to provide more personal data than necessary. Here, users are pushed to provide unnecessary data by being repeatedly asked about it.
Examples:

  • continuous prompting: repeatedly asking users to provide unnecessary data or consent to a processing purpose,
  • privacy maze: making the user navigate through too many pages,
  • too many options: too many options to choose from leaving the user to overlook some settings or giving up data protection preferences

Skipping

Deceptive designs that distract users from worrying about the protection of their personal data. Here, the most invasive features and options are already enabled by default.

Examples:

  • deceptive snugness: most data invasive features are pre-selected by default,
  • look over there: distract users.

Stirring

Wordings or visuals that are presented in a way that influences users’ emotional state to lead them to act against their data protection interests. This deceptive pattern has a higher impact on children and other vulnerable categories of data subjects. For example, users are more likely to overlook or have difficulty reading small font sizes or text written in colors that do not contrast sufficiently.

Examples:

  • emotional steering: patterns, wordings, or visual styles that strongly influence the users’ emotional state and lead them to act against their best interests i.e. by making users feel scared or guilty, or safe or rewarded,
  • hidden in plain sight: visual styles that nudge users toward less restrictive and more invasive options.

Obstructing

Providing misleading information to users to either push them to provide unnecessary personal data or influence their decision by holding them up and questioning their initial choices.

Examples:

  • dead end: while users are looking for information or control, they end up not finding it as a redirection link is either not working or not available,
  • longer than necessary: unnecessary steps required to activate data protection options,
  • misleading information: when social media providers inform users that an action on their part would lead to certain consequences, and the action actually results in a different outcome.

Fickle

Unclear designs that make it hard for the user to navigate the different data protection control tools or understand the purpose of the processing.

Examples:

  • lacking hierarchy: redundancy of information,
  • decontextualising: a data protection control is located on a page that is out of context,
  • inconsistent interface: an interface is not consistent across varied contexts or with users’ expectations,
  • language discontinuity: information not provided in the official language of the country where users live.

Left in the dark

Interfaces that hide information or data protection tools or leave users unsure of how their data is processed and what controls they have regarding the exercise of their rights.

Examples:

  • conflicting information: making the information unclear and unintelligible or misleading users by not matching their expectations.
  • ambiguous wording or information: vague wording or making data subjects unsure of how data will be processed or how to have control over their data.

As per the Guidelines, social media platform providers and designers should avoid the use of any deceptive patterns and ensure to provide a clear choice to users with respect to their personal data. The Guidelines further recommend that all data protection principles under the GDPR should be complied with during the design of user interfaces of online applications.

Some of the best practices recommendations as per the Guidelines are:

  • Ensure the processing of personal data is not detrimental, discriminatory, unexpected, or misleading to the data subject.
  • Ensure to obtain consent as per the requirements of the GDPR, i.e. consent must be freely given, specific, informed, and unambiguous wherever consent is required for data processing.
  • Provide information to data subjects about their rights with respect to their personal data or any communication in a concise, transparent, intelligible, and easily accessible form and language.
  • Lack of consent should be considered the default state.
  • Allow easy consent withdrawal.
  • Do not ask for additional and unnecessary personal data that is not required for the particular processing.
  • Ensure data protection by design by avoiding the use of any deceptive or manipulative language in designs and presenting all information in an objective and neutral manner.
  • Ensure data protection by default by pre-selecting and/or highlighting the least data invasive features and options by default.
  • Design a privacy dashboard, allowing users to centralize their privacy settings.
  • Be able to demonstrate compliance by documenting consent records.
  • Provide relevant links to further information, settings, or actions wherever helpful for users to navigate online interfaces.
  • Put options with the same processing purpose together so users can set their desired privacy settings more easily.
  • Provide an overview of a privacy policy at the top/start of such policy and include hyperlinks to each section. Clearly state the company contact address in the policy. Moreover, whenever any changes are made to the privacy policy, make the previous versions available and highlight the changes in the updated policy.
  • Clearly highlight the identity of the relevant supervisory authority and provide a link to its website.
  • Use consistent wordings and definitions.
  • Provide definitions for technical words or jargon. Moreover, provide examples to explain different processing purposes.
  • Make any data protection related actions or elements conspicuous on an interface that is not directly dedicated to the matter.
  • Whenever users create a new account on an online platform, provide them key information on data protection.
  • Use notifications to keep users continually aware of all aspects, changes or risks related to the processing of their data.
  • Use sticky navigation or provide a ‘return to top’ button to help ease navigation for users.
  • Provide neutral explanations for any consequences the users may face for their actions, such as withdrawing consent, or activating or deactivating an account.
  • Ensure cross-device consistency in relation to settings and information provision.
  • Provide a directory with all data protection related actions and information.
  • Use self-explanatory URLs for pages related to data protection settings or information.
  • Provide a dedicated form that helps users understand their rights and how to enforce them.

Ask for a demo today to understand how Securiti can help you achieve compliance with the provisions of the GDPR. In today’s digital world, it is important for organizations to implement privacy-compliant user interfaces and website designs and obtain consent as per the applicable legal requirements. Securiti’s Consent Management Solution enables you to design consent banners as per the applicable geographical requirements and avoid the use of deceptive patterns that can manipulate a user’s choice.

Frequently Asked Questions

Dark patterns are manipulative design choices that push users toward actions they might not take if they had a clear choice. For example, hiding privacy settings, making “accept all” buttons more visible, or making it hard to delete an account.

Dark patterns are manipulative design choices that steer users toward actions they might not choose if options were clearly presented. Examples include hiding privacy settings, highlighting “accept all” buttons, or making it difficult to delete an account.

The EDPB identifies several common examples:

  • Overloading users with too many confusing options
  • Skipping by pre-selecting privacy-intrusive choices
  • Stirring emotions to influence consent
  • Obstructing users from opting out or withdrawing consent
  • Hiding or obscuring information about how data is used

Any company that offers online services to EU users should follow these guidelines, not just social media platforms. This includes apps, e-commerce websites, and other digital service providers.

Analyze this article with AI

Prompts open in third-party AI tools.
Join Our Newsletter

Get all the latest information, law updates and more delivered to your inbox


Share

More Stories that May Interest You
Videos
View More
Rehan Jalil, Veeam on Agent Commander : theCUBE + NYSE Wired: Cyber Security Leaders
Following Veeam’s acquisition of Securiti, the launch of Agent Commander marks an important step toward helping enterprises adopt AI agents with greater confidence. In...
View More
Mitigating OWASP Top 10 for LLM Applications 2025
Generative AI (GenAI) has transformed how enterprises operate, scale, and grow. There’s an AI application for every purpose, from increasing employee productivity to streamlining...
View More
Top 6 DSPM Use Cases
With the advent of Generative AI (GenAI), data has become more dynamic. New data is generated faster than ever, transmitted to various systems, applications,...
View More
Colorado Privacy Act (CPA)
What is the Colorado Privacy Act? The CPA is a comprehensive privacy law signed on July 7, 2021. It established new standards for personal...
View More
Securiti for Copilot in SaaS
Accelerate Copilot Adoption Securely & Confidently Organizations are eager to adopt Microsoft 365 Copilot for increased productivity and efficiency. However, security concerns like data...
View More
Top 10 Considerations for Safely Using Unstructured Data with GenAI
A staggering 90% of an organization's data is unstructured. This data is rapidly being used to fuel GenAI applications like chatbots and AI search....
View More
Gencore AI: Building Safe, Enterprise-grade AI Systems in Minutes
As enterprises adopt generative AI, data and AI teams face numerous hurdles: securely connecting unstructured and structured data sources, maintaining proper controls and governance,...
View More
Navigating CPRA: Key Insights for Businesses
What is CPRA? The California Privacy Rights Act (CPRA) is California's state legislation aimed at protecting residents' digital privacy. It became effective on January...
View More
Navigating the Shift: Transitioning to PCI DSS v4.0
What is PCI DSS? PCI DSS (Payment Card Industry Data Security Standard) is a set of security standards to ensure safe processing, storage, and...
View More
Securing Data+AI : Playbook for Trust, Risk, and Security Management (TRiSM)
AI's growing security risks have 48% of global CISOs alarmed. Join this keynote to learn about a practical playbook for enabling AI Trust, Risk,...

Spotlight Talks

Spotlight
Future-Proofing for the Privacy Professional
Watch Now View
Spotlight 50:52
From Data to Deployment: Safeguarding Enterprise AI with Security and Governance
Watch Now View
Spotlight 11:29
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Watch Now View
Spotlight 11:18
Rewiring Real Estate Finance — How Walker & Dunlop Is Giving Its $135B Portfolio a Data-First Refresh
Watch Now View
Spotlight 13:38
Accelerating Miracles — How Sanofi is Embedding AI to Significantly Reduce Drug Development Timelines
Sanofi Thumbnail
Watch Now View
Spotlight 10:35
There’s Been a Material Shift in the Data Center of Gravity
Watch Now View
Spotlight 14:21
AI Governance Is Much More than Technology Risk Mitigation
AI Governance Is Much More than Technology Risk Mitigation
Watch Now View
Spotlight 12:!3
You Can’t Build Pipelines, Warehouses, or AI Platforms Without Business Knowledge
Watch Now View
Spotlight 47:42
Cybersecurity – Where Leaders are Buying, Building, and Partnering
Rehan Jalil
Watch Now View
Spotlight 27:29
Building Safe AI with Databricks and Gencore
Rehan Jalil
Watch Now View
Latest
View More
Building Sovereign AI with HPE Private Cloud AI and Veeam Securiti Gencore AI
How HPE Private Cloud AI, NVIDIA acceleration, and Veeam Securiti Gencore AI support secure, governed enterprise AI with policy enforcement across RAG, assistant, and agentic workflows.
View More
Securiti.ai Names Accenture as 2025 Partner of the Year
In a continued celebration of impactful collaboration in DataAI Security, Securiti.ai, a Veeam company, has honored Accenture as its 2025 Partner of the Year....
Largest Fine In CCPA History_ What The Latest CCPA Enforcement Action Teaches Businesses View More
Largest Fine In CCPA History: What The Latest CCPA Enforcement Action Teaches Businesses
Businesses can take some vital lessons from the recent biggest enforcement action in CCPA history. Securiti’s blog covers all the important details to know.
View More
AI & HIPAA: What It Means and How to Automate Compliance
Explore how the Health Insurance Portability and Accountability Act (HIPAA) applies to Artificial Intelligence (AI) in securing Protected Health Information (PHI). Learn how to...
View More
Minimize What You Expose: Privacy Guardrails for AI Agents and Copilots
Minimize data exposure in AI agents and copilots. Apply privacy guardrails like data minimization, access controls, masking, and policy enforcement to prevent leakage and...
View More
From Data Visibility to AI Velocity
Access the whitepaper and discover how unified DataAI security turns data governance into a business enabler, boosting AI innovation with visibility, compliance, and risk...
View More
Agent Commander: Solution Brief
Learn how Agent Commander detects AI agents, protects enterprise data with runtime guardrails, and undoes AI errors - enabling secure, compliant AI adoption at...
Compliance with CCPA Amendments with Securiti View More
Compliance with CCPA Amendments with Securiti
Stay compliant with 2026 CCPA amendments using Securiti, covering updated consent requirements, expanded sensitive data definitions, enhanced consumer rights, and readiness assessments.
View More
Take the Data Risk Out of AI
Learn how to prepare enterprise data for safe Gemini Enterprise adoption with upstream governance, sensitive data discovery, and pre-index policy controls.
View More
Navigating HITRUST: A Guide to Certification
Securiti's eBook is a practical guide to HITRUST certification, covering everything from choosing i1 vs r2 and scope systems to managing CAPs & planning...
What's
New