Veeam Completes Acquisition of Securiti AI to Create the Industry’s First Trusted Data Platform for Accelerating Safe AI at Scale

View

EDPB Guidance on Deceptive Patterns in Social Media Interfaces

Published March 8, 2023
Author

Maria Khan

Data Privacy Legal Manager at Securiti

FIP, CIPT, CIPM, CIPP/E

Listen to the content

On 14 March 2022, the European Data Protection Board (EDPB) released guidelines titled Dark Patterns in Social Media Platform Interfaces: How to recognize and avoid them. These Guidelines provided best practice recommendations to designers and social media platform providers on how they can assess and avoid dark / deceptive patterns in social media interfaces that violate the requirements of the GDPR. Subsequently, the EDPB adopted version 2.0 of these Guidelines on 14 February 2023, titled Deceptive design patterns in social media platform interfaces: how to recognise and avoid them.

Deceptive patterns are interfaces and user experiences that lead social media users into making unwilling and potentially harmful choices, which are often geared towards the benefit of a social media platform and the detriment of the user, in relation to the processing of their personal data. Deceptive patterns aim to hinder users’ ability to make a conscious choice with respect to their personal data and ultimately take away users’ control for the protection of their personal data.

The EPDB has recognized six major categories of deceptive patterns

Overloading

Users are provided with too much information to push them to provide more personal data than necessary. Here, users are pushed to provide unnecessary data by being repeatedly asked about it.
Examples:

  • continuous prompting: repeatedly asking users to provide unnecessary data or consent to a processing purpose,
  • privacy maze: making the user navigate through too many pages,
  • too many options: too many options to choose from leaving the user to overlook some settings or giving up data protection preferences

Skipping

Deceptive designs that distract users from worrying about the protection of their personal data. Here, the most invasive features and options are already enabled by default.

Examples:

  • deceptive snugness: most data invasive features are pre-selected by default,
  • look over there: distract users.

Stirring

Wordings or visuals that are presented in a way that influences users’ emotional state to lead them to act against their data protection interests. This deceptive pattern has a higher impact on children and other vulnerable categories of data subjects. For example, users are more likely to overlook or have difficulty reading small font sizes or text written in colors that do not contrast sufficiently.

Examples:

  • emotional steering: patterns, wordings, or visual styles that strongly influence the users’ emotional state and lead them to act against their best interests i.e. by making users feel scared or guilty, or safe or rewarded,
  • hidden in plain sight: visual styles that nudge users toward less restrictive and more invasive options.

Obstructing

Providing misleading information to users to either push them to provide unnecessary personal data or influence their decision by holding them up and questioning their initial choices.

Examples:

  • dead end: while users are looking for information or control, they end up not finding it as a redirection link is either not working or not available,
  • longer than necessary: unnecessary steps required to activate data protection options,
  • misleading information: when social media providers inform users that an action on their part would lead to certain consequences, and the action actually results in a different outcome.

Fickle

Unclear designs that make it hard for the user to navigate the different data protection control tools or understand the purpose of the processing.

Examples:

  • lacking hierarchy: redundancy of information,
  • decontextualising: a data protection control is located on a page that is out of context,
  • inconsistent interface: an interface is not consistent across varied contexts or with users’ expectations,
  • language discontinuity: information not provided in the official language of the country where users live.

Left in the dark

Interfaces that hide information or data protection tools or leave users unsure of how their data is processed and what controls they have regarding the exercise of their rights.

Examples:

  • conflicting information: making the information unclear and unintelligible or misleading users by not matching their expectations.
  • ambiguous wording or information: vague wording or making data subjects unsure of how data will be processed or how to have control over their data.

As per the Guidelines, social media platform providers and designers should avoid the use of any deceptive patterns and ensure to provide a clear choice to users with respect to their personal data. The Guidelines further recommend that all data protection principles under the GDPR should be complied with during the design of user interfaces of online applications.

Some of the best practices recommendations as per the Guidelines are:

  • Ensure the processing of personal data is not detrimental, discriminatory, unexpected, or misleading to the data subject.
  • Ensure to obtain consent as per the requirements of the GDPR, i.e. consent must be freely given, specific, informed, and unambiguous wherever consent is required for data processing.
  • Provide information to data subjects about their rights with respect to their personal data or any communication in a concise, transparent, intelligible, and easily accessible form and language.
  • Lack of consent should be considered the default state.
  • Allow easy consent withdrawal.
  • Do not ask for additional and unnecessary personal data that is not required for the particular processing.
  • Ensure data protection by design by avoiding the use of any deceptive or manipulative language in designs and presenting all information in an objective and neutral manner.
  • Ensure data protection by default by pre-selecting and/or highlighting the least data invasive features and options by default.
  • Design a privacy dashboard, allowing users to centralize their privacy settings.
  • Be able to demonstrate compliance by documenting consent records.
  • Provide relevant links to further information, settings, or actions wherever helpful for users to navigate online interfaces.
  • Put options with the same processing purpose together so users can set their desired privacy settings more easily.
  • Provide an overview of a privacy policy at the top/start of such policy and include hyperlinks to each section. Clearly state the company contact address in the policy. Moreover, whenever any changes are made to the privacy policy, make the previous versions available and highlight the changes in the updated policy.
  • Clearly highlight the identity of the relevant supervisory authority and provide a link to its website.
  • Use consistent wordings and definitions.
  • Provide definitions for technical words or jargon. Moreover, provide examples to explain different processing purposes.
  • Make any data protection related actions or elements conspicuous on an interface that is not directly dedicated to the matter.
  • Whenever users create a new account on an online platform, provide them key information on data protection.
  • Use notifications to keep users continually aware of all aspects, changes or risks related to the processing of their data.
  • Use sticky navigation or provide a ‘return to top’ button to help ease navigation for users.
  • Provide neutral explanations for any consequences the users may face for their actions, such as withdrawing consent, or activating or deactivating an account.
  • Ensure cross-device consistency in relation to settings and information provision.
  • Provide a directory with all data protection related actions and information.
  • Use self-explanatory URLs for pages related to data protection settings or information.
  • Provide a dedicated form that helps users understand their rights and how to enforce them.

Ask for a demo today to understand how Securiti can help you achieve compliance with the provisions of the GDPR. In today’s digital world, it is important for organizations to implement privacy-compliant user interfaces and website designs and obtain consent as per the applicable legal requirements. Securiti’s Consent Management Solution enables you to design consent banners as per the applicable geographical requirements and avoid the use of deceptive patterns that can manipulate a user’s choice.

Frequently Asked Questions

Dark patterns are manipulative design choices that push users toward actions they might not take if they had a clear choice. For example, hiding privacy settings, making “accept all” buttons more visible, or making it hard to delete an account.

Dark patterns are manipulative design choices that steer users toward actions they might not choose if options were clearly presented. Examples include hiding privacy settings, highlighting “accept all” buttons, or making it difficult to delete an account.

The EDPB identifies several common examples:

  • Overloading users with too many confusing options
  • Skipping by pre-selecting privacy-intrusive choices
  • Stirring emotions to influence consent
  • Obstructing users from opting out or withdrawing consent
  • Hiding or obscuring information about how data is used

Any company that offers online services to EU users should follow these guidelines, not just social media platforms. This includes apps, e-commerce websites, and other digital service providers.

Analyze this article with AI

Prompts open in third-party AI tools.
Join Our Newsletter

Get all the latest information, law updates and more delivered to your inbox


Share

More Stories that May Interest You
Videos
View More
Mitigating OWASP Top 10 for LLM Applications 2025
Generative AI (GenAI) has transformed how enterprises operate, scale, and grow. There’s an AI application for every purpose, from increasing employee productivity to streamlining...
View More
Top 6 DSPM Use Cases
With the advent of Generative AI (GenAI), data has become more dynamic. New data is generated faster than ever, transmitted to various systems, applications,...
View More
Colorado Privacy Act (CPA)
What is the Colorado Privacy Act? The CPA is a comprehensive privacy law signed on July 7, 2021. It established new standards for personal...
View More
Securiti for Copilot in SaaS
Accelerate Copilot Adoption Securely & Confidently Organizations are eager to adopt Microsoft 365 Copilot for increased productivity and efficiency. However, security concerns like data...
View More
Top 10 Considerations for Safely Using Unstructured Data with GenAI
A staggering 90% of an organization's data is unstructured. This data is rapidly being used to fuel GenAI applications like chatbots and AI search....
View More
Gencore AI: Building Safe, Enterprise-grade AI Systems in Minutes
As enterprises adopt generative AI, data and AI teams face numerous hurdles: securely connecting unstructured and structured data sources, maintaining proper controls and governance,...
View More
Navigating CPRA: Key Insights for Businesses
What is CPRA? The California Privacy Rights Act (CPRA) is California's state legislation aimed at protecting residents' digital privacy. It became effective on January...
View More
Navigating the Shift: Transitioning to PCI DSS v4.0
What is PCI DSS? PCI DSS (Payment Card Industry Data Security Standard) is a set of security standards to ensure safe processing, storage, and...
View More
Securing Data+AI : Playbook for Trust, Risk, and Security Management (TRiSM)
AI's growing security risks have 48% of global CISOs alarmed. Join this keynote to learn about a practical playbook for enabling AI Trust, Risk,...
AWS Startup Showcase Cybersecurity Governance With Generative AI View More
AWS Startup Showcase Cybersecurity Governance With Generative AI
Balancing Innovation and Governance with Generative AI Generative AI has the potential to disrupt all aspects of business, with powerful new capabilities. However, with...

Spotlight Talks

Spotlight 50:52
From Data to Deployment: Safeguarding Enterprise AI with Security and Governance
Watch Now View
Spotlight 11:29
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Not Hype — Dye & Durham’s Analytics Head Shows What AI at Work Really Looks Like
Watch Now View
Spotlight 11:18
Rewiring Real Estate Finance — How Walker & Dunlop Is Giving Its $135B Portfolio a Data-First Refresh
Watch Now View
Spotlight 13:38
Accelerating Miracles — How Sanofi is Embedding AI to Significantly Reduce Drug Development Timelines
Sanofi Thumbnail
Watch Now View
Spotlight 10:35
There’s Been a Material Shift in the Data Center of Gravity
Watch Now View
Spotlight 14:21
AI Governance Is Much More than Technology Risk Mitigation
AI Governance Is Much More than Technology Risk Mitigation
Watch Now View
Spotlight 12:!3
You Can’t Build Pipelines, Warehouses, or AI Platforms Without Business Knowledge
Watch Now View
Spotlight 47:42
Cybersecurity – Where Leaders are Buying, Building, and Partnering
Rehan Jalil
Watch Now View
Spotlight 27:29
Building Safe AI with Databricks and Gencore
Rehan Jalil
Watch Now View
Spotlight 46:02
Building Safe Enterprise AI: A Practical Roadmap
Watch Now View
Latest
View More
DataAI Security: Why Healthcare Organizations Choose Securiti
Discover why healthcare organizations trust Securiti for Data & AI Security. Learn key blockers, five proven advantages, and what safe data innovation makes possible.
View More
The Anthropic Exploit: Welcome to the Era of AI Agent Attacks
Explore the first AI agent attack, why it changes everything, and how DataAI Security pillars like Intelligence, CommandGraph, and Firewalls protect sensitive data.
HIPAA PHI Explained: Identifiers, De-identification & Compliance Checklist View More
HIPAA PHI Explained: Identifiers, De-identification & Compliance Checklist
Discover what PHI is under HIPAA. Understand what is considered PHI as per HIPAA Rules, the list of 18 identifiers, and what happens to...
Red Teaming View More
What is AI Red Teaming? Complete Guide
AI red teaming tests AI systems for security, safety, and misuse risks. Learn how it works, common techniques, real-world use cases, and why it...
View More
Australia’s Privacy Overhaul: Landmark Reforms in Privacy, Cyber Security & Online Safety
Access the whitepaper and gain insights into Australia’s Privacy Law landscape, CSLP, Social Media Minimum Age Act, and how Securiti helps ensure swift compliance.
View More
CNIL’s €475 Million Cookie Consent Enforcement: Key Lessons for Organizations
Download the whitepaper to learn about CNIL’s €475 million cookie consent enforcement fine. Discover key lessons for organizations and how to automate compliance.
View More
Solution Brief: Microsoft Purview + Securiti
Extend Microsoft Purview with Securiti to discover, classify, and reduce data & AI risk across hybrid environments with continuous monitoring and automated remediation. Learn...
Top 7 Data & AI Security Trends 2026 View More
Top 7 Data & AI Security Trends 2026
Discover the top 7 Data & AI security trends for 2026. Learn how to secure AI agents, govern data, manage risk, and scale AI...
View More
Navigating HITRUST: A Guide to Certification
Securiti's eBook is a practical guide to HITRUST certification, covering everything from choosing i1 vs r2 and scope systems to managing CAPs & planning...
The DSPM Architect’s Handbook View More
The DSPM Architect’s Handbook: Building an Enterprise-Ready Data+AI Security Program
Get certified in DSPM. Learn to architect a DSPM solution, operationalize data and AI security, apply enterprise best practices, and enable secure AI adoption...
What's
New